AI Product Management - Complete Course - 3.5 hours - Masterclass | AI Agents, RAG, Evals, LLMs.
By HelloPM
Summary
Topics Covered
- LLMs Are Next-Token Predictors
- Transformers Enable Contextual Understanding
- GenAI Value Stack Layers
- RAG Solves Context Limits
- LLM-as-Judge Enables Evaluations
Full Transcript
If you are someone who's looking to get into AI product management, then this is a mustwatch video for you. Few days ago, I conducted a free master class for
above 500 people who joined this session and then they were able to understand how AI product management is done. Now
I've decided to make this video available for everyone so that everyone can take the benefit. My only request is that before you go ahead and start watching this video, keep a notebook and pen with you so that you are able to
take notes and have your full attention in the session. I'm sure this is going to help you out and if it helps, go ahead and let me know in the comments.
Thank you and happy learning.
So, we have divided this two-day master class into three components. The first
component is geni fundamentals where we'll talk about the geni value tech types of AIPM and nature of large language models. Then we will talk about
language models. Then we will talk about inside EIPM. We'll talk about context
inside EIPM. We'll talk about context engineering, rag, prompt engineering, finetuning, AI agents and AI events.
Even if you're not able to understand anything right now, it is my promise that by the end of tomorrow, you should be able to understand these topics. And
then we are going to talk about giving you actual examples how products such as chat PR granular gamma AI notebook LM are actually built. Right? So this is going to be the session agenda for these
two days.
And the motivation or the purpose of this session is not to make you an expert in AI product management because that cannot happen in 4 hours. But the
motivation of this session is to make sure that you feel motivated after completing these sessions that yes I have a clear road map. If I can learn
this much I can also go ahead and learn and explore the entirety of AI product management. So this is to give you the
management. So this is to give you the confidence so that you are able to get started. Right? And tomorrow and today
started. Right? And tomorrow and today and in today's session I'm going to ask you in the end that whether you are feeling feeling motivated after the session or not right
now let's talk about the Gen AI revolution.
Right? The companies that you are seeing on the screen, they are now among the world's most valuable companies and many
of these companies have able to get this kind of amazing transition just after 2022.
So from 2022 to 2023, this company Open AI has got a valuation of 500 billion. This anthropic company which was not existing few years ago now
has a valuation of $183 billion and then we have something called as XAI Elon MRI which has the valuation of $200 billion.
Not only the new companies are doing well but the older companies which are betting on AI as a revolution which is Microsoft has gone from so Microsoft was
founded in Yes. Microsoft
Yes. Microsoft was founded in about 1980s 1983 and from then to 2021 they were able to get a
market cap of 1.8 trillion but in the last 3 years they have grown from 18.8 to 3.8 almost twice increase. Similarly
Google after betting on AI they have grown their valuation almost 2.5x similarly meta has grown too much from 0.3 billion or 300 billion to 1.8 8
trillion and then Nvidia which is the most which is the company which has most benefited from this AI revolution it has gone almost 10x of their valuation right
and then we have also seen some very new age companies which are actually created in this period only so ker the company did not existed few years ago is now worth $9 billion glean is worth $7
billion 11 labs is worth $6 billion and perplexity is is worth $15 billion So now there is a sharp change that has
happened in the market recently because of which people are able to understand that there is this new technology called Genai which can go ahead and do so many things together. Okay, but understand as
things together. Okay, but understand as many people have predicted this can well be a bubble. So how do you check if something is a bubble or not? It might
be possible that all of these companies are overvalued in some of the other way.
But let me ask you a very honest question. Okay, everyone go ahead and
question. Okay, everyone go ahead and answer me in the chat.
When was the last time that you have used chat GP and Gemini?
Yeah, I'm sure for many of you your one of the tab is chat GP right now.
Yes. Cool.
Right. So, understand there have been many technologies that have came and gone. For example, there was a
gone. For example, there was a uh there was a revolution on blockchain, there was a revolution on cryptocurrency, there was a revolution on NFTs. Although you were listening
on NFTs. Although you were listening about all of these things on news on multiple websites and Twitter but they could not become a part of your daily
life right so the verdict is that it might be possible that these things that we are talking about the valuation and all these could be larger than life valuation that is completely possible I'm not disagreeing with that but we
also have to understand that genai is a technology that has been able to prove itself that it is actually very helpful Because technology if it is only
increasing the valuation that might mean that yes it could be a bubble but if it is generally helping you in your daily life then it just means that it is just a matter of time that we are able to find its relevant use cases and utilize
it for work and more of our personal life right but how has it happened like what is there in the genai revolution so understand the genai revolution has
happened because of a innovation in technology which is called as large language model. Okay. So what is a large
language model. Okay. So what is a large language model? A large language model?
language model? A large language model?
A large language model is a neural network trained on massive amounts of text to learn patterns in language enabling it to understand, generate and reason with natural language by
predicting the next more likely most likely word or token in the context.
Right? This is the Wikipedia definition of large language models. But what
exactly is a large language model? Let
me go ahead and tell you. Please
everyone focus on the session super important part.
Yes. So we are talking about large language model.
So a large language model is it is kind of a code or an algorithm.
It is kind of a code or algorithm or a piece of technology that is very good at predicting the next word in the sequence. That is get very
good in predicting the next word of the sequence.
Right? For example, I just take this sentence.
I will take this sentence. I will
increase the size.
Right? So let's say if I enter this particular thing to a large language model, if I tell a large language model that a large language
model is a then it is going to return me neural. It is going to return to me the
neural. It is going to return to me the next word. If I give it a complete
next word. If I give it a complete sentence that a large language model is a neural then it is going to return me network and then I will keep on iterating so that I'm able to get the
complete sentence. So what is a large
complete sentence. So what is a large language model? Large language model is
language model? Large language model is the next word or token predictor.
And when you do this repeated number of times, you get the complete sentence, the complete book, the complete essay, the complete summary. Right? So this is the basic fundamental how a large model
works. You give it some sentence and
works. You give it some sentence and complete sentences. It will try to
complete sentences. It will try to complete it. And how this how does it
complete it. And how this how does it completes it? Let me go ahead and tell
completes it? Let me go ahead and tell you. Okay. So understand first of all if
you. Okay. So understand first of all if you guys are able to observe I have written token in the bracket. So what
happens is whenever you give a word to a large language model. Okay. So let's say I
language model. Okay. So let's say I give a word to language model that he lives near the river bank.
He lives near the river bank. Okay. So
large language model is a machine. It's
an algorithm. It's a code.
it will not be able to understand what are these exact words. In order to make sure that the machine is able to understand everything, what we do is
we convert this into numbers. So in
order to make sure that the machine is able to understand things efficiently, we convert these into numbers. These
numbers are called as tokens.
Let me show you how it works. So I'll go to a website called as tick tokenizer where if I write this sentence
the sentence is he lives near the river bank right now the large language
model has converted this the large language model has converted this into these tokens. So these are the tokens
right? Every model can generate
right? Every model can generate different kind of tokens. Every model
has different kind of combination of tokens. Okay. The tokens are generated
tokens. Okay. The tokens are generated to make sure that the large language model is able to efficiently utilize its memory and context because numbers are
easier to interpret for machines rather than the words.
Right? And in total in English language we have around 26 alphabets and we would have so many maybe millions of
applications, millions of spellings, right? Vocabulary. In GPT4 we have 100k
right? Vocabulary. In GPT4 we have 100k tokens, right? So what is a token? Token is a
right? So what is a token? Token is a numerical representation of your text.
Token is a numerical representation of your text. Right. And at a high level,
your text. Right. And at a high level, three words in English is equal to sorry
at a high level in English two words are equal to three tokens at a high level like not this is not
completely true in all the cases but for most of the times on an average two words is equal to three tokens. Why
tokens are important? to understand if you go to any of these application let's say let's understand open AI pricing so if you look at pricing of any of
these models you'll be able to understand that they work on tokens how many tokens are input how many tokens are generated how many tokens are cashed
right so for every 1 million tokens that you give to charge GP it is going to ask you for the input it is going to take $1.25 $25 from you. Right? So what is a
token? Token is a
token? Token is a conversion of your words into mathematical numbers so that the model is able to understand them very
very efficiently.
Right? How many tokens are possible in a model is called as the vocabulary size of the model.
How many tokens are possible in a model is called as the vocabulary size of the model.
Right? So this is this is called as the tokenization. Right? Now what
the tokenization. Right? Now what
happens the purpose of an LLM is that if you give it a set of tokens let's say 1 2 3 4 5 it will try to predict what is
the next token then if you run it again then this becomes your full sentence up to six then it will generate seven and then you have new sentence 1 to 7 and
then it will generate eight. So it is going to go ahead and complete the sequence. Right? That is at the most
sequence. Right? That is at the most fundamental level how large language models work.
Right?
Everyone a quick yes or no if you're able to understand everything so far.
A quick yes or no before I could go ahead.
Wonderful. Great. Now let me go ahead.
But this is a simple explanation. But
how does it happen? How does it predicts? Okay. So what happens is
predicts? Okay. So what happens is this is how a large language model works. Okay. So
there are three stages. Whenever you we create whenever we create a large language model,
we use these three stages. Every large
language model is created by going through these three stages. The first
stage is called as a pre-training stage.
The second stage is called as the training stage. And the third training
training stage. And the third training is called as a post training stage. Right? So in the
training stage. Right? So in the pre-training state, understand what happens is the models are able to predict the next token very accurately
because they are very good with pattern recognition. So what we do is we need to
recognition. So what we do is we need to train this model on a lot of data. We
need to train these models on a lot of data. Okay. But before that let me tell
data. Okay. But before that let me tell you let me help you understand why this data is necessary. Okay. So how many of you
can tell me what is this?
Tell me in the chat what is this guys? Go ahead and tell me in the chat
guys? Go ahead and tell me in the chat what is this?
Wonderful. That means you guys have studies in the uh in the elementary mathematical classes, right? So y= mx is
nothing. It is a very simple equation.
nothing. It is a very simple equation.
It is the equation of a line. any
straight line. This is y-axis. This is
x- axis. This is the slope and maybe this is the intercept.
Right? So this is a simple equation that is y = mx + c. Right? Now if I want to know
what is the value of y at this value of x. If I want to know what is the value
x. If I want to know what is the value of y at this value of x. So x is the input and y is the output.
So now if I want to know what is the value of x at this what is the point of y at this value of x because there are two things right now I might not be able
to understand because there are two parameters which are unknown to me I will not be able to understand but what if I already know two values of input and two values of output so let's say
somehow I was able to find that if x were our one if x is equal to
1 then y = 2 and when x =0 then y = 1.
Right? So let's say somehow it is given to me that this is true. This is called as training data. Training data is
training data. Training data is something where the output and the input is known to me. Right? Right now what
I'll do is I will try to find out the value of m and c. So y = mx + c. I'll
put the value of x and I'll put the value of y. So 2 is equal to 1 into m +
c. 2 = m + c.
c. 2 = m + c.
So c = 2 - m.
Right? I also have the second value. So
what I'll do is y = mx + c. That means 1 is equal to
m into 0 + c. c = 1. Right? Now c = 1
and c = 2 - m which means m = 2 - c. 2 -
1 is equal to 1. So c= 1 and m= 1. Now
the equation becomes I have found all the parameters. I was
able to find the value of M. I was able to find the value of C. So C is 1 and M is also one. So the equation is Y = X + 1. Now you give me any kind of input in
1. Now you give me any kind of input in the world. Give me any kind of output in
the world. Give me any kind of output in the world. Give me any kind of input in
the world. Give me any kind of input in the world and I'll be able to tell you what is going to be the output.
Right? You give me any kind of input in the world and I'll be able to give you the output. Right? Now if you go ahead
the output. Right? Now if you go ahead and amplify the same thing, if you go ahead and amplify the same concept to billion number of times. So now what
happens is a large language model is a equation like this
but it contains almost greater than a billion parameters.
Right? So in order to find two parame in order to find two value of parameters m and c I needed to
have two values of input and output also called as training data. Training data
that I already know right in the case of a large language model a large language model is nothing. It is a very complicated mathematical equation
plus the equation contains a lot of parameters. How many parameters? Maybe
parameters. How many parameters? Maybe
billions of parameters.
Right? So in order to determine these parameters, what I need to do? I need to have billions of training data, which is I need to have
large training data, right? And then this is repeated
right? And then this is repeated multiple number of times to get the values of these parameters. So how do we get this training data? We get this training data from the pre-training. So
before we train any model, what we do is we crawl the internet.
We look at the forums. We take the books that are written already, literature, well-written books. And then
we convert this data, we clean this data, and then we train on this data.
How do we train? How do we train on this data? Understand?
data? Understand?
Any piece of text that you have that piece of text can be converted into two parts the input and the output. For example,
let's say in the training data I was able to get this.
So I have written a blog around a IML product management. Okay. So let's say I
product management. Okay. So let's say I have this sentence. So I will collect
sentence. So I will collect I will collect a lot of data. I will
break down this data into small parts so that I'm able to feed it to the large language model and that then what I'll do is let's say one of the training data is this
one of the training data is this. So
what I'll do is I know that this sentence is correct. So
what I'll do is I will just feed this part I will feed this part as the input and this part as the output
because what does a large level model does if you go ahead and give it if you go ahead and give it any sentence it will try to complete it. So I'll give
multiple billions of rows and billions of uh let's say components of data in the form of input and output and when the large language model will learn it will learn from these patterns and then
it will be able to give you almost real kind of reasoned answers right a quick yes or no everyone if you are able to understand this
I have tried to help you understand large language models with the help of a very simplistic A linear equation.
Yes. Cool. Now let's go ahead.
Yes.
Cool. Now we have understood this. Now
let me go ahead and take you one step ahead. Okay. So now everyone please look
ahead. Okay. So now everyone please look at this very carefully.
So how does a model trains itself. What
happens is this is what the structure is. You have the training data. In the
is. You have the training data. In the
training data, let's say you have this sentence. You are going to break this
sentence. You are going to break this sentence into two parts.
One part is this the initial part of the sentence. Other part is this the later
sentence. Other part is this the later part of the sentence. The initial part of the sentence is the input and the later part of the sentence is the output.
So ideally what you want is you want the model you want the model to behave in a way that whenever you write how LLMs
it should complete with work. Okay, this
is what you want. So what you'll do is you will break the sentence. This is one sentence. This other sentence you are
sentence. This other sentence you are going to tokenize it. Tokenize as in you are going to convert it into maybe some kind of numbers. Okay. And
every model that you are running they would have their own method. So there
are some free and available APIs available through which you can go ahead and tokenize your content into any kind of numbers. Okay. So we have done the
of numbers. Okay. So we have done the tokenization.
Now whatever are the input tokens tokens for how LLMs are going to be feed as an input
right and the output token is going to be here I'll take I'll tell you how it is going to work right so now it will be feed into a large language model okay a
large language model is a specific kind of uh I would say uh right now we are talking about a large language model which is a neural network and the Type of the neural network is transformer.
I'll talk about what a transformer is in brief in tomorrow's session. Okay. So a
transformer what is a transformer? It is
nothing but a mathematical expression a non- mathematical expression understand it thousand times more complicated than the line equation that I have given you. And it contains the
parameters the constant that I have told you right. And the purpose of this large
you right. And the purpose of this large language model training is to identify the value of these parameters to identify the value of M, C and other
parameters that it has. It can have billions of parameters. Right now what will happen? It will initially when it
will happen? It will initially when it is not trained, it will try to generate anything random. Okay? And also
anything random. Okay? And also
understand one thing that the large language models are not deterministic.
Listen to this very carefully. This is
super important to understand the language of large language models. I
told you that they are next token predictors. Predictors means they will
predictors. Predictors means they will not tell you that this is a token after this. They will give you some
this. They will give you some probabilities. So
probabilities. So he is a this is the input and the large language model is going to give you
probabilistic outputs. So it is going to
probabilistic outputs. So it is going to give you many options. It is it will tell you he is a man, he is a boy,
he's a father, he's a singer. So there are multiple probabilities, right? And then he's it
probabilities, right? And then he's it is also going to tell you the L language model that what is the probability. So
let's say man the probability is 0.5 boy the probability 0.4 father probability is 0.05 single probability is 0.05.
Right? So it will not tell you the one answer. It will not just give you one
answer. It will not just give you one answer. It will give you the
answer. It will give you the probabilities of what other words could be there and then you can go ahead and choose the top word or second word whatever your large language model can
go ahead and pick up. Okay. So always
remember the large language models are not deterministic. They are stochastic
deterministic. They are stochastic which means they can go ahead and pick random words which are going to be correct in some
context but if you go ahead and give same prompt to same model at the same time it can go ahead and give you different kind of outputs
right so this is important nature of an LLM to know and this is also going to be of foundation of when we go ahead and learn about AI evaluations Right?
And how does it generates this? It
generates on the basis of training. So
it would have seen a lot of data and based on that data it generates that yes I think this should be the probability based on what I have learned.
Right?
Yes.
Cool. Now
going back to a diagram. So what will happen here is what we have done so far is we have done the tokenization we have taken the input part we have feed it into the transformer the transformer has
generated some random results and then what will happen we have the real actual token the actual token is going to be
the meaning of work like the the the numerical token for work. Okay. And here
it is going to generate something. Now
it will be compared that work should have a probability of one or work should be at the top of the predicted numbers or the predictive tokens and then we are
going to calculate the difference that it should be word it should be work but it is not work. So it is going to share this feedback with the transformer and then what
transformer will do is transformer will correct the parameters inside and then this model will run on multiple training sets for billions and trillions
of time and then you are going to go ahead and get a model that is going to accurately predict what should be the next probable word.
Right? So what we do is we give it the training model. We have the output. We
training model. We have the output. We
have the input. Then we check we check the difference between the real output and the expected output. Right? Then we
calculate the difference which is called as a loss function which is called as a mathematical loss function. This is passed as a feedback
function. This is passed as a feedback feedback to the large language model.
And then when it keeps on doing this billion number of times it is able to set a very accurate value of these parameters.
Right? If you run it a number of times, you will have a pretty accurate model or a large language model. Right?
Understand? If the number of parameters are in a race of billions, this is called as a large language model. If it
is less than billions, maybe in the ratio of millions, it will be called as a small language model. Example of a large language model. Example of a large language model is Gemini chat GPT spell
uh example of a small language model is Google's Gemma right yes quickly everyone a quick yes or no
in the chat if you're able to understand this at least at a high level wonderful great you guys are intelligent great amazing So loss function understands the
difference between what is the predicted token and what is the actual token and then it pass on this feedback to the transformer. The
transformer adjusts everything accordingly.
Right now a bigger question that is that should aride arise in your mind is that why all of these things are happening after 2021
like uh why all of these things are happening very recently why not before okay so understanding language
understanding language is not an easy task for machines why because of this reasons so let's say my sentence is he lives
near the river bank.
Right?
Initially, traditionally what people used to do was if they want to create any kind of AI, if they want to create any kind of predictive model or intelligent model, what they will do is
they will work on word by word.
You might have seen autocorrect in your mobile phones, right? Whenever you try to write something, it will only catch the last word and then based on it, it is going to predict. It will not look at the entire sentence. Okay. But people
who are able to understand, people who understand NLP or natural language processing or programming, they were
able to understand that it's not enough.
They were able to understand that if if we want to really understand language, we need to understand the meaning of each word with respect to that sentence.
For example, in this particular sentence I have written he lives near the river bank.
So in this bank means the side of the river.
Right? But if I did not know this sentence, I could have as well concluded that this bank is a financial institution.
So in 2017, a group of computer researchers created a research paper, published a research paper. They all were working in
research paper. They all were working in the Google brain and deep mind team. And
the name of the paper was attention is all you need.
And they told they came up with the hypothesis that if we really want to truly understand the language models, if we truly want to understand the language
rather than looking at things token by token, we should try to find out what are the relationships of these word with entirety of the sentence. Initially
these these predictive models can only look at one word at a time but now with this approach we were able to go ahead and consume all the sentences and essays and whatever is
the context window at once. So every
word is now able to map it relationship with other words. So these words are created into tokens and vectors and then the relationship between all the vectors
is calculated real time and then as a whole we are able to understand what does this paragraph what does this sentence what does this context says
that was a very big shift of how genai was born okay but now because of that one problem occurred the problem was the sequential
processing of data was easy It can be done on a CPU but if you need to process multiple sentences multiple parts of the sentences at once understand when I'm
talking about the sentence he lives near the river bank
at once the algorithm or the computer has to calculate this relationship of he with lives with
near with the with river with bank. Then
parallelly it needs to understand the relationship of lives with he near the river bank and then it also needs to understand the relationship of this with
this. So now understand this is
this. So now understand this is complicated maths.
This is resource intensive maths. Resource intensive
intensive maths. Resource intensive maths. It will take time. It will take
maths. It will take time. It will take effort. It will take a lot of processing
effort. It will take a lot of processing power, right? So in order to create a
power, right? So in order to create a good solid large language model, you need two things. First, you need a lot of data on which you can train so that
AI is able to find patterns. Plus, you
need super huge processing power.
And now people started thinking where do we get this processing power from?
because quantum computers are not yet invented at least at the scale for practical purposes and CPUs are very much limited.
Then people thought that these kind of metric multiplication like one to many multiplication also happen at gains
and games are running on GPUs.
So people understood that yes transformer is an architecture. So the
neural networks or the algorithms that operate in this way by calculating the uh the relationship of one word with
every other word. These kind of algorithms are called as transformers.
They work on the mechanism of attention which is one word is paying attention to every other word in the sentence.
Attention means one word is paying attention to every other word in the sentence to derive its meaning. Right?
And then the neural networks or the algorithms that use attention as a mechanism to predict are called as transformers.
Right? And transformers are best run on the same technology that runs these complicated games with is GPUs.
Right? So this is a history about how GPUs were best suited for running large language models. And that is also the
language models. And that is also the reason why we were able to get a huge jump here. We were able to get a huge
jump here. We were able to get a huge jump here.
Right everyone, a quick yes or no if you're able to understand this.
Yes. If I see one more LinkedIn link in the chat, I'm going to block you. Don't
distract other guys, please.
That's so large language models run on GPUs. They are trained on GPUs.
GPUs. They are trained on GPUs.
Right? Now,
so we have understood this. Now let's go ahead. Okay. Now the third part, I have
ahead. Okay. Now the third part, I have told you about the two parts already.
The third interesting part of a large language model.
So we have talked about pre-training, we have talked about pre-training, we have talked about training. Now the
third part is post-training.
Now after you generate, after you train a large language model on all the internet data that is available, you get a personality like a baby. A baby has
the ability to think. It has the ability to perceive its surroundings and they will also be able to go ahead and maybe blabber a few words, right? But they
don't have a personality yet. So
depending upon a baby is born in a Hindu family or a Muslim family or a particular culture, they are going to get their nature. So if you are reading Quran, you are going to think this is the God. If you are going to read maybe
the God. If you are going to read maybe Gita, you are going to understand that maybe this is the God. Okay? Reality
could be something else altogether.
But based on how you are conditioned, based on how you are trained after you are in the world, that is called as postraining. That gives you a
postraining. That gives you a personality. So postraining is very
personality. So postraining is very important. It converts
important. It converts the next token predictor machine next token predictor machine to a
helpful assistant.
It converts the next token predictor m machine to a helpful assistant. Right?
So what we do is we hire real human laborers like what what do we call it? People who go ahead and label the data. So these
people give accurate information to the large language models and then this data is feed again into the training and then the large language model gets a
personality. For example, there is one
personality. For example, there is one thing that large language models have struggled for a long period of time.
Okay. So large language models generally tend to hallucinate. Although the
hallucination has reduced very uh very recently but still there is let's say a lot of hallucination still.
Okay. So Meta came up with this very interesting perspective. What they did
interesting perspective. What they did was they just give some sample input and output to the large language models.
In the input they told some random question like question for which the answer is not possible right for example they created some
random names and ask GPT or ask let's say llama that who is this person initially before this training llama or chat GPT they are going to give you random answers even if the person does
not exist they will try to frame something that is called a hallucination they are going to give you wrong answer but they'll give you with complete confidence right so what meta did was it gave them some questions which did not
have any answer and in the output they mentioned they asked the llama model to give the answer as I do not know about this I do not know about this and then
once the large language model started understanding that yes I can give the answer that I do not know about this that is okay right to show you some other sample data
okay so this is how the data looks like for example we want to understand about summarization we want to make sure that our model is able to summarize. So we go give this kind of input. Summarize the
following passage about neural network in one paragraph. A neural network is composed of layers of neurons that transforms inputs. Okay. So we'll give
transforms inputs. Okay. So we'll give this prompt and then this is the desired output. This will also be feeded into
output. This will also be feeded into the large language model and then we will be able to retrain it again and again in this data to make sure that we
are able to get the right answer.
Right? we are able to go ahead and get the right answer. Okay. And then these are the methods in which we are going to test. It has to be clear factual summary
test. It has to be clear factual summary and neutral to right. So we give all of these feeds to large language model after it is created to the base model and then we are able to get a more
intelligent model.
Right? Now guys, all of this data, all of these files, all of these presentations, the notes and everything would be given to you if you remain till the end of the session today and then we are going to go ahead and uh and you go
ahead and give the submit the feedback form. Okay.
form. Okay.
Yes.
Cool. So I hope everyone was able to understand how large language models works. Everyone please tell me on a
works. Everyone please tell me on a scale of 1 to five if you're able to understand whatever I have told you so far. On a scale of 1 to five, please.
far. On a scale of 1 to five, please.
One you did not understand, five you understood it well.
Cool. We have 821 people. Amazing.
Great.
Super. Super guys. Amazing. Amazing.
Thank you.
Yes.
People who are not able to understand, just look at the recording. I'm sure
you'll be able to go ahead and understand this well. Yes. Cool. People
are not able to understand, please focus better. Please focus better in the
better. Please focus better in the session.
Yes.
Cool. Let me just go ahead.
Yes. So now based on whatever I have told you now let us go back to economics. Okay.
So we have talked about mathematics, we have talked about computer science, we have talked about psychology of a model.
Now let's back go back to economics and the business. Okay. So
the business. Okay. So
as a product manager you need to understand at what levels is the value created in the large language models by the large language
models. Okay. So the first layer is the
models. Okay. So the first layer is the infrastructure layer. This is the layer
infrastructure layer. This is the layer on top of which large language models operate.
Okay. What they do is here we have companies such as Nvidia and Google's Vert.x AI which provides the tensors or
Vert.x AI which provides the tensors or the GPUs where these models can run. So
they provide computation data platform GPUs and all right. So the core technology that you need which is infrastructure which is uh how where do you deploy
these models is provided by these infrastructure layer companies.
The second layer is the model layer companies who are creating and owning these models. Companies such as OpenAI,
these models. Companies such as OpenAI, Anthropic Meta Grock uh and Deep Seek these are the companies which are owning these models and they will give you large language models,
small language models. They will also give you fine-tuning. What is
finetuning? We are going to discuss in the coming sessions. Right now, this is the core. Now, on top of this, there is
the core. Now, on top of this, there is a bigger opportunity that you as a product manager can utilize all of these
technologies like these models in order to build some amazing use cases. So,
lovable gamma chat GPT Gemini they are all based on this model and the infra layer. So you as a product manager
infra layer. So you as a product manager can play a very important role in the application. Application layer is how
application. Application layer is how are you able to utilize the power of a large language model in order to solve the problems for your customers.
Right? So internet is provided by someone else. The routers are provided
someone else. The routers are provided by someone else. But Amazon has created an application on top of internet which is able to give you the e-commerce at your doorstep.
Tinder has created an application on the top of internet so that you are able to find your next state. Similarly in AI you should be able to create a lot of applications on these on the top of
these large language models so that you are able to create a lot of value right and the fourth layer is you can
use these tools you can use these large language model based tools such as chat GPT such as GMA such as Kurser in order to make your clients and make yourself
more productive. So service agency
more productive. So service agency companies such as TCS, Realast and all they will use these tools and then they will help the company achieve their outcomes.
Right? So this is the whole value chain.
People create infra then some people create model some people create application based on model and some people create services. Okay. But this
is where you can drive a lot of value without investing a lot of money.
And this is value at scale.
Right? All of the companies here, all of these companies are same companies. They don't own models. What they do is they work on top
models. What they do is they work on top of models created by other companies, right? And they are able to generate a
right? And they are able to generate a lot of value.
Right? Now, based on this gen AI value stack, we can divide multiple kind of AIPM jobs. Okay? So, first of all,
AIPM jobs. Okay? So, first of all, understand that every PM is an AIPM. If
you are using any kind of AI technology in your life, GPT, Gemini or anything, you are an AIPM. So I divide AIPM into two parts. The first part is the AI
two parts. The first part is the AI enabled PMS. These are people who are using products like Chat GPT, Lovable, Jira or any
other AI product in order to make themsel more productive in order to solve problems for their customers. in order to save time in
customers. in order to save time in order to provide more impact. So anyone
any PM who is a user of any of these AI tools they are an AI enabled team. So
100% of all of us are AI enabled PMs. Right. The second is AI product PM. Now
Right. The second is AI product PM. Now
this is where the value is. This is what we are going to talk about which we have talked in the last sess in the uh today's session as well initially and then we are going to talk about today more and in the next session as well.
Okay. So AI product PMs can be divided into two parts. One is the core AIPM.
These core AIPMs need to understand how the technology behind AI works. So
training pre-training post-raining how do you create a large language model? How do you store? How do you make
model? How do you store? How do you make it efficient? How do you consider memory
it efficient? How do you consider memory of a large language model? All of these core things are the responsibility of a core AIPM. So people who are creating
core AIPM. So people who are creating the infra or the model are the core AIPMs. These people need to be technical and they should have a solid understanding of machine learning in
order to become a core IP.
Whatever I have told you so far about how GP works and all this is a work of a core IPM.
The second kind of PMS where most of you should be going is applied AIPM. Now
applied AIPM are the people who are creating useful applications by using these core technologies. So people who are building notion AI, Grammarly, lovable, chat, GPT, people who are
building these products are applied AI PMs. Applied AI means the people who are actually using AI tools like AI
technologies such as models and infra in order to build something helpful for their customers.
Okay. So these are the only the users.
These are the core technology creators and these are the product creators.
Right? So can you guys tell me where do you fall right now?
What is your best description right now?
Where do you fall right now?
Yes, wonderful. So all of you fall under the
wonderful. So all of you fall under the AI enabled part right now but eventually you should go ahead and Yeah guys, everyone is trying to focus and
learn. Please do not put any kind of
learn. Please do not put any kind of links in the chat. I've already blogged a couple of people. Do not let me to do it more. Let us go ahead and focus on
it more. Let us go ahead and focus on the session.
Cool. Now
this is the differentiation. Everyone is
here.
If you are coming from already a very strong technical background, you should try to move here. But everyone, even if you are coming from a technical or nontechnical background, you can always move here. This is where I want you to
move here. This is where I want you to move.
Okay. Now let's go ahead. Okay. So tell
me everyone.
Tell me everyone if you are able to understand this meme. Look at this very carefully and tell me if you are able to understand this meme.
Yes, cool. So understand there are like so much valution that has been created in all of these things. Okay, this is sitting on the top of LLMs. Okay, so I'm
going to give you some examples of how various kind of technology the various kind of technologies are working on top of these large language models. Okay,
but we are right on time. We will take a quick break here. We will take a quick break of five minutes and then we will come back. Okay.
come back. Okay.
So now let me go ahead and tell you how do So now you have understood that yes a large language model. Okay. So
there is something called as a large language model.
which is provided by companies such as Open AI, Anthropic, Meta, Google and many more. Okay, these are
the large language models that are provided by so many companies. Now,
right now, this large language models can do three things for you.
Okay, the first thing is they can understand content.
They can understand what you are trying to say because they have seen so many patterns of data. They're able to understand what you are trying to say.
Okay. So if you ask it to generate a summary, it will be able to understand the whole paragraph and then it will generate the summary. Okay. The second
is it is able to transform content.
Transform means if you ask it to create a summary of some content, it will transform that into summary. If you ask it to create a quiz on some content, it will be able to transform that into
code. Okay. And the third part is it is
code. Okay. And the third part is it is able to generate content. So if you want it to generate images, code anything it because it is data it should be able to generate because it has seen enough code
in its life right. So it will be able to go ahead and generate. So as a large language model the capabilities of a so as a product manager you should understand that there are three core capabilities of a large language model.
The first is it can understand second is it can transform and third is it can generate data.
Right? So you can use it for multiple ways. You can use it
ways. You can use it to generate summary. You can use it to generate answers for your questions. You
can use it to generate code. And you can generate for almost everything that you want, right? And because they have been
right? And because they have been trained on tons of data, tons of terabytes of data, generally the accurations are very accurate. So now
what happens? So far whatever the software that we used to create so far whatever the software that we were used to create these softwares were generally of four
types. Okay. So any software that we
types. Okay. So any software that we have used so far what they will do is they will generally create a data they will read data they will update the data or they will delete the data. So all the software that we have seen let's say
Facebook on Facebook what happens you go ahead and create a profile and then on your profile some database element is created and then your people can see your profile some people can go ahead and upload some videos and then you are
able to see these videos and entertain yourself so generally most of the application before the AI world were actually create read update and delete
this is called as application but now AI has been able to give a brain to your softwares. So as a product
manager, you need to understand what this large language model can do and how can I utilize it to create better products.
Right? So
I'm going to give you this understanding with the help of few examples. Okay? So
there is a product called as granola.
Okay. And this product is similar to many other products. So it's a very simple product that if you go ahead and use granula, it will go ahead and get added to your calendar
and whenever you are joining the meeting, it will take the transcription and recording of that meeting and it is going to summarize the meeting for you so that you do not have to pick up a notebook every time. What happens is
whenever we are joining the meetings we have to write down that what are the things that are discussed so that we do not forget and then we create a minutes meeting like the minutes of the meeting right so there are multiple tools like
granola there is fireflies otter fathom most of you most of them which you can see also on the uh the attendee list right because many people have added
their note takers so now how these noteers are created what is the value behind them so I'm going to show you a simple tear down Okay.
So, everyone a quick yes or no in the chat if you understand what a AI noteaker is.
Amazing. Great.
Yes.
Yes. So now let us try to understand how this product works and this is like a very important case. It will help you understand how AI products are built.
Okay. So first of all understand everyone whenever you create any product as a product manager you have to look at five angles.
Don't just go ahead and just think about technology. Technology is one important
technology. Technology is one important part but it is not the most important part. But AI is just a piece of
part. But AI is just a piece of technology. So do not focus on just
technology. So do not focus on just using AI. Focus on solving the problem
using AI. Focus on solving the problem of the user. So whenever you look at any product, please ask yourself five questions. Right? The five questions are
questions. Right? The five questions are what problem does it solves in people's life?
Right? What problem does it solves in people lives? Okay. So
people lives? Okay. So
can you guys tell me? to think about it for a moment and then tell me in the chat what do you think what are the problems that granola solves in people's life
think about this as a product manager this is the most important value that you need to add that how am I being useful to the users what problem does it solve so tell me what is a problem that
these note takers solve Yes. Very good. So
Yes. Very good. So
at a task level, it helps you record and summarize your meetings.
But that is just an output. The outcome
is that it is able to save you time and it will help you be more accurate because when you are writing something it might be possible that yes you are missing something and you also have to
invest a lot of time and attention when you are writing things right so granular saves you time it makes you more efficient and accurate always remember
it is okay if you're not using AI in your product AI is not a silver bullet that will improve your But if the product does not have a value, if you are not able to clearly define why this
product should exist, then please do not create the product. Okay, this is the most effective value that you add as a product manager. Okay, now the second
product manager. Okay, now the second part which is a business value.
Why should the business per pursue this opportunities?
Okay, think about this for a moment everyone and then tell me in the chat why should the business pursue this opportunities?
How do you find this out?
Let us keep it collaborative so that everyone is able to learn accordingly.
Yes.
So now what we'll do is now what we will do is we will try to understand that if we are creating a meeting assistant
how many people in the world are first looking for this kind of solution.
So you will have a prediction that people who are a part of small businesses or large businesses they are going to love this kind of solution because they are going to save time and then you are going to understand how
many people are there who are working in these small businesses. how many small businesses are there and then you are going to size a market opportunity.
Okay, that is one way that it is going to help the business because there is a large market opportunity and in order to do this research you can go ahead and use chat GPT notebook LM Gemini or anything right that is the business
value now the other kind of business value is let's say we are not talking about granola we are talking about zoom right so because zoom already has a lot
of users who are using its platform to conduct meetings it is an strategic opport opportunity for Zoom. It is a strategic business opportunity for Zoom
to create an inbuilt meeting assistant.
Right? So it aligns with the strategy and it also aligns with the market opportunity.
Right? So understand everyone as a product before being an AI product manager, you are a product manager and the purpose of a product manager is to bring the user value and to bring the
business value.
Right? Now the third thing is how do we make sure that the product is intuitive for the user to use. How do we make sure that the
use. How do we make sure that the product is very intuitive.
Right? So in all of these tools you would have found that you do not have to do anything. You just install the tool
do anything. You just install the tool you add it to your calendar.
Automatically whenever a meeting occurs the person the the meeting note taker would be automatically added. Granola
goes to the next level that right now this meeting is recorded with granular but you do not know why because it is installed on my computer. It is taking over the microphone and the video and it is doing it by itself. So now you have
to do optimizations on the design part so that you are able to reduce the friction for the user. If I have to add a granular every time, if I have to add uh uh this fireflies every time when I'm
doing a meeting that is going to be resistance for me right? So this is called as being a proactive software.
So they are able to get meeting into meetings automatically. So you don't
meetings automatically. So you don't have to do the work automatically it is happening. Right? So this is the
happening. Right? So this is the intuitiveness and then they have created an interface so that I'm able to see what are the meetings that I have conducted. I can go ahead and ask
conducted. I can go ahead and ask questions based on my meetings. Right?
So this is the value part, business part and the design part. The next part that we are going to understand today is how will this be engineered.
Engineered part is feasibility part. But
please remember this line. User value
and business value as a product manager should be your first focus and feasibility should be the next focus.
Feasibility should not be the number one focus of a product manager right now.
Please start with the user problem and the business problems. That is what you add the most to the company. That is where you are being most impactful rather than only talking about engineering and
feasibility because there are people who are more eligible than you who can do the task. Okay. So how will this AI
the task. Okay. So how will this AI product be orchestrated? But before that everyone a quick yes or no in the chat if you are able to understand the three
parts user value, business value and design value.
Amazing.
Wonderful.
Yes. So there is one person who has said no four number of times.
Great. So
yes. So now we have understood the three parts. Now let me give you a bit about
parts. Now let me give you a bit about engineering. Okay. So this is how this
engineering. Okay. So this is how this product looks at the back end. Now look
at this very carefully. Super important.
If you just under understand this one product, I can assure you that you can reverse engineer almost any AI product.
So please focus on this slide. Okay. So
behind granula whenever you do any meeting on zoom on meet or anyone
granola or fireflies or otter or any of the meeting recording tools even the zoom tool they will be connecting to the meeting either they can join the meeting
as a bot or they will capture the content from your screen whatever you are speaking they will take over your mic or they are going to ask the API. API as
in they will ask Zoom or Google Meet to share the recording or the transcript if you have given the permission. Okay?
So they will get the audio of the meeting, right? Once they get the audio of the
right? Once they get the audio of the meeting, they are going to convert that audio into text transcription.
There are multiple APIs available for the same. So, OpenAI has an API called
the same. So, OpenAI has an API called as whisper where you give it audio. It
will be able to give you the transcript of that audio. There is also Google text to speech like uh speech to text where you can just go ahead and give it a
audio. It will be able to convert into
audio. It will be able to convert into accurate transcript.
Right? There are some smaller players as well which are assembly.ai and other players who can also do this job. Okay?
So now what we have is we have a transcript.
Now this transcript is available with us. What these people will do is what
us. What these people will do is what granula or fireflies will do is they will take this transcript and then they
will add a system prop.
Okay. So this is a system prop.
So now everyone please look at this very carefully. Look at this very very
carefully. Look at this very very carefully. So it is going to add this
carefully. So it is going to add this prompt. The prompt is you are an expert
prompt. The prompt is you are an expert chief of staff and noteaker.
Given the following raw meeting transcript, generate a structured and actionable summary that includes the following sections. meeting title,
following sections. meeting title, data and participants, date and participants, key decision made, action items, open questions, summary, notable quotes and then you have written some
more instructions.
Right? So this prompt might be created by a product manager at granula which will convert a raw transcript that
is generated into very helpful meaning notes, meeting notes, right? And this
prompt plus this transcript would then be sent to a large language model.
Right? This will be sent to a large language model such as Gemini or Jad GPD, right? And then it will do its
GPD, right? And then it will do its magic. It will create summary
magic. It will create summary actionables and tasks. Okay? So this is how these meeting tools work. And then
what happens? You can take it to the next step. You can make it more agentic.
next step. You can make it more agentic.
Agentic means they will take action. So
what can happen is they can also go ahead and push these tasks to your slack to your calendar or to your Jira. So now
you are just adding a meeting notetaker like a meeting assistant to your calendar and then it will get the recording. It will do the transcript. It
recording. It will do the transcript. It
will go ahead and create the summary with the help of system prompts and the transcript. then it can go ahead and convert into multiple task and then it can also if you give it the permission it can also go ahead and add
it to your Slack calendar or Jira right so this simplistic architecture which is dependent on multiple LLMs and things is now
a possibility because large language models are possible right everyone a quick yes or no if you are able to understand This
right? So I want you guys to think in
right? So I want you guys to think in this manner. Okay. I want you guys to
this manner. Okay. I want you guys to think in this manner.
Everyone a quick yes or no in the chat please.
Yes. So now large language models have unlocked intelligence for you. Okay. Now
you can use them like an expert in order to go ahead and do a lot of things.
Right? So now this is like a simple orchestration. Now you can think about
orchestration. Now you can think about how you can innovate on top of the same right. So the history of granular is
right. So the history of granular is that that before granular there were players like fireflies, autoi and fedum.
What happened because of these was they will keep themselves added on. They will
keep themselves added on.
Yes. They will keep themselves added on multiple meetings and people generally get irritated if so many bots are joining the meeting. Okay. And it was also not very privacy intense in
incentive for intensive for other people. Okay. That people do not want
people. Okay. That people do not want your bots to join their meetings. Okay.
So Granla came up with a AI with a UI innovation. With Granla you don't have
innovation. With Granla you don't have to add anything. It will operate in your computer. It will listen to the audio
computer. It will listen to the audio and then it will it is going to automatically transcribe the meeting for you and give you the context right so you can do these optimizations in order to make sure that you are able to go
ahead and yes you are able to go ahead and create a lot of things right now the magic that you can add as a product manager is understand AI is
not difficult you just have to understand the concept that I'm telling you after that the next real skill that you have to develop is customer empathy Understanding
what customers want and giving them as soon as possible in the best design possible. That is the value that you add
possible. That is the value that you add as a product manager.
Okay. Now we have talked about granular.
Now let's go to the next level. Okay. So
granular is a very simple example. Okay.
The next example that we need to understand is okay. Okay. So one part is left.
okay. Okay. So one part is left.
That part is understand just building a product which is useful and business wants to build it and it is good design it is good feasibility doesn't make sure that product is
actually very very helpful right doesn't make sure that the product is very very helpful
so there is one more thing okay there is one more thing that is needed to make a product achieve its potential and generate revenue out
of the for the company which is how do we grow this product in this market which is creating a go to market strategy building
a product that is actually loved by the users creating distribution for the product making sure that people are actually using the product that is called as the growth part so how do you
grow in this market so all of these meeting assistants they are viral by default So let's say I and you did a meeting
together and because I believe in productivity I do not believe in taking notes manually. I have added a noteaker
notes manually. I have added a noteaker called as fireflies. Now when you see this that in the meeting there is another thing which is fireflies. You
are going to get curious that what is this? I should also go ahead and check
this? I should also go ahead and check what is this. Okay. So by using virality they are going ahead and creating a distribution right. Similarly whenever
distribution right. Similarly whenever the meeting has ended most of these people are going to send you emails. Let
me show you one of the emails. Just give
me a moment.
Let me check if I have those kind of emails. What is the name? Fireflies.
emails. What is the name? Fireflies.
Yes. Cool. So, let's say cool. So,
cool. So, this is the link. Okay. So generally
when we do the meeting this is the link that we received after that. Okay. So
someone added me to their meeting and they also added fireflies because they were using fireflies and after the meeting has ended it always sends me this thing that this is what you have done. This is a summary this is the
done. This is a summary this is the action item and it also tells me to go ahead and if I click on view the meeting and all it will ask me to sign up.
Right? So these are some growth tactics that these people are implementing in order to make sure that it is able to grow. The product is actually able to
grow. The product is actually able to succeed. So marketing and growth is also
succeed. So marketing and growth is also a good responsibility or a must have responsibility for a product manager, right? And understand guys, product
right? And understand guys, product management is multi-dimensional.
Not that you just create a great product and put all of your efforts into it and it is going to work. You need marketing, you need stakeholder management, you need a bit of understanding of technology, you need a bit of understanding of analytics so that you
know if you are moving in the right direction. It's a multi-dimensional job.
direction. It's a multi-dimensional job.
Right? So this is the story behind granola. Now let me go ahead and give
granola. Now let me go ahead and give you some more complicated examples.
Okay? So
tell me everyone if you ask chat GPT if you ask chat GPT that what is the most important item in my Google
calendar tomorrow would it be able to answer?
If you ask GP that what is the most important item in my calendar tomorrow would it be able to answer yes if you have not connected a Google
calendar it would not be able to answer right and if I take example of let's say this particular product called as calendar is connected to Google but if I
ask about notion then Gemini will also be not be able to answer Right.
So now we have so much data in our calendar, in our to-do list, in our Jiraa dashboard, in our stripe document like stripe dashboard. But how do we make sure that all of this content is
available to large language models so that they can go ahead and give us the answers. So rather than asking like
answers. So rather than asking like rather than thinking that where is the PD what is the task that I need to do today from my Jira or notion I should just ask a query that what are the three
things that I need to get done today and then it should give me the answer. Okay
if I want to go to this level what I can do is a very simple approach that I can do is I can just go ahead and give all
the data from my notion or stripe or atlation to the large language model.
Okay. But when I give a lot of data to the large language model, what happens?
The model runs out of context.
Okay? Because these models cannot carry infinite context. Context
means the kind of input that you can provide to these models.
Right? So now we have to work we have to work hard to make sure that these models work. Okay? And also although companies
work. Okay? And also although companies are trying to increase the context window as much as possible, there has been research that is done by great people at Chroma which is a
vector-based database company. They were
able to find that as you increase the context, as you increase the context, the performance of the model goes down.
Even if these people claim that our model can work in so much context, when these people actually tested multiple models, they were able to find in almost all the models that as the context or
the input size increases, if you increase the size of your prompts, then the quality of performance actually
reduces after a certain level. Right? So
now understand as a product manager, so even if someone is not able to understand this or this, at least you'll be able to understand this. Okay. So
when you give concise input context to the large language models, they appear to be very intelligent and very insightful and very careful. But when
you go ahead and give them too large input, then they don't know what they are doing and they get confused.
Okay. So now how to solve this? What we
do is we have another methodology called as context engineering. Context
engineering means how do you carefully curate what should go as an input to the large language model so that it is accurate so that you have given the
right kind of information so that you are able to get the right kind of insight or output from the large language model right so this is called as context
engineering so context engineering I am going to quote the person who has acquainted this term which is and Karpathi. So he says that context
Karpathi. So he says that context engineering is the delicate art and science of filling the context window with just the right information for the next step.
Art and science means you have to apply your mind. You have to think creatively
your mind. You have to think creatively to give enough context to the large language model so that it is able to answer you in the most accurate format.
Right? So this is context engineering and in context engineering we do multiple things. Okay. One of the point
multiple things. Okay. One of the point of context engineering is prompt engineering. How well you write the
engineering. How well you write the prompt is a part of context engineering.
Okay. But apart from prompt what happens is you need to give some data about your software. For example,
software. For example, if you want to build a software called as notion AI. Okay. So
let's say you will to you want to build a software called as not not motion AI but let's say you want to build a Bloomberg chatbot.
What is a Bloomberg chatbot? Bloomberg
is an organization which has access to a lot of financial data available in the world.
Right? So now what you want to do is that whatever the data is available with Bloomberg you should be able to send a question to a large language model in a natural language and then it should be
able to give you answer based on the content that is available with Bloomberg.
Right? So here the problem is you cannot store all the content that Bloomberg has in the context window. You cannot have all this content as the input. Why?
because it is going to run out of context. So what you will do is you will
context. So what you will do is you will carefully build something called as a R a pipeline. Now this is a very important
pipeline. Now this is a very important concept.
Even if you remotely want to become an AI product manager or AI practitioner, you need to understand this. This is
like a very basic thing that every person who calls themsel AI enabled has to understand. Okay. So rag means
to understand. Okay. So rag means retrieval, augment and generate.
retrieval, augment and generate.
Okay. So what happens when someone comes to your house and they
tell you that I need some sugar. Okay.
When they tell you, "I need some sugar."
You don't go to your study room and you do not look for sugar in you do not start by looking for sugar in your cupboard or in your
wardrobe or in let's say your library or maybe under your sofa. What you do is you straight away understand that sugar is an item that might be available
in my kitchen in a particular shelf. It
might be possible you don't know that which part of shelf that it is put on but you know a particular shelf or maybe you at least know that it is going to be found in the kitchen. Similarly, if your kid asks for a book or if they ask for
money, you do not go ahead and go anywhere else in the house. You know
where the money is put. You are going to just pick the room where the money is.
Then you are going to pick a particular wallet and then you are going to see in the wallet that yes in all of these four to five pockets there could be a place where I have put the money. Right? So at
a high level you know that where this information is the same analogy can be drawn to rag. So in rag what we do is
the first part is we get all the data. We get all the data. For example, we have PDFs or a
data. For example, we have PDFs or a better example is we have Bloomberg data.
So Bloomberg data is they have a lot of files. So we are going to break down
files. So we are going to break down files into small small chunks maybe in the form of paragraphs. Okay. So we have one chunk, we have another chunk, we have another chunk. We are going to
create all of these chunks which are small enough and meaningful enough.
Okay. We are going to break down this data.
Now in this data what we'll do is what we'll do is now we have the chunking of data. Now what we'll do is we will convert these data chunks into something called as vector embeddings.
Now listen to this very very carefully.
Super important. If you miss this you will miss rag. Okay I'll repeat this again.
You will take the data whatever data is available you will break it down into small chunks and then when you have these small chunks you will create you will convert these chunks into vector
embedding. So what is a vector
embedding. So what is a vector embedding? So quick oneonone on vector
embedding? So quick oneonone on vector embedding that like large language models we have also created algorithms
which are able to create any form of text into a numerical
representation numerical representation as in let's say I have a word called as cat right so I have created a mathematical
representation of cat okay what is a mathematical representation I have assigned some numbers okay so let's say the number is 1a 2a 3 this is let's say
x-axis y ais z ais okay right now to show you because we can only look at 3D I'm showing you three dimensional but in reality it can be n dimensional which is n is a very large number okay so we take
words we take sentences Then we generate vectors from these sentences. Vector is this like this
sentences. Vector is this like this point right? And then what we do is
right? And then what we do is we keep on generating these vectors.
Just a moment.
Yes. So we keep on generating these vectors and the beauty of these vectors is that somehow when you search for cat and when you
search for lion lion okay because these belong to a same family they are going to be very closer
to each other right? When you plot something like a
right? When you plot something like a pat it is going to be much closer but when you go ahead and plot something like a table then it is going to be far
from cat.
So mathematically computer scientist has been able to create this vector thing. So what this vector thing does is it will take a sentence or a piece of information. It
will convert it into a numerical identity which is a vector and then these vectors are created in such a way that if you go ahead and compare these vectors with other words based on the
distance you will understand how much this word is similar to that word or not in the meaning right. So this is vector embedding.
right. So this is vector embedding.
Okay. So what is happening here? We have
had the content and the text. We have
converted them into long sentences of vector embedding which is numerical numbers. Okay. Now what happens?
numbers. Okay. Now what happens?
And then we are going to store these vector embeddings into a vector database.
Right? So now we have the data ready. We
have data ready to be searched. So now
what will happen whenever just a moment where is that?
Yes. So this is rag.
Look at this very carefully everyone. So
there are two parts of rag. First we are creating the knowledge base. What we
have done is we have taken all the content whether it is text, PDF, transcript code or any Bloomberg data.
We have converted that into small small chunks, small information of data and after that we have created vector embeddings and then once we have created vector embeddings understand there are
standard programming code or functions to create these vector embeddings.
Okay, once the vectors embeddings are created they are stored in vector storage. There are multiple databases
storage. There are multiple databases that give you by default vector storing which is pine cone, superbase, AWS S3, vector storage. It gives you all of
vector storage. It gives you all of that. Okay. Now what happens? A user
that. Okay. Now what happens? A user
gives a query. The query is how has Tesla performed in the last quarter?
Okay. Now what will happen?
This query would be converted into an embedding. It
will again be converted into vector. Now
what will happen? We have two vectors.
One is the vectors that are stored.
Other is this vector. So if we want to see what are the vectors which are similar to this vector which is where this information is stored. Right? Where
is that let's say what where is that I would say wardrobe or where is that place where where in my kitchen the sugar might be stored. Okay. So it is
going to go ahead and understand this embedding and then you are going to search for things which are similar to this embedding in this database.
Okay. And you are going to pick maybe 1 2 3 5 10 things based on your algorithm based on your needs which are most similar to this particular vector. So
you will have things which are most similar to this particular vector. Then
what you will do is you will have some information. The information is you will
information. The information is you will go ahead and get this information and then this will be your final prompt. So
now you have taken the vector you have found what are other similar vectors.
Now you have created a list of the vectors which you have found. Now
finally you will get this.
Now finally you will have this. So you
will give this prompt to the large language model. Just a moment.
language model. Just a moment.
Yes, you will give this prompt to the large language model.
Please observe this very carefully.
Answer the question using the context below. This is the context. So, chunk
below. This is the context. So, chunk
one, chunk two, whatever has been retrieved from there. Whatever
information, okay, that information in the text form, it will be written there.
Plus this is the question from the user which is the query which is this how has Tesla performed in the last quarter right and then eventually you are going
to go ahead and get give this prompt to the large language model cloud Germany or anything and then you are going to go ahead and get your answer.
This is rag in a nutshell. So we first create a vector knowledge base then we get a query. We match that query for all the similar terms in the vector
database. Then we get all the
database. Then we get all the information. We create a prompt. Then we
information. We create a prompt. Then we
send it to the LLM and then we get the answer. So now with rag what happens is
answer. So now with rag what happens is you do not have to store all this information into your model or in the context. You can just cherrypick the
context. You can just cherrypick the information that is very relevant to that particular information the prompt and then you can get all the information at once.
Okay, everyone a quick yes or no if you are able to understand this.
Amazing. Great. Super.
Yes, people are not able to understand.
uh if you just go ahead and uh submit the feedback form in the last like in the next 10 minutes I will give you the recording so that you can go ahead and do a revision. Cool.
Yes.
Cool. Uh guys, tokens and embeddings are different. Embedding is actually a
different. Embedding is actually a meaningful representation of that particular sentence. Okay. But token is
particular sentence. Okay. But token is just let's say a vocabulary. Okay. It is
just a numerical representation of particular word. Okay. Now this is rag.
particular word. Okay. Now this is rag.
Now in order to help you understand better I have created a small app that I have white coded very recently. Let me
show it to you.
Yes. So this is a simple rag process visualizer which I created in just 10 minutes using cursor and chat GPD. Okay. And in the program that we offer at LOPM, we are
going to teach you how to do all of this. Okay. So let me show you what
this. Okay. So let me show you what happens in that. So I've told you that in DAG, what happens? You have a piece of information. You first convert it
of information. You first convert it into chunks, right? After that you create embeddings.
After that you convert it into and save it to vector.
And once the vector is created, you are going to write a query, you are going to write a query and then from that query, you are going to match what are the matching chunks, how many
chunks you want to match and then you will go ahead and compose the query as in the prompt is going to be composed and then you will generate the answer.
Okay. So let us see how it works. So I
will take this thing. So I have written like a this log a very long time ago. So
let me go ahead and do this.
So I have taken this piece of information. Let's say till this I can
information. Let's say till this I can take any kind of information.
I will go to this. I will paste this content.
Now the chunk size is 350 which is it is going to take 350 characters in a particular chunk and the overlap of 80 which is the there would be overlap of
80 characters between the first and the second and the third chunk. Okay. So I
will just go ahead and create chunk document.
Now look at this chunks are created.
This is the first chunk. This is the second chunk. This is the third chunk.
second chunk. This is the third chunk.
This is the four chunks. Understand this
chunk was removed here. So it is in this detail 4,000 words guide setting the context how to build then rather than how. Okay. So now there is overlap in
how. Okay. So now there is overlap in this chunk. So 50 characters are over
this chunk. So 50 characters are over overlapped so that all the chunks have a meaningful information. So now we have
meaningful information. So now we have created the chunks.
Okay. So what is chunking? Breaking down
information into smaller parts so that they are meaningful and smaller enough.
Okay. Understand? You cannot use the same chunking method methodology for every application. If you're using rag
every application. If you're using rag for code in order to retrieve code and in order to generate code the chunking has to be done mostly at the functional level. If you are creating for a book
level. If you are creating for a book then maybe it can be done at a chapter level. If you're doing it for a meeting
level. If you're doing it for a meeting it can be done at a maybe uh at a paragraph level. So chunking depends
paragraph level. So chunking depends upon the product that you are creating and tomorrow I'm going to tell you an evaluation that is how to decide your chunking strategy. Okay. So this is
chunking strategy. Okay. So this is chunk. Now I will create embeddings.
chunk. Now I will create embeddings.
I'll create on embeddings and then the embeddings are created.
Okay. So if I look at this these are the embeddings guys.
These are the embeddings. Okay. Okay. So
for every chunk I have created vector embeddings. Okay. And this graph that
embeddings. Okay. And this graph that you are able to see let me zoom it up a bit.
So this graph that you are able to see is a representation of the vector. Now I
will press the next button to save to vector DB and then it will be saved locally. Okay. This is not
saved locally. Okay. This is not connected to any server. So this is going to be saved locally. And after
this, yes. And after this
yes. And after this I will write a query. The query is what is AI product
management.
Okay. And I want to retrieve three chunks. Retrieve relevant chunks. And
chunks. Retrieve relevant chunks. And
then these are the chunks.
So these chunks because there is product management and all it is is being accessed here. So these are the most
accessed here. So these are the most relevant piece of information that I need to store here. So this is chunk number eight, chunk number two, chunk number three. If I click on this, I'll
number three. If I click on this, I'll be able to see the complete chunk. Okay,
this is the cosine which is how distant these things are from each other and from my vector.
Okay. And then eventually what I'll do is I'll just create on compost query.
Now the query is composed.
And this is the query. Answer the
following query based on only the provided data. User query. What is data
provided data. User query. What is data product management? What is I product
product management? What is I product management? And these are
management? And these are these are the chunks three chunks.
Okay. And then when I click on this, this will be sent to a large language model and then I'll get the answer. But
I have not connected the LM because I don't want to pay. And then we have this answer. Okay. But I can always go ahead
answer. Okay. But I can always go ahead and put my open AI key here API key in order to do this. Then the answer would be generated.
Okay. Everyone tell me on a scale of 1 to five if you're able to understand rag now.
Cool.
Wonderful.
People who are four, I would recommend you to watch the recording once more and then you'll be able to understand.
Yes.
Cool. So, this was about rag. Super
important concept everyone needs to understand. It will also improve your
understand. It will also improve your problem solving capabilities. Now, the
next part is prompt engineering. Okay. So guys, I'll
prompt engineering. Okay. So guys, I'll only take 10 more minutes and then I will share the feedback form. So if you have been here in this session for so long, please remain for another 10
minutes so that we can share the feedback form and you can learn the prompt engineering as well. Okay.
Yes.
Cool. So now we'll talk about prompting.
So understand guys, you need to understand how to give instructions to your model.
In order to give the right data, you have understood you have to use rag.
But in order to give instructions, you need to understand that you need to learn prompt engineering. Prompt
engineering, no matter what the world says, it is a super important skill in today's world. Okay? because you are
today's world. Okay? because you are using let's say AI and LLMs for so much of your time in a day. It is worth investing a couple of hours in order to
learn prompt engineering and also if you are working as a product manager as an AI product manager prompt engineering is your way to interact with the LLMs.
Okay. So it is worth investing few hours
Okay. So it is worth investing few hours of your time in order to build great products. Okay. So this is also a guide
products. Okay. So this is also a guide on prompt engineering. So understand
this products such as Fireflaws, Firefly, Gong, Granola which are almost a billion dollar companies or at least multi-million dollar companies. They are
based on these kind of prompts. So you
should know how to learn these prompts and I'm going to go ahead and give you some I'll also share this guide with you. But right now let me share with you
you. But right now let me share with you few tips to improve the prompt.
Okay. So right now everyone tell me out of 1 to5 how would you rate yourself on your prompt engineering skills? How would you rate yourself on a
skills? How would you rate yourself on a scale of 1 to five on your prompt engineering skills?
Three. Don't give yourself three. Three
is a very diplomatic answer.
Yes.
Many modest answers. Yes. any
reflective answers. Okay. So, first of all is first is please take a step back.
So, I'll not give you a very uh exhaustive guide that you're not able to even read.
I'll give you a very simple guide and this guide follows the parto principle.
If you invest what is like if you learn these 20% of the skills as a prompt engineer 80% of your efforts like problems are going to get solved. The
first piece of advice is please take a step back. Okay, understand the value of
step back. Okay, understand the value of prompts.
What happens is the unfortunate reality is that when you go to chat GPT or Gemini when you give it an very awful
prompt, it will still be able to give you an answer most of the time because it has given you an answer. Okay, answer
you go ahead and move forward from there and then you start working on whatever you are working on. Okay. But understand
always remember in many phrases in many places in your life good
can be the enemy of best.
Okay. In many places in your life good can be the enemy of best. which means you have to
of best. which means you have to understand and recognize what is the value of this task and is it worth investing a couple of more minutes or five seven more minutes in order to
write this prompt well. Okay. So the
first advice that I have for you is please take a step back and understand if it is worth investing time in that particular prompt if it is giving you that many results. Okay. The second part
is right role. Whenever you go ahead and start using an NLM please try to give it a role. role means try to say it, try to
a role. role means try to say it, try to tell it that act as a person with 160 plus IQ or act as a seasoned product
manager or act as someone who is a seasoned psychological therapist. Right? So give
psychological therapist. Right? So give
these personas to the LLM so that it is able to perform well. If you want to experiment, you can share two prompts with chat GPT or china or anything. In
one give a very like give it a very solid persona in one don't give it a persona. You'll be able to observe the
persona. You'll be able to observe the answers.
Right? So the first piece of advice is please give it a right. The second piece of advice is please write the output expectation. Output expectation as in
expectation. Output expectation as in you want JSON as an output. You want a table. You want to make sure that these
table. You want to make sure that these mistakes are not repeated. Please write
it explicitly. You have to be very very specific. If you're not specific, it is
specific. If you're not specific, it is going to do anything and then you cannot go ahead and control it. Right? The
third part is multistep. Multistep means
rather than telling your large language model to give you answer at once, ask it to think in a step and give you a step-by-st answer. If you know the step,
step-by-st answer. If you know the step, you can tell it the step. If you do not know the step, go ahead and ask it to tell you the step. For example, rather than asking chat GPD that how would I
improve this product LinkedIn, I would ask it that I want to improve LinkedIn.
Please follow these steps in order to suggest me the improvements. First
understand who are the different users.
Then understand their important problems. Then give me what are the uh what are the top priorized three problems. Then give me the solution of these problems. Then prioritize these problems and prioritize the solution and
then give me a particular solution. Now
when you are focus like forcing chat GPD to work in this manner you are more probably like to get a better answer.
Okay. So please do the thinking. Let it
do the thinking in multiple steps. Okay.
After this give examples. This is also called as few short prompting.
Few short prompting means you are giving it few examples.
You are giving it few examples that this is the format that you should follow or these are the examples that make sure that you are able to go ahead and get the right content. Right? And then we
have constraints. So in example what you
have constraints. So in example what you can do is whenever you are writing any piece of blog or any piece of article that you want to learn from
go ahead copy the best author that you know copy their blog and then give it to Chad Gigi. Okay. For example, what I do
Chad Gigi. Okay. For example, what I do is whenever I write some piece of content because I have put a lot of efforts in writing this particular guide, I copy this guide and I go to
chat GPT and I ask that go ahead. So let
me show it to you.
So this is Chad GPT. I am asking Chad GPT to write a detailed article on AI
on mistakes that people make while learning
AI product management.
Right now we will see the output.
Yes. So you can see that it is writing.
It is doing like a pretty interesting job right?
In the meanwhile, I'll also open a new chat.
Right? So it has written this. You can
see that it is uh well created. This is
point number one. Point number two, point number three, four, this, this, this, this, this. Right? I'm still
writing. Right? This is good one. Right?
Now, what I'll do is I will copy this article.
I will copy this complete article.
Maybe a Ctrl C.
Right. So, okay.
First I'll go to this then I will write. So I've written the sim similar prompt. Now I'm writing
make sure to follow the similar structure voice and helpfulness.
from this example.
Okay. And I can go ahead and either I can go ahead and uh give it the link. It
will be able to crawl the link or I can just go ahead to be assured I can go ahead and give it this content.
Okay.
So now I have given an example. Let us
see how it works.
Yes. Now you look at this. Okay. Rather
than straight away jumping to what you should do, what you should not do, rather than straight away jumping to actionable, it is first setting the context. Why do you need this? This is a
context. Why do you need this? This is a quick context and then it is able to give you multiple mistakes and then how you can
go ahead and avoid these mistakes.
Right? This is a slightly better structure than you could have got in the first way. So this is called as few
first way. So this is called as few short prompting where you give examples to your large language model as in how they should operate.
Right? And
the other part is the similar example which is constraints.
Please give constraints. Constraint as
in be helpful like this. Okay. Or
constraint as in if you want constraint in terms of what do we call it? If you want large language model to not do something then
you should always include them as explicit constraints. Okay. For example,
explicit constraints. Okay. For example,
sometimes large language models assume a lot. Large language models actually
lot. Large language models actually assume a lot.
Large language models assume a lot. So
you can go ahead and tell them that please do not assume only take check the information that is available in this context that I have shared with you.
Similarly, you'll understand more about constant when we talk about iteration.
So understand prompt engineering is called engineering. Prompt engineering
called engineering. Prompt engineering is called engineering because it is iterations which is you tweak your way around. You
give it a prompt. You understand that I got the output but the output does not follow these rules or the output is not good in this particular way. So you are going to understand the output and then
you are going to give it the constraint that boss make sure that you are not taking unnecessary assumptions. Make
sure that you are not using examples which are let's say very much vague or generic. Use specific examples. Okay. So
generic. Use specific examples. Okay. So
when you go ahead and apply all of these steps and when you keep on iterating on top of all of these six steps you are going to learn how to design better
products.
Okay. Now even after I give you all of this I know that most of the people are lazy and they are not going to follow this. So now I'm going to give you a
this. So now I'm going to give you a prompt guide for busy people. Okay. So
four things that everyone should learn and then we are going to wrap up the session. First is please build a prompt
session. First is please build a prompt directory. Whenever you come across good
directory. Whenever you come across good prompts have a notion document or a Google doc where you are storing all of these things. Okay? So that you don't
these things. Okay? So that you don't have to find your way through good prompts every time. Right? Second is
quick shortcuts. Okay? Keep a bookmark where you are able to access this library as soon as possible. So what I do is whenever I want to write something I do not want to find a writing pad.
What I do is I just click on this new tab and then I have all of my notes from notion available here.
Okay, Chrome new tab is the name of the extension and then it is going to load here. Right now I think notion is down
here. Right now I think notion is down so it is not loading but now it has loaded. Okay, so all of my notes and
loaded. Okay, so all of my notes and everything is here. So what I've done is I have set up a Chrome extension where whenever I click on this new tab button it will automatically load my dashboard.
so that I don't have to run here and there. I can just go ahead and click on
there. I can just go ahead and click on this and then I can start writing my notes and my prompt directory is also there. Right?
there. Right?
The other piece of advice is use a prompt manager. Okay? So there are
prompt manager. Okay? So there are multiple prompt managers. You can use any one of them. And you can also use this prompt wild card. Okay. If you are too lazy to write your own prompt, you
can also get this wild card which is analyze the following prompt idea.
Analyze the following prompt idea.
Insert prompt here. Rewrite the prompt for clarity and effectiveness. Identify
potential improvements or additions.
Refine the prompt based on identified improvements. Present the final
improvements. Present the final optimizing prompt. Okay. And also there
optimizing prompt. Okay. And also there is a prompt creator from anthropic. Okay. So there
is a prompt creator called as anropic prompt creator. Okay, where you can go ahead
creator. Okay, where you can go ahead and create prompts. So it would be somewhere or have to login in order to see this. But you should be able to go
see this. But you should be able to go ahead and generate this from cloud console.
Okay. Once you log in, it will give you a prompt uh creator so that you can go ahead and create your own prompts which are much better than what you would have gone ahead and written by yourself.
Okay. So this is about the guide to prompt engineering for busy people. Now in tomorrow's session what
people. Now in tomorrow's session what we'll do is we will talk about notebook LM how does it works. We will talk take the similar kind of case studies. We'll
detail it enough. Then we'll talk about fine-tuning. Then we'll talk about how
fine-tuning. Then we'll talk about how to choose between fine-tuning, prompt engineering and rag. And then we'll talk about AI agents. Then I'll give you how products such as lovable, bolt,
emergent, and visero works. And then we are going to go ahead and talk about MCPS. And then eventually I'll give you
MCPS. And then eventually I'll give you some examples of how to write AI evaluations. Okay, this is going to be
evaluations. Okay, this is going to be tomorrow. But guys understand whatever
tomorrow. But guys understand whatever I'm telling you in these two sessions this is just 10% of what we offer in our bigger program at hello. Okay. So at
hello we have created the world's most detailed AI product management program.
Okay. If you want to understand anything about the program just go to this website called hellopm.co click on this video and then I'll explain you what that content is. Okay.
This is the most detailed, the most practical and the most long-term support AI product management program that you can find. The content that I've told you
can find. The content that I've told you today is just 10% of what we teach in the program just in the first two classes. We will make sure that tomorrow
classes. We will make sure that tomorrow you'll get a much more power pack session. So thank you everyone. Bye and
session. So thank you everyone. Bye and
take care. If you have any queries, you can use the same form to publish these queries. Tomorrow we are going to meet
queries. Tomorrow we are going to meet at the same time. Make sure that you also go ahead and bring your friends.
Thank you everyone. Bye and take care.
You have been a great audience. I hope
you got some value from this session and the two hours that you have invested.
We'll meet everyone tomorrow. Thank you
everyone. Take care. Today is the day two of AI product management masterass and we are going to give you almost everything that you need to get started
in your journey as an AI product manager. This is the foundation. Now you
manager. This is the foundation. Now you
can go ahead and explore the whole canvas in order to go into detail. But
these are the major building blocks that you need to absolutely know if you really want to call yourself as an AI product management enthusiast. Right?
And because AI product management is so new, even if you are coming from a very senior background or you are just aspiring to be a product manager just straight out straight out of college, this content is going to be equally
useful for you. So yesterday what we have done, we have talked about the agenda. I talked to you about what is
agenda. I talked to you about what is the GI revolution. lot of companies are building value on the top of this geni revolution then even in our normal lives
AI has found its way right yes cool I will stop the chat for some time so
that you guys are not getting distracted yes cool so AI has also found the way onto our daily lives and almost every
day we are using chat GPT or Gemini or some of the other AI tools tools in order to make ourselves more productive and grow the impact of our work. Right?
And then I talked about the most fundamental unit why this AI boom has started which is because of a particular technology called large language models.
These large language models, these large language models are algorithms that are trained on massive amount of text and they have learned the patterns of
language really well and they are now acting as the brain of the computers.
And now what is happening? They are
enabling us to understand, generate and reason with natural language by predicting the next token most likely or the token in the context. And I have told you yesterday how all of these things actually work. I have given you
the whole architecture of what an LLM is. The content is like you will go
is. The content is like you will go ahead and take all the content as a training data from the internet. You
will break it down into input and output. Then you are going to tokenize
output. Then you are going to tokenize it. You are going to send the input from
it. You are going to send the input from here. The neural network or the
here. The neural network or the transformer is going to predict the next token. And then whatever is the right
token. And then whatever is the right token from the uh training data, you are going to compare it. Then whatever is the difference, the feedback is going to be sent to the transformer again. And
then it will keep on iterating. When
this cycle happens billions and trillions of time, you will have a very mature model, right? But this model is only a base model. It is trained on some
random data from the internet. So it it has certain kind of biases. It
hallucinates a lot. So what we do is we go ahead and do something called as post training. Questioning means we are going
training. Questioning means we are going to have some human labelers or we are going to have some data analytors who are going to give some specific input
and output maybe in this particular format.
Wait.
Yes. Maybe in this particular format where we are going to insert this kind of data like well articulated inputs and well articulated outputs so that the human
can the uh uh algorithm can go ahead and learn from the same. Okay. This is
called a supervised learning. And one
part of that learning is also called reinforcement learning through human feedback which is humans are telling how what is correct and what is incorrect.
Right? Humans are reinforcing the correct things by giving the rewards to the LLM. Now after that
the LLM. Now after that we talk about the genai value stack. So
there is multiple kind of values. For
example there are companies who are creating infrastructure such as these big Nvidia and Google. Then there are company who are creating the models. You
know about anthropic and open AI. Then
there are companies who are building on top of these large language models. This
is where most of the value unlock is.
And then we have companies and people who are actually using these AI tools in order to make their clients happy, solve their problems and also solve their own
problems. Right? And then I discussed
problems. Right? And then I discussed about type of AI PMs. Everyone who is using AI is AI enabled PM AI tools such as chat GPD and all. Anyone who is using AI tools is a AI enabled PM. And then we
have other type of PM which are more serious PMs. These are PMs who are either working on core technologies such as databases, models, infrastructure and then there are people who are working on the application layer and these are
people who are actually using these LLMs and models and everything in order to build some truly AI products. Agents are
one of the examples. Chat GP is one of the examples. Grammarly is one of the
the examples. Grammarly is one of the example. Lovable is one of the examples.
example. Lovable is one of the examples.
So if you are building these products not building with these product you are building these products then you are an applied AIPM and the purpose of this session is to help you understand the core AIPM part and the applied AIPM
part. Okay, for this part we have
part. Okay, for this part we have already uploaded a lot of content on our YouTube channel that you can see right then
I talked about how this application these meeting assistant such as granola works okay and not only granola but if you take any of the meeting assistant such as fireflies
fed them or anything everyone works in a similar format the format is whenever you want to create a product as a product manager manager understand please do not look
only from the engineering or the AI angle. You are the product manager who
angle. You are the product manager who is responsible to bring the user outcomes and the business outcomes. So
please look at these five things. The
first thing is what is the value that is being generated for the users? Why would
the users use it? Second is what is the business value that is generated. Then
there is how do you make design so intuitive that people are able to use it easily and then how it is feasible with the help of engineering. So all the system architecture is going to come
here and then growth and adoption. In
granular we knew that this product helps us save time by recording our meetings and sending us the summary. How does it helps a business? The business should
invest in granular in creating granular because there is a large opportunity in meeting apps. So it is going to make
meeting apps. So it is going to make money. Similarly, how do we make this
money. Similarly, how do we make this intuitive to the users? By making the design very intuitive. How do we make it intuitive? We make the design intuitive
intuitive? We make the design intuitive by making sure that just after logging and connecting connecting granular fireflies. You don't have to do
fireflies. You don't have to do anything. Whenever a meeting is
anything. Whenever a meeting is scheduled, it is automatically going to join that meeting. It is automatically going to generate the summary. It is
automatically going to send the summary to everyone who was invited to that meeting. Right? With just few clicks,
meeting. Right? With just few clicks, you should be able to do so without your involvement. So that is the
involvement. So that is the intuitiveness. Similarly, how will this
intuitiveness. Similarly, how will this be engineered? So I have told you about
be engineered? So I have told you about this.
So let's say Zoom and Google meet or any kind of platform is the place where you are conducting this meeting. Then Fireflies or granola
this meeting. Then Fireflies or granola is going to either use the API for Zoom or Google or they are going to do the
audio capture through a bot. Okay, like
you can see that there are a lot of bots in this meeting who are actually recording the meeting on the behalf of their administrators. Right? After that
their administrators. Right? After that
that audio is captured and then there are tools called as audio to text transcription machine learning models.
Right? These models such as whisper by open AAI, speech to text by Google, they are going to convert the audio into text
transcription right and then you are going to attach a system prompt. So you
as a product manager of granular is going to add a system prompt and then you are going to add a transcript. Let
me show you that system prompt that I have shown you yesterday.
So this is the prompt that you will paste. The prompt is very simple. You
paste. The prompt is very simple. You
are an expert chief of staff and note taker given the following things. Please
go ahead and summarize the summary of the please go ahead and summarize the given transcript in this particular format.
Right? So now what will happen? We have
this sub we have the system prompt and then the system prompt is going to be combined with the transcript and then it will be sent to large language models
such as Gemini or Google Gemini or open uh open AI's APIs right understand this is happening at the back end you do not know what is happening you are just using granular at the back end granular is doing all of this and then it will
generate the summary actionables and the task and you can take a next step you can make it into an agent by actually sending the tasks on a
calendar on your Jira or on your slack.
So this is how any application which is a note takingaking assistant can work right even if you go ahead and understand a bit of by coding you can also go ahead and create your own
personal AI meeting assistant. This is
the simplest architecture that teams use, right? And after that we talk about
use, right? And after that we talk about a very important part that yes in this example we were able to work with prompt engineering but how do we go ahead and
build applications which are able to work on the large set of data such as data which is on your Jira dashboard, data which is on notion
AI, data which is on stripe.
So what we do is we cannot let's say I have a notion database I want to ask some questions on the top of the database. So I cannot just go ahead and
database. So I cannot just go ahead and put all of these content in the text in the context. Why? Because the LLM is
the context. Why? Because the LLM is going to go out of context and then there are researches that are conducted which have proven that with increasing
context the accuracy of the LLM incre decreases. With increasing context, the
decreases. With increasing context, the accuracy of the LLM decreases. And after
this, we talked about the simple thing that if you have if you are working with LLMs, try to make your inputs concise otherwise they can go ahead and hallucinate too much and they are going
to run out of context and give you irrelevant answers. And then we
irrelevant answers. And then we presented you to con context engineering. In context engineering, we
engineering. In context engineering, we first talk about rag. So what is rag?
Super simple.
What we do is rather than putting all the documents let's say I want to build a functionality where on my notion document or on my Google drive I ask a
question that what is the main what are the tasks that I need to do tomorrow okay and somewhere in one of the PRDs one of the documents it is
written that I have to do these tasks tomorrow okay so in a non ideal scenario what will happen I will try to get all the content of the Google Drive I will put it in the LLM context of the LLM and
then I'll ask the question among all of this content what is something that I need to do tomorrow okay and then what will happen because it is lot of content the LM the LLM is going to run out of
context so what we'll do is rather than going ahead and taking all the data at once to the LLM we are going to only find the content that is relevant right
so what we do in the what we do here We simply go ahead.
We simply go ahead first whatever data that we have we break that data. We
break that data into chunks.
And after chunking the data into small enough and meaningful chunks, we are going to convert it into vector embeddings. And after we have converted
embeddings. And after we have converted into vector embeddings, we are going to store in a vector storage. Once the
vectors are stored, we will go ahead and get the user prompt, find all the relevant content from the vector storage, then convert it into text and then ask this question to the LLM that
based on these information please try to answer this question and then it will give you answer answer the question from the limited context that you have given.
Right? And I have also shown you a visual example with a rag model that we have created like a rag vibe app that we have created.
Right? So this is about rank and then we talked about prompt engineering. Okay, I have shared
prompt engineering. Okay, I have shared with you some very simple but very effective tips which is try to understand the value of prompt engineering. Make sure that you are
engineering. Make sure that you are giving it the right role so that it is more authoritative and you are getting the right answers. Then mention the output expectation very clearly. give it
do something called as chain prompting where you are going to mention ask the LLM to think in steps so that it is able to go ahead and give you what is the reason behind this thinking and then give examples this is called as few
short prompting when you don't give example it is zero short prompting when you give example it is few short prompting right and then mention the constraints understand all of this is called as iterations because once you
give it a prompt you will understand that maybe there is something missing I should have mentioned this clearly and then you are going to iterate on the prompt and then you'll understand what are the constraints that you need to put okay and then I have shared with you
some advices and some wild card prompts so that you can go ahead and get better at prompt engineering and writing better prompts right understand prompts are your doorway to seize this AI
opportunity right it is not just plain English you need to go ahead and think deeper about the prompts so as to get the better answers right you have a very
o uh a very obedient and a very powerful individual at your disposal. If you ask it like uh awful questions, you are going to get awful answers. Now it's
your responsibility to make sure that you're asking it better kind of questions. Right? Yes. So this is what
questions. Right? Yes. So this is what we have done so far. I think we have spent 15 minutes just in the revision.
Now everyone please tell me go ahead in the chat if you have understood everything so far. How many of you are aware about this product called as notebook?
So in case someone has not used I'll just give you a glimpse. Okay. So let's
say I want to learn about AI product management. Okay. This is the topic for
management. Okay. This is the topic for today.
But sadly I do not have time. I do not want to invest time. So what I'll do is I'll go to this platform notebook.google.com.
notebook.google.com.
Okay. I will click on new notebook.
Right? And now what I can do is I can give it lot of documents, lot of resources and then I can ask questions and summarize these documents from this
interface right on chat GBT when you give so many document documents it runs out of context but it is this particular platform notebook is working on rag that I have just told you so it will be able
to answer let me show you how it works okay so let's say I want to understand about a IML product management so I'm going to insert some resources okay so the resources At hello PM we have gone ahead and done
something called as the AI sprint which has everything that you need to understand as an AI product manager. So
I will click on this.
Yeah. So I'll click on my channel and then I'll click on playlist and then I'll click on the AI sprint >> and then >> okay so this is one video. So what I'll
do is I'll copy the URL.
I will go to notebook. I am now collecting the resources. Okay. So my
resource are these URLs. So I'll click on YouTube. I will paste the URL here.
on YouTube. I will paste the URL here.
Insert. So this would be inserted in some time. Second. So now what is
some time. Second. So now what is happening the Google uh uh Google notebook alm is actually taking the transcript of the video it is breaking
the transcript into small small parts and then it is vectorizing all of these transcripts okay all of the contents out there. Second,
there. Second, I will take maybe uh uh the second video.
I will put it again in the YouTube part.
Insert. Now I want a blog. So a IML this. So I've written a blog as well. So
this. So I've written a blog as well. So
I will go ahead and just add it here.
And then I would put a link of website.
I can put multiple URLs as well. Okay.
So now I have all of this context ready.
These are long 1 and a half hour 2 hours videos and this is a very long I would say 3 4,000 words of vlogs. Okay. So now
you understand that yes I have all the context available. I can go ahead and
context available. I can go ahead and ask it any kind of question and then these questions would be answered from the data that I have given. Right? So I
will ask a question how do I what skill does a
AI product manager needs.
Now this answer would be grounded in what is written in these sessions.
Right? So if you want to do any research of the world you can just catch up some YouTube videos, some interesting podcasts from the founders and industry leaders. put up some blogs, put up some
leaders. put up some blogs, put up some reports in PDFs from Mackenzie and other companies and then you should be able to do this. Okay, now it has given me the
do this. Okay, now it has given me the answer. But this is not the beauty. This
answer. But this is not the beauty. This
is okay. This is cool. But it also has some wonderful features which is it can help me create mind map. Right? So if I click on this, it will create a beautiful mind map from all the
concepts. Okay.
concepts. Okay.
Another important part is reports. So if
I click on the reports, these are all the report that it will create. It will create a complete blog
create. It will create a complete blog post which I can read. It will create a study guide so that it will be having a short answer, quizzes, some uh revision material, everything. Then a briefing
material, everything. Then a briefing doc so that I can go ahead and just understand what is important in these particular three four resources. And
then there are multiple things. Okay,
concept explainer, process overview and everything. So it is now able to
everything. So it is now able to understand the content and based on content it is suggesting me that this kind of reports are possible from this
and I can also give my own report right.
So let's say I want to get a briefing.
So now it will take some time in order to generate this particular report.
Okay. And then I can also go ahead and create mindm flashcards quiz and everything. And also eventually I can
everything. And also eventually I can also create an audio overview. Okay. So
this is what this product does. This is
almost a number one research tool based on AI which almost every product manager and every individual should use. Okay.
So this is notebook LM for you. There
are so many use cases. You just think about it and notebook LM should be able to do these things. Right. So my
recommendation to everyone would be that after this session go ahead and try to play around with notebook. Super helpful
tool. Right? Yes. Everyone a quick yes or no if you're able to understand this tool.
Amazing. Thank you. Yes. So now
let us try to understand and also what I'm telling you is not just how notebook LM works but also any of the tools any of the AI tools which are
based on documents. So in notebook we are giving it a lot of documents and transcripts and it is working on the same. On notion AI we are actually
same. On notion AI we are actually giving it all of our databases the notion document that we have created it is working on the same in Salesforce Einstein we have all the document that we have created in our salesforce all
the sales record all the CRM all the emails all the customer conversations and then there is an LLM on top of it.
So the mental model is that if you have a lot of data and if you want to put a LLM on top of it in order to make it more queryable with the help of a natural language then this is what you should
do. This is the same architecture for
do. This is the same architecture for all of these products. Okay. Yeah. So
what we do is we follow the same format. Question
number one is what problem does it solve in people lives? This is something that you need
lives? This is something that you need to answer as a product manager before even building the product. So why does notebook help why is notebook helping helpful? Because it helps people conduct
helpful? Because it helps people conduct research by helping them read so many documents and summarize these documents at once.
It helps people read and learn about exhaustive full topics within a lot less period of time.
It also helps people create some effective nodes for their learning, for their research, for building any product or for doing anything. Okay. So, there
is definitely a value in the product, right? So, it will help people become
right? So, it will help people become more productive and take better decisions in shorter period of time, right? What is the business value? Now
right? What is the business value? Now
Google has understood that people are using chat GPD a lot and Google has understood that now if people keep on using chat GP and perplexity the research that people used to do on
Google is now going to get reduced. So
they have to create a product that will make sure that they are getting their lost foothold. And now notebook alum has
lost foothold. And now notebook alum has become popular that people are have started using it. So it is a strategic bet for Google to create this kind of product. Right. Third is how do we make
product. Right. Third is how do we make it intuitive to use? So Google has made it very simple. So you can see the UI.
It is very simple. It tells you. So
Google might have thought that why not we go ahead and allow people to just enter content. Okay. But content could
enter content. Okay. But content could also be in the form of YouTube. So they
explicitly mentioned that yes people want to learn from YouTube so you can enter YouTube. People want to upload
enter YouTube. People want to upload their PDFs. So they are allowing you to
their PDFs. So they are allowing you to upload the PDFs. So they understood that from what kind of resources people would want to ideally get the information and they have enlisted all these resources.
Similarly they were able to understand that most of the people are very lazy in terms of prompt engineering. So they are giving you some automatic templates of reports. So rather than you thinking
reports. So rather than you thinking that what is the question that I asked from this document they will give you automated question in the form of report that I have just shown you.
Right? So this is how you make the product more intuitive. Right? And if
you guys want to learn how to make the product more intuitive, a simple exercise is go ahead look at 10 top AI products and then reverse engineer what are the things that they are doing in
order to make it more intuitive. You can
follow the similar kind of principles.
You will be ahead of 80% people who do not know anything about UI, right? And
then we have how will this be engineered, right? How will this be
engineered, right? How will this be engineered? This is the structure.
engineered? This is the structure.
Please look at this very very carefully.
Okay.
Yes, we have 510 people in the chat.
Everyone please write focus if you are ready to go through this. Please write
focus if you are ready to go through this.
Amazing. Great. Cool. So focus here everyone. So here what is happening is
everyone. So here what is happening is first is the resources.
Resources are in the format of articles, videos, PDFs and text which is videos it will not be able to understand. So what
it does is it takes a YouTube videos and YouTube videos comes with a transcript.
So it is understanding the transcript and not the video. Okay. Second is PDFs to convert PDF into text.
We can either if the PDF is created digitally, we can go ahead and convert it into text like through some PDF extractors. However, if it is an image
extractors. However, if it is an image that we have taken photos of and then we have converted the PDF, then we have something called as OCR or optical image recognition, character recognition,
right? So these things would be creative
right? So these things would be creative to understand it is not just that it building an AI product is only about implementing the LLM. You have to do this dirty work as well which is making
sure that all the content is in the right format which is text right after this after this
and even if let's say we have audio right now I don't think it allows audio but if it allows audio what it will do is it will get the audio from you it will give it to the back end like from
the back end it will give it to Google speechtoext API or whisper API and then it is going to get the transcript and then the transcript is going to be added here. Okay. So as soon you add the
here. Okay. So as soon you add the sources the sources are extracted they are chuned and then they are chuned at two levels document
level and the paragraph level. Right? So
it tells you that from which document and which part of document is this content coming from. So whenever you are storing these chunks there would be a link to that document whole document like the source and there will also be a
link to that particular part of the document which is the chunk right I have shown you this example yesterday with that thing right after this
we have embedding so now this will be stored in Google's own database for embedding and understand whenever you change the content whenever you update the content this is going to get refreshed
right till Here we have not used the genai so far. We are only using engineering computer science engineering so far and some machine learning
algorithms. Right now what happens your platform notebook is going to understand what this content
is all about. So it will try to go ahead and create highle summaries of all the documents that you have created like highle summary for every document automatically and then it will try to understand what are the different kind
of reports that I can generate and for generating these reports it would have some kind of templates already. So let
me show you so this is what Google might have created at the back end. Okay. So let's
say if we think that there there could be a summary that we can create. So it will go ahead and give this kind of prompt which is summarize the key ideas in the
plain English focus on main findings and implications. if they need to create a
implications. if they need to create a study guide. So here what is happening
study guide. So here what is happening is we have these report prompts right it will show you that these are the things that you can do with this report and
then the user input whatever the user has inputed the prompt and the retrieved context. Retrieve context means whatever
context. Retrieve context means whatever is matching with that prompt. Whatever
the words that you have mentioned in that prompt, whatever is matching in the vector database that is going to be fetched and then that is be that will be sent to the LLM. If it is Google, the
LLM is most probably going to be Gemini.
So this is the prompt library. Okay. So
if the summary type the report type is summary, then this is the prompt. If it
is a study guide, then generate a structured outline with key component, definitions, and questions for self- testing.
Right? Mind map. Organize the main ideas hierarchically with branches for non for subconcepts and examples.
Right?
And then reports would be generated and then reports could be summary, mind web, guides, FAQs and then what will happen?
You as a user will give you the feed will give the feedback that it is good or it is bad and then it is going to go ahead and autocorrect these prompts and you can also go ahead and put your own
prompts. So what is happening this is
prompts. So what is happening this is augmenting this part is retrieval and this part is generation.
Right? So we are first creating the documents then we are generating then we are uh going ahead and giving it to the LLM and then we are generating and going giving the output to the user right so
this is how notebook LM works in a nutshell it is a pure play rag application right everyone a quick yes or no if you are able to understand
H now one question to everyone question to everyone okay let's say if I ask the question to notebook alm
that summarize all the contents summarize everything that you are like from all the resources summarize Now what do you think will happen? Think
about it for a moment and tell me in the chat what will happen then? Let's say if I give it a query, give me a summary of this content. What will happen
this content. What will happen now? So summarization. Yes.
now? So summarization. Yes.
Cool. So understand summarization is a non-typical task. Why? Because
non-typical task. Why? Because
summarization is the most I would say popular task but it is a non-trivial task. Because in the prompt what would be mentioned?
Summarize this document.
When this is mentioned in the prompt when it will search in the rag model when it will search in the embedding what will it search for? Summarize
article. And now how will it find that where is summarize and all because it is doing a vector like search. So it might find some chunks where summarize brief executive summary these kind of things
are written but summary would not be accurate.
Right? So what is the solution? The
solution is we do not implement this normal rag pipeline in these kind of applications. What we do is
applications. What we do is we have something called as hierarchal hierarchal rag.
I'm not sure what the spelling is. So
hierarchal rag what it is is that I have source one, I have source two, I have source three.
Okay. So now first of all understand the problem clearly. Okay, I have created
problem clearly. Okay, I have created chunks. Let's say there are 10 chunks
chunks. Let's say there are 10 chunks that are from this document. There are
10 chunks from this document and four chunks from this document. If I ask LLM the question that please summarize this, then what will happen?
It will make the vector of summarize, right? And summarize vector might be
right? And summarize vector might be equal to brief.
Okay.
or maybe some other things like synonyms of summarize right and it will go ahead and try to find summarize brief and some things in these chunks and whatever whatever are the chunks that are relevant it will go ahead and give these
chunks let's say one chunk from here one chunk from here and one chunk from here so we have got three chunks right but understand these chunks are incomplete they have mentioned summary
or brief or in short something but they do not contain the true summary they just contain these words right. So what we will do here is we
right. So what we will do here is we maintain a summary we maintain a summary at a high level.
So for all of these documents as soon as you upload this document in notebook LM it is first going to generate the summary of the whole document. So it
will send it to the LLM. It will
generate the complete summary so that it understands what the report is. Then for the other it is going to
is. Then for the other it is going to generate the summary. For other it is going to generate the summary without you saying anything. And then whenever you send a query of summary it will understand what the equation is. And
then it is going to answer by concatenating or copying or taking all of these summaries into the context.
That way context is also not increased but there would be some more LLM calls.
Right? So how we have solved this problem? We have taken all the sources
problem? We have taken all the sources automatically without even the user saying so we have taken the summaries of every individual source.
Right? And once we got the summary from the LLM now whenever a user is asking question about summary we are going to combine all of these summaries in order to get a better answer.
Right?
Yes, everyone a quick yes or no if you're able to understand this and like this guys although rag is simple okay rag is not a
rocket science but you have to understand the nuances every product is unique so you have to understand these kind of nuances before going ahead and
calling yourself an AI product manager okay so this is about how not works and all the rag applications works in the similar format you have the knowledge base, you convert the knowledge base
into the vector database. After that,
you go ahead and get the user query, you match the pattern, you get the uh uh top similar chunks and then you put the LLM on top of it along with the system
block. Right? Yes.
block. Right? Yes.
Now in the context engineering we have so far talked about rag we have talked about uh prompt engineering and the third part of context engineering is actually
fine-tuning. So sometimes what happens
fine-tuning. So sometimes what happens is sometimes what happens is you have to retrain the model you have to change the parameters in the
model for your use case. For example,
Google has this amazing model called as alpha. Okay, Google alpha and
alpha. Okay, Google alpha and Google alpha fold or something is the name alpha fold I think. Okay, so what it does is it is able to identify the
patterns in protein and then it is able to suggest that what is going to be let's say the consequences which is the diseases which are identified from protein patterns and so many things.
Okay, but understand when Chad GPT or Gemini was trained, it does not have any information about proteins and all. It
has the information about how normal natural processing works, but it does not have any understanding about how to understand the protein models. Okay. So
now what happens is we are going to take some new data which only contains information about these protein folds and then we are going to
go ahead and retrain our large language model again. So that now we have
model again. So that now we have contextual information about this. So if
there is some information that is completely lacking that cannot be compensated with rag then we use fine-tuning plus sometimes when you have
to change the personality of the model for example this is a real use case that one of the bigger banks they used chat GPT in order to answer the customer
queries right and then they were able to find that chat GP is able to answer well but it does not have that personality of how their people were actually going ahead and answering. So what they did was they actually taken all of their
past records of the chat which humans have done they have fine-tuned GPT and other models on the top of the same and then they were able to get a new model
which is now able to act nicely right similarly Bloomberg has created a product called as Bloomberg GPD where they have trained complete Bloomberg on
top of uh the a GPT on top of Bloomberg data so that if it has that kind of context Right? It kind that kind of information
Right? It kind that kind of information and knowledge. So finetuning is adopting
and knowledge. So finetuning is adopting a pre-trained large language model which is a chat GP or any base model with a domain specific data to make it more useful for a particular task. Right? So
what we do is we take a base model llama GT sonnet. Then we do the fine tuning
GT sonnet. Then we do the fine tuning with curated data and examples and then we have a fine-tuned error. Okay. It is
one line of code which you have to use in the openi terminal. OpenAI API
fine-tune create t data.json JSON
whatever the data that you have then this is the model that you want to train one line of code can go ahead and do the finetuning but understand finetuning is the costliest of all the three method
that we have talked about finetuning takes a lot of money a lot of resources so unless you are not sure you should not start with finetuning first do prompt engineering then try to
understand if the problem can be solved with rag if it is lack of information problem if it is not solved then you should go ahead and go to finetuning Right. And
Right. And generally finetuning is of two types.
One is full finetuning. Second is
parameter efficient finetuning or PFP.
Okay. So in full finetuning what happens is whatever are the billion parameters that we have in the large language models we are going to try to update
them all. So it is very intensive right.
them all. So it is very intensive right.
So this is going to give you great results but it is also going to be very very intensive on the CPU and GPU and it is going to cost you a lot of money.
Right? Similarly we have parameter efficient finetuning. Okay. Where what
efficient finetuning. Okay. Where what
we do is rather than so I have talked about parameters in the last class. In
parameter efficient finetuning rather than going ahead and uh uh optimizing and changing every parameter we only change few parameters to make
sure that middle model is able to get that behavior without spending a lot of money on training right so it updates the subset of parameters it has lost cost and store
right and the f the training can be done faster the examples are adopters and lora Right. So this is called as fine tuning
Right. So this is called as fine tuning in LMS. Now a major question is when should I choose what to do whether I should do prompt engineering or rag or
fine-tuning. How should I choose? So
fine-tuning. How should I choose? So
this is a simple decision tree for you.
Okay. If you want to decide the right contextualization method, this thing is called contextualization or context engineering. If the base LLM is good
engineering. If the base LLM is good enough with careful prompting, then go with prompting. Experiment. Try with
with prompting. Experiment. Try with
prompt first. Before going to advanced method, try with prompt first. If it is giving you the answer, go ahead and do this. There are multiple products that
this. There are multiple products that are based on the same. For example,
granola and chat PD, they are only working on prompt engineering.
Right? Second is if it is not working, prompt is not enough. Then if the problem is only about missing or dynamic knowledge then go ahead use rag is try
notion AI notebook lm they are using rag right and the third is if the issue is about style tone or a domain specific behavior that you'll understand by
asking questions to the LLM then you should go with this and examples are dual indo max and bloomberg GP that I have already given you examples of right so guys this was context engineering.
Everyone tell me yes or no in the chat if you are able to understand everything so far.
We talked about prompt engineering, we talked about rag, we also talked about fine-tuning, right? And the question is the question
right? And the question is the question is that what you should use what you should use is
usually decided by your context in your problem. Okay? Don't think that if some
problem. Okay? Don't think that if some company was able to use something and get the results, you will also get it.
Please try to understand the user problems, the context, the limitations, the trade-offs and then choose whether you should go with prompt engineering, rag or fine-tuning. You should always
start with prompt engineering. If
there's a issue with data, go with rack and then if the issue is with personality or the base data, then go with fine-tuning. Right? There is also
with fine-tuning. Right? There is also an interesting thing that is coming across these days. I'll just give you a hint. We talk about this in detail in
hint. We talk about this in detail in our bigger program which is an amazing thing which is called as transfer learning. Okay, this is an
transfer learning. Okay, this is an upcoming very I would say very promising field and many companies are going to adopt this. Just a moment.
adopt this. Just a moment.
Yes. So there is something called as transfer learning.
So what we do is the issue with large language models is that more intelligent the model okay more it is going to cost
and more it is going to have latency which is it is going to take more time to respond. So if you go to the OpenAI
to respond. So if you go to the OpenAI API, okay, OpenAI API cost, okay, so if I go to pricing,
understand that this GPT 5 is their best model and the cost of GPT5 is for output $10 per 1 million tokens, right? And
then they have a small model which is GP nano which is $0.41 $4 for 1 million tokens
almost 25x difference. Right? So models
which are small models have less number of parameters are more efficient and cost effective but they do not have the right kind of they are not trained on more data and they do not have more parameter that means their output could
be a bit here and there. Okay. So we
have something called a transfer learning. What we do is we pick a small
learning. What we do is we pick a small language model such as Gemma or a combination of or something between the small and the large language model
such as nano which is GP nano 5 nano or 4 nano whatever you want to choose. Okay, these
things have less cost. What you will do is you will go ahead generate some synthetic data from the top models which
is GPT5 or Gemini 2.5 flash okay you are going to use Gemini in order to generate some synthetic data and then you are going to generate that data and then you are going to fine-tune these models on
that data so that these these models although they will not get as intelligent as GPT5 or or or Gemini 2.5 but still if you give on specific use cases. For example, if you are using for
cases. For example, if you are using for customer support, if you are using for coding, if you are using for coding in a particular language, if you are using for a particular use case, then you can generate enough data and then you can
retrof these models with fine-tuning and then there is going to be one-time cost of finetuning. But then eventually your
of finetuning. But then eventually your overall cost of API and inference will reduce. This is called as transfer
reduce. This is called as transfer learning. So you have a small language
learning. So you have a small language model which is efficient which has less cost but it does not know about certain knowledge. So you will generate that
knowledge. So you will generate that knowledge artificially from bigger models put that knowledge into fine-tuning data for these small language models then the small language models will become more powerful for
your context.
Right? So imagine these large language models as swords.
But sometimes for some purposes you don't need a sword. You just need a needle and then transfer learning gives the power of sword to this particular needle.
Right? A quick yes on everyone if you're able to understand the power of transfer learning.
You can explore more about this. We talk
about this in more detail in our bigger program. super helpful skill for any
program. super helpful skill for any product manager because implementing LM and understanding is is is is easy but getting the most out of from this is
super important.
Yes. Cool. Now with that being said, let us go ahead and talk about the interesting piece in the room. Okay. Which is AI agents.
the room. Okay. Which is AI agents.
So far the AI agent was like your most interesting, your most uh
intelligent manager. They are telling
intelligent manager. They are telling you that this is what you should do, this is what you should do, this is what you should do. But they are not taking action. And nothing is more frustrating
action. And nothing is more frustrating than a handsoff manager which has they don't know how to do the work but they always try to give you knowledge and gam okay and LM are the same. You give them
the input, they give you the output and they appear the most intelligent part.
But they cannot act. Act means
you do not want the LLM just to write an email. You want them to send the email
email. You want them to send the email and take the follow-ups. You want them to do your work rather than just being your uh uh uh your brainstorming buddy.
So in order to give agency to the agent in to the AI, we have something called as AI agents. So AI agent is nothing but
a very simple concept. Okay, which is we give access to some actions to our
agents. We get our AI to act. So what we
agents. We get our AI to act. So what we do is a simple example of agent is let's say I post very frequently on LinkedIn.
Okay. And if someone wants to go ahead and be updated with what is happening in the product management world, what they can just do is go to my profile, look at
what I have recently posted and then you should be able to understand what is happening in the AI product management world. Okay, but this thing you have to
world. Okay, but this thing you have to do manually. What if there would be a
do manually. What if there would be a large language model? What if there would be a software or an agent that can go to my profile every day in the
morning look at the five recent post that I have done let's say not every day but let's say in a way in a week in that week whatever I have posted they should collect all that post and then they
should give all of this post to a large language model which can summarize all the important content and then it can send me an email. This is an example of a large language model. This is an
example of an AI agent. Okay, which is I am choosing a time where I am going to crawl a profile, LinkedIn profile. I'm
going to get all of their post. Then I'm
going to summarize this content and then I'm going to get them to send an email.
The summarize part is the intelligent part that is done by AI. That is why it is called as an AI agent.
Right now I can go ahead and tweak this AI agent more. What I can do is I can allow tell the AI agent that rather than getting the post from someone, you only go ahead and send me the
message rather rather than crawling all the post, you should only go ahead and crawl the post where something is written about AI product management. So
now the AI agent will go ahead and take some calls based on or take some decisions based on the content that it finds. So when the AI agent is taking
finds. So when the AI agent is taking some decisions based on its own context then it is called an autonomous AI agent when it is just going ahead and following the instruction it is a
workflow agent right a quick yes everyone if you're able to understand the example of AI agent cool and we'll actually create that
agent today okay so intelligence autonomy and uh action Right? And there
are multiple platform guys understand building AI agents these days is not rocket science. It is commoditized
rocket science. It is commoditized information. How to build great agents
information. How to build great agents is something that is something that you need to learn. Okay. So if you go to platforms such as agents.ai, ZPR, NAN and OpenAI has very recently created
their agent kit. You can go ahead and create multiple agents. Okay. However,
there has been a report recently. There
has been a report recently and this report got way too popular and it has mentioned that 95% of most of the AI initiative most of them are AI agents
have failed and the reason being there was a learning gap. Learning gap means these project did not fail because these people were failing to engineer agents.
They were able to create agents but the problem was these agents were actually not solving any good enough problem and it was not learning from the
feedback right so anytime you want to go ahead and build any kind of product please remember this if you are creating AI agent if you are creating a normal
traditional product or even if you are creating an offline product please make sure that you start from these two If you read this report, I'll I'll attach this report in the resources
section when we are sending this to you, you will understand why not having the right product sense was a major problem why these things fail rather than just not knowing about the AI. So
understanding AI is easy but using AI at a scale is a different beast altogether.
Right? Yes. So this is simple about AI agents. I'm going to help you understand
agents. I'm going to help you understand how AI agents are built. So,
a quick yes or no everyone if you have ever built an agent.
A quick yes or no if you have ever built an agent.
Cool. A lot of nos.
Cool. So, I I'll Okay. So, understand is also a fantastic flow, but I'll show you something even simpler. Okay. So you can go to this app called relay.app.
This is free with enough credit. So I'm
using this. Okay. And it is very simple.
Okay. So I'll click on login.
I might be logged on already. Okay. So
I'm already logged in. Okay. So what
I'll do is I'll create a simple flow.
Okay. So I'll click on new workflow.
Now every agent starts from a understand guys. uh this agent I am telling you
guys. uh this agent I am telling you these agents that you can build the production grade agents are different the products that I'm going to give you examples in just a moment they are also
agent but right now we are using like a agent that you can also build on your own okay this is for your own productivity okay so the agent is
yes so the agent is will do the simple thing at 8:00 a.m. In the morning it is going to crawl my profile or any other profile that you want to. It will take
all the posts and then it will summarize these post and then it will send me an email. That's it. Okay. Summarization
email. That's it. Okay. Summarization
will happen with an AI. So add a trigger. Trigger means when should this
trigger. Trigger means when should this event happen? When should this event
event happen? When should this event happen? Okay. So I will just go ahead
happen? Okay. So I will just go ahead and click on schedule trigger. It should happen
schedule trigger. It should happen tomorrow or at today's 26th today at 8 a.m. So this time has already passed. I
a.m. So this time has already passed. I
can make it tomorrow.
Okay, it can run at this time and it should run daily. I can also make it any other time. Okay, this is how the event
other time. Okay, this is how the event is triggered. Okay, and then I'll click
is triggered. Okay, and then I'll click on done.
To make it simple, I have chosen this frequency. But you can also execute your
frequency. But you can also execute your agent on any other event. For example,
when you send a message to Slack, when you send and get an email from someone, when you go ahead and maybe send a message on Telegram, when you go ahead and click a button, when you go ahead
and get a new lead from your CRM, you can go ahead and add any kind of triggers. Okay, this is just to make it
triggers. Okay, this is just to make it simple. Keep it simple, right? Then
simple. Keep it simple, right? Then
we'll add a step. The other step is I want to add LinkedIn.
So on LinkedIn I will just click on add this tab. I will search for LinkedIn and
this tab. I will search for LinkedIn and then get LinkedIn comments person post. I want some people
some person's post. I'll click on this and then I'll select a URL. So what is my LinkedIn URL? I'll go to this LinkedIn slashin
/ ankit.
So this is my LinkedIn profile.
I'll just click on this.
I will enter manually. I will enter this
enter manually. I will enter this and then I want to get the top 10 posts and then if posts are not returned then
continue without a result and then done.
Second step is done. The third step is I have to summarize.
So I will summarize and then what will I summarize? I will summarize posts. Next
summarize? I will summarize posts. Next
it will automatically tell me what you want to summarize based on what I have sent from the last step. Okay. And then
this is the prompt.
Okay. I can do it better. I can go to chat GPT and I can ask chat GPT that write me an
effective prompt to summarize my LinkedIn posts.
Right. So I will go ahead and copy this.
Right. I'll go ahead and copy this.
Okay. But to keep it simple, I'm not using this. But you can just go ahead
using this. But you can just go ahead and copy this and paste it there. Okay.
I'll go ahead with the simple one which is summarize the attached post. Use
fewer than 100 words. Okay. Then access
to this this this this is this this.
Okay. Plain text. Okay.
Done.
Right. And then we have plus button which is I want to send myself an email whenever this happens.
Email send email to yourself. What is
the subject?
Summary from Ankit's post add body. If I write X I should if I rect the rate I should be able to get the sum AI output
summary is above okay and then I'll go ahead and click on done right and now okay so everyone so far
write yes or no in the chat if you're able to understand the flow so far it is super simple Right.
Correct. And now what we'll do is we will just go ahead and try to test this start now. I can
test this, right? I don't have to wait for morning tomorrow because you guys would not be there. So I will test this summarize with AI and then an email to yours. So I'll
click on this. This is the email. The
post primarily promotes a free multi session AIPM master class which drew over 800 live attendees. The master
class focuses on mastering LLM. This
this this and this. I hope you guys remember this right now guys tell me confident in the chat if you are confident in creating your own agents now
tell me confident in the chat. Yes. So
understand this tooling was never difficult.
This tooling was never difficult. What
one thing that AI has done is it has gone ahead and made things simpler for everyone. It has gone ahead and made
everyone. It has gone ahead and made things super simpler for everyone. Okay.
As a person, you need to understand that now you have so much power and you have the power to exercise this power. You
have the ability to exercise this power.
Okay. Now you should focus on the right use cases. Okay. So what I'll also do is
use cases. Okay. So what I'll also do is I will share with you a list. So Google
has created an amazing guide.
Google has created an amazing guide which talks about what are the different kind of use cases in different kind of industries for AI agents. I will share this guide
also with you in the slides so that you are able to understand and then create agents there. First whenever you create
agents there. First whenever you create any product understand the users and the business. An AI product manager is a
business. An AI product manager is a product manager before they are an AI product manager. Okay tools are easy. I
product manager. Okay tools are easy. I
have already taught you the tools. they
are not so difficult. Important thing is understanding to solve the right problems. Okay. And then when you create an AI
Okay. And then when you create an AI agent, these are the things that you need. So this is a general stack that
need. So this is a general stack that you guys can go ahead and take a screenshot of. In an agent, you need
screenshot of. In an agent, you need reasoning so that you can summarize, you can take decisions, you can call certain kind of tools, right? And generally LLMs are used. LM's large language models are
are used. LM's large language models are GPT, Claude, Llama and others. And small
language models are F3 and JMA.
Right? Similarly,
you need memory. For example, if you in chat GPT, it remembers everything that you have told. But when APIs are called in in agents, they might forget about your older conversation. So, you need to
save your memory. Okay. So, understand
here what has happened in this model.
What is memory? Let me tell you. So, in
this model, in this flow, it found 10 post then it gathered the data. It sent the data to the other step
data. It sent the data to the other step where the AI summarized it and then it was able to get that memory that data that summarization to the next step.
This is called as memory. Making sure
that LLM and or the whole uh model the whole agent is able to remember what you just said. Right? This is called as
just said. Right? This is called as memory. And for memory we have multiple
memory. And for memory we have multiple tools.
For memory we have vector databases such as pine cone, chroma, rag and feedback loops. Right. And then there are tools.
loops. Right. And then there are tools.
Tools can be accessed with the help of databases, visual agents and APIs. APIs
are available. There is something also called as visual agents. Visual agents
is it will if it does not have an API, it will act like a computer. It will
browse, it will click buttons in order to get the data for you. Okay. And then
so open world operator is example of a visual agent.
Right. Then we have orchestrators platform on which you can create these kind of tools these kind of agents. So
langraph crew AI NA10 is these platforms. I have given you a new example today which is array.app and
then we have guard rails. So now
understand AI agents are like fire. You
can use them to cook your own food. You
can use them to show away danger animals. And you can also use it to burn
animals. And you can also use it to burn your own hands and burn your own home.
So guardrails are super important.
Guardrails make sure that your content is relevant, your agents are safe, you have you are not abusing the tools and then you are also validating the output
before presenting it to users. Right? In
our major big program we tell you about in detail of all of these things so that you are able to build some secure agents, production grade agents,
right? Yes. And uh these are some
right? Yes. And uh these are some examples of agents which are useful.
Customer support agents which will read tickets, drafts, replies and then update the CRM meeting productivity such as fireflies, fireflies and granular that we have discussed yesterday and bizops
business ops which is sepia or enterain agents that monitor incoming data and execute automations. Right? So this is
execute automations. Right? So this is in a nutshell what agents are. Okay. Now
in the next part of the class we are going to talk about how platforms such as V coding platforms such as Lovable Berchant on Vero works. Okay. But right
now let's take a quick break. Okay. So
before that everyone tell me on a scale of 1 to five how good are you able to understand everything so far?
Great. Cool. Amazing. Cool. So now
everyone a quick break of 4 minutes and then we are going to come back.
Quick break of 4 minutes and then come back.
I think we are right on time. How much
is the time?
Oh 911. Oh 911 is the time. So
great behind lovable.
Cool.
I'm sure many of you might have gone ahead and tried your hands or at least heard about these apps like these very high growth apps called as lovable,
bolt, emergent, fezero. What they do is you give them a prompt then they work in the back end in order to create a full-fledged application for you. Right?
What these app do at the back end is something that I'm going to tell you today. Okay, I'll tell you at a high
today. Okay, I'll tell you at a high level how these things are supposed to work.
But before that we should follow the follow the rules. Okay, the first thing is what is the user value? The user
value here is that as an entrepreneur, as a product manager, as an engineer, I either do not know how to code or I do
not have enough time to spend on a unproven idea. So what I do is I give it
unproven idea. So what I do is I give it a prompt. I give I want a tool that can
a prompt. I give I want a tool that can help me build a full-fledged app that can help me test my idea. Before these
kind of apps, before AI, what used to happen? I will hire a developer. I'll
happen? I will hire a developer. I'll
build something and then it will take me a lot of time. Then it will take me a lot of money and then maybe I'll know that this thing is not working. Right?
Why? Because you only know if a product is working when you go ahead and launch it. So it was very difficult to build
it. So it was very difficult to build full-fledged apps without the involvement of a developer. These low
code tools that used to happen before that like web flow and all they were not very effective for any non-engineer.
They still used to take a lot of time.
So now with the possibility of AI we have some tools with which we can just go ahead and give it a prompt. It will
be able to go ahead and generate a full website or an app. Right? So value
unlock is definitely there. Okay. But
you should define what is the market that you are playing in. Is it engineers or hobbyist or entrepreneurs or non-coders that will define the UI and the UX right and the strategy. Why
should the business pursue this opportunity? There are so many people
opportunity? There are so many people who want to build their own products, right? And traditionally these people
right? And traditionally these people have not do not have the power to do this. Why? Because technology could not
this. Why? Because technology could not support it. Now what has happened right
support it. Now what has happened right now is that there is a technology called as Gen AI through which this is possible now. Okay. So now there is a huge
now. Okay. So now there is a huge business opportunity out there. You can
do the sizing of the market in terms of tam but there is a big market opportunity right and many product managers many entrepreneurs they want to build quick prototypes before they can
go ahead and full-fledged take the business the engineering resources there is a business value the third part is design how do we make it intuitive to use so if you look at the platform such
as lovable you'll be able to find multiple things okay first is
these These platforms give you a very good interface where you can just write a prompt like a chat plus they will let you to add
any screenshots. So most of the people
any screenshots. So most of the people they are inspired. So they will go to a website look at the website it is looking good they will take a screenshot and they want to make something like
that right and then they are also going to give you examples of how other people have done this. So you are getting some inspiration. Okay. Plus they also make
inspiration. Okay. Plus they also make sure that there is a community which can go ahead and help you. So you
can go ahead and get added to this community and then you can go ahead and learn about how other people are using Lovable. So it makes sure that you are
Lovable. So it makes sure that you are able to get on boarded on the platform very very easily.
Right? And then they also have some learning tutorials and all that makes sure that it is easier for you to use.
And they have also made sure that they have a free pricing plan.
Yes, I already have a paid plan. So they
do have a premium plan through which they can you can go ahead and try the platform before you go ahead and start spending your money. Right? So this
makes sure that the platform is usable.
The ultimate usability is coming from the simple fact that an LLM can allow you to generate like convert your simple prompt into a full-fledged website or a
web app. Okay. So this is the usability
web app. Okay. So this is the usability and then how does it work?
Okay, how does it work? So this is the simple architecture of these kind of apps. Okay, the user will give the
apps. Okay, the user will give the input. Okay, the user will give a input
input. Okay, the user will give a input that please create a clone for Airbnb.
Simple but they are going to enhance it with a system prompt. So they will write a prompt like you are a senior level
system architect and a backend developer and a front end or a fullstack developer and now you have to create the app the product that the user is requesting.
Make sure that you are going ahead and following up with uh you are going ahead and following all the engineering principles and using these kind of files and they would go ahead and give more context. Okay. So your prompt whatever
context. Okay. So your prompt whatever you are giving will be enhanced with the system prompt and then as a result what you will have first it will create a prompt to plan the whole product. Step
number one is create this file. Step
number two is create this file. Step
number three is create this function then create an API. So it is going to lay down the complete plan. After that
it will create files.
It has access to a virtual environment.
So now it is becoming agentic. Why? It
has an access to a tool. What is the tool? The tool is a virtual environment.
tool? The tool is a virtual environment.
So what will happen a virtual environment is nothing a piece of computer that is given to lovable where it can go ahead and deploy your apps.
Okay. So it has access to a file system it has access to an automatic deployment. Okay. And then it is going
deployment. Okay. And then it is going to keep on running the file on itself so that it is able to understand if there are any kind of mistakes and then because it's an agent it is going to automatically correct these mistakes and
eventually you are going to get the output.
Okay. So lovable, emergent, vzero are nothing. They are just agents which work
nothing. They are just agents which work in a loop to correct the mistakes and they also make sure that they have access to some kind of file system where they can host these files so that you as
a user can go ahead and access these files.
Okay. So let me give you a moment. Okay.
This is chat GPT.
I am giving a simple prompt to charge GBD. Okay.
GBD. Okay.
I want to create a clone of Okay. help me create a system prompt
Okay. help me create a system prompt that I can use in my vibe coding app
where user can enter their idea of app and the system
prompt will make it detail. scale
detail. scale effective to be able to create a
dependable modern app with let's say JavaScript as the primary
language. Most of the apps you will see
language. Most of the apps you will see that they are written in JavaScript.
Okay, I'll just click on this and then I'll get a very detailed system prompt.
Yes. So here what will happen? While the
prompt is being generated, what will happen here?
I will create a system.
If I want to create a clone of lovable, I will create a system. This is the box.
User will input something. This request
will be added with my system prompt and the user input will be added and then it will be sent to an LLM and then I'll get the complete set of instructions.
Right? So this prompt I have taken. Now
this is a prompt system prompt. This you
are a Yoder. This this this is this this this. And now you understand everything
this. And now you understand everything is mentioned here.
Right. This now will be sent to the LM.
Now based on this files would be generated. So it will create a plan. In
generated. So it will create a plan. In
the plan there would be some files, there would be some testing, there would be some functionality and then this will be again sent to the prompt.
Right? You can actually create your own lovable, your own emergent with the help of relay app because there also you are executing instructions step by step.
Right?
So this is how lovable and every platform works on the back end. So agent
is nothing. It is thinking and accessing tools in a loop. So they have orchestrated them in such a way that they are able to go ahead and execute
things again and again in a methodical or a step-wise manner. Right? Now agents
are built on multiple patterns. One
pattern is that there is one agent who is doing multiple task. Then
let's say if I give you people an assignment that tomorrow you have to create a case study on lovable five people in the same team. Okay. Now what
will happen? I believe that if five people are working on the same team the quality should improve. That is why I'm putting five people and I'm giving you less time. For an individual I would
less time. For an individual I would have given one week but for five people I'm giving two days. So now because I think that efficiency will increase. So
now what you guys will do is you will divide the work among yourself that I will look do the research you do the direct someone is going to do the design someone is going to do something else.
Okay. And when you go ahead and come back you are going to go ahead and combine all the out outputs and then you are going to create something which is a one case study. So this is how agents
also do the orchestration. So there is also design patterns in agents which define how different agents are going to work with each other.
Okay. So this is in a nutshell what agents are. Okay. Everyone tell me on a
agents are. Okay. Everyone tell me on a scale of 1 to five you're able to understand agents.
Yes. So please everyone please understand things from first principles which is agent is nothing. Agent did not existed. AI agent did not existing 4
existed. AI agent did not existing 4 years ago. Okay. So it's very simple. If
years ago. Okay. So it's very simple. If
you are creating a workflow which can have an AI LLM in between which can execute actions for you and do some intelligent task that is an AI agent.
Right now after this there is one problem with the agents.
The problem is that the agents in order to act they need to connect with tools.
LinkedIn is one platform, Gmail is one platform, Google Drive is one platform, maybe your SMS is one platform, your Telegram is one platform and there are so many platform, your Slack is one platform. So they need to
connect with many tools. And if you are someone who is building agents where LLM needs to connect with the agents with the tools, then it's a nightmare to read
every API and connect your product with the API. Okay. So if so many let's say
the API. Okay. So if so many let's say like one LM has to connect with multiple products then everyone has their unique API. So that is a nightmare. So now what
API. So that is a nightmare. So now what these people have done is enthropic the people behind claude they have created a simple thing which is they told everyone
that if you want to they have told Gmail slack Jira everyone and the whole world that if anyone wants the LLM to use their APIs please go ahead and create something
called as model context protocol or MCP.
So what is an MCP? MCPS is these people have created one more versions of their APIs or a layer over their APIs which can be understood by large language
models.
Right? So maybe yesterday only or a couple of days before Zumato has launched their own MCP. So Zumato is an Indian website where you can is an Indian app or website where you can go
ahead and order food. So this is the detailed article on the same where is that? Yes. So now you can go ahead and
that? Yes. So now you can go ahead and do so many things from your cloud. So on
your cloud desktop or your chat GPT desktop you can go ahead and install this uh uh this cla this summato MCP and then you should be able to go ahead and
write these kind of queries Indian food in your Indraagar pizza under 200 rupees and then you should be able to make an order from there only. Right. So now
Zumato has an API which already existed for a long period of time but you cannot have conversation with an API through the LLM through natural language. Now
they have created an MCP. What does an MCP do? MCP converts the natural
MCP do? MCP converts the natural language helps the LLM to talk with the API in natural language.
Right? That is an NCP. Okay. And there
are multiple I'll share this link with you so that you are able to go and understand how it is created. Understand
it's not difficult to create MCP. What
you need to do is if you are a product like Zumat or any other product in the world, you would have an API documentation. Take that API
documentation. Take that API documentation, go to chat GPT and ask it to convert into an MCP. It will give you a file that you have to host on your server and then you are done. You have
to do a couple of testing as well but that is done right. So that is model context
right. So that is model context protocol. So in model context protocol
protocol. So in model context protocol we have three components. The first part is the AI host where you are going to send query. So it could be claude,
send query. So it could be claude, cursor, chat, GPT. This is where your MCP client is going to be there. This is
your MCP client. Okay. Then there's an MCP server that is hosted at the person who has given you the MCP which is Zumato or Razer Pay or something. And
then there is a data source. This is
stored at Zumato or somewhere where you can go ahead and have the data or you can take actions. Okay. So any MCP server will give you a small file that is called as client configuration file.
You have to enter it into your cloud or any of the LLM that you are using and then your agent will be able to talk to these tools without understanding the API.
Right? So this is MCP. You can go ahead and read more about it. Right? Yes. So
we have talked about so many things so far.
Cool.
Everyone tell me how would you describe the session so far in one word so that I know if we are going in the right direction.
Tell me how would you explain this session so far in one word so that I can go ahead and tell you something more.
Yes.
Great. Now let me just go ahead. Thank
you. Thank you everyone. Thank you for the nice words. Now I think we are on time. So we have some more time. So I'll
time. So we have some more time. So I'll
give you one more thing. Okay, which is evaluations.
Look at this very carefully. Okay, so
there are some challenges that happen in building AI products. The
challenges are they hallucinate. They
have certain kind of biases and they are indeterministic.
Okay. But understand nowadays you people would be understanding that language models are becoming more and more intelligent. Okay. If you find that
intelligent. Okay. If you find that there is an error, you can ask it to correct that error and it corrects that error. Okay. But still large language
error. Okay. But still large language models are these powerful beast. If you
leave them without a leash then they can do a lot of disaster to your customers.
Okay. So this is what AI believes like.
Okay. It believes that it has all the content in the world and it can give answers like wrong answers with complete confidence.
Okay. So that is why we have something to tame these beast. The leash that works on your LLMs is called as evaluations.
Okay. And what happens is whenever your LLM is not working and if you're not getting the right kind of output, there are certain issues and you can resolve these issues with three things. Okay. So
when you are building any kind of GI product, you have to take three important decisions. The one important
important decisions. The one important decision is what is the model that you are going to choose. If you choose a very high quality model then you have to pay more cost. If you choose a very low
cost model then you have an issue of accuracy.
Right? Similarly rag instrumentation in rag there are multiple components.
If your chunk size is not correct then the information that it is giving you would be incorrect. If you understand I have told you that there is a vector embedding but for vector embedding also there are multiple algorithms which
algorithm should you use okay it depends on the use case. So what is going to be the instrumentation of your rag? What is
the vector algorithm? What is the size of chunking? What are the different kind
of chunking? What are the different kind of model that you will use? That is also very important.
Okay. These are the decision that you need to take. And then context engineering. What should be the prompt?
engineering. What should be the prompt?
How long should be the context? how you
should give the instructions. These
things are going to determine whether your geni app is working rightly or not.
And this is actually engineering.
Engineering why? Because there is going to be a lot of course correction.
Okay. So in order to do these things, in order to understand whether our geni application is working well or not, we have something called as evaluations. So
what evaluations do is they check the output for correct syntax and formats for bias for correctness relevance and other unnecessary checks. Okay. So for
example if on notebook LM al on granular we generated a summary we want to understand whether this summary is actually aligned with whatever has been
said in the meeting or not. So how do we test this? We test this with the help of
test this? We test this with the help of large language models. Okay. So what we do is to help you understand I'm going to give
you a case study on the AI first job website. Okay. So tell me in the chat
website. Okay. So tell me in the chat how many people are looking for jobs.
Tell me in the chat yes or no if you're looking for a job.
Yes guys even if you don't know your colleagues and your bosses might be present in this meeting. Now
I will give you an idea. Okay. And I'm
sure many people would have faced this.
Okay. So let's say I am creating a new kind of job website where what I'll do is I will crawl all the websites from the all I will crawl all the jobs from
major websites from the internet. So
I'll go to hired, I'll go to LinkedIn, I'll go to no, I'll go to Instaire, I'll go to multiple websites. I'll go to indeed and other websites in order to crawl jobs. scroll as in I will go to
crawl jobs. scroll as in I will go to this website automatically I will get the content and I'll scrape the jobs so now I have a job database I have
database of all the job descriptions titles companies and everything right now what I'll do is I will enhance this
job with LLM I will enhance this job with LLM and by enhancing what I mean is I will give it a prompt and then what I'll do is I will generate this output
so I will enter a job description and then the LLM will give me summary of the job description possible interview question. So from the job description in the company it will
try to understand what are the question that can be asked skills that are needed to do this job learning guide for this job and quiz for assessment which is a small quiz that
will help me understand whether I'm ready to fill or apply for this job or not. Okay, this is a simple website that
not. Okay, this is a simple website that I need to create. Okay, tell me everyone will this kind of website will be helpful for you or not? People are
looking for jobs.
Yes. So, it will give you some more context about the job description.
Right. So, now and this can be created.
You can go ahead and create this website today also like by using labable or something. Okay. But now there is an
something. Okay. But now there is an issue. How do I know this job
issue. How do I know this job description, interview question, skill needed, learning guide, everything is good or not. Okay. So I'm going to introduce one method of so understand
these things that I have told you which is all of these things they are done with multiple methods. One of the method
to run your evaluations is called as LLM as a judge. So we run our outputs through another more intelligent LLM in
order to understand whether it is correct output or not. Okay. So here the issue is that although I'm generating good information but this information can be wrong.
Okay. So what I'll do is I'll do this.
Okay. So I have crawled a job from the internet. I have stored them in my
internet. I have stored them in my database. I have ran them through
database. I have ran them through LLMs and then I have created this content.
I have created this content.
Now what I'll do is I will get all of these things again go through another LLM in order to understand which is called as evaluation
that whether this content is correctly created or not. This is called as LLM as a judge and then I will find out whether the
content is correct or not. This is a simple method of evaluation. Let me show you how it works.
Yes. So this is the evaluation. Look at
this very carefully guys. This is the prompt that I'll go to the I'll give to the another LM. Okay. You are an AI content quality evaluator for product management job listings.
Given the following, this is the original job description. This is the AI generated summary. This is the interview
generated summary. This is the interview question. This is the output from the
question. This is the output from the other LLM. Right? These are the skills.
other LLM. Right? These are the skills.
These are the concept. This is a quiz.
Evaluate the content using the following checklist. Is the summary accurate?
checklist. Is the summary accurate?
Interview questions correct? Listed
skills aligned? Are the concept useful?
Is the quiz challenging? is any
information correct or hell needing and then give me the response in this particular format okay then I'll have a dashboard where for every job description this score
would be mentioned then I would know whether my LLM is working correctly or not okay so let's say if I get poor or fair or low or weak what I'll do is I
will go back I will again try to tweak these things if my quality is add then I will maybe choose change the model or I'll change
the rag instrumentation or I'll go ahead and change the context engineering which is the prompt that I have given okay that is how you improve upon AI product with the help of evaluation this is a
super simplistic example of evaluations okay there are codebased evaluations as well there are LLM based evaluations as well generally this evaluation LLM that
you use so this is LLM1 and this is LLM2. This LLM is generally a more intelligent LLM which is able to assess.
Right? So this is an example of evaluation. Everyone a quick yes or no
evaluation. Everyone a quick yes or no if you're able to understand this.
Yes, there are many open source LLM guys. OpenAI OSS is an open source LLM.
guys. OpenAI OSS is an open source LLM.
Lama is an open source LLM.
Right? So this is there. Now we have almost completed the session. This is
some set of tools. If you guys are curious, you can go ahead and start using this in your product management workflow. There is also a detailed video
workflow. There is also a detailed video on this on our YouTube channel. Okay. AI
for PM toolkit discovery for discovery for discovering what to build chat GP notebook perplexity mix panel textsql.
Okay. for delivering at intelligence which is nothing but Jira, Asana, Vzero, Kurser, Postbot, Figma which you can use to create wireframes prototypes very quickly. Okay. And then for distribution
quickly. Okay. And then for distribution you have craftable, chatbased, notion, genex, moage that can help you create marketing materials, collaborate with your team and look at analytics.
Okay. So try to explore these tools at your own time so that you are able to become an AI enabled PM as well.
Whatever we have discussed so far was for AI like applied AIPM, right? And
then one more important thing guys that in this program like so far what I have done is I have given you information
I have given you information about evaluations about agents about various tools about rack about prompt engineering about finetuning but information is not
helpful for you information is just a couple of prompts away what you need is knowledge.
And how do you convert information into knowledge? With the help of action
knowledge? With the help of action and with the help of feedback and reflection.
Right? So what you need to do is please do not limit yourself to only this knowledge.
Go ahead and do something. Doing as in go ahead and build something. Okay? You
have so much information available across the internet. Utilize and try to go ahead and build something right. If
you guys are looking for mentorship then we also have a very detailed program. So
understand whatever I have told you in this last two session that is not even 10% of what we teach in our bigger program. So if you really found this
program. So if you really found this information genuine useful I teach a bigger program which is a 15week program on AI product management. Okay you can go to the website you can get all the
details out there. But if you have to ask me let's say three reasons why you should join LOPM program against other programs is very simple.
First it is the most detailed AI product management program. We first talk about
management program. We first talk about the traditional product management then we talk about AI product management. You
can go to the website you can compare the curriculum. You can compare
the curriculum. You can compare curriculum to any other course in the world. You will not be able to find a
world. You will not be able to find a more detailed program. First point.
Second point is everything is practical.
We believe in doing things. So in the end of the program, you're actually going to go ahead and build your own product. Okay. So if you go ahead and go
product. Okay. So if you go ahead and go to the website, if you go ahead and go to the website hellom.co, you should be able to find in the curriculum section that this is the
whole curriculum. Everything is
whole curriculum. Everything is supported with a lot of case studies and tools.
Plus in the end of the sessions, we are also going to help you build your own AI product.
This helps you prove to other people and to yourself that you are not an information seeker. You are actually
information seeker. You are actually someone who understands how to build products and this is a big opportunity for you. Right. And we also help you
for you. Right. And we also help you prepare for interviews and all that is given. And then
given. And then yes and then the third part is we are the only program that supports you for the long term. Even if you have
completed the program, we understand that some people are able to get the job within the program, some people get the job after the program. We support you for a period of two years. So in these two years, if you have any questions, any feedback, anything that you want to
get help from us, you should be able to talk to us. Okay. And every week, we keep on doing sessions to make sure that you are able to learn. Okay. The
Yes.
Yes. And you can go to the website in order to check all the other details that are out there. The cost of the program stands at 75,000 rupees or $1,000 if you are outside from India.
Okay. But it is my assurance that if you go ahead and take this program, this is going to go ahead and give you a lot of content, right? We don't give any kind of
right? We don't give any kind of discounts, but this is the context. You
if you want to get more information about this cohort, just go to the website hello.co and then click on this video. You should be able to get all the
video. You should be able to get all the information right now an important part. Okay, that is why I've given you this context.
Yes, classes are conducted on Saturday and Sunday evening 8:30 to 10:30 for ISD and morning for the US batches. Okay. So now
everyone understand if you really want to get into AI product management you have to build a strong portfolio so that people understand that you have what it takes.
Okay. You are actually not a talker but a doer. Okay. So what you should do is
a doer. Okay. So what you should do is in order to create a portfolio you can start very small. You start by commenting and start by creating content
about geni. Understand for yourself.
about geni. Understand for yourself.
Okay. Then do tear down in case studies.
Then do product improvements and go ahead and build some like try to go ahead and build some hypothesis of the products. After that build some side
products. After that build some side projects and then if you have already done some past work you can include all of these things and then you can use in your portfolio. You can go to the hello
your portfolio. You can go to the hello channel there we have created. So if you go to the hello channel on YouTube
we have an amazing playlist called as summer of product. Okay. Or this is the playlist getting started with product management or summer of product where I have explained how to create a portfolio.
Okay. And then the last part is this.
This is the most important slide. What
you should do is please do not limit this information that I have told you so far to just this session. Okay? Don't
forget about everything as soon as you leave the session. Please take action.
What you should do is reverse engineer top AI products in the same way that I have told you. Understand gaps,
frustrations in your personal and professional workflow. Build AI tools
professional workflow. Build AI tools and agents to solve these problems. Start writing what you are learning. Tag
me. I'll be able to like and post and depost. Right? And then keep a track of
depost. Right? And then keep a track of AI companies what they are doing. Help
them improve, enhance, expand and then follow up. Try to find companies who are
follow up. Try to find companies who are do really doing good and then understand their product. Reverse engineer the
their product. Reverse engineer the product. Suggest improvements. reach out
product. Suggest improvements. reach out
to the founders and then follow up with the founders. Most of the time you are
the founders. Most of the time you are going to get some appreciation, attention and maybe a job.
Okay, there are many people that we have helped came from completely non- tech background, some senior background, some junior background and they were able to become product managers. You can go ahead and look at this playlist called as transition to product management
where you can read their stories.
Okay. So this is it about today's session guys. Yes. Uh the flagship
session guys. Yes. Uh the flagship program is four weeks of foundation PM content which is strategy and product sense.
8 weeks of going ahead and creating like understanding AI in detail. So
understand if I can give you this much information in 2 days what can I do in 8 weeks or maybe 15 weeks and then there are three weeks of building your own
product.
So flagship program covers the entire AIPM syllabus.
If some of you wants to check how does a hello cohort looks like, let me show you something. Give me a moment.
Yes, I have shared a link in the chat. This
is the detailed video where we have shared the kickstarting call for a cohort. So at hello PM, we try to keep things as transparent as
possible. We love that. We love being
possible. We love that. We love being honest in this really corrupted tech world. So I have gone ahead and shared a
world. So I have gone ahead and shared a link where you can understand who are the kind of people who join, what are their backgrounds so that you are able to go ahead and take a better call.
Yes. And there is a lot of free content that is available on our YouTube channel. You guys should absolutely go
channel. You guys should absolutely go ahead and check it out. The placement
support that we offer is including four things. We offer you we help you create
things. We offer you we help you create a better resume.
We help you build a strong portfolio.
We help you prepare better for the interviews and then we also have an internal job board where alumni or network companies and other companies that we seek out to also list your jobs
and for many of these jobs referrals and recommendations might also be available.
We do not guarantee placements but we help you out in all the possible ways that we can and we also uh are available across the gloes. So we have a lot of
people who are going to join from UK, from the US, from Canada, Australia, New Zealand and as well from India.
And this happens because of a large alumni network that we have created.
You can use self-hosted LLMs without having internet, but you should have a powerful computer.
Cool. Thank you everyone. I'll just go ahead and end the session now. Thanks a
lot for choosing to spend your weekend with me and with Hello PM. I believe I was able to go ahead and add some kind of value as available from the feedback and the LinkedIn comments. We are very
soon going to conduct more sessions. The
next batch is going to start from 9th of November. So take your time. Understand
November. So take your time. Understand
you do not have to take this decision in FOMO. Go ahead, take the right decision
FOMO. Go ahead, take the right decision and understand it is not just a time invest. It is not just a money
invest. It is not just a money investment but it's an important time investment. Okay? So take your decision
investment. Okay? So take your decision by understanding everything. We want you to win but once you come in the session we make sure that you are working hard.
Yes. Cool. Yes. Everyone if you have any custom questions you can go to the website. There's a shoulduling call
website. There's a shoulduling call link. You can go ahead and schedule a
link. You can go ahead and schedule a call and then the team should be able to help. And we keep on doing these
help. And we keep on doing these counseling sessions so that you can ask your questions as well. Okay. Cool.
Yes. Cool. Thank you everyone. Thank you
everyone. Bye. Take care. You have been a great audience. Thanks again for choosing to spend your time with me. I
really feel good that I was able to be somewhat helpful for you guys. Thank you
everyone. Bye and take care. It was my pleasure hosting you guys. Thank you
guys. Take care.
Loading video analysis...