Giving the LLM/Agents a mind for long term planning and remembering stuff associatively. The memory is the half agi within the generative multiagentic system where the LLM is the context processor.
I specialize in Retrieval-Augmented Generation (RAG). Your introduction is good, but it lacks technical depth. You glossed over chunking and how to use it correctly based on the data. Pinecone is good, but it's not necessarily better than vector databases built in Rust or Go, like Qdrant and Weaviate (which are free and open source). It's also important to explain in-memory vector database solutions using tools like FAISS or on-disk solutions like Qdrant and Pinecone, and to discuss the pros and cons of each. A significant omission is not addressing implicit behavior or implicit data versus explicit data, and their relationship with graph databases. Rerankers might be too advanced a concept; often, you can achieve better results by optimizing chunking, similar to how tokenization is used for semantic understanding. Often, agents are unnecessary, and having a chain-of-thought agent before sending to the LLM can be a waste. Additionally, discussing the similarities between the internals of a transformer and a vector database is intriguing. Overall, the video feels like a Pinecone sponsorship. Regarding fine-tuning, it's about improving the understanding or behavior of an LLM in a specific domain at the cost of losing understanding in other areas. You should only fine-tune if the model does not seem to understand. Use RAG when the model lacks knowledge or when you want to reduce hallucinations, but relying solely on vector databases is a missed opportunity. One micro aspect you did not touch on is tokenization. The two biggest things people often overlook are chunking and tokenization, and there are massive gains to be made when these are properly understood.
one good use is ecommerce products for conversational shopping...creating new experiences...built a few prototypes of this as mvps for pitches...its a night and day experience
@@FunwithBlender Great comment! What is your go to open source RAG pipeline? I am beginning to learn and discover all these tools. It is pretty amazing.
You can ask Claude 3.5 create a locally run vector database. It will manage it in a day and you will avoid having to pay for another clouded service. I did it and it worked.
Great video! I've bee following you for awhile and have set up some edge LLM's using your tutorials. RAG is the future for any business wanting to truly utilize their data. to the fullest. I think that a lot of companies aren't even sure how they can implement their data for the greater good of the business while saving money at the same time. Videos like this help clarify the subject. Please do a video on Pinecone. I'm sure there is a lot of us that would like to see it's capabilities. Keep up the great work.
Thanks you for making this Video. I am a Non Techie trying to get easy to understand method of querying my documents using RAG with open source LLMs. Would eagerly await your full tutorial on this topic .
Be interested to see best practices for keeping the RAG database up to date. For example if a new PDF is dropped into a watched folder the PDF gets submitted to the embedding model automatically. Likewise for PDFs that are out of date and removed which should them be dropped from the vector database.
You could add a useage count, entered date, last accessed date, etc and have a background thread check for old info. Like say 2-3 years unless its something your llm wouldn't know
In RUclips, there are hundreds of channels baffling buzzwords and lame tutorials about these concepts without putting real effort on creating meaningful videos. And this channel is not one of those. I appreciate your videos Matt, thank you for the great content
Oh and please publish both tutorials , Picone and more RAG applications - those are the future and using agents with that is golden for the near future for all of us
I would also like more tutorials on RAG and techniques to improve chatbots. Thanks Matthew for this content. I like your posts on news but tutorials are also useful and appreciated given your ability to communicate such concepts.
Just discovered this channel and it quickly leapfrogged others as one of my favorite AI channels. I'm a Data Scientist starting to work in the LLM arena and these videos are super helpful. I'd love a full tutorial on RAG!
Yes! Please set up a full tutorial for us. This is powerful. I have a Custom GPT business and I’ve always known I need to incorporate RAG in the most pragmatic way possible to advance my capabilities. So it sounds like Pinecone is the way to go. Thanks so much for your help.
I've heard about RAG before, but this video helped me understand it much better. Thank you for sharing your knowledge! I would greatly appreciate it if you could make another video demonstrating how to use it with a real-life example
Yes, we need a full tutorial please. This is great knowledge and a very simple to understand video! I actually have a pinecone account, and started using it when I first started playing around with Auto-GPT, but I haven't used it since. I'm interested in developing some new projects soon, and RAG sounds like something I need to be thinking about.
RAG requires a knowledge graph DB as well in order to find information not directly mentioned which is a limitation of RAG, a tutorial incorporating both would be amazing
00:02 An intro to RAG and its misunderstood nature 01:51 RAG is efficient for continually providing new knowledge to large language models 03:42 RAG enables adding external knowledge to AI models 05:29 RAG allows AI to access and incorporate new information into its responses. 07:25 Utilizing embedding models to enhance AI understanding 09:12 RAG enhances AI by providing external knowledge sources 11:10 Utilizing external knowledge for AI searches 12:57 RAG simplifies retrieval augmented generation process
Claude's new Projects feature is like a simple RAG. I've given it all the knowledge about a novel I'm working on and it has been surprisingly good at understanding all the nuances. Way better than a normal conversation.
Yes definitely need to expand on RAG, vector database and pinecone. Full end to end process for incorporating specific business data sets to generate highly customized content. Creative/marketing use case if possible.
I'm defintely interested in doing RAG but more so in doing it locally. Especially with all the important information I can't trust a service for storing it, if there is a local way of doing it I'd be very interested in building a RAG pipeline. Great video for explaining the basics of it.
The OpenAI Dev Days from last year had a great session on optimizing LLMs. Their progression was to try few-shot, then RAG, then fine-tuning - and their description of fine-tuning was that it was a good way to provide "intuition" to the model, but not knowledge.
* 00:00:00 Introduction to Retrieval Augmented Generation (RAG) * 01:02:22 Misunderstandings about RAG and Large Language Models * 02:11:44 RAG as an external source of information for large language models * 03:13:22 Context window limitations * 04:12:11 RAG for chatbot conversation history * 04:51:17 RAG for access to internal company documents * 05:39:22 RAG to update large language models with new information * 06:02:22 How Retrieval Augmented Generation Works * 07:32:22 Workflow with RAG for finding relevant information * 08:22:22 Embedding model and Vector database * 10:11:22 RAG with agents for iterative approach * 12:13:22 Pine Cone for Vector database * 13:11:22 Conclusion * 13:47:22 Outro
Slight pet peeve of mine - I think presenting it this way makes it sound like you must use an embedding model/vector db to do RAG. The basic version of RAG is just that idea of passing additional, retrieved info with the prompt to the LLM. Yes, the embedding model w/ vector db is a very efficient way of doing that - especially with large amounts of data. But it is not the only way to accomplish it, and may not even be the best way to do it, depending on the use case.
It is essential to conduct a thorough preprocessing of the documents before entering them into the RAG. This involves extracting the text, tables, and images, and processing the latter through a vision module. Additionally, it is crucial to maintain content coherence by ensuring that references to tables and images are correctly preserved in the text. Only after this processing should the documents be entered into a LLM.
I've been dreaming about using RAG to compile the summary of key references I use in my profession (Geophysical interpretation). Obviously, professionals may not utilize every key learning from published materials and some information may be conflicting with other published materials in the same field. What would be immensely useful is a method of adding weights to information you utilize on a daily basis and to identify where an AI finds conflicts in logic. If a conflict is found, a model can be taught which path to follow.
I would be grateful to see the full tutorial on embedding large documents using some of your favorite tools, storing it in Pinecone, and building an AI app using RAG. Have you recorded any tutorials of that kind already? Thank you!
Great video as always, Matthew. Thanks a lot. I really would love to have a detailed video on how to set up RAG. Trying to establish an AI retrieval system for all teachings of our Buddhist teacher. About 16 min words. The seems to be the perfect system
Yes please create a tutorial video showcasing step by step instructions around practical techniques for RAG, local open source vector databases, and automations
I have following doubts bro please clarify, it would be most for me:- 1) Which should I learn first? RAG or AI agent 2) which should I learn first LLM or NLP? 3) which programming language is necessary for learning all of them? 4) While writing code for any of these for example if we are writing a code in tensorflow pytorch in which pattern should we write a code? 5) what is the scope for prompt engineering? Is it just a hype or is it really worth?
open prompt language model. No limit to the prompt input of a language model. You can basically add an additional large language model of data within you prompt. :)
I would like to see a more in depth RAG tutorial. Pinecone is great, but maybe at the end show how to use a local vector db for those of us who want it completely private. Thanks!
A deeper dive into RAG and embeddings would be a great help for developers like me. I work in C# with GPT4o and I use REST rather than Python, but then OK, you can't always get what you want 🙂
Ok that was awesome. Of course I’d like to know more! I’ve had a hard time understanding rag til now for some odd reason. Would also love a tutorial on pinecone and embedding.
I am a fire investigator/researcher. I have an enormous amount of data that I would like to store in a format for LLMs to search against. This was a great intro but I need to figure out how to specifically set up a RAG, how to embed the info so it’s in a vector space and then how to use a LLM to conduct a query considering all of that. Is there a place to go learn this??? Thanks!!!!!
LM Studio and GPT4ALL have this RAG (local document) feature, you provide your document and the chosen model responds only based on the information received.
Please if you can make a tutorial to use the Rag combined to a chatbox. Where the system decides whether to use the knowledge stored in the vector database, or if the LLm has the information and decides to use only its knowledge. Thank you very much...
I would love to see a video on how to do all of this with open source software that I can run locally. A project combining RAG with Ollama models would be awesome
Please make a tutorial on automatic and repeatable adding labeling and improving clarity and context in my documents before indexing. Guidance on chunking would also be appreciated.
God I wish I had this like a 18 months ago, it was kinda hard for me to jump into it and figure it out. I'm glad I can at least confirm my process was at least successful.
A clever way to make an ad, here for Pinecone, by delivering knowledge. It's much more acceptable this way. Well done, and thanks for the intro to RAG :) The people @Pinecone must be proud of this video. I've just to say that, it's more about giving AI an optimized context than truly giving them a "memory". The title feels a bit misleading. A real memory would be a workable space where the AI stores itself the required data for later retrieval, and which becomes part of its infrastructure. This is not it.
What's your favorite use case for RAG?
Giving the LLM/Agents a mind for long term planning and remembering stuff associatively. The memory is the half agi within the generative multiagentic system where the LLM is the context processor.
I specialize in Retrieval-Augmented Generation (RAG). Your introduction is good, but it lacks technical depth. You glossed over chunking and how to use it correctly based on the data. Pinecone is good, but it's not necessarily better than vector databases built in Rust or Go, like Qdrant and Weaviate (which are free and open source). It's also important to explain in-memory vector database solutions using tools like FAISS or on-disk solutions like Qdrant and Pinecone, and to discuss the pros and cons of each.
A significant omission is not addressing implicit behavior or implicit data versus explicit data, and their relationship with graph databases. Rerankers might be too advanced a concept; often, you can achieve better results by optimizing chunking, similar to how tokenization is used for semantic understanding. Often, agents are unnecessary, and having a chain-of-thought agent before sending to the LLM can be a waste. Additionally, discussing the similarities between the internals of a transformer and a vector database is intriguing. Overall, the video feels like a Pinecone sponsorship.
Regarding fine-tuning, it's about improving the understanding or behavior of an LLM in a specific domain at the cost of losing understanding in other areas. You should only fine-tune if the model does not seem to understand. Use RAG when the model lacks knowledge or when you want to reduce hallucinations, but relying solely on vector databases is a missed opportunity. One micro aspect you did not touch on is tokenization. The two biggest things people often overlook are chunking and tokenization, and there are massive gains to be made when these are properly understood.
Using my local scanned (searchable) PDF documents in RAG.
one good use is ecommerce products for conversational shopping...creating new experiences...built a few prototypes of this as mvps for pitches...its a night and day experience
@@FunwithBlender Great comment!
What is your go to open source RAG pipeline? I am beginning to learn and discover all these tools. It is pretty amazing.
A vector database tutorial would be great! Excellent content.
You can ask Claude 3.5 create a locally run vector database. It will manage it in a day and you will avoid having to pay for another clouded service. I did it and it worked.
Great @@gabrielsandstedt
I would like a deeper dive into RAG and an end to end pinecone tutorial! Thanks for the great video!
You could use pinecone but Claude 3.5 can build you a custom vector search algorithm that will work and you can store locally using sqlite
PLEASE do the how-to on setting this up. It is a key piece to the puzzle, for sure. Thank you for all the great content!
Great video! I've bee following you for awhile and have set up some edge LLM's using your tutorials. RAG is the future for any business wanting to truly utilize their data. to the fullest. I think that a lot of companies aren't even sure how they can implement their data for the greater good of the business while saving money at the same time. Videos like this help clarify the subject. Please do a video on Pinecone. I'm sure there is a lot of us that would like to see it's capabilities. Keep up the great work.
I would love to see a tutorial on how to use RAG! I was just thinking on how to solve some of this knowledge problem on a small project I'm working on
A full tutorial would be great - thanks so much 👍
Very well explained : short and clear with good examples, thanks!
Thanks you for making this Video. I am a Non Techie trying to get easy to understand method of querying my documents using RAG with open source LLMs. Would eagerly await your full tutorial on this topic .
This is one of the best videos I watched from you as a junior AI engineer 👌🏼 BEAUTIFUL
Be interested to see best practices for keeping the RAG database up to date. For example if a new PDF is dropped into a watched folder the PDF gets submitted to the embedding model automatically. Likewise for PDFs that are out of date and removed which should them be dropped from the vector database.
You could add a useage count, entered date, last accessed date, etc and have a background thread check for old info. Like say 2-3 years unless its something your llm wouldn't know
YESSS DO ITT PLEASE 🙏
In RUclips, there are hundreds of channels baffling buzzwords and lame tutorials about these concepts without putting real effort on creating meaningful videos. And this channel is not one of those.
I appreciate your videos Matt, thank you for the great content
Oh and please publish both tutorials , Picone and more RAG applications - those are the future and using agents with that is golden for the near future for all of us
I would also like more tutorials on RAG and techniques to improve chatbots. Thanks Matthew for this content. I like your posts on news but tutorials are also useful and appreciated given your ability to communicate such concepts.
Just discovered this channel and it quickly leapfrogged others as one of my favorite AI channels. I'm a Data Scientist starting to work in the LLM arena and these videos are super helpful. I'd love a full tutorial on RAG!
Yes! Please set up a full tutorial for us. This is powerful. I have a Custom GPT business and I’ve always known I need to incorporate RAG in the most pragmatic way possible to advance my capabilities. So it sounds like Pinecone is the way to go. Thanks so much for your help.
I've heard about RAG before, but this video helped me understand it much better. Thank you for sharing your knowledge! I would greatly appreciate it if you could make another video demonstrating how to use it with a real-life example
What a great commercial
Yes! Please go through a full demo! would love to see it!
Yes, we need a full tutorial please. This is great knowledge and a very simple to understand video! I actually have a pinecone account, and started using it when I first started playing around with Auto-GPT, but I haven't used it since. I'm interested in developing some new projects soon, and RAG sounds like something I need to be thinking about.
A tutorial would be amazing! It’s exactly what I need for something I wanted to experiment with
RAG requires a knowledge graph DB as well in order to find information not directly mentioned which is a limitation of RAG, a tutorial incorporating both would be amazing
100% on board with seeing a full tutorial. Also highly interested in seeing a fully open-sourced setup.
00:02 An intro to RAG and its misunderstood nature
01:51 RAG is efficient for continually providing new knowledge to large language models
03:42 RAG enables adding external knowledge to AI models
05:29 RAG allows AI to access and incorporate new information into its responses.
07:25 Utilizing embedding models to enhance AI understanding
09:12 RAG enhances AI by providing external knowledge sources
11:10 Utilizing external knowledge for AI searches
12:57 RAG simplifies retrieval augmented generation process
Claude's new Projects feature is like a simple RAG. I've given it all the knowledge about a novel I'm working on and it has been surprisingly good at understanding all the nuances. Way better than a normal conversation.
Great videos man! Listening them every day now.
This is by far the best AI educational video!!
Please share more RAG solution , this will be very very useful for your audience !!
I like your style of explaining things. Thank you for your videos as I've learned a lot from you.
Finally understood RAG . thank you!
Finally an explanation without using complex terminologies. Thank you Matthew. Lets do one with RAG + Agents
Yes definitely need to expand on RAG, vector database and pinecone. Full end to end process for incorporating specific business data sets to generate highly customized content. Creative/marketing use case if possible.
Thank you for your hard work Mathew! Please do videos on all suggestions that you made in this video.
I'd be very happy to see the whole process presented in a video ♥
This topic is the kind of knowledge everyone thinks they have and brush over.. thanks Matthew
Yes please, more information about Pinecone and RAG! Great content, thanks!
I'm defintely interested in doing RAG but more so in doing it locally. Especially with all the important information I can't trust a service for storing it, if there is a local way of doing it I'd be very interested in building a RAG pipeline. Great video for explaining the basics of it.
The OpenAI Dev Days from last year had a great session on optimizing LLMs. Their progression was to try few-shot, then RAG, then fine-tuning - and their description of fine-tuning was that it was a good way to provide "intuition" to the model, but not knowledge.
Would absolutely love to see a tutorial on this. Thanks for doing something more technical like this, Love it!
Great knowledge!! Please create a new video about pinecone..
* 00:00:00 Introduction to Retrieval Augmented Generation (RAG)
* 01:02:22 Misunderstandings about RAG and Large Language Models
* 02:11:44 RAG as an external source of information for large language models
* 03:13:22 Context window limitations
* 04:12:11 RAG for chatbot conversation history
* 04:51:17 RAG for access to internal company documents
* 05:39:22 RAG to update large language models with new information
* 06:02:22 How Retrieval Augmented Generation Works
* 07:32:22 Workflow with RAG for finding relevant information
* 08:22:22 Embedding model and Vector database
* 10:11:22 RAG with agents for iterative approach
* 12:13:22 Pine Cone for Vector database
* 13:11:22 Conclusion
* 13:47:22 Outro
please continue to educate and show us the RAG vectoring tutuorial. Great video!
Yes ! To all of the walk through on setting up local rag llms and mixed agents
Slight pet peeve of mine - I think presenting it this way makes it sound like you must use an embedding model/vector db to do RAG. The basic version of RAG is just that idea of passing additional, retrieved info with the prompt to the LLM. Yes, the embedding model w/ vector db is a very efficient way of doing that - especially with large amounts of data. But it is not the only way to accomplish it, and may not even be the best way to do it, depending on the use case.
Rag tutorial please, especially use case of local open source llm. Thanks!
With long term memory implementation, as well. All open source, please.
Great explanation of RAG and how it differs from fine-tuning and prompt engineering
It is essential to conduct a thorough preprocessing of the documents before entering them into the RAG. This involves extracting the text, tables, and images, and processing the latter through a vision module. Additionally, it is crucial to maintain content coherence by ensuring that references to tables and images are correctly preserved in the text. Only after this processing should the documents be entered into a LLM.
I have said that it's less about compute power and now about organization of data and mimicking the brain.
This is one way to do it
A full tutorial is NEEDED
For search, there are two ways to do it: lexical or semantic search. RAG can also be used with lexical search
This is a great and very easy to understand explanation. Please make a full tutorial!
I've been dreaming about using RAG to compile the summary of key references I use in my profession (Geophysical interpretation). Obviously, professionals may not utilize every key learning from published materials and some information may be conflicting with other published materials in the same field. What would be immensely useful is a method of adding weights to information you utilize on a daily basis and to identify where an AI finds conflicts in logic. If a conflict is found, a model can be taught which path to follow.
Would love a full RAG tutorial. Thanks for the great video.
I would be grateful to see the full tutorial on embedding large documents using some of your favorite tools, storing it in Pinecone, and building an AI app using RAG. Have you recorded any tutorials of that kind already? Thank you!
Great video as always, Matthew. Thanks a lot. I really would love to have a detailed video on how to set up RAG. Trying to establish an AI retrieval system for all teachings of our Buddhist teacher. About 16 min words. The seems to be the perfect system
Yes, please do a step-by-step guide!!!
Thank you!
Yes, please, show advanced RAG solution including ranking and SQL usage.
Come up with an index, store data as a BLOB, then use SQL to retrieve it and add it to prompt.
Yes please create a tutorial video showcasing step by step instructions around practical techniques for RAG, local open source vector databases, and automations
Yes, full tutorial on rag and pinecone. Provide details on keeping private data private.
Brilliant!! Yes, a deeper dive will help
I have following doubts bro please clarify, it would be most for me:-
1) Which should I learn first? RAG or AI agent
2) which should I learn first LLM or NLP?
3) which programming language is necessary for learning all of them?
4) While writing code for any of these for example if we are writing a code in tensorflow pytorch in which pattern should we write a code?
5) what is the scope for prompt engineering? Is it just a hype or is it really worth?
Yess we want a tutorial! Amazing content thank you !
Looking forward for the Tutorial 🎉!!
This is exactly what I've been looking for! Thanks so much for this
Great video. A step-by-step video on RAG and Pinecone would be great! 👍
amazing explanation of RAG thank you!!
Thanks, Matt, interesting concept. A video tutorial would be great!
Love to see a full tutorial.!
Do it, but without pinecone, with opensource, locally working tools only.
open prompt language model. No limit to the prompt input of a language model. You can basically add an additional large language model of data within you prompt. :)
This was very interesting and a full step by step video would be very helpful!
🏆 Very helpful, with just the main points... love it! As with other, looking forward to more details.
I would like to see a more in depth RAG tutorial. Pinecone is great, but maybe at the end show how to use a local vector db for those of us who want it completely private.
Thanks!
YeYe, do the Tutorial
A deeper dive into RAG and embeddings would be a great help for developers like me. I work in C# with GPT4o and I use REST rather than Python, but then OK, you can't always get what you want 🙂
A deeper dive on how to set-up RAG with Pinecone and an embedding model would be great!
Ok that was awesome. Of course I’d like to know more! I’ve had a hard time understanding rag til now for some odd reason. Would also love a tutorial on pinecone and embedding.
It will be great to do a full tutorial. If you add multimodal RAG and agents functionalities it will be even better.
Great video, please do a deeper dive into RAG and later DSPy video as well.
Yeah, it would be awesome to get know to vector base actually works and how to connect it to the model
I am a fire investigator/researcher. I have an enormous amount of data that I would like to store in a format for LLMs to search against. This was a great intro but I need to figure out how to specifically set up a RAG, how to embed the info so it’s in a vector space and then how to use a LLM to conduct a query considering all of that. Is there a place to go learn this??? Thanks!!!!!
LM Studio and GPT4ALL have this RAG (local document) feature, you provide your document and the chosen model responds only based on the information received.
An excellent tutorial I would really like you to do a deeper dive into RAG and show how you would set it up.
I have been working on a chromadb vector database sothis is awesome! Thanks!
Yes Please ! we are team RAG show us the way :)
Yes please do Pinecone RAG demo. Thanks!
Thank you for this wonderful explanation on RAG, very informative. Just a note regarding Claude's Context Window: it's 200K and not 100K.
Great top view of RAG concept, please give us a detail walk-through on a concrete coding example, many thanks! 🙏
yes please! I’d like to see a full tutorial on how to do the whole process
Would love to see both the tutorial and deeper dive using RAG
PLEASE DO A FULL RAG SETUP TUTORIAL!! 🔥
Please if you can make a tutorial to use the Rag combined to a chatbox. Where the system decides whether to use the knowledge stored in the vector database, or if the LLm has the information and decides to use only its knowledge. Thank you very much...
Hi Matthew,
This video is very informative about basic RAG,
Please provide a tutorial on Pinecone
I would love to see a video on how to do all of this with open source software that I can run locally. A project combining RAG with Ollama models would be awesome
Please make a tutorial on automatic and repeatable adding labeling and improving clarity and context in my documents before indexing. Guidance on chunking would also be appreciated.
Yes. A full RAG tutorial would be great. Thank you.
God I wish I had this like a 18 months ago, it was kinda hard for me to jump into it and figure it out. I'm glad I can at least confirm my process was at least successful.
How about using faiss instead of vector db and something else instead of openai embeddings and maybe use any other llm different to openai
Hey it would be very very helpful if you drop a detailed video on rag setting up and usage!
A clever way to make an ad, here for Pinecone, by delivering knowledge. It's much more acceptable this way. Well done, and thanks for the intro to RAG :) The people @Pinecone must be proud of this video.
I've just to say that, it's more about giving AI an optimized context than truly giving them a "memory". The title feels a bit misleading. A real memory would be a workable space where the AI stores itself the required data for later retrieval, and which becomes part of its infrastructure. This is not it.
Would love to see a complete tutorial on Pinecone and RAG.