Having programmed since 1963, it looks like a return to COBOL. Programming today is 90% getting things hooked up. 10% getting things done. 100% being clueless as to what's going on under the covers. Follow the bouncing ball programming.
A few months following their courses and their YT channel. I believe it's as useful (if not more) than my university degree, which took a lot of money and years.
I just completed their certification course in responsive web design and actually learned a lot from any other course and only in 40 days (300-350hours )and everything for free too
Ania, that was freaking amazing! You simplified all the concepts without going too high-level and dumbing it down altogether. You told us what happens and showed us HOW they happen. I found this very informative and you answered so many questions I have been pondering. I'm not a developer or an AI person, I'm a network engineer. So, thank you... By the way, I used an embedding model to map your face, and the semantic engine returned the words "gorgeous," "lovely," "beautiful," etc... 🙂
Wow the instructor for this vid was actually amazing. I only clicked on it because It was 30 minutes long, having no real intention to actually learn and just have play in the background while I read a textbook for fizz. The instructor was phenomenal, I understood everything she said, every instruction was clear to follow although I only really know some JavaScript and cpp. I actually learned a few things. Before the video began like I said no real intentions of implementing this but since I actually learned and understood pretty much all of it I could see myself actually implementing it on some project and adding it to my resume. Would be cool. Thanks
🎯 Key Takeaways for quick navigation: 00:00 📘 Anikubo's course covers vector embeddings using OpenAI's GPT-4. 01:49 🖥️ Vector embeddings transform various data types into numeric forms for algorithm processing. 06:12 📈 Numbers can represent complex data, and cosine similarity helps compare them. 08:04 🌐 Embeddings find applications in recommendation systems, NLP tasks, and more. 14:04 🛠️ LangChain, an open-source framework, enhances AI interactions, chaining models and data. 23:25 🛠️ The tutorial walks through setting up a Python environment and key scripting steps. 24:22 ⚙️ Essential packages and tools are installed for AI development. 33:17 🤖 The AI assistant, using vector-based search, fetches relevant documents from a database. Made with HARPA AI
I did this in my Numerical Analysis course using Maple and MatLab. Then i did some analysis on images when i took Fourier analysis. Shame i never got much chance to use it professionally, as i tend to work with financial data.
Dumb question: are they interoperable? Like creating embeddings for a dataset with GPT-3 but then comparing them to a new embedding created by a different model
This is a very good video but I would like to understand why do we need datastax and store in DB if the intension is just to use as prompt and get answer. We can get directly from the OPenAI with key and prompt without storing or anything to do with vector embedding and those will be internal to OpenAI, I wanted to understand the use case of approach.
I believe this method is better than fine-tuning and significantly superior to using prompts, especially when you have a lot of information; the chat will provide much better answers.
I think this is more for AI to answer questions on your data..hence she downloaded data from hugging face . But this could also be your own data vectorized..stored in db and queried . I may be wrong but this what I infer.
Thank you for tutorial. does some llms models and tools like chatgpt handles all the tasks related to from storing data to vectorbases and querying relevant data ? doesnt openai provide any database for storing the embedded text so that we used cassandra for this purpose?
Thank you for so clearly and articulately presenting these lessons for us for free! Your eyes, your smile, and your beauty all together it is incredibly distracting to me though haha! I mean you are one of the most gorgeous women I've ever seen in my life and I appreciate your time. Sorry to come off like a creep but you're stunning.
This tutorial is very interesting, except that without a paid account on OpenAI we cannot really put it into practice. But I can't afford to pay 20 euros per month just to set up a tutorial.
You don't need to sign up for OpenAI Plus in order to create an API Key, they are billed separately. You also get a free 3-month API credit when you first create an account, the amount varies, I think they've decreased it now to about $5 (unfortunately I missed out on my credit, since I created my account last year and wasn't coding anything)
actually the 20$ for chatgpt is different from the api, for the api you can for example buy a 10$ credit and use it for as much as you would like, as long as it's used to finetune it will last long
1 indicates an identical vector, or very close semantic meaning, or even identical text. (Note that it's the similarity of the vector's direction only, not scale.) 0 indicates an orthogonal relationship, ie, unrelated semantically. -1 in theory represents complete semantic opposition, but in practice, a perfect -1 is rare in natural language contexts.
How do ML models create embeddings for new, or novel, words? For example, what if I fed it "Hexamethylenetetramine" (an organic compound)? My brain is frying thinking about this...
10:51 THIS is a 'golden nugget' right here: "The core advantage of vector embeddings..." Such a great summation of exactly what an ai model really is. Thanks for such a fantastic video. I love it !
LangChain: "chaining" resulted in two answers for each prompt: "I don't know" and the headlines. The first answer came from OpenAI's LLM, the second answers (the headlines) came from the vector-DB (Astra/Cassandra) that she set up outside OpenAI. LangChain was the bridge between the two. It's a simple little example without much relevance, but it shows the bones. There is a lot more work to make something useful. For instance, you could use a pre-trained LLM to perform the organizational tasks and composition (the language parts) using current data from a real-time source. For instance, "what kind of activities would be good at Los Angeles beaches today?" The LLM could contextualize the meaning of the question using pre-trained data (an understanding of what constitutes beach sports and the condition necessary for each is something that won't much change over time), and then use an external source (weather channels, surf sites, diving data sites, sailing sites, etc.) to search for real-time conditions at LA Baches. The LLM - using the current real-time data - could then look for nearness matches based on how the conditions match up to certain sports. So instead of a generic pre-trained answer like, "people like to sunbathe, swim, dive and surf at beaches" you can get a specific answer such as, "The conditions at Redondo Beach suggest it's a good surf day, but rip tide warnings suggest it is a bad day for people to be swimming. There is a water quality alert for bacteria in Santa Monica Bay." The LLM used the external real-time data to give accurate point-in-time suggestions that it would otherwise never have using data from training months earlier. That's where LangChain can help - merging "new" or custom data into the pre-trained contextual LLM model. Hope this helped.
Yeh it seems to have moved location, open your table and on the right side under 'Database Details', 'Region' click the three dots and select download SCB :)
did anyone run into this error llm = openai(openai_api_key=OPEN_API_KEY) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ TypeError: 'module' object is not callable
4:40 "...Joe is 38 on the 0 to 100 scale... so -.4 on the -1 to 1 scale...": How is that? I get -.24. If it's -.4 on the -1 to 1 scale, that's 30 on the 0 to 100 scale. Please fix my math.
Context has two flavors, near and "not near". Joe is 38% near. Maybe Alice is 40% "not near", which would equate to a negative value (-.4). So context is more than "this one is like the other", it's also "but it's not like this other thing". If we just used a single dimension, then literally everything would be "like" everything else, which makes it a little difficult to differentiate. "The school bus is yellow. A banana is yellow. A bus is NOT LIKE a banana." Dimensions in dimensions.
Great video, but @AniaKubow, if you do not mind, you could have had a very successful career in modelling. Refreshing to see that you chose computing and specifically AI.
🎯 Key Takeaways for quick navigation: 00:14 📉 *Sam Altman was fired by OpenAI's board for not being consistently candid in his communications, leading to implications of lying by omission.* 01:11 🤯 *Various theories circulated, including speculation about dangerous AI developments, financial ties with Saudis, and a letter from former employees alleging dishonesty.* 01:53 🌐 *OpenAI employees expressed discontent, with over 500 threatening to quit, potentially joining Microsoft to dominate the AI space.* 02:21 🔄 *After negotiations with Microsoft failed, Altman and Brockman formed a new AI research team at Microsoft, but Altman eventually returned to OpenAI as CEO on November 21st.* 03:02 ❓ *Uncertainty remains about the true reasons behind Altman's firing, with speculation about conflicts of interest, AI commercialization, or a possible publicity stunt involving the board, Microsoft, and Altman.* Made with HARPA AI
It is disappointing that you have taken the route of OpenAI which is far from open - not only have they not open-sourced their models, their payment methods are limited to a few options and setting up billing is required for their API to work. It would have been much better to use free, open-source models to demonstrate LangChain.
Ania looks fake. I don't know if it is the lighting or the background or what but it just looks fake. It's freaking me out. The voice sounds natural at least.
you did not explain cosine similarity and why this is a good definition of similarity when it comes to comparing vectors. Without understanding this, everything you say after it is incomprehensible
yawn why is modern programming just gluing together packages. not this channel or the creator of the video's fault but i just hate what it's all become
So much love for this incredible community! Hope you like this video
Thanks for the work, hope you gain 1 million subs🎉
seriously, you guys have no idea how much I've learned from this channel. The value is incalculable. Thank you so much!
You look like A.i generated model😅
Thank you for all your work 🙏 you are an inspiration 😊 I hope one day I can be as good as you
@@rameezalam1968 😂
Can we stop just for a moment and appreciate her!!! learned loads of thing from you!!! hats off 🎓🎓 ❤❤ love you from Ethiopia
She's an actor. There's tonnes of people behind the scene to write this content.
@maciej2320 wow amazing me too from Ethiopia ❤
How do you know she's creating the content? You learned loads of stuff but is it making your more money?
@@maciej2320 I made this myself :)
@@jay_wright_thats_right I made it myself :)
do you agree that this legendary channel is better than most paid courses for coding out there?
Having programmed since 1963, it looks like a return to COBOL. Programming today is 90% getting things hooked up. 10% getting things done. 100% being clueless as to what's going on under the covers. Follow the bouncing ball programming.
A few months following their courses and their YT channel. I believe it's as useful (if not more) than my university degree, which took a lot of money and years.
I just completed their certification course in responsive web design and actually learned a lot from any other course and only in 40 days (300-350hours )and everything for free too
what, do you mean there is such thing as a paid course?? I am already 100% clueless. .
@@Jd-zd6bh I said it's free means it is free all courses on their website with certificate
This lecture on vector embedding is undoubtedly one of the best I've encountered! Huge thanks to Ania and FCC! Kudos to you all!
Ania, that was freaking amazing! You simplified all the concepts without going too high-level and dumbing it down altogether. You told us what happens and showed us HOW they happen. I found this very informative and you answered so many questions I have been pondering. I'm not a developer or an AI person, I'm a network engineer. So, thank you...
By the way, I used an embedding model to map your face, and the semantic engine returned the words "gorgeous," "lovely," "beautiful," etc... 🙂
You have a talent to deliver complex information in a very interesting manner! Waiting for more videos!
Wow. Probably the best lecture to meaningfully explain what vector embeddings are, and how cosine similarity works. Thank you very much!
Nice introduction to vector embeddings with clear explanation. One of the best lecture/videos encountered on you tube related to this topic.
You're dazzling and wise, a true blend of grace and intellect.
Wow the instructor for this vid was actually amazing. I only clicked on it because It was 30 minutes long, having no real intention to actually learn and just have play in the background while I read a textbook for fizz. The instructor was phenomenal, I understood everything she said, every instruction was clear to follow although I only really know some JavaScript and cpp. I actually learned a few things. Before the video began like I said no real intentions of implementing this but since I actually learned and understood pretty much all of it I could see myself actually implementing it on some project and adding it to my resume. Would be cool. Thanks
ok simp
This girl is an ideal perfect educator!
Simply love this presentation! That vec math (King - man + woman = Queen) just blow my mind!
🎯 Key Takeaways for quick navigation:
00:00 📘 Anikubo's course covers vector embeddings using OpenAI's GPT-4.
01:49 🖥️ Vector embeddings transform various data types into numeric forms for algorithm processing.
06:12 📈 Numbers can represent complex data, and cosine similarity helps compare them.
08:04 🌐 Embeddings find applications in recommendation systems, NLP tasks, and more.
14:04 🛠️ LangChain, an open-source framework, enhances AI interactions, chaining models and data.
23:25 🛠️ The tutorial walks through setting up a Python environment and key scripting steps.
24:22 ⚙️ Essential packages and tools are installed for AI development.
33:17 🤖 The AI assistant, using vector-based search, fetches relevant documents from a database.
Made with HARPA AI
You're a lifesaver
thank you. a TIMELY course for my projects!
This is a masterclass!!! Thanks so much! Really appreciated!
So well done and well put together! Thanks for the value ❤
I did this in my Numerical Analysis course using Maple and MatLab. Then i did some analysis on images when i took Fourier analysis. Shame i never got much chance to use it professionally, as i tend to work with financial data.
Those skills are usefull for grad and research
Fun fact : GPT3 has vector embedding sizes of around 12,888 which is 100 times more than tiny models and 25 times more than normal NLP models.
Dumb question: are they interoperable? Like creating embeddings for a dataset with GPT-3 but then comparing them to a new embedding created by a different model
This channel is goldmine
Incroyable !
Ania is the best.
Great video very detailed 🎉 0:35
Wow mashallah you are amazing freecodecam we love you from Ethiopia 🇪🇹
I was take certificate of responsive web development that was amazing ❤❤❤
I just studying on this...Thanks
This is a very good video but I would like to understand why do we need datastax and store in DB if the intension is just to use as prompt and get answer. We can get directly from the OPenAI with key and prompt without storing or anything to do with vector embedding and those will be internal to OpenAI, I wanted to understand the use case of approach.
Valid question
I believe this method is better than fine-tuning and significantly superior to using prompts, especially when you have a lot of information; the chat will provide much better answers.
I think this is more for AI to answer questions on your data..hence she downloaded data from hugging face . But this could also be your own data vectorized..stored in db and queried . I may be wrong but this what I infer.
Thank you so much for this amazing video! I learnt a lots from it!
Thank you for sharing the information and knowledge.
Any educational course should always first explain what prerequisites are necessary to understand and learn the course material.
Think its safe to assume if youre here that you know a bit of CS at least.
Love it !
Thank you! this was really insightful.
@AniaKubow you're great at explaining. The only thing lacking in this otherwise excellent video is Poetry ;)
Great tutorial, thanks! Lol at those answers it was spitting out though...
Thanks for making this video
i like the course which is less than 1 hour
Exactly my thought
This came just in time as I just discovered Flowise which is just a code-less LangChain and wanted to play around with long term memory for my models
Thank you for tutorial. does some llms models and tools like chatgpt handles all the tasks related to from storing data to vectorbases and querying relevant data ? doesnt openai provide any database for storing the embedded text so that we used cassandra for this purpose?
Brilliant!
Amazing explanation! How could I use an existing Access database for my data set? It actually contains text reports and keywords for each report.
Awesome Video to get started in AI. Any reason why you used datastax instead of a vector DB like pinecone?
Thank you for so clearly and articulately presenting these lessons for us for free! Your eyes, your smile, and your beauty all together it is incredibly distracting to me though haha! I mean you are one of the most gorgeous women I've ever seen in my life and I appreciate your time. Sorry to come off like a creep but you're stunning.
tks a lot amazin video
say you wahnt the computah to scan this for whads...
i appreciate the video, and your accent.
This tutorial is very interesting, except that without a paid account on OpenAI we cannot really put it into practice. But I can't afford to pay 20 euros per month just to set up a tutorial.
You don't need to sign up for OpenAI Plus in order to create an API Key, they are billed separately. You also get a free 3-month API credit when you first create an account, the amount varies, I think they've decreased it now to about $5 (unfortunately I missed out on my credit, since I created my account last year and wasn't coding anything)
@@donaldoalmazan7338 thx for info
actually the 20$ for chatgpt is different from the api, for the api you can for example buy a 10$ credit and use it for as much as you would like, as long as it's used to finetune it will last long
I'm first woohoo tho I can only write "hello world" 🙂
Damn. Jay took one for the team learning vector embeddings 💀
When calculating cosine similarity, does a value closer to 1 mean more similar and a value closer to -1 mean less similar?
1 indicates an identical vector, or very close semantic meaning, or even identical text. (Note that it's the similarity of the vector's direction only, not scale.) 0 indicates an orthogonal relationship, ie, unrelated semantically. -1 in theory represents complete semantic opposition, but in practice, a perfect -1 is rare in natural language contexts.
When we do ask question using the vector database like "what are the biggest questions in science" does it consume token from open api also?
Wow. .Lovely
I'm new to this. At 13:18 what terminal is she using to input the code?
That is the simple Terminal. Later she uses Visual Studio Code
It is the linux terminal....curl is a linux command..but if u are on windows ..u may use wsl to use linux terminal on windows
Can we do this for image search? Can we see embeddings of images? Can langchain do that? Thanks
a git with that would be awesome
we have to have a GPT-4 prenium suscription fo follow this course ?
How do ML models create embeddings for new, or novel, words? For example, what if I fed it "Hexamethylenetetramine" (an organic compound)? My brain is frying thinking about this...
9:50 "...we also use it for information retrieval...": How does it deal with misspellings, either in the query or in the training data?
With a lower score than expected
The model has seen misspellings before, and it knows they are related to the correct spelling more than other words.
Amazing. How could it know? @@technolus5742
Is there a repo with the code?
Hey, I would love you could somehow make a video on bun and scyllaDB, been trying to learn them but theres no source 😥
10:51 THIS is a 'golden nugget' right here: "The core advantage of vector embeddings..." Such a great summation of exactly what an ai model really is. Thanks for such a fantastic video. I love it !
What is the prerequisites of this course?
passion to learn and explore
Link to a gist file with the code would be helpful please.
Is it possible to connect this to a custom GPT for the openai store?
Is there a tutorial on doing these using LLaMA?
Is python good language to learn dsa?.. because on the internet there are lot of guys who are telling that you should learn java/c++
Why to use OpenAI instead of some open source code?
download secure bundle is not showning and where to get the client secret id , please help
someone
I did not understand how the LLM (OpenAI) uses the embeddings stored in the DB.
LangChain: "chaining" resulted in two answers for each prompt: "I don't know" and the headlines. The first answer came from OpenAI's LLM, the second answers (the headlines) came from the vector-DB (Astra/Cassandra) that she set up outside OpenAI. LangChain was the bridge between the two.
It's a simple little example without much relevance, but it shows the bones. There is a lot more work to make something useful.
For instance, you could use a pre-trained LLM to perform the organizational tasks and composition (the language parts) using current data from a real-time source. For instance, "what kind of activities would be good at Los Angeles beaches today?"
The LLM could contextualize the meaning of the question using pre-trained data (an understanding of what constitutes beach sports and the condition necessary for each is something that won't much change over time), and then use an external source (weather channels, surf sites, diving data sites, sailing sites, etc.) to search for real-time conditions at LA Baches. The LLM - using the current real-time data - could then look for nearness matches based on how the conditions match up to certain sports.
So instead of a generic pre-trained answer like, "people like to sunbathe, swim, dive and surf at beaches" you can get a specific answer such as, "The conditions at Redondo Beach suggest it's a good surf day, but rip tide warnings suggest it is a bad day for people to be swimming. There is a water quality alert for bacteria in Santa Monica Bay."
The LLM used the external real-time data to give accurate point-in-time suggestions that it would otherwise never have using data from training months earlier. That's where LangChain can help - merging "new" or custom data into the pre-trained contextual LLM model.
Hope this helped.
help with this error on this part of code anyone?
llm = openai(openai_api_key=OPENAI_API_KEY)
TypeError: 'module' object is not callable
It seems secure bundle (20:36) option is no longer available. Can someone acknowledge it if it is so?
Yeh it seems to have moved location, open your table and on the right side under 'Database Details', 'Region' click the three dots and select download SCB :)
so fine
used the python code to get the embed, ran with no errors but returned no results?
Embeddings vs Fine tuning?
Too bad nowadays openAi won't get you use api key unless you are a paying customer
I was going to say, I cannot run the initial vector embedding program because of billing issues.
please guys make video on laravel react js
Vectors of 2704 table matrix
Id like for her to embed my vector
thank you ! sadly you dont go in deep into the needed data...hoy big are the documents etc...but still good thanks!
Can non-coder take this course?
did anyone run into this error
llm = openai(openai_api_key=OPEN_API_KEY)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
TypeError: 'module' object is not callable
does someone knows cheaper alternatives to OpenAI GPT's API ?
Google bard API
Make video about godot for unity users
How to use chat GPT API key
lol i am not able to find api keys and lclient secret keys for astra
4:40 "...Joe is 38 on the 0 to 100 scale... so -.4 on the -1 to 1 scale...": How is that? I get -.24. If it's -.4 on the -1 to 1 scale, that's 30 on the 0 to 100 scale. Please fix my math.
Context has two flavors, near and "not near". Joe is 38% near. Maybe Alice is 40% "not near", which would equate to a negative value (-.4).
So context is more than "this one is like the other", it's also "but it's not like this other thing". If we just used a single dimension, then literally everything would be "like" everything else, which makes it a little difficult to differentiate.
"The school bus is yellow. A banana is yellow. A bus is NOT LIKE a banana." Dimensions in dimensions.
Somebody having the problem of "You exceeded your current quota, please check your plan"
Great video, but @AniaKubow, if you do not mind, you could have had a very successful career in modelling. Refreshing to see that you chose computing and specifically AI.
perfect 👌
There are a lot of gaps in this tutorial, especially in the programming part. Good for conceptual learning, but don't recommend the implementation.
Whats wrong of her explanation?
She doesn't explain the basics. What is the terminal that she uses?
you can use a vscode terminal
@@chidiebere I'm using VS Code and I'm getting errors about the API key I created. Is there a way to validate a key?
Hello please help me how can run .jar file in HTML ,js please
It made me really laugh how she speaks about word and calls it "text". LOL
i don't code don't how to, i am here due to the thumbnail
👸🌟💝🌹💕🌹🌹🌟🌟
🎯 Key Takeaways for quick navigation:
00:14 📉 *Sam Altman was fired by OpenAI's board for not being consistently candid in his communications, leading to implications of lying by omission.*
01:11 🤯 *Various theories circulated, including speculation about dangerous AI developments, financial ties with Saudis, and a letter from former employees alleging dishonesty.*
01:53 🌐 *OpenAI employees expressed discontent, with over 500 threatening to quit, potentially joining Microsoft to dominate the AI space.*
02:21 🔄 *After negotiations with Microsoft failed, Altman and Brockman formed a new AI research team at Microsoft, but Altman eventually returned to OpenAI as CEO on November 21st.*
03:02 ❓ *Uncertainty remains about the true reasons behind Altman's firing, with speculation about conflicts of interest, AI commercialization, or a possible publicity stunt involving the board, Microsoft, and Altman.*
Made with HARPA AI
It is disappointing that you have taken the route of OpenAI which is far from open - not only have they not open-sourced their models, their payment methods are limited to a few options and setting up billing is required for their API to work. It would have been much better to use free, open-source models to demonstrate LangChain.
Ania looks fake. I don't know if it is the lighting or the background or what but it just looks fake. It's freaking me out. The voice sounds natural at least.
you did not explain cosine similarity and why this is a good definition of similarity when it comes to comparing vectors. Without understanding this, everything you say after it is incomprehensible
Very educative. Thank you for this video. We have prepared a Russian version: ruclips.net/video/q5dx9jiYFPM/видео.html.
how did you prepared a russian version?
are you Ai generated or an actual person.... say hi if you are AI and say dhinchak-dhinchak if you are actual persona...yeah audio will be better
yawn why is modern programming just gluing together packages. not this channel or the creator of the video's fault but i just hate what it's all become
most of this stuff is expensive to create
It takes time to create some of them if you were to code them from scratch. Also, you don't want to "reinvent the wheel"
Imagine kissing her
I need a code grAm that remembers geo locations to work with satelite for live location accuracy hitmarks