THE 🐐 I became the Python developer I am today because of this channel. From learning Python for my AS level exams in 2020, to an experienced backend developer. From the bottom of my heart, Thank You Tim. I'm watching this video because I have entered a Hackathon that requires something similar. This channel has never failed me.
The context will fill up the context windows very fast. You can store the conversations embedings with the messages in a vector database and pull the related parts from it.
Yes but that's a bit beyond this video. But I guess he should quickly mention there is a memory limit. But storing in vector is a whole other beast I'm looking to get into next with langxhain 😂
@@Larimuss It is not that hard. I coded it locally and store them in json file. You just store the embeding with the messages then you create the new message embedings and with cosine distance you grab the most matching 10-20 messages. It is less then 100 lines. this is the distance fucntion: np.dot(v1, v2)/(norm(v1)*norm(v2)) . I also summerize the memories with llm too so I can get a shorter length.
New to the world of coding. Teaching myself through YT for now and this guy is clearly S Tier. I like him and Programming with Moshs' tutorials . Any other recommendations? I'd prefer more vids like this with actual walkthroughs on my feed.
idk but I never understood anything from programming with Mosh videos. Tim is a way better explainer for me, especially that 9 hour from beginner to advanced video.
Wow thanks! This is really simple, straightforward guide to start me getting into writting the python rather than just using peoples UI. Love the explanations.
Thanks to your tutorial I recreated Jarvis with a custom GUI and using llama3 model to make Jarvis, i use it in italian cuz i'm italian, but you can also use it in english and other languages.
@@akhilpadmanaban3242with llama yes as they run locally and well you are not using apis. But they are pretty resource consuming..tried it and they couldn't run
Hey man thanks a Lot, could you explain how to implement own Data, PDF, web sources etc. for giving answers when I need to give it some more detailed knowledge about certain internal information about possible questions regarding my use Case?
Adding a context, of course, generates interesting results: context": "Hot and Humid Summer" --> chain invoke result = To be honest, I'm struggling to cope with this hot and humid summer. The heat and humidity have been really draining me lately. It feels like every time I step outside, I'm instantly soaked in sweat. I just wish it would cool down a bit! How about you? ...🥵
Thanks, super useful and simple! I just wondered with the new Llama model coming out, how I could best use it - so perfect timing xD Would have added that Llama is made by Meta - so despite being free, it's compareable to the latest OpenAI models.
Hi Tim, I recently completed your video on django-react project , but i need an urgent help from your side if you can make a video on how to deploy django-react project on vercel,render etc. or other known platform (this would really be helpful as there are many users on the django forum still confused on deployment of django-react project to some popular web deployment sites. Kindly help into this.
If you read my message, thank you for teaching and would you mind teaching me more about fine-tune? What should I do? (I want Tensorflow) and I want it to be able to learn what I can't answer by myself. What should I do?
hello tim! this video is awesome, but the only problem i have is that the ollama chatbot is responding very slowly, do you have any idea on how to fix this?
Great video learned a lot. Can you advise me the route I would take if I wanted to build a chatbot around a specific niche like comedy. build an app that I could sell or give away for free. I would need to train the model on the specific niche and that niche only. Then host it on a server I would think. An outline on these steps would be much appreciated.
Please Tim help me how to add long term (infact ultra long) memory to my cool AI agent using only ollama and rich library. May be memgpt will be nice approach. Please help me!
not na ai expert so i could said something wrong: you mean the ai remeber things from messagrd way back in the conversation? if so thats called context of the ai, and is limited by the training and is also an area of current developpement, on the other hand tim is just making an intrface for already trained ai
@@birdbeakbeardneck3617 I know that bro but I want custom solutions for what I said, like vector database or postgre, the fact is I don't know how to use them, the tutorials are not streight forward unlike Tim's tutorial also docs are not able to provide me specific solution. Yes I know after reading docs I will be able to do that but I have very little time (3 Days), and under these days I will have to add 7 tools to the AI agent. Otherwise I'm continuously trying to do that. ❤️ If you can help me through any article or blog or email, please do that 🙏❤️
Great video! Is there any way to connect a personal database to this model? (So that the chat can answer questions based on the information in the database). I have a database in Postgre, already used RAG on it, but I have no idea how to connect the db and the chat. Any ideas?
Hey, Tim! Thanks for your tutorial. A haver a problem. Bot isn't responding to me? Mabe someone else have the same problem. Give me some feedback, please
which version of the python did you use ? trying to navigate version 3.13. in the line it shows for when inside the terminal C:\Users\Computer\.ollama> I may have done it wrong.
hello do you know if its possible to use this model as a "pre trained" one, and add some new let say.. local information to the model to use it for an specific task?
I have not implemented myself, but I have doubt, you are using langchain where the model is llama 3.1, langchain manages everything here, then what's the use of Ollama ?
You may find it 'amusing' or 'interesting' that when I (nihilistically) prompted with "Hello Cruel World!', 'llama3.1:8b' responded: " A nod to the Smiths' classic song, 'How Soon is Now?' (also known as 'Hello, Hello, How are You?') " !?!?!🤣
I had been using the program Ollama on my laptop, and it was utilizing 101% of my CPU's processing power. This excessive usage threatened to overheat my device and decrease its performance. Therefore, I decided that I would discontinue using the program.
It was a great tutorial and I follow it properly but still I am getting an error : ConnectError: [WinError 10061] No connection could be made because the target machine actively refused it I am running this code on my office machine which has restricted the openai models and Ai site
If you need to work with large amounts of data OpenAI performance still can't be matched locally, unless you spend a ridiculous amount on your computer build.
Hey how can i show this as a ui I want to create a chatbot which can provide me the programming related ans with user authentication otp Please tell me how can i create this by using this model And create my UI i am a full stack developer ane new to ml please reply
PS C:\Windows\system32> ollama pull llama3 Error: could not connect to ollama app, is it running? what seems to be wrong? (sorry for hte noob question)
THE 🐐 I became the Python developer I am today because of this channel. From learning Python for my AS level exams in 2020,
to an experienced backend developer. From the bottom of my heart, Thank You Tim. I'm watching this video because I have entered a Hackathon that requires something similar. This channel has never failed me.
Whenever I get a idea this guy makes a video about it
Me too 😂
You are BRILLIANT @umeshlab987
We are one
That's right!
*an idea
already used it to set up a Q&A system to answer customers' common questions. Thank you so much for the sharing and demo.
Thanks for saving the day. i been following your channel for four years now
The captions with keywords are like built-in notes, thanks for doing that
for some Window users, if all of the commands don't work for you, try source name/Scripts/activate to activate the venv.
The context will fill up the context windows very fast. You can store the conversations embedings with the messages in a vector database and pull the related parts from it.
Yes but that's a bit beyond this video. But I guess he should quickly mention there is a memory limit. But storing in vector is a whole other beast I'm looking to get into next with langxhain 😂
@@Larimuss It is not that hard. I coded it locally and store them in json file. You just store the embeding with the messages then you create the new message embedings and with cosine distance you grab the most matching 10-20 messages. It is less then 100 lines. this is the distance fucntion: np.dot(v1, v2)/(norm(v1)*norm(v2)) . I also summerize the memories with llm too so I can get a shorter length.
@@krisztiankoblos1948 This would be awesome to learn how to implement. Do you have any recommendations on tutorials for this?
@@krisztiankoblos1948Hi! Do u have a repo to share? Sounds interesting!
@@krisztiankoblos1948 Brother, you are beautiful.
Awesome that was "the tutorial of the month" from you tim !!! because you didn't use some sponsored tech stack ! they usually are terrable !
New to the world of coding. Teaching myself through YT for now and this guy is clearly S Tier.
I like him and Programming with Moshs' tutorials . Any other recommendations? I'd prefer more vids like this with actual walkthroughs on my feed.
idk but I never understood anything from programming with Mosh videos. Tim is a way better explainer for me, especially that 9 hour from beginner to advanced video.
Bro code is GOAT 🐐
@@M.V.CHOWDARI Appreciate it!
Wow thanks! This is really simple, straightforward guide to start me getting into writting the python rather than just using peoples UI. Love the explanations.
Thanks to your tutorial I recreated Jarvis with a custom GUI and using llama3 model to make Jarvis, i use it in italian cuz i'm italian, but you can also use it in english and other languages.
these models beiung completely free?
@@akhilpadmanaban3242with llama yes as they run locally and well you are not using apis. But they are pretty resource consuming..tried it and they couldn't run
@@akhilpadmanaban3242 Yes
Very much enjoyed your instruction style - subscribed!
This just inspired me saving GPT Costs for our SaaS Product, Thanks Tim!
hey i'm into saas too did u make any project yet?
Hey man thanks a Lot, could you explain how to implement own Data, PDF, web sources etc. for giving answers when I need to give it some more detailed knowledge about certain internal information about possible questions regarding my use Case?
Do you have a video on fine-tuning or prompt engineering? I don't want it to be nameless please.😅
This is what I need right now!!! Thank you CS online mentor!
Wow, so cool ! You really nailed the tutorial🎉
Awesomesauce! Tim make more vids covering LangChain projects please and maybe an in depth tutorial! ❤🎉
Fantastic explanation - thank you for this
Timmy! Great explanation, concise and to the point. Keep 'em coming boss =).
Tech With Tim is my favorite.
Can I ask who is in 2nd and 3rd?
@@WhyHighC 1: tim 2: tim 3: tim
Could you please tell us about how to create a fine tunning chatbot using our own dataset.
Hi Tim - Now we can download Llama3.1 too... By the way can u also convert this to UI using streamlit
Adding a context, of course, generates interesting results: context": "Hot and Humid Summer" --> chain invoke result = To be honest, I'm struggling to cope with this hot and humid summer. The heat and humidity have been really draining me lately. It feels like every time I step outside, I'm instantly soaked in sweat. I just wish it would cool down a bit! How about you? ...🥵
very helpful video Tim !
Thanks, super useful and simple!
I just wondered with the new Llama model coming out, how I could best use it - so perfect timing xD
Would have added that Llama is made by Meta - so despite being free, it's compareable to the latest OpenAI models.
Thanks Tim, ran into buncha errors when running the sciprt. Guess who came to my rescue, chatGPT :)
lol thumbnail had me thinking there was gonna be a custom UI with the script
If you combine this with a webview you can make a sorta of artifact in your local app
Simple and useful! Great content! :)
Awesome.....i really needed a replica of chatbot for a project and this worked perfectly....thank you
Thank you very much for the video, i'm gonna try that :)
This is what i need thank you bro ❤
Hi Tim,
I recently completed your video on django-react project , but i need an urgent help from your side if you can make a video on how to deploy django-react project on vercel,render etc. or other known platform (this would really be helpful as there are many users on the django forum still confused on deployment of django-react project to some popular web deployment sites.
Kindly help into this.
This is Very useful content Keep it up
If you read my message, thank you for teaching and would you mind teaching me more about fine-tune? What should I do? (I want Tensorflow) and I want it to be able to learn what I can't answer by myself. What should I do?
I love how you make it easy for us.
After that we need an UI and bingo.
Btw, does it keep the answers in memory after we exit? Don't think so, right?
based on the code, no. only a single runtime
Great video, thank you very much!
hello tim! this video is awesome, but the only problem i have is that the ollama chatbot is responding very slowly, do you have any idea on how to fix this?
Great video learned a lot. Can you advise me the route I would take if I wanted to build a chatbot around a specific niche like comedy. build an app that I could sell or give away for free. I would need to train the model on the specific niche and that niche only. Then host it on a server I would think. An outline on these steps would be much appreciated.
Please Tim help me how to add long term (infact ultra long) memory to my cool AI agent using only ollama and rich library. May be memgpt will be nice approach. Please help me!
not na ai expert so i could said something wrong:
you mean the ai remeber things from messagrd way back in the conversation? if so thats called context of the ai, and is limited by the training and is also an area of current developpement, on the other hand tim is just making an intrface for already trained ai
@@birdbeakbeardneck3617 I know that bro but I want custom solutions for what I said, like vector database or postgre, the fact is I don't know how to use them, the tutorials are not streight forward unlike Tim's tutorial also docs are not able to provide me specific solution. Yes I know after reading docs I will be able to do that but I have very little time (3 Days), and under these days I will have to add 7 tools to the AI agent. Otherwise I'm continuously trying to do that. ❤️ If you can help me through any article or blog or email, please do that 🙏❤️
Thx. Tim ! Now, llama3.1 is available under Ollama, It generates great results and has a large context memory !
@@davidtindell950 But bro my project is accordingly that can't depend on the LLM's context memory. Please tell me if you can help me with that!
@@siddhubhai2508 I have found FAISS vector store provides an effective and large capacity "persistent memory" with CUDA GPU support.
...
This is great! thanks
Great video! Is there any way to connect a personal database to this model? (So that the chat can answer questions based on the information in the database). I have a database in Postgre, already used RAG on it, but I have no idea how to connect the db and the chat. Any ideas?
Cool!! Could I get this to summarize my e-library?
This is swag, how can we create a custom personality for the llama3 model?
Thank you so much!!
Hey there, is your VSCode theme public? It's really nice, would love to have it to customize
Will this run on an android tablet?
Useful. keep doing
Amazing!
how can this be moved from locally to on an internal website
should i install ollama in a virtural env?
Hey, Tim! Thanks for your tutorial. A haver a problem. Bot isn't responding to me? Mabe someone else have the same problem. Give me some feedback, please
Can You teach us how to implement it in GUI form, i don't want to run the program every time i want help of this type things
Nice one
Hello! Tim when i run ollama directly there is no delay in response but using script with langchain some delay appear. Why is that? How to solve it?
which version of the python did you use ?
trying to navigate version 3.13.
in the line it shows for when inside the terminal C:\Users\Computer\.ollama>
I may have done it wrong.
Can you show us how to do RAG with llama3?
hello do you know if its possible to use this model as a "pre trained" one, and add some new let say.. local information to the model to use it for an specific task?
thank you.
Thank You.
I have not implemented myself, but I have doubt, you are using langchain where the model is llama 3.1, langchain manages everything here, then what's the use of Ollama ?
the langchain simplifies interactions with LLM's, it doesn't provide the LLM. We use Ollama to get the LLM
Is there any way to make python script to automatically train a locally-ran model?
How much ram required to make this program running well? Cause i have 4GB ram only
How do you get Local LLM to show? I don’t have that in my VS Code
You may find it 'amusing' or 'interesting' that when I (nihilistically) prompted with "Hello Cruel World!', 'llama3.1:8b' responded: " A nod to the Smiths' classic song, 'How Soon is Now?' (also known as 'Hello, Hello, How are You?') " !?!?!🤣
Why does microsoft publisher window keep popping up saying unlicensed product and will not allow it to run?
Where do you get all this stuff from
Hi, I have tried this and its working, but the model is taking long response time anything I can do for reducing that?
I had been using the program Ollama on my laptop, and it was utilizing 101% of my CPU's processing power. This excessive usage threatened to overheat my device and decrease its performance. Therefore, I decided that I would discontinue using the program.
is it possible to host this in a cloud server? so that i can access my custom bot whenever i want?
how coincidental, i made this project just 2 days ago
A dummy question.. Where is used the template ?
can i train this model? give him information that he can answer to me before?
do i need to install longchain?
It was a great tutorial and I follow it properly but still I am getting an error :
ConnectError: [WinError 10061] No connection could be made because the target machine actively refused it
I am running this code on my office machine which has restricted the openai models and Ai site
what's your pc specs sir?
Tim this ollama is running on my cpu and hence really slow can I make it run on my GPU somehow?
If you need to work with large amounts of data OpenAI performance still can't be matched locally, unless you spend a ridiculous amount on your computer build.
it can be matched by running llama 3.1 405B model !.
i dont know what is happening when i run python file in cmd it shows me hello world then the command ends
thx ;)
Hey how can i show this as a ui
I want to create a chatbot which can provide me the programming related ans with user authentication otp
Please tell me how can i create this by using this model
And create my UI i am a full stack developer ane new to ml please reply
Does anybody know what type of data the llama software is exchanging ?
Hi. Is there a way to uninstall llama3 again?
Does respose speed of AI bot depend on gpu like llama ?
YES
Nice
how can we stream output ??
what's the minimum hardware requirement? thank you!
8GB RAM
@@gunabaki7755 no discrete GPU needed?
do i need vram 4 this ?
How do I activate this on windows ??
Anyone knows Tim HW specs?
PS C:\Windows\system32> ollama pull llama3
Error: could not connect to ollama app, is it running?
what seems to be wrong? (sorry for hte noob question)
you need to run the ollama application first, it usually starts when u boot up ur pc
@@gunabaki7755 will try ths thanks bro!
Hello there ... Can i e-mail you ? I am facing a problem and really could use some help
Sadly even though I have 32GB of ram, the 7B "llama3" takes up to 1 minute to answer
this context thing is not working, the bot does not know what was earlier in the conversation
❤🎉
How do i deploy it to my website?
I think u can try to convert it into API using FASTAPI and call the API from frontend
where is the script bro?
can i embed this chatbot into a website? doing this for an assignment
does it require gpu ?
depends on size of the model, smaller models don't require