Zen van Riel
Zen van Riel
  • Видео 18
  • Просмотров 34 712
Speed up local AI by 50% using all your devices at once
Learn how to boost your local AI model performance by 50% using hardware you already own! This works for many large language models including Llama, Deepseek and more. In this project showcase, I'll demonstrate a local compute cluster using Exo. You will learn how to combine multiple hardware devices for faster AI processing.
⌛ Timestamps:
0:00 Introduction
0:25 Setting up a single node cluster
3:48 Testing two node cluster
6:53 Limitations and frustrations
🔧 Technology Stack:
Exo
github.com/exo-explore/exo
Просмотров: 1 432

Видео

Run AI models locally without an expensive GPU
Просмотров 9 тыс.День назад
Learn how to run state-of-the-art AI models completely free on your local machine - no expensive GPU required! In this step-by-step tutorial, I'll show you how to set up your own local AI lab environment using Docker. You will learn: - How to configure Docker for running AI models without a GPU - Setting up a lightweight but powerful 3GB language model - Running AI completions locally without a...
Build your private Google: self-hosted AI search in 10 minutes
Просмотров 4,4 тыс.14 дней назад
Learn how to create your own AI-powered search engine using Perplexica's open-source stack. In this step-by-step tutorial, I will teach you: - How to deploy a private search engine with source validation - Configure multiple search backends with SearXNG (Google, Bing, DuckDuckGo) ⌛ Timestamps 0:00 Introduction 0:12 Setting up Perplexica 2:10 Trying out the AI native search engine 4:41 Customizi...
Use AI to turn your resume into a job-landing portfolio
Просмотров 16421 день назад
Landing a top tier tech job in 2025 is not easy. I will help you by teaching you how to transform your static resume into a modern React portfolio site in minutes. Using Lovable AI, you'll get a production-ready React project you can actually build upon - not just another basic HTML template. In this tutorial, I'll show you: - How to bootstrap a portfolio site from your resume using AI - Why th...
Create AI teachers from ANY book using Docling
Просмотров 2,7 тыс.Месяц назад
Turn any ebook into an accurate AI-powered teacher/learning tool with Docling and Python! Full source code available at: github.com/Zenulous/booktutor-ai In this video I teach you how to: - Use Docling to process complex book content, including tables and diagrams - Create a local vector database for lightning-fast responses - Get AI-generated answers accurate to the book's actual content ⌛ Tim...
The secret to finding the best AI tools on GitHub
Просмотров 240Месяц назад
Most AI tools on GitHub are useless for real projects. Here's my engineer's guide to finding the ones that actually work. Timestamps ⌛ 00:00 Why most AI tools don't help 00:53 Pro tip 1: Use exact keywords 1:37 Pro tip 2: Filter on active repositories and stars 3:03 Pro tip 3: Make sure repos have open licenses 3:59 Selecting our tool of choice Key technical aspects covered: - Effective reposit...
Build your fast & free private ChatGPT AI clone in Javascript
Просмотров 266Месяц назад
Learn how to build your own AI chat interface like ChatGPT, Claude, or Bard using vanilla JavaScript code. This tutorial shows you how to create a professional chat UI with real-time streaming responses that works with any AI model. - Build a flexible chat UI that works with any AI model - Implement real-time streaming for instant responses - Written in vanilla JavaScript (but easily adaptable ...
How this AI gave $47k to a human
Просмотров 6062 месяца назад
I'll be talking about how Freysa AI gave prize money to a human as part of an interactive game. By using clever prompting, the prompt engineer managed to convince the artificial intelligence system to hand over the prize. But is everything as it seems, and is this story actually true? Let's look at it together. This video and my channel as a whole will help you understand how AI systems are bro...
Why you should use Aider for AI coding
Просмотров 6 тыс.2 месяца назад
This is how I use AI as a pair programmer to stop wasting time and make room for learning more as an engineer. Learn how I Aider so I can focus on becoming a better engineer. Check out Aider via aider.chat/ 📌 Timestamps: 0:00 - Why you should use an AI pair programmer 0:44 - Setting up Aider for your terminal 1:23 - Adding your code as context 2:53 - Aider explaining your code 4:30 - Pairing on...
NEW DeepSeek R1 model vs. Human Reasoning
Просмотров 2092 месяца назад
In this video, I test DeepSeek's new R1 model and see if it can solve a difficult brainteaser which requires reasoning to solve. Is it able to crack the puzzle? To test DeepSeek yourself, check out www.deepseek.com/. 📌 Timestamps: 0:00 - Introduction 0:32 - Testing the model's reasoning with a brainteaser Music track: Creamy by Aylex Source: freetouse.com/music Royalty Free Music for Videos (Sa...
Achieve responsible AI by fixing AI safety risks with Azure AI Studio
Просмотров 762 месяца назад
In this video I test my own AI system by exploiting it with Azure AI Studio's conversation simulator to ensure my model does not return malicious output. You should do the same if you are serious about getting your AI model into production! 📌 Timestamps: 0:00 - Introduction 0:28 - What is responsible AI? 1:37 - Exploiting the model with simulations 4:24 - Investigating the simulation output 5:2...
4 tips for AI at work as a Software Engineer
Просмотров 2952 месяца назад
Did you know that Samsung banned ChatGPT and other systems for AI coding after sensitive code was leaked? Companies are banning and blocking AI systems left and right. However, AI still offers great value for us as software engineers. With all these threats, you might wonder how you can still use AI at work. In this video, I'll share survival tips for using AI safely and respectfully at your jo...
Fast AI coding with Qwen 2.5 and Cursor
Просмотров 9 тыс.2 месяца назад
In this video, I will teach you how you can combine the AI native code editor Cursor with one of the best open-source AI models for coding today: Quen 2.5 30b! Other powerful local models are also supported. To learn about setting up LM Studio in 10 minutes, check out ruclips.net/video/f40iM0mt4ww/видео.html To download Ngrok, see ngrok.com/ 📌 Timestamps: 0:00 - Introduction 1:00 - Powerful mod...
Free local AI setup in 10 minutes (Llama, Mistral, and other LLMs)
Просмотров 6452 месяца назад
In this video I will teach you how you can download and run your own AI models for free on your local laptop/machine in only 10 minutes! The AI can be used for many automation tasks and can be programmed with as well. This video uses lmstudio.ai/ Models like Llama 3.2, Mistral, Phi 3.1, Gemma 2, DeepSeek 2.5, Qwen 2.5, and more are supported. 📌 Timestamps: 0:00 - Introduction & Setting Up LM St...
Measure AI costs to save money - GPT Cost Management Tutorial
Просмотров 473 месяца назад
In this video, you will learn how to effectively manage and measure the costs associated with using GPT and AI services. To make sure you understand the basics, I explain what tokens are and how you can measure the token usage of your AI application. Want to learn more? If you are interested in more AI content, leave a comment with your suggestions. #gpt #ai #softwareengineer
Personalized AI at scale with Vector Databases on Azure CosmosDB
Просмотров 1033 месяца назад
Personalized AI at scale with Vector Databases on Azure CosmosDB
How I use AI with my own data and Python as an AI Engineer
Просмотров 1813 месяца назад
How I use AI with my own data and Python as an AI Engineer
AI Solutions from POC to Production - Channel Intro
Просмотров 1383 месяца назад
AI Solutions from POC to Production - Channel Intro

Комментарии

  • @Adi-nf5kf
    @Adi-nf5kf Час назад

    Thank you. This is an ideal combination of hardware. Does the workflow gets distributed if one of the nodes go down?

  • @bigbrotherr
    @bigbrotherr 2 часа назад

    It's nice

  • @rahuldinesh2840
    @rahuldinesh2840 15 часов назад

    Is it any better than cursor or windsurf?

  • @pondopondo1497
    @pondopondo1497 18 часов назад

    Considering amd cards do not have nvlink analogs would be great to combine different radeon powered devices through exo

    • @zenvanriel
      @zenvanriel 17 часов назад

      You can definitely do that! I think newer NVIDIA chips also lack nvlink nowadays so there's use for both.

  • @matthewbond375
    @matthewbond375 2 дня назад

    Perplexica is awesome! To save some of you time, I've got best results using mistral-nemo 12B, with a 16K context window, as both the chat model and embedding model, via Ollama. Have fun! Great vid!

  • @gorevyoneticisi
    @gorevyoneticisi 3 дня назад

    Algorithm got me there. But this is perfect idea I liked it!

  • @SavunCheam
    @SavunCheam 3 дня назад

    Well done. But I am still wondering if there is a way to change model for completions too (code suggestion on typing). Thanks

  • @ConservativeWatchParty
    @ConservativeWatchParty 5 дней назад

    This video is hard to follow. I feel it would be easier if you cleared your screen and had less going on so that the viewers Eyes knew where to go. Showing unnecessary code all over your screen makes it very confusing for beginners to identify what they are supposed to be looking at. Also you may consider being more clear about which tabs you are pressingand navigating through your screen. I appreciate the video and I look forward to a less confusing one.

    • @zenvanriel
      @zenvanriel 5 дней назад

      Thanks for the constructive feedback! And I do agree, in my newer videos I tend to try and keep only two tabs open across the main content of the video. And if I do need more context I try to switch slower. Of course I can also start showing the shortcuts I'm pressing but just keeping the context smaller is a good idea. Just like with an LLM sometimes 😉

  • @Gatterman
    @Gatterman 6 дней назад

    Hello, im confused whether I can run any model on my Ryzen 3 3200G PC, with 16gb of ram or not?

    • @zenvanriel
      @zenvanriel 6 дней назад

      Did you try it out for yourself with the repository in the description? I expect Phi 3.5 to work "okay" on that machine, but not very fast.

  • @samueladam7693
    @samueladam7693 7 дней назад

    I don't ever usually comment, though the video is very informative, want to support you, keep going :)

  • @toby645pa
    @toby645pa 9 дней назад

    I always start my prompting with the question “Was 2024 a leap year”. On simple local machines it’s telling you fairy tales. That’s disappointing. But thanks for your good explanation 👏

    • @zenvanriel
      @zenvanriel 9 дней назад

      Yeah, for local devices a RAG implementation makes more sense. I'll explain that later on.

  • @ultrasound1459
    @ultrasound1459 9 дней назад

    The voice is not synchronised with your video

    • @zenvanriel
      @zenvanriel 9 дней назад

      Oops, looks like it is slightly off indeed. Will make it better next time.

  • @엠케이-p3p
    @엠케이-p3p 10 дней назад

    phi...ok but not that powerful, i thought there might be something i don't know in this video...i would rather use openwebui or anythingllm with ollama if you just need a chat interface

    • @zenvanriel
      @zenvanriel 9 дней назад

      That's right, I will show in later content that LocalAI allows you to use other models not just LLMs. If you just want a language model, you can use ollama

  • @nufh
    @nufh 10 дней назад

    Hey, thanks for the video.

  • @BadalPradhan-h1v
    @BadalPradhan-h1v 10 дней назад

    Kindly make a video how to setup and make own deepseek r1 api

    • @NightHawk_Pro
      @NightHawk_Pro 9 дней назад

      It's like 0.0014 per prompt just get it lol things unbelievable I was stuck for 3 days with clide and chatgpt and cascade. All of them. Could t handle the size of my algo any more. Thjs thing did what I was trying to figure out in three days. Did it in 3 prompts

    • @zenvanriel
      @zenvanriel 9 дней назад

      Deepseek r1, definitely the full model you can access via web/api, is really difficult to host on your own hardware. You'll need to start with a smaller model like Phi-3.5/4 or a smaller Llama model.

    • @NightHawk_Pro
      @NightHawk_Pro 8 дней назад

      I bought it and it's the best model hands down. Nothing compares ​@@zenvanriel

    • @BadalPradhan-h1v
      @BadalPradhan-h1v 7 дней назад

      @@zenvanriel I have installed deepseek r1 7b using lm studio, but how to create api to access in side my program

    • @zenvanriel
      @zenvanriel 5 дней назад

      LM studio has a server mode in it, check the main menu icons. It is the green one that looks like a CLI

  • @zensajnani
    @zensajnani 10 дней назад

    nice name zen

    • @zenvanriel
      @zenvanriel 10 дней назад

      We share the same first name based on your handle? If so, love it!

    • @zensajnani
      @zensajnani 10 дней назад

      @ yes we do 🤝🏼

  • @CarlosRodriguez-xk9ot
    @CarlosRodriguez-xk9ot 11 дней назад

    How can you make an interface or make it receive input files?

  • @CarlosRodriguez-xk9ot
    @CarlosRodriguez-xk9ot 11 дней назад

    Awesome content, could you make a video where you customize a reasoning model It further like connecting It to a folder of PDF files as database

    • @zenvanriel
      @zenvanriel 10 дней назад

      This is indeed all in the backlog of the next videos! Stay tuned.

  • @Ai-inarabic
    @Ai-inarabic 11 дней назад

    How can i contact u zen

    • @zenvanriel
      @zenvanriel 10 дней назад

      In a few weeks I'll make sure we can all get in touch. Stay tuned.

  •  11 дней назад

    the most comprehensive video on the subject !!!! thank you so much for this.. my question is. would it be possible to work in composer mode of cursor too?

    • @zenvanriel
      @zenvanriel 10 дней назад

      Unfortunately not at the moment.. I'm also hoping this will become available.

    •  10 дней назад

      @@zenvanriel too bad.. would it be somewhere in the source code, in an editable config file of cursor, where the model decides which api it will use in composer or chat? maybe thats where the hack is.

  • @alwayssmile1778
    @alwayssmile1778 11 дней назад

    Will try it surely!

  • @clemngene2313
    @clemngene2313 11 дней назад

    What's the difference between using docker and ollama? If you could give a detailed explanation, I'd like that

    • @zenvanriel
      @zenvanriel 11 дней назад

      The important distinction here would be that this guide shows you how to use LocalAI with Docker. Similarly, you can use Ollama with Docker. Docker is just meant to give you an easy one-command way to start an environment consistently. So I'm going to take the liberty to answer the difference between Ollama and LocalAI, because those are actually the systems you are interested in comparing. For large language models, Ollama is actually just as good if not even better supported by the community. You can use Ollama just fine in Docker as well: hub.docker.com/r/ollama/ollama However, LocalAI supports more types of AI models than just language models in one package. I aim to show more of those possibilities later on hence I focus on LocalAI. If you just want to use language models, definitely try out Ollama!

    • @clemngene2313
      @clemngene2313 11 дней назад

      @zenvanriel oh, LocalAI. I understand now. Thanks

  • @zenvanriel
    @zenvanriel 12 дней назад

    How does this model perform on your device? Let me know!

  • @1ndev
    @1ndev 12 дней назад

    Thank you for the heads up!

  • @thefoxialbunny
    @thefoxialbunny 15 дней назад

    May I do it not been a programmer/dev/coder?

  • @KresimirPendic
    @KresimirPendic 15 дней назад

    thanks for great video,, but can I use it with llama for example? (if running locally with ollama for example)

    • @zenvanriel
      @zenvanriel 15 дней назад

      Yes, ollama also exposes a server you can use. Check out github.com/ollama/ollama/blob/main/docs%2Fapi.md

    • @KresimirPendic
      @KresimirPendic 15 дней назад

      thank you very much@@zenvanriel I'll play with it - but I guess it will be less great for code generation as for example paid models..

  • @tanyongsheng4561
    @tanyongsheng4561 15 дней назад

    I am actually finding ways on how to self host it on cloud hosting via docker and traefik for quite some time, but still fails... 😂

  • @michaelrichey8516
    @michaelrichey8516 16 дней назад

    open webui has been able to do this for a while, with the added bonus of being able to connect to multiple ollama instances in a load-balanced connection. When the query/response is complete, you can turn the web search feature off and refine or discuss the content.

  • @sevenaac4783
    @sevenaac4783 16 дней назад

    "Clear and easy to understand, learn a lot in a short time. You're amazing!"

  • @tomwawer5714
    @tomwawer5714 16 дней назад

    I have been using it for a month and it’s great. Recent docker pull gives me server error though:( It’s also quite easy to access searxng container directly using simple api so no limit

    • @zenvanriel
      @zenvanriel 16 дней назад

      What docker command are you using?

    • @tomwawer5714
      @tomwawer5714 15 дней назад

      docker-compose up

  • @musprodev
    @musprodev 16 дней назад

    Nice bro, works like perplexity

  • @envt
    @envt 17 дней назад

    waarom gebruik je geen vscode?

    • @zenvanriel
      @zenvanriel 16 дней назад

      In a way, Cursor is just a vscode fork so I am using it 😄

  • @iham1313
    @iham1313 17 дней назад

    the image search found an image of a mockup, not the actual 5090 - which is a problem… not of the tool, but in general

    • @zenvanriel
      @zenvanriel 17 дней назад

      Yeah, that's true; as you say though that's the nature of image search these days. Not sure if you've noticed but a lot of the times when a movie comes out a new 'sequel' trailer is quickly generated with AI video tools. Screenshots from those also show in image search :/

  • @mohdjibly6184
    @mohdjibly6184 18 дней назад

    Great video...thanks for sharing bro

  • @gu9838
    @gu9838 18 дней назад

    ai agents are awesome!

  • @tecnopadre
    @tecnopadre 18 дней назад

    Why not better change epub to Markdown?

    • @zenvanriel
      @zenvanriel 18 дней назад

      That's also possible, as Docling supports markdown too. But generally when converting to Markdown you do get rid of complex formatting from the book which might be used for important content. It just depends on the book what the right approach is. Try it out I'd say!

  • @Алексей-й4з6ш
    @Алексей-й4з6ш 18 дней назад

    Wow what a lovely alternative! Nice structured video, like and sub!

  • @gubrans2091
    @gubrans2091 20 дней назад

    Can it help with other languages like c#

  • @burbilog
    @burbilog 22 дня назад

    careful with adding all files, because with any sufficiently large project you will run out of the context...

  • @GeneMathews
    @GeneMathews 23 дня назад

    I've always wondered how to properly showcase my projects. This seems like a great way to build a portfolio quickly.

  • @DUSTYPruitt
    @DUSTYPruitt 23 дня назад

    This tutorial is super helpful. I've been struggling to make my resume stand out, so the idea of using AI to create a portfolio is interesting.

  • @SahilMalik-ss6qi
    @SahilMalik-ss6qi 23 дня назад

    Great content, best video on vector database with example. I was searching for vector db integration in azure cosmos db and randomly found this video, thank you so much.

    • @zenvanriel
      @zenvanriel 23 дня назад

      Thanks so much! Any other topics you'd love to see covered?

    • @SahilMalik-ss6qi
      @SahilMalik-ss6qi 22 дня назад

      @zenvanriel Training on GEN AI with real world example implementation

  • @phoenyfeifei
    @phoenyfeifei 23 дня назад

    3:02 ok, that makes sense, i was wondering why you revert forward your ports since you can alerady run a serve locally

  • @b.c.2177
    @b.c.2177 25 дней назад

    I installed Qwen/Qwen2.5-Coder-3B-Instruct-GGUF for test and it replies with strange code. Looks like encrypted information :(

    • @zenvanriel
      @zenvanriel 24 дня назад

      Perhaps get a slightly larger model than 3b. What programming language are you writing code with?

  • @hamidsadeghijabali
    @hamidsadeghijabali 27 дней назад

    Hi, I did what you told in this video but I get the following errors! ❓ Ask a question: what is the name of the book 🔄 Processing your question... Traceback (most recent call last): File "C:\Program Files\Python312\Lib\site-packages\httpx\_transports\default.py", line 101, in map_httpcore_exceptions yield File "C:\Program Files\Python312\Lib\site-packages\httpx\_transports\default.py", line 250, in handle_request resp = self._pool.handle_request(req) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpcore\_sync\connection_pool.py", line 256, in handle_request raise exc from None File "C:\Program Files\Python312\Lib\site-packages\httpcore\_sync\connection_pool.py", line 236, in handle_request response = connection.handle_request( ^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpcore\_sync\connection.py", line 101, in handle_request raise exc File "C:\Program Files\Python312\Lib\site-packages\httpcore\_sync\connection.py", line 78, in handle_request stream = self._connect(request) ^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpcore\_sync\connection.py", line 124, in _connect stream = self._network_backend.connect_tcp(**kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpcore\_backends\sync.py", line 207, in connect_tcp with map_exceptions(exc_map): ^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\contextlib.py", line 158, in __exit__ self.gen.throw(value) File "C:\Program Files\Python312\Lib\site-packages\httpcore\_exceptions.py", line 14, in map_exceptions raise to_exc(exc) from exc httpcore.ConnectError: [WinError 10061] No connection could be made because the target machine actively refused it The above exception was the direct cause of the following exception: Traceback (most recent call last): File "C:\Program Files\Python312\Lib\site-packages\openai\_base_client.py", line 996, in _request response = self._client.send( ^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpx\_client.py", line 914, in send response = self._send_handling_auth( ^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpx\_client.py", line 942, in _send_handling_auth response = self._send_handling_redirects( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpx\_client.py", line 979, in _send_handling_redirects response = self._send_single_request(request) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpx\_client.py", line 1014, in _send_single_request response = transport.handle_request(request) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpx\_transports\default.py", line 249, in handle_request with map_httpcore_exceptions(): ^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\contextlib.py", line 158, in __exit__ self.gen.throw(value) File "C:\Program Files\Python312\Lib\site-packages\httpx\_transports\default.py", line 118, in map_httpcore_exceptions raise mapped_exc(message) from exc httpx.ConnectError: [WinError 10061] No connection could be made because the target machine actively refused it The above exception was the direct cause of the following exception: Traceback (most recent call last): File "C:\My Git\booktutor-ai\booktutor.py", line 213, in <module> main() File "C:\My Git\booktutor-ai\booktutor.py", line 200, in main result = qa_system.invoke( ^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 170, in invoke raise e File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 160, in invoke self._call(inputs, run_manager=run_manager) File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\conversational_retrieval\base.py", line 170, in _call answer = self.combine_docs_chain.run( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain_core\_api\deprecation.py", line 182, in warning_emitting_wrapper return wrapped(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 611, in run return self(kwargs, callbacks=callbacks, tags=tags, metadata=metadata)[ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain_core\_api\deprecation.py", line 182, in warning_emitting_wrapper return wrapped(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 389, in __call__ return self.invoke( ^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 170, in invoke raise e File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 160, in invoke self._call(inputs, run_manager=run_manager) File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\combine_documents\base.py", line 138, in _call output, extra_return_dict = self.combine_docs( ^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\combine_documents\stuff.py", line 259, in combine_docs return self.llm_chain.predict(callbacks=callbacks, **inputs), {} ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\llm.py", line 318, in predict return self(kwargs, callbacks=callbacks)[self.output_key] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain_core\_api\deprecation.py", line 182, in warning_emitting_wrapper return wrapped(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 389, in __call__ return self.invoke( ^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 170, in invoke raise e File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 160, in invoke self._call(inputs, run_manager=run_manager) File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\llm.py", line 126, in _call response = self.generate([inputs], run_manager=run_manager) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\llm.py", line 138, in generate return self.llm.generate_prompt( ^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain_core\language_models\chat_models.py", line 786, in generate_prompt return self.generate(prompt_messages, stop=stop, callbacks=callbacks, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain_core\language_models\chat_models.py", line 643, in generate raise e File "C:\Program Files\Python312\Lib\site-packages\langchain_core\language_models\chat_models.py", line 633, in generate self._generate_with_cache( File "C:\Program Files\Python312\Lib\site-packages\langchain_core\language_models\chat_models.py", line 851, in _generate_with_cache result = self._generate( ^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain_openai\chat_models\base.py", line 717, in _generate response = self.client.create(**payload) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai\_utils\_utils.py", line 279, in wrapper return func(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai esources\chat\completions.py", line 859, in create return self._post( ^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai\_base_client.py", line 1283, in post return cast(ResponseT, self.request(cast_to, opts, stream=stream, stream_cls=stream_cls)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai\_base_client.py", line 960, in request return self._request( ^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai\_base_client.py", line 1020, in _request return self._retry_request( ^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai\_base_client.py", line 1098, in _retry_request return self._request( ^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai\_base_client.py", line 1020, in _request return self._retry_request( ^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai\_base_client.py", line 1098, in _retry_request return self._request( ^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai\_base_client.py", line 1030, in _request raise APIConnectionError(request=request) from err openai.APIConnectionError: Connection error.

    • @zenvanriel
      @zenvanriel 26 дней назад

      It seems it couldn't connect to the language model. Did you start a local language model server using LM Studio? You can have a look at ruclips.net/video/f40iM0mt4ww/видео.htmlsi=LWXqRGFWNZszCzch to set this up

  • @toolbelt
    @toolbelt 27 дней назад

    Wow. That it so cool! Thank you for sharing this. I love it.

  • @sammyvandenburg
    @sammyvandenburg 27 дней назад

    Zen, you just keeps casting pearls. You deserve so many more subscribers.

  • @andrew.derevo
    @andrew.derevo 28 дней назад

    Thanks for a video. what model do you use with aider? looks really fast

    • @zenvanriel
      @zenvanriel 27 дней назад

      In this video I am using Azure OpenAI's gpt4o model! :) What is your daily driver?

  • @nofchi
    @nofchi 29 дней назад

    Well done, thanks :) One out of scope question - you have nicely optimized 'ls' command. You use it like 1:10 of the video. Could you please share your settings?

    • @zenvanriel
      @zenvanriel 29 дней назад

      I got the config from a fellow creator, specifically this video: ruclips.net/video/LnZdaNfQ86o/видео.htmlfeature=shared

    • @nofchi
      @nofchi 25 дней назад

      @@zenvanriel Amazing, thank you! Thank you for your videos. They are very useful.

  • @yokuna-b5m
    @yokuna-b5m 29 дней назад

    Hi guys! Thank you for your great video. And have a question , my macbook pro m1 (RAM : 16GB) , download Qwen 2.5 32b, but can't load this model in LM studio. please hep me. I wan know your pc or laptap spec.

    • @zenvanriel
      @zenvanriel 29 дней назад

      You'll likely need more RAM, I've got 64GB for this demo video. How about trying a smaller sized model? For example 3B.