- Видео 18
- Просмотров 34 712
Zen van Riel
Нидерланды
Добавлен 8 окт 2024
Hi, I'm Zen! I help you get AI solutions from concept to production.
Views expressed are my own and do not reflect those of my employer.
Views expressed are my own and do not reflect those of my employer.
Speed up local AI by 50% using all your devices at once
Learn how to boost your local AI model performance by 50% using hardware you already own! This works for many large language models including Llama, Deepseek and more. In this project showcase, I'll demonstrate a local compute cluster using Exo. You will learn how to combine multiple hardware devices for faster AI processing.
⌛ Timestamps:
0:00 Introduction
0:25 Setting up a single node cluster
3:48 Testing two node cluster
6:53 Limitations and frustrations
🔧 Technology Stack:
Exo
github.com/exo-explore/exo
⌛ Timestamps:
0:00 Introduction
0:25 Setting up a single node cluster
3:48 Testing two node cluster
6:53 Limitations and frustrations
🔧 Technology Stack:
Exo
github.com/exo-explore/exo
Просмотров: 1 432
Видео
Run AI models locally without an expensive GPU
Просмотров 9 тыс.День назад
Learn how to run state-of-the-art AI models completely free on your local machine - no expensive GPU required! In this step-by-step tutorial, I'll show you how to set up your own local AI lab environment using Docker. You will learn: - How to configure Docker for running AI models without a GPU - Setting up a lightweight but powerful 3GB language model - Running AI completions locally without a...
Build your private Google: self-hosted AI search in 10 minutes
Просмотров 4,4 тыс.14 дней назад
Learn how to create your own AI-powered search engine using Perplexica's open-source stack. In this step-by-step tutorial, I will teach you: - How to deploy a private search engine with source validation - Configure multiple search backends with SearXNG (Google, Bing, DuckDuckGo) ⌛ Timestamps 0:00 Introduction 0:12 Setting up Perplexica 2:10 Trying out the AI native search engine 4:41 Customizi...
Use AI to turn your resume into a job-landing portfolio
Просмотров 16421 день назад
Landing a top tier tech job in 2025 is not easy. I will help you by teaching you how to transform your static resume into a modern React portfolio site in minutes. Using Lovable AI, you'll get a production-ready React project you can actually build upon - not just another basic HTML template. In this tutorial, I'll show you: - How to bootstrap a portfolio site from your resume using AI - Why th...
Create AI teachers from ANY book using Docling
Просмотров 2,7 тыс.Месяц назад
Turn any ebook into an accurate AI-powered teacher/learning tool with Docling and Python! Full source code available at: github.com/Zenulous/booktutor-ai In this video I teach you how to: - Use Docling to process complex book content, including tables and diagrams - Create a local vector database for lightning-fast responses - Get AI-generated answers accurate to the book's actual content ⌛ Tim...
The secret to finding the best AI tools on GitHub
Просмотров 240Месяц назад
Most AI tools on GitHub are useless for real projects. Here's my engineer's guide to finding the ones that actually work. Timestamps ⌛ 00:00 Why most AI tools don't help 00:53 Pro tip 1: Use exact keywords 1:37 Pro tip 2: Filter on active repositories and stars 3:03 Pro tip 3: Make sure repos have open licenses 3:59 Selecting our tool of choice Key technical aspects covered: - Effective reposit...
Build your fast & free private ChatGPT AI clone in Javascript
Просмотров 266Месяц назад
Learn how to build your own AI chat interface like ChatGPT, Claude, or Bard using vanilla JavaScript code. This tutorial shows you how to create a professional chat UI with real-time streaming responses that works with any AI model. - Build a flexible chat UI that works with any AI model - Implement real-time streaming for instant responses - Written in vanilla JavaScript (but easily adaptable ...
How this AI gave $47k to a human
Просмотров 6062 месяца назад
I'll be talking about how Freysa AI gave prize money to a human as part of an interactive game. By using clever prompting, the prompt engineer managed to convince the artificial intelligence system to hand over the prize. But is everything as it seems, and is this story actually true? Let's look at it together. This video and my channel as a whole will help you understand how AI systems are bro...
Why you should use Aider for AI coding
Просмотров 6 тыс.2 месяца назад
This is how I use AI as a pair programmer to stop wasting time and make room for learning more as an engineer. Learn how I Aider so I can focus on becoming a better engineer. Check out Aider via aider.chat/ 📌 Timestamps: 0:00 - Why you should use an AI pair programmer 0:44 - Setting up Aider for your terminal 1:23 - Adding your code as context 2:53 - Aider explaining your code 4:30 - Pairing on...
NEW DeepSeek R1 model vs. Human Reasoning
Просмотров 2092 месяца назад
In this video, I test DeepSeek's new R1 model and see if it can solve a difficult brainteaser which requires reasoning to solve. Is it able to crack the puzzle? To test DeepSeek yourself, check out www.deepseek.com/. 📌 Timestamps: 0:00 - Introduction 0:32 - Testing the model's reasoning with a brainteaser Music track: Creamy by Aylex Source: freetouse.com/music Royalty Free Music for Videos (Sa...
Achieve responsible AI by fixing AI safety risks with Azure AI Studio
Просмотров 762 месяца назад
In this video I test my own AI system by exploiting it with Azure AI Studio's conversation simulator to ensure my model does not return malicious output. You should do the same if you are serious about getting your AI model into production! 📌 Timestamps: 0:00 - Introduction 0:28 - What is responsible AI? 1:37 - Exploiting the model with simulations 4:24 - Investigating the simulation output 5:2...
4 tips for AI at work as a Software Engineer
Просмотров 2952 месяца назад
Did you know that Samsung banned ChatGPT and other systems for AI coding after sensitive code was leaked? Companies are banning and blocking AI systems left and right. However, AI still offers great value for us as software engineers. With all these threats, you might wonder how you can still use AI at work. In this video, I'll share survival tips for using AI safely and respectfully at your jo...
Fast AI coding with Qwen 2.5 and Cursor
Просмотров 9 тыс.2 месяца назад
In this video, I will teach you how you can combine the AI native code editor Cursor with one of the best open-source AI models for coding today: Quen 2.5 30b! Other powerful local models are also supported. To learn about setting up LM Studio in 10 minutes, check out ruclips.net/video/f40iM0mt4ww/видео.html To download Ngrok, see ngrok.com/ 📌 Timestamps: 0:00 - Introduction 1:00 - Powerful mod...
Free local AI setup in 10 minutes (Llama, Mistral, and other LLMs)
Просмотров 6452 месяца назад
In this video I will teach you how you can download and run your own AI models for free on your local laptop/machine in only 10 minutes! The AI can be used for many automation tasks and can be programmed with as well. This video uses lmstudio.ai/ Models like Llama 3.2, Mistral, Phi 3.1, Gemma 2, DeepSeek 2.5, Qwen 2.5, and more are supported. 📌 Timestamps: 0:00 - Introduction & Setting Up LM St...
Measure AI costs to save money - GPT Cost Management Tutorial
Просмотров 473 месяца назад
In this video, you will learn how to effectively manage and measure the costs associated with using GPT and AI services. To make sure you understand the basics, I explain what tokens are and how you can measure the token usage of your AI application. Want to learn more? If you are interested in more AI content, leave a comment with your suggestions. #gpt #ai #softwareengineer
Personalized AI at scale with Vector Databases on Azure CosmosDB
Просмотров 1033 месяца назад
Personalized AI at scale with Vector Databases on Azure CosmosDB
How I use AI with my own data and Python as an AI Engineer
Просмотров 1813 месяца назад
How I use AI with my own data and Python as an AI Engineer
AI Solutions from POC to Production - Channel Intro
Просмотров 1383 месяца назад
AI Solutions from POC to Production - Channel Intro
Thank you. This is an ideal combination of hardware. Does the workflow gets distributed if one of the nodes go down?
It's nice
Is it any better than cursor or windsurf?
Considering amd cards do not have nvlink analogs would be great to combine different radeon powered devices through exo
You can definitely do that! I think newer NVIDIA chips also lack nvlink nowadays so there's use for both.
Perplexica is awesome! To save some of you time, I've got best results using mistral-nemo 12B, with a 16K context window, as both the chat model and embedding model, via Ollama. Have fun! Great vid!
Algorithm got me there. But this is perfect idea I liked it!
Glad you liked it!
Well done. But I am still wondering if there is a way to change model for completions too (code suggestion on typing). Thanks
This video is hard to follow. I feel it would be easier if you cleared your screen and had less going on so that the viewers Eyes knew where to go. Showing unnecessary code all over your screen makes it very confusing for beginners to identify what they are supposed to be looking at. Also you may consider being more clear about which tabs you are pressingand navigating through your screen. I appreciate the video and I look forward to a less confusing one.
Thanks for the constructive feedback! And I do agree, in my newer videos I tend to try and keep only two tabs open across the main content of the video. And if I do need more context I try to switch slower. Of course I can also start showing the shortcuts I'm pressing but just keeping the context smaller is a good idea. Just like with an LLM sometimes 😉
Hello, im confused whether I can run any model on my Ryzen 3 3200G PC, with 16gb of ram or not?
Did you try it out for yourself with the repository in the description? I expect Phi 3.5 to work "okay" on that machine, but not very fast.
I don't ever usually comment, though the video is very informative, want to support you, keep going :)
I always start my prompting with the question “Was 2024 a leap year”. On simple local machines it’s telling you fairy tales. That’s disappointing. But thanks for your good explanation 👏
Yeah, for local devices a RAG implementation makes more sense. I'll explain that later on.
The voice is not synchronised with your video
Oops, looks like it is slightly off indeed. Will make it better next time.
phi...ok but not that powerful, i thought there might be something i don't know in this video...i would rather use openwebui or anythingllm with ollama if you just need a chat interface
That's right, I will show in later content that LocalAI allows you to use other models not just LLMs. If you just want a language model, you can use ollama
Hey, thanks for the video.
Kindly make a video how to setup and make own deepseek r1 api
It's like 0.0014 per prompt just get it lol things unbelievable I was stuck for 3 days with clide and chatgpt and cascade. All of them. Could t handle the size of my algo any more. Thjs thing did what I was trying to figure out in three days. Did it in 3 prompts
Deepseek r1, definitely the full model you can access via web/api, is really difficult to host on your own hardware. You'll need to start with a smaller model like Phi-3.5/4 or a smaller Llama model.
I bought it and it's the best model hands down. Nothing compares @@zenvanriel
@@zenvanriel I have installed deepseek r1 7b using lm studio, but how to create api to access in side my program
LM studio has a server mode in it, check the main menu icons. It is the green one that looks like a CLI
nice name zen
We share the same first name based on your handle? If so, love it!
@ yes we do 🤝🏼
How can you make an interface or make it receive input files?
There is a opensource webuis
This is upcoming in a next video!
Awesome content, could you make a video where you customize a reasoning model It further like connecting It to a folder of PDF files as database
This is indeed all in the backlog of the next videos! Stay tuned.
How can i contact u zen
In a few weeks I'll make sure we can all get in touch. Stay tuned.
the most comprehensive video on the subject !!!! thank you so much for this.. my question is. would it be possible to work in composer mode of cursor too?
Unfortunately not at the moment.. I'm also hoping this will become available.
@@zenvanriel too bad.. would it be somewhere in the source code, in an editable config file of cursor, where the model decides which api it will use in composer or chat? maybe thats where the hack is.
Will try it surely!
What's the difference between using docker and ollama? If you could give a detailed explanation, I'd like that
The important distinction here would be that this guide shows you how to use LocalAI with Docker. Similarly, you can use Ollama with Docker. Docker is just meant to give you an easy one-command way to start an environment consistently. So I'm going to take the liberty to answer the difference between Ollama and LocalAI, because those are actually the systems you are interested in comparing. For large language models, Ollama is actually just as good if not even better supported by the community. You can use Ollama just fine in Docker as well: hub.docker.com/r/ollama/ollama However, LocalAI supports more types of AI models than just language models in one package. I aim to show more of those possibilities later on hence I focus on LocalAI. If you just want to use language models, definitely try out Ollama!
@zenvanriel oh, LocalAI. I understand now. Thanks
How does this model perform on your device? Let me know!
Thank you for the heads up!
May I do it not been a programmer/dev/coder?
Of course!
thanks for great video,, but can I use it with llama for example? (if running locally with ollama for example)
Yes, ollama also exposes a server you can use. Check out github.com/ollama/ollama/blob/main/docs%2Fapi.md
thank you very much@@zenvanriel I'll play with it - but I guess it will be less great for code generation as for example paid models..
I am actually finding ways on how to self host it on cloud hosting via docker and traefik for quite some time, but still fails... 😂
open webui has been able to do this for a while, with the added bonus of being able to connect to multiple ollama instances in a load-balanced connection. When the query/response is complete, you can turn the web search feature off and refine or discuss the content.
"Clear and easy to understand, learn a lot in a short time. You're amazing!"
I have been using it for a month and it’s great. Recent docker pull gives me server error though:( It’s also quite easy to access searxng container directly using simple api so no limit
What docker command are you using?
docker-compose up
Nice bro, works like perplexity
waarom gebruik je geen vscode?
In a way, Cursor is just a vscode fork so I am using it 😄
the image search found an image of a mockup, not the actual 5090 - which is a problem… not of the tool, but in general
Yeah, that's true; as you say though that's the nature of image search these days. Not sure if you've noticed but a lot of the times when a movie comes out a new 'sequel' trailer is quickly generated with AI video tools. Screenshots from those also show in image search :/
Great video...thanks for sharing bro
ai agents are awesome!
Why not better change epub to Markdown?
That's also possible, as Docling supports markdown too. But generally when converting to Markdown you do get rid of complex formatting from the book which might be used for important content. It just depends on the book what the right approach is. Try it out I'd say!
Wow what a lovely alternative! Nice structured video, like and sub!
Thanks for the sub!
Can it help with other languages like c#
Yes!
@zenvanriel thank you, I will try it
careful with adding all files, because with any sufficiently large project you will run out of the context...
I've always wondered how to properly showcase my projects. This seems like a great way to build a portfolio quickly.
This tutorial is super helpful. I've been struggling to make my resume stand out, so the idea of using AI to create a portfolio is interesting.
Great content, best video on vector database with example. I was searching for vector db integration in azure cosmos db and randomly found this video, thank you so much.
Thanks so much! Any other topics you'd love to see covered?
@zenvanriel Training on GEN AI with real world example implementation
3:02 ok, that makes sense, i was wondering why you revert forward your ports since you can alerady run a serve locally
I installed Qwen/Qwen2.5-Coder-3B-Instruct-GGUF for test and it replies with strange code. Looks like encrypted information :(
Perhaps get a slightly larger model than 3b. What programming language are you writing code with?
Hi, I did what you told in this video but I get the following errors! ❓ Ask a question: what is the name of the book 🔄 Processing your question... Traceback (most recent call last): File "C:\Program Files\Python312\Lib\site-packages\httpx\_transports\default.py", line 101, in map_httpcore_exceptions yield File "C:\Program Files\Python312\Lib\site-packages\httpx\_transports\default.py", line 250, in handle_request resp = self._pool.handle_request(req) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpcore\_sync\connection_pool.py", line 256, in handle_request raise exc from None File "C:\Program Files\Python312\Lib\site-packages\httpcore\_sync\connection_pool.py", line 236, in handle_request response = connection.handle_request( ^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpcore\_sync\connection.py", line 101, in handle_request raise exc File "C:\Program Files\Python312\Lib\site-packages\httpcore\_sync\connection.py", line 78, in handle_request stream = self._connect(request) ^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpcore\_sync\connection.py", line 124, in _connect stream = self._network_backend.connect_tcp(**kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpcore\_backends\sync.py", line 207, in connect_tcp with map_exceptions(exc_map): ^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\contextlib.py", line 158, in __exit__ self.gen.throw(value) File "C:\Program Files\Python312\Lib\site-packages\httpcore\_exceptions.py", line 14, in map_exceptions raise to_exc(exc) from exc httpcore.ConnectError: [WinError 10061] No connection could be made because the target machine actively refused it The above exception was the direct cause of the following exception: Traceback (most recent call last): File "C:\Program Files\Python312\Lib\site-packages\openai\_base_client.py", line 996, in _request response = self._client.send( ^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpx\_client.py", line 914, in send response = self._send_handling_auth( ^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpx\_client.py", line 942, in _send_handling_auth response = self._send_handling_redirects( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpx\_client.py", line 979, in _send_handling_redirects response = self._send_single_request(request) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpx\_client.py", line 1014, in _send_single_request response = transport.handle_request(request) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\httpx\_transports\default.py", line 249, in handle_request with map_httpcore_exceptions(): ^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\contextlib.py", line 158, in __exit__ self.gen.throw(value) File "C:\Program Files\Python312\Lib\site-packages\httpx\_transports\default.py", line 118, in map_httpcore_exceptions raise mapped_exc(message) from exc httpx.ConnectError: [WinError 10061] No connection could be made because the target machine actively refused it The above exception was the direct cause of the following exception: Traceback (most recent call last): File "C:\My Git\booktutor-ai\booktutor.py", line 213, in <module> main() File "C:\My Git\booktutor-ai\booktutor.py", line 200, in main result = qa_system.invoke( ^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 170, in invoke raise e File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 160, in invoke self._call(inputs, run_manager=run_manager) File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\conversational_retrieval\base.py", line 170, in _call answer = self.combine_docs_chain.run( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain_core\_api\deprecation.py", line 182, in warning_emitting_wrapper return wrapped(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 611, in run return self(kwargs, callbacks=callbacks, tags=tags, metadata=metadata)[ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain_core\_api\deprecation.py", line 182, in warning_emitting_wrapper return wrapped(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 389, in __call__ return self.invoke( ^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 170, in invoke raise e File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 160, in invoke self._call(inputs, run_manager=run_manager) File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\combine_documents\base.py", line 138, in _call output, extra_return_dict = self.combine_docs( ^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\combine_documents\stuff.py", line 259, in combine_docs return self.llm_chain.predict(callbacks=callbacks, **inputs), {} ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\llm.py", line 318, in predict return self(kwargs, callbacks=callbacks)[self.output_key] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain_core\_api\deprecation.py", line 182, in warning_emitting_wrapper return wrapped(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 389, in __call__ return self.invoke( ^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 170, in invoke raise e File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\base.py", line 160, in invoke self._call(inputs, run_manager=run_manager) File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\llm.py", line 126, in _call response = self.generate([inputs], run_manager=run_manager) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain\chains\llm.py", line 138, in generate return self.llm.generate_prompt( ^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain_core\language_models\chat_models.py", line 786, in generate_prompt return self.generate(prompt_messages, stop=stop, callbacks=callbacks, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain_core\language_models\chat_models.py", line 643, in generate raise e File "C:\Program Files\Python312\Lib\site-packages\langchain_core\language_models\chat_models.py", line 633, in generate self._generate_with_cache( File "C:\Program Files\Python312\Lib\site-packages\langchain_core\language_models\chat_models.py", line 851, in _generate_with_cache result = self._generate( ^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\langchain_openai\chat_models\base.py", line 717, in _generate response = self.client.create(**payload) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai\_utils\_utils.py", line 279, in wrapper return func(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai esources\chat\completions.py", line 859, in create return self._post( ^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai\_base_client.py", line 1283, in post return cast(ResponseT, self.request(cast_to, opts, stream=stream, stream_cls=stream_cls)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai\_base_client.py", line 960, in request return self._request( ^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai\_base_client.py", line 1020, in _request return self._retry_request( ^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai\_base_client.py", line 1098, in _retry_request return self._request( ^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai\_base_client.py", line 1020, in _request return self._retry_request( ^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai\_base_client.py", line 1098, in _retry_request return self._request( ^^^^^^^^^^^^^^ File "C:\Program Files\Python312\Lib\site-packages\openai\_base_client.py", line 1030, in _request raise APIConnectionError(request=request) from err openai.APIConnectionError: Connection error.
It seems it couldn't connect to the language model. Did you start a local language model server using LM Studio? You can have a look at ruclips.net/video/f40iM0mt4ww/видео.htmlsi=LWXqRGFWNZszCzch to set this up
Wow. That it so cool! Thank you for sharing this. I love it.
Zen, you just keeps casting pearls. You deserve so many more subscribers.
Thanks for a video. what model do you use with aider? looks really fast
In this video I am using Azure OpenAI's gpt4o model! :) What is your daily driver?
Well done, thanks :) One out of scope question - you have nicely optimized 'ls' command. You use it like 1:10 of the video. Could you please share your settings?
I got the config from a fellow creator, specifically this video: ruclips.net/video/LnZdaNfQ86o/видео.htmlfeature=shared
@@zenvanriel Amazing, thank you! Thank you for your videos. They are very useful.
Hi guys! Thank you for your great video. And have a question , my macbook pro m1 (RAM : 16GB) , download Qwen 2.5 32b, but can't load this model in LM studio. please hep me. I wan know your pc or laptap spec.
You'll likely need more RAM, I've got 64GB for this demo video. How about trying a smaller sized model? For example 3B.