Volko Volko
Volko Volko
  • Видео 99
  • Просмотров 23 374
o1 is dead, long life to o3 ! AGI is CONFIRMED by ARC PRIZE!
Open ai did they 12nd (and last one) announcement.
They released o3, which is a huuuge performance gain.
The performance jump is higher going from o1 to o3 than from GPT4o to o1 !!!
You really have to see the video to fully understand
Просмотров: 44

Видео

Gemini 2.0 Flash became ... o1 ? Gemini 2.0 Flash Thinking 🤯
Просмотров 4694 часа назад
In this video, we are going to take a look at the new Gemini 2.0 Flash Thinking that just released
This new feature of Gemini 2.0 is ... SCARY
Просмотров 3789 часов назад
In this video, we are going to take a look at Multimodal Live API with Gemini 2.0 available in aistudio.google.com for free. He is able to hear you, talk to you and see your entire SCREEN !
Gemini 2.0 vs GPT4o mini vs Claude 3.5 Haiku ! Gemini did SOOO BAD 😨
Просмотров 97421 час назад
Gemini is reaaally not that great in real use cases. I don't really see why it is ranked so high in the benchmarks
Let's test QwQ, the new opensource alternative to o1
Просмотров 1,3 тыс.День назад
In this video, I'm going to test the performances of QwQ, a new preview model released by Qwen that works the same way as o1 from OpenAI
Llama 3.3 vs Llama 3.2 ! HUGE IMPROVEMENTS !
Просмотров 1,7 тыс.14 дней назад
Let's compare these two models !
Quantization is Simple! Here is how it works
Просмотров 39121 день назад
In this video, I'm going to show you how quantization is working under the hoods
BAD vs GOOD prompting
Просмотров 775Месяц назад
Let's see in this video if we still need to make good prompting nowadays and if there is a difference, at what point is it different. Feel free to leave comments, they be negative if they are constructives.
Is Bigger Better ?
Просмотров 277Месяц назад
In this video, we are going to test the performances of each model size and compare it against other sizes
Qwen2.5 Coder 32B vs GPT4o vs Claude 3.5 Sonnet (new)
Просмотров 5 тыс.Месяц назад
Let's see which model is the best
Ultimate Guide: Easily Quantize Your LLM in Any Format
Просмотров 156Месяц назад
Today, I'm going to show you how I quantize all my models : Link : colab.research.google.com/drive/1b6nqC7UZVt8bx4MksX7s656GXPM-eWw4
New Chinese LLM beat Llama3.2 !!
Просмотров 128Месяц назад
Let's test a new chinese model and see how it perform against LLama 3.2 and proprieteray models (new Claude 3.5 sonnet, GPT4o)
Avoir ChatGPT sans Connexion | FR
Просмотров 182Месяц назад
Bienvenue, Aujourd'hui, je vais vous apprendre à installer une IA locale similaire à ChatGPT.
[Tuto] [FR] | Comment installer et créer une Machine Virtuelle avec VirtualBox (2024)
Просмотров 317Месяц назад
Aujourd'hui, je vais vous montrer comment créer vous même des machines virtuels sur votre ordinateur afin que vous ayez un environnement sécurisé et propre
How to repair iPhone that Keeps Restarting
Просмотров 202 месяца назад
How to repair iPhone that Keeps Restarting
How to upgrade your RAM on Mac
Просмотров 1132 месяца назад
How to upgrade your RAM on Mac
This is how to solve the Folder Icon on mac
Просмотров 312 месяца назад
This is how to solve the Folder Icon on mac
J'ai recréé Joyca en Jeu Vidéo
Просмотров 1502 месяца назад
J'ai recréé Joyca en Jeu Vidéo
Is LLama 3.2 any good ?
Просмотров 1582 месяца назад
Is LLama 3.2 any good ?
How to access o1 (Strawberry) API & chat without tier 5
Просмотров 3473 месяца назад
How to access o1 (Strawberry) API & chat without tier 5
Who ?
Просмотров 2233 месяца назад
Who ?
Blender Animation Cycles
Просмотров 74 месяца назад
Blender Animation Cycles
OpenAI released GPT4o Mini | Let's test it !
Просмотров 1345 месяцев назад
OpenAI released GPT4o Mini | Let's test it !
MathΣtral First Test ! Quite impressive results ! Mistral AI
Просмотров 745 месяцев назад
MathΣtral First Test ! Quite impressive results ! Mistral AI
Make a DIY Onewheel - $200! Episode 2
Просмотров 245 месяцев назад
Make a DIY Onewheel - $200! Episode 2
I Recreated The Finals, but it's Open Source !
Просмотров 2705 месяцев назад
I Recreated The Finals, but it's Open Source !
Gemma2:27 Ollama Correction ! Now Incredible !
Просмотров 3195 месяцев назад
Gemma2:27 Ollama Correction ! Now Incredible !
Gemma2:27B First Test ! How Can it be THAT Bad ?!
Просмотров 3155 месяцев назад
Gemma2:27B First Test ! How Can it be THAT Bad ?!
Gemma2 First Test ! Incredible Results for a 9B model
Просмотров 1525 месяцев назад
Gemma2 First Test ! Incredible Results for a 9B model
UNDERVOLT, make your pc COOL AGAIN !
Просмотров 185 месяцев назад
UNDERVOLT, make your pc COOL AGAIN !

Комментарии

  • @mimocode
    @mimocode 11 часов назад

    hi good vd just wanted ask if after exchanging messages with gemini it give a something went wrong error and you need to make a new chat

    • @volkovolko
      @volkovolko 4 часа назад

      Hi, I'm sorry but no, I did not encounter a similar issue. Maybe reach the Google support to be helped 🤷 Thanks a lot for this kind comment ^^

  • @sonOfLiberty100
    @sonOfLiberty100 День назад

    You can type in, "provide the whole file" and you can also use "no yapping" then it will only write what you need, just code

  • @sasori-wm4rd
    @sasori-wm4rd 7 дней назад

    merci car je suis parano d installe des trucs surtout apres qu un virus c est multiplie 800k fois sur mon ancien pc

    • @volkovolko
      @volkovolko 6 дней назад

      De rieng ^^ Profite bien de ta Machine Virtuelle

  • @sLavoncheg
    @sLavoncheg 7 дней назад

    why 3.5 haiku? sonnet has more scrores

    • @nashh600
      @nashh600 6 дней назад

      Its the smaller model for fair comparison. You wouldn't compare it with o1

    • @volkovolko
      @volkovolko 6 дней назад

      Yes, it is true. But the released was for the Gemini 2.0 flash. I needed to compare it against equal size/price models as @nashh600 correctly pointed out.

  • @alexg.994
    @alexg.994 9 дней назад

    Interesting; what persona did you select?

    • @volkovolko
      @volkovolko 8 дней назад

      We need to select a persona for Gemini ?

    • @alexg.994
      @alexg.994 8 дней назад

      @ I don’t know Gemini, but the others, I would strongly recommend it to get a better result.

  • @UCs6ktlulE5BEeb3vBBOu6DQ
    @UCs6ktlulE5BEeb3vBBOu6DQ 9 дней назад

    btw QwQ can totally do multi-turn. Set it to 32k context and 16k output tokens so its thinking isn't cut before he's done. llama.cpp has much more settings.

    • @volkovolko
      @volkovolko 9 дней назад

      Oh okay, I didn't knew that. I thought it cannot do multi turn because it's single turn only in the QwQ Space ^^ Thanks a lot for the precision !

  • @UCs6ktlulE5BEeb3vBBOu6DQ
    @UCs6ktlulE5BEeb3vBBOu6DQ 9 дней назад

    Tetris game is often my coding test and they all struggle with it.

    • @volkovolko
      @volkovolko 9 дней назад

      Yes, tetris is quite difficult for LLMs. Only Claude 3.5 Sonnet and Qwen2.5 Coder 32B got it right on my tests. Even gpt4o didn't got it in my test (but i think it has more related to luck)

  • @SoM3KiK
    @SoM3KiK 12 дней назад

    hey! Would it work with a 3060ti and 32gb ram?

    • @hatnis
      @hatnis 11 дней назад

      I mean, you can't fit the required 24 gb of VRAM on your graphics card, but hey, only one way to find out if it works right.

    • @SoM3KiK
      @SoM3KiK 11 дней назад

      @@hatnis well, it was free to ask 😅

    • @volkovolko
      @volkovolko 11 дней назад

      Yes, but you will have to offload a lot in your CPU/RAM. It will run pretty slow but it will work 👍

    • @volkovolko
      @volkovolko 11 дней назад

      In the video, I ran it in my 24Go of VRAM. I think it is q4_k_m

    • @Timely-ud4rm
      @Timely-ud4rm 11 дней назад

      I was able to get it working on my new Mac mini base m4 pro chip model. QwQ-32B-Preview-GGUF bartowski repo. IQ3_XS quantization. the only one I could download as this one is 13.71 gb of ram. Note because I am using a Mac mini apples ram is unified so my 24gb of ram is shared between the gpu and cpu. if I spent spent a extra 300$ from the 1.4k I spent for the m4 pro model I could of loaded the max quantization model but I don't really do AI locally as I use online Ai services more. I hope this helps!

  • @SpaceReii
    @SpaceReii 14 дней назад

    Was really fun to see how 3.3 made Tetris, worked amazingly. Was sad to see how 3.1 failed badly at making Tetris though...

    • @volkovolko
      @volkovolko 12 дней назад

      Yeah, I wasn't expecting 3.1 to be THAT bad neither 😅

  • @Maisonier
    @Maisonier 14 дней назад

    Great video, liked and subscribed.

    • @volkovolko
      @volkovolko 14 дней назад

      Ohhh that's awesome man !!! Thanks a lot ^^

    • @volkovolko
      @volkovolko 14 дней назад

      I liked and subscribed you vidéos too 😉

  • @atmosphere14874
    @atmosphere14874 14 дней назад

    no difference except for hair style

  • @petchlnwzaaa
    @petchlnwzaaa 14 дней назад

    Great and really quick content man! Keep up the good work

    • @volkovolko
      @volkovolko 14 дней назад

      Thanks a lot man 🙏👍👍

  • @ikorin
    @ikorin 15 дней назад

    What is your build specs to run 3.3 w/ this speed ?

    • @volkovolko
      @volkovolko 15 дней назад

      I'm using Groq, they have specialized LPUs (specialized hardware for LLM inference) that allows insane speeds and it is free to use (and you even have an api). There is also sambanova that offers the same thing

  • @alparslankorkmazer2429
    @alparslankorkmazer2429 15 дней назад

    Where are you from?

  • @samandarkhan2431
    @samandarkhan2431 15 дней назад

    how much vram required?

    • @volkovolko
      @volkovolko 15 дней назад

      24go at the very least (with insane quantization like q2-q3). You absolutly need at least 1 3090/4090. I would recommand running it in your RAM with CPU and have at least 64Go of RAM and run it as q4-q5

    • @volkovolko
      @volkovolko 15 дней назад

      unquantized it is about 130go.

    • @volkovolko
      @volkovolko 15 дней назад

      But you can use it for free on Groq and SambaNova

    • @tekratek4077
      @tekratek4077 15 дней назад

      @@volkovolko I'm going to run it with 8GB vram and 32GB ddr4 ram as usual. Q2 is around 26 Gigs so it shouldn't go to swap

    • @14supersonic
      @14supersonic 15 дней назад

      I'd highly doubt you'd be able to run it with that. I have a RTX 3070 and 48GB of DDR4 RAM, and I tried running Llama 3.2 11b vision and it doesn't even load.

  • @bigbrotherr
    @bigbrotherr 20 дней назад

    Not a great test to me because these models have been trained with these games before and the codes are in there. Let's try something custom and let's see how it can reason, create and solve problems. That will make it a good model. Also Claud 3.5 sonet is the best coder and very hard to make mistakes when coding.

    • @volkovolko
      @volkovolko 15 дней назад

      I would be happy to test with any prompt you give to me ^^

  • @user-ub8mc6td1p
    @user-ub8mc6td1p 27 дней назад

    btw your thumbnail isnt readable, you should make it bigger

    • @volkovolko
      @volkovolko 26 дней назад

      Hummm, I don't really understand 😅😅

  • @vanhans6215
    @vanhans6215 27 дней назад

    en Français c'est possible?

    • @volkovolko
      @volkovolko 27 дней назад

      Oui en soit c'est possible, c'est simplement que la majorité de mon audience est anglaise

    • @vanhans6215
      @vanhans6215 27 дней назад

      @@volkovolko mouais.., non mais c'est super intéressant, ce qui serait cool aussi ça serait de comparer les different types: GGUF, AWQ, GPTQ, bon après on comprend que c'est toujours une manière de rationaliser les virgules flottantes (et donc la quantité de bits et donc la taille du modèle), mais je pense que les gros progrès de l'IA maintenant ça va être sur la quantization, réussir à condenser au maximum les modèles pour les rendre, certes un peu moins précis mais beaucoup plus pertinent (utilisables sans avoir des racks de GPU). Enfin, en tout cas merci.

    • @volkovolko
      @volkovolko 25 дней назад

      Je partage tout à fait cet avis

  • @SpaceReii
    @SpaceReii 27 дней назад

    Theres a little issue with the video cutting and such, but I did get a lot of information from it!

    • @volkovolko
      @volkovolko 27 дней назад

      Yes, I had some issues with Clipchamp, which I use to edit the video. But I'm glad it helped you !!

  • @cot_clan
    @cot_clan 27 дней назад

    commen ton installe le iso stp

    • @volkovolko
      @volkovolko 27 дней назад

      Ici : www.microsoft.com/fr-fr/software-download/windows11 ^^

  • @riodjajoesman4989
    @riodjajoesman4989 27 дней назад

    what are the factors that make a good prompt. would be great if u reply

    • @volkovolko
      @volkovolko 27 дней назад

      Quite verbose, explain exactly what features you want, well ordered and formated, use prompting techniques like Chain of Thought etc. In practice, I first ask chatgpt or any other LLM to create a good prompt for me then, I regive it to him. I gives way better results

  • @cydeYT
    @cydeYT 29 дней назад

    Now make it create a Minecraft JNI client with a basic autosprint module. Don’t think it can do that yet.

    • @volkovolko
      @volkovolko 28 дней назад

      I don't even know what it is 😅😂😂

  • @GameRoMan
    @GameRoMan 29 дней назад

    Knowing how to make good prompts is a useful skill nowadays

  • @DOCTOR-FLEX
    @DOCTOR-FLEX Месяц назад

    Thank you for this demonstration. In the future, please work on more complex apps. I’m happy you tried Tetris instead of only the snake game.

    • @volkovolko
      @volkovolko 28 дней назад

      They the issue is that we need to balance the complexity of the tasks. If it's too easy all models get it right so we cannot compare them If it's too difficult all models fails so we cannot compare them. Tetris and Pac man games seems currently a good fit for SOTA and aren't that tested so that's why I use them

  • @barberousse5112
    @barberousse5112 Месяц назад

    Même si les LLM deviendraient extrêmement puissant, et pourrait parfaitement répondre à un prompt, le prompt engineering resterait important. L'une des choses les plus importantes, c'est de faire comprendre notre intention au LLM, et c'est pas forcément évident, notamment si y a pas suffisamment de contexte, ou alors qu'il y a du contexte qui parasite la réponse. L'exemple que j'ai en tête, c'est quand on lui demande d'écrire un mail de manière formelle, et qu'il utilise plein de mots compliqués qui rende la réponse peu naturelle.

    • @volkovolko
      @volkovolko Месяц назад

      Oui, cependant il y a souvent des options "par défaut". Si je lui demande de me créer un programme en python, il va systématiquement me faire un hello world. Donc certains détails "par défaut" on pourrait penser que tu n'as pas besoin de les donner (comme ce serait le cas pour un humain) mais en réalité ca améliore la qualité de la réponse. Cependant, je pense qu'à l'avenir on aura pas besoin de préciser ces chses "Par défauts"

  • @SpaceReii
    @SpaceReii Месяц назад

    This really does show a lot. I think the issue was that the FPS was 60 in the second one, which caused the ghosts to move real fast. Overall, i think this showed that good prompting will always give better results depending on the question.

    • @volkovolko
      @volkovolko Месяц назад

      Yeah, before this test I thought that nowadays we don't need good prompting anymore. Especially due to the system message prompting of ChatGPT, the fact that it ask the model to reformulated the question before anwsering However it turns out that I was completely wrong

  • @Zippnix
    @Zippnix Месяц назад

    You must good and spesific prompting to generate pac man maybe

    • @volkovolko
      @volkovolko Месяц назад

      Yes, you're right. Providing a better, longer and more precise prompt would have surely improve the quality of the LLM answer. However, the video would have been longer and I think in real life, most people are lazy and don't provide to the LLM a well done prompt. Maybe I will make a video comparing how big is the performance hit between a good and a bad prompt

  • @andrepaes3908
    @andrepaes3908 Месяц назад

    I understand all models are quantized at 4 bits in this video? Can you do another video with 8bit quantization for the 3b, 7b and 14bi models? Maybe a higher quant may improve their quality. It can be very useful info!

    • @volkovolko
      @volkovolko Месяц назад

      Yes, they are 4bits as recommanded by Ollama. However, I can make a new video trying with 8bits as you asked. But I think the results will be pretty much the same. Maybe I should rather make a video comparing each quant of the 7b

    • @andrepaes3908
      @andrepaes3908 Месяц назад

      @volkovolko yes in my experiments with llms so far it looks like for small models (<8b params) quantization does have an impact on model output quality while for larger models the impact is less noticeable.

  • @nashh600
    @nashh600 Месяц назад

    Thanks for the comparison but this was painful to watch. Please cut the parts that are not relevant to the subject or at least add timestamps

    • @volkovolko
      @volkovolko Месяц назад

      I'm trying to do my best. When I made this video. I didn't had any speakers so I couldn't test the audio nor make great cuts

  • @SpaceReii
    @SpaceReii Месяц назад

    Upgrading MacBook RAM Speedrun (WORLD RECORD)

  • @kobi2187
    @kobi2187 Месяц назад

    If you do a real software project, you'll find out claude sonnet new is the best, and gpt4 is very good at organizing.

    • @volkovolko
      @volkovolko Месяц назад

      I do real software projects as I'm a developer. While Claude and GPT4o are still better for big projects, qwen is a good alternative for just little prompting to avoid going to stack overflow for quick and simple questions

  • @electroheadfx
    @electroheadfx Месяц назад

    amazing, thanks for the test

  • @SpaceReii
    @SpaceReii Месяц назад

    Now this is my kind of stuff. Personally, 0.5b is incapable of any sort of coding. In my experience, I've seen it get really bad results compared to 3b, 7b, etc. as it's made for small coding tasks and fill-in for code. Thanks for showing this comparison though! I also recommend benchmarking their processes at making Discord bots, as that would show a variety of new things. 3b is capable of making a lot more than GPT-4, as GPT forgets variables, portions of code, etc. when Qwen is capable of creating Discord bots which run with no error on start. Love your videos, don't ever stop making them ❤

    • @volkovolko
      @volkovolko Месяц назад

      For you a 3B Qwen can be better for discord bot coding than GPT4o ? (not a judgment, I just want to know)

    • @SpaceReii
      @SpaceReii Месяц назад

      @volkovolko Yeah, it does a lot better at importing packages, using Python properly, and even does well at managing errors gracefully.

  • @oguzhan.yilmaz
    @oguzhan.yilmaz Месяц назад

    Nice video but i think Claude is still better. If i compare these models at first i always say to myself "If these models are slightly close to each other (In terms of technical specifications) it is okay to compare but if it's not what is the point? Like i understand comparing between open source models like Qwen and Llama or closed source models like Gpt4-o and Claude 3.5 Sonnet

    • @volkovolko
      @volkovolko Месяц назад

      Yes, the results of the tests I made in this video seems to show that : GPT4o < Qwen2.5coder32b < Claude 3.5 Sonnet (new)

    • @sthobvious
      @sthobvious Месяц назад

      The point is to compare quality... simple as that. Once you know quality, you can consider other factors like speed, price, availability, and of course confidentiality. The fact that Qwen2.5-Coder-32B is even close to Claude while being a _small_ open-weight model is amazing. Of course other factors can matter more than just quality. Speed and price are just as important. But limiting it to "Only compare quality when technical specs are comparable" makes no sense.

    • @oguzhan.yilmaz
      @oguzhan.yilmaz Месяц назад

      @@sthobvious actually makes sense because if you think to compare gpt-3.5 and gpt-o1 or gpt-4o, do you really think this is fair? Gpt-3.5: 😭 Gpt-4o & gpt-o1: 🗿🗿

  • @owonobrandon8747
    @owonobrandon8747 Месяц назад

    the error produced by gpt was minimal; a "hallucination"

  • @IgnatShining
    @IgnatShining Месяц назад

    Sweet. I remember, when Chat GPT just appeared, feeling very pessimistic that this tech would be locked in big companies datacenters. Glad I was wrong

    • @volkovolko
      @volkovolko Месяц назад

      Yes, it's so awesome they this technology is going toward open sourcing 👍

  • @Rolandfart
    @Rolandfart Месяц назад

    you should ask for physics demos like softbody , particles, fluid particles, cloth. Anything math heavy pretty much.

    • @volkovolko
      @volkovolko Месяц назад

      Okay, I will try in the next video

  • @mathiasmamsch3648
    @mathiasmamsch3648 Месяц назад

    Why do people do these stupid tests where the code can be found 1000 times on the internet.

    • @volkovolko
      @volkovolko Месяц назад

      As explained in the video, I'm looking for other original tests. If you have one that you want me to try in a following video feel free to leave it in a comment so that I can try it in the following video

    • @mathiasmamsch3648
      @mathiasmamsch3648 Месяц назад

      @@volkovolko If you are testing how to write a snake game, then you are basically testing knowledge retrieval, because that code exists in 1000 variants on the Internet. It gets interesting if you demand variations, like 'but the snake grows in both directions' or 'random obstacles appear and disappear after some time in not too close proximity of the snakes'. Think of whatever you want, but if you can do Tetris or snake is hardly a test for llms these days.

    • @mathiasmamsch3648
      @mathiasmamsch3648 Месяц назад

      @5m5tj5wg The 'better' model is not one that can retrieve known solutions better, but the one that can piece the solution to a unheard but related problem better. If you can find the question and the answer on the net then comparing a model with 32B params to a Multi-hundred-billion parameter model like GPT4o or sonnet makes even less sense, because of cause they can store more knowledge. You need to ask for solutions to problems where you cannot find the answer on the Internet to evaluate how good a model will be in practical use.

    • @volkovolko
      @volkovolko Месяц назад

      Yes, there is a part of true. However, I think you can all agree that you don't want a 50+ min video. Also most of the code you will ask it to make in the real world will also be knowledge retrieval. As developper we very often have to remake what as already been made. And the Snake game isn't that easy for LLMs. The Tetris game is very difficult and I didn't ever see a first try fully working

    • @volkovolko
      @volkovolko Месяц назад

      And it is interresting to see that the Qwen model did better on these "retrieval" questions than GPT and Anthropic despite being way smaller in terms of parameters. It indicates that knowledge can still be compress a lot more than what we thought

  • @mnageh-bo1mm
    @mnageh-bo1mm Месяц назад

    try a next js app.

    • @volkovolko
      @volkovolko Месяц назад

      Okay, I will try in the next video

  • @tpadilha84
    @tpadilha84 Месяц назад

    Funny thing: I tried the same tetris example locally with the q8 and fp16 versions of Qwen coder 2.5 32b and it generated buggy code in both cases. When I tried with the default quantization (q4_k_m if I'm not mistaken) it got perfect the first time (properly bounded and you could lose the game too). I guess there's a luck factor involved.

    • @volkovolko
      @volkovolko Месяц назад

      Yeah, it might be because of the luck factor. Or maybe the architecture of qwen is optimised for high quantizations levels 🤷‍♂️ Or maybe your q8 version wasn't properly quantized, I think they updated their weight at one moment

    • @66_meme_99
      @66_meme_99 Месяц назад

      luck it's called temperature nowadays :D

    • @volkovolko
      @volkovolko Месяц назад

      Yeah, I now. Top_k also right ? ​@@66_meme_99

  • @cerilza_kiyowo
    @cerilza_kiyowo Месяц назад

    I think you should ask qwen 2.5 coder 32B again to make the tetris game better so it will be fair .. In my opininion In tetris game qwen literally win .. even claude generate better after error , but offcource it failed at first

    • @volkovolko
      @volkovolko Месяц назад

      Yeah, for me the win was for Qwen. But okay, for the following videos, I will always let one second chance for all models. I will soon make a video comparing each size of qwen2.5 coder (so 0.5B vs 1.5B vs 3B vs 7B vs 14B vs 32b) So subscribe if you want be notified ^^ I also started to quantize each model in GGUF and EXL2 on HuggingFace for those who are interested : huggingface.co/Volko76

    • @renerens
      @renerens Месяц назад

      Seems very interesting I will try it tomorrow, for me nemotron 70b was the best but even on my 4090 I can't run it locally.

    • @volkovolko
      @volkovolko Месяц назад

      I made the video comparing sizes : ruclips.net/video/WPziCratbpc/видео.htmlsi=o3eKo-3pGY78wmMr

    • @volkovolko
      @volkovolko Месяц назад

      Yes, 70B is still a bit too much for consumer grade GPUs

  • @cot_clan
    @cot_clan Месяц назад

    c'est du fake?

  • @Kusmoti
    @Kusmoti Месяц назад

    nice vid! what's your 3090 setup my guy

    • @volkovolko
      @volkovolko Месяц назад

      Asus ROG STRIX 3090 32Go ddr4 3200MHz i9 11900kf

  • @SpaceReii
    @SpaceReii Месяц назад

    This is pretty cool to see! It's nice to see how the models compare between each other. For me, even the 3B model was amazing at making a Python snake game. Thanks for the comparison, it really does show the difference.

    • @volkovolko
      @volkovolko Месяц назад

      Yeah, I totaly agree. The Qwen series (especially the coding one for me) are just so amazing. I don't know why they aren't as known as the llama ones.

    • @volkovolko
      @volkovolko Месяц назад

      Do you want me to make a video comparing the 3B to the 32B ?

    • @SpaceReii
      @SpaceReii Месяц назад

      ​@@volkovolko Yeah, that would be really cool to see! I'd love to see how the models perform.

    • @volkovolko
      @volkovolko Месяц назад

      Okay, I will try to do it tomorrow

  • @Regis-sn7pw
    @Regis-sn7pw Месяц назад

    Ta la même voix que UNITY FR mdrrr

    • @volkovolko
      @volkovolko Месяц назад

      Ah mdrrr ben ecoute tant mieux, j'aime bien sa voix xD

  • @CrownPointeChroma
    @CrownPointeChroma Месяц назад

    Germany countryhuman: too much too much oh yeah

  • @CrownPointeChroma
    @CrownPointeChroma Месяц назад

    Germany countryhuman: Hello Napoleon!

  • @yahyarahil3082
    @yahyarahil3082 Месяц назад

    Comment peut-on uploader un fichier ou une image ? Merci.

    • @volkovolko
      @volkovolko Месяц назад

      Il suffit d'installer la version 4 (qui est en alpha) de ollama : github.com/ollama/ollama/releases/download/v0.4.0-rc5/ollama-windows-amd64.zip Ensuite tu peux installer llama3.2 et utiliser des images. Si tu le souhaites sinon tu peux créer un rag (mais c'est plus complexe)

  • @EncoreUnNerd
    @EncoreUnNerd Месяц назад

    LLM

    • @volkovolko
      @volkovolko Месяц назад

      Yes, it is. ChatGPT is also an LLM

  • @robertcumordi
    @robertcumordi Месяц назад

    Salut et surtout merci beaucoup pour votre vidéo 🙏🙂

    • @volkovolko
      @volkovolko Месяц назад

      Ahh ça fait toujours vachement plaisir d'aider les gens 🙏🙏