Build your own local o1 - here’s how

Поделиться
HTML-код
  • Опубликовано: 26 окт 2024
  • Wanna start a business with AI Agents? Go here: www.skool.com/...
    Work with David directly: gvw0h8ku6fc.ty... (limited to 5 people)
    Get early access to David's startup: forms.gle/SpuE...
    Ollama: ollama.com/
    Nemotron: ollama.com/lib...
    Cursor: www.cursor.com/
    Follow me on Instagram - / thedavit
    Follow me on Twitter - x.com/DavidOnd...
    Subscribe if you're serious about AI.
    Here's how to build your own 100% local o1 assisstant.

Комментарии • 75

  • @DavidOndrej
    @DavidOndrej  18 часов назад +6

    Wanna build your own AI Startup? Go here: www.skool.com/new-society

    • @startingoverpodcast
      @startingoverpodcast 11 часов назад

      Why aren't you using Msty?

    • @aaaaaaaaooooooo
      @aaaaaaaaooooooo 3 часа назад

      Wait, my data is not private with o1? I didn't know that. Where can I check this? Where is this notified to the user, or did they bury it in small text?

  • @indiemusicvideoblog
    @indiemusicvideoblog 15 часов назад +20

    Great! Now build a local agent with lama that can control your computer like Antropic

    • @orthodox_gentleman
      @orthodox_gentleman 13 часов назад +4

      Very doable with Open-Interpreter which is open source and free

    • @Bllakez
      @Bllakez 13 часов назад +1

      @@orthodox_gentleman How much should I pay someone to setup for me?

    • @alexrayoalv
      @alexrayoalv 11 часов назад +4

      I literally did this 6 months ago.

    • @anubisai
      @anubisai 6 часов назад

      You build it.😂

    • @marilynlucas5128
      @marilynlucas5128 4 часа назад

      Skyvern!

  • @DCinzi
    @DCinzi 4 часа назад +3

    There is a model called Llama3.3B-Overthinker. I think it would fit the task quite nicely.

    • @JackGamerEuphoriaDev
      @JackGamerEuphoriaDev 2 часа назад

      Is there available in Ollama or hugging face? If you don't mind the question. Thanks by the way for giving directions..

  • @eado9440
    @eado9440 18 часов назад +9

    🎉 you actually made it. Thanks

  • @Luxcium
    @Luxcium 15 часов назад +4

    😂 I love the way you have called out your mistake 4:00 it was just so delightful to see you handle it like a boss that I have had to replay it more than 3 times to enjoy the moment... You are definitely a smart man!!! I am eager to see the evolution over time!!! 😅

  • @MiNiD33
    @MiNiD33 Час назад

    "Comments are apologies in code." - Robert C Martin.
    Cursor is helping you.
    Also for the price of the spec of this machine, you can buy an insane number if tokens from anthropic or openai. It might be worth getting people started using a hosted service.

  • @chrystofferaugusto1194
    @chrystofferaugusto1194 16 часов назад

    Btw, the concept you reached in this video of undetermined number of agents is far superior than it was from a video from 5 days ago. Really awesome 👏🏻

  • @avi7278
    @avi7278 3 часа назад +1

    Oh yeah im sure openai is quaking in their boots, bro.

  • @MrMoonsilver
    @MrMoonsilver 18 часов назад +1

    Cool new format with the presentation man

  • @szebike
    @szebike 5 часов назад

    Nice, your contribution to the open source community is awesome!

    • @ysh7713
      @ysh7713 4 часа назад

      opensource?

    • @szebike
      @szebike 4 часа назад

      @@ysh7713 Well kind of ~ better than giving all you data to a faceless big company who wills steal your data 100%.

  • @FuZZbaLLbee
    @FuZZbaLLbee 2 часа назад

    You can also use the ollama streaming output to generate text. This way you know what’s the generator is doing.
    Also I think that GPT o1 does more then split up a task and let agents fix the individual tasks. But nevertheless, a nice tutorial on making agents.

  • @olivert.7177
    @olivert.7177 15 часов назад +3

    There is also an nemotron-mini model which is only 4b.

    • @samuelgarcia1802
      @samuelgarcia1802 14 часов назад

      How good it is? In hugging face I saw nematron was in a bad place

    • @orthodox_gentleman
      @orthodox_gentleman 13 часов назад

      Really??? Omg that is great

  • @yacinebenhamou8143
    @yacinebenhamou8143 5 часов назад

    exactly what I needed thank you so much David🎉

  • @mihaitanita
    @mihaitanita Час назад

    So, you've used Claude 3.5 (2024 october update) within Cursor AI Editor to develop a (simple) python script that run some agenting on a 70b model on ollama?
    Where's the o1 in here?

  • @EtH-xf6br
    @EtH-xf6br 5 часов назад

    What a beast Macbook you need to have to get such a fast response. I have 7800x3D and 4080 rtx and its waaay slower.

  • @mariomanca7546
    @mariomanca7546 16 часов назад

    If you instruct the agent to use the fewest possible lines, it's likely to eliminate comments, which is suboptimal but expected.

  • @samimejri8079
    @samimejri8079 14 часов назад

    I just used Llama 3.2 locally and asked about starting a 3d printing business as a 3D beginner. It gave a similar output of what you spent a good time building in this video... Maybe do it the next time, show a before and after response from an LLM.

  • @erictidmore8047
    @erictidmore8047 11 часов назад

    You gotta do a video on Aider if you haven't. I think you'd like it better than Cursor plus it's free and does the same thing.

  • @skulltrick
    @skulltrick 2 часа назад

    Very inspiring! Thanks

  • @jayhu6075
    @jayhu6075 3 часа назад

    What a great explanation. Thnx

  • @Visualife
    @Visualife 14 часов назад

    You should use Anything LLM and docker / Open WebUI

  • @orthodox_gentleman
    @orthodox_gentleman 13 часов назад +1

    Dude, there are very few people that can run nemotron locally….

  • @hotlineoperator
    @hotlineoperator 16 часов назад +1

    I have test o1 - and it is not so smart. People still need to quide its selections. Big problem with models is censorship, someone else have select what you can do and not to do with these tools.

  • @VinceOmondi
    @VinceOmondi 17 часов назад

    Good stuff, Ondrej!

  • @dark_cobalt
    @dark_cobalt 11 часов назад +1

    Already have it lol. Running it on my RX 7900XTX with q4m, but i think ill buy myself 1-2 Radeon W7900 Pro to gain a lot more performance. Alsp you don't need Ollama for it, because it's available in LM Studio and it's downloading from Huggingface.
    Btw what PC hardware specs do you have?

    • @rhadiem
      @rhadiem 8 часов назад

      He's clearly using a 128gb Macbook Pro which can use the memory as vram. He's running un-quantized. How much vram do you have on your gaming gpu? Nobody asked about your hardware bro.

    • @dark_cobalt
      @dark_cobalt 8 часов назад

      @@rhadiem Every PC can use the RAM as VRAM. It's how computers work. It's called virtual memory. If the VRAM fills up, the computer uses the RAM as backup memory, to stay stable and not crash. But the RAM is waaaaaaay slower than the VRAM, that's why I am asking him what specs he has. My GPU has 24GB of VRAM and even with the Quant 4M (around 32GB) model of Nemotron 70B my VRAM gets filled up completely and my RAM also to 50GB, which slows down the model to such an amount, that it's painfully slow. He is using a way bigger model, without any issues. If he has a GPU with this huge amount of VRAM, this would be totally understandable, but with the RAM? I don't understand why lol. 😄

  • @MrAndrew535
    @MrAndrew535 33 минуты назад

    I want to preserve a million-word dialogue between myself and my ChatGPT on multiple threads while upgrading to your recommendations. How do I achieve that?

  • @eviv8010
    @eviv8010 7 часов назад +4

    nice clickbait

  • @KiranMohan-dpthinkr
    @KiranMohan-dpthinkr 3 часа назад

    Hey David, how can we reassure clients that their data is secure and won't be shared with the LLM provider for internal training purposes? What steps can we take to ensure their data privacy and address any concerns they might have?

  • @michaeltse321
    @michaeltse321 5 часов назад

    You downloade nemotron and not the 70b version which is why you had the error

  • @MrMoonsilver
    @MrMoonsilver 17 часов назад +6

    Also, I hope the bruise on your nose heals soon. Been a long time now.

    • @Tetardo
      @Tetardo 15 часов назад +1

      I think it’s a medical device that helps him breathe

  • @SjarMenace
    @SjarMenace 16 часов назад +4

    why do you have that thing on your nose?

    • @babyjvadakkan5300
      @babyjvadakkan5300 15 часов назад

      For correcting the nasal path/nose bridge (or something like that

    • @INeedMeme
      @INeedMeme 13 часов назад

      More oxygen bro

    • @cdunne1620
      @cdunne1620 4 часа назад

      Soccer players used to wear them years ago for example Robbie Fowler for Liverpool

  • @aaaaaaaaooooooo
    @aaaaaaaaooooooo 3 часа назад

    Are my prompts on o1-preview used to train the AI even if I opt out? Where do I find this information?

  • @costatattooz840
    @costatattooz840 18 часов назад +3

    locally what hardware do you need to run this at minimum? i have a 64gb ram + 3060 12gb

    • @ticketforlife2103
      @ticketforlife2103 18 часов назад

      Watch the video

    • @H3XM0S
      @H3XM0S 17 часов назад +5

      You'll need over 40gb vram so like 2 x rtx 4090 might be a good option. No idea what hardware is being used in the video. Anyone saying 'watch the video' should provide a timestamp.

    • @bollvigblack
      @bollvigblack 16 часов назад

      this guys is rich. not even joking so

    • @chrystofferaugusto1194
      @chrystofferaugusto1194 16 часов назад +3

      He is on a MacBook Pro bro…

    • @neomatrix2669
      @neomatrix2669 16 часов назад

      64GB RAM + 4070 Ti Super (16 VRAM) = Run Nemotron-70b-instruct-q2_K

  • @supermandem
    @supermandem 5 часов назад

    Bro llama is nowhere near o1 wtf

  • @gauravrewaliya3269
    @gauravrewaliya3269 8 часов назад

    How to make local ai with backpropogation feature ( if got wrong stuff, CEO instruct what's wrong and it improve sub local agent by time )

  • @SCHaworth
    @SCHaworth 11 часов назад

    No. Not quite. You have to split the turns.

  • @chrystofferaugusto1194
    @chrystofferaugusto1194 16 часов назад

    You should have a discord community to people share projects and business

    • @chrystofferaugusto1194
      @chrystofferaugusto1194 16 часов назад

      Never mind, now I got the business model on skool. Nice call, thinking about joining it

  • @claxvii177th6
    @claxvii177th6 9 часов назад

    1 token per second is too slow for any pratical use...

  • @gustavramedies2901
    @gustavramedies2901 3 часа назад

    David i would like to create sales agents,lead generators,receptionist,appointment setters and I want to sell them.Can you help 😢

  • @danieleduardo9800
    @danieleduardo9800 17 часов назад

    How’d you get composer in the sidebar?

  • @andreapinaroli974
    @andreapinaroli974 24 минуты назад

    43GB on a computer? That's not a good computer, that is an almost non existent computer. You're pretty much lying there

  • @aljosja3353
    @aljosja3353 15 часов назад

    Which computer u can use for local llm

  • @ShishuSud
    @ShishuSud 17 часов назад +1

    😇

  • @Álvaro-o5e
    @Álvaro-o5e 20 минут назад

    99% of free stuff sucks. One of them is this video. 20 minutes to answer "why is the sky blue?"

  • @HimaLoubi
    @HimaLoubi 11 часов назад

    😂 you need a graphic card with a price of a Tesla car to run that module locally ; btw you talk like 10.000word/min , 😅

  • @surendarreddys7298
    @surendarreddys7298 18 часов назад +2

    1st one to comment 😄