ArrrZero: Why DeepSeek R1 is less important than R1-Zero

Поделиться
HTML-код
  • Опубликовано: 8 фев 2025
  • While everyone's talking about DeepSeek R1, the real game-changer is R1-Zero. In this video, I break down how this model eliminated multiple steps in traditional AI training, going straight from base model to reasoning chatbot in one giant leap.
    We'll cover
    How traditional LLM training from base model to helpful chatbot assistant
    Why current methods require extensive human annotation
    How R1-Zero bypasses these limitations using math and code problems
    A live demo of a simplified R1-Zero style training process
    Links mentioned:
    State of GPT talk by Andrej Karpathy: • State of GPT | BRK216HFS
    RAGEN replication: github.com/Zih...
    TinyZero replication: github.com/Jia...
    Willccbb replication: gist.github.co...
    💡 Want to understand AI better? Check out my "Spreadsheets Are All You Need" class where you learn to implement a real LLM entirely in Excel! maven.com/spre...
    #AI #MachineLearning #DeepLearning #AIEducation

Комментарии • 21

  • @emport2359
    @emport2359 6 дней назад +9

    Reallllly good video, reward: 2

  • @russellmiller3396
    @russellmiller3396 7 дней назад +3

    Really helpful learning why R1 0 was such a breakthrough. Made a lot more sense once you walked us through the intermediate steps they removed

  • @selvakumars6487
    @selvakumars6487 7 дней назад +1

    Thanks a ton! I heard all these terms in bits and pieces but couldnot quite wrap my head around them. You’ve done an amazing job of putting everything together and explaining the magic behind this model

  • @rajneesh31
    @rajneesh31 7 дней назад +2

    This was great Ishan, Thank you for the effort.

  • @proctussibelius8552
    @proctussibelius8552 6 дней назад +2

    Thank you for your humble explanation. You should get in more details in the future. View numbers are disappointing. Don’t worry. More people will appreciate for your work in future.

  • @DistortedV12
    @DistortedV12 6 дней назад +2

    Great channel!! You definitely know your stuff.

  • @MichaelLaFrance1
    @MichaelLaFrance1 8 дней назад +3

    Great video, clear and easy to understand. Will this efficiency boost keep open source models competitive with foundation models? Are billions & billions of dollars in GPUs still critical to AI advancement?

    • @Spreadsheetsareallyouneed
      @Spreadsheetsareallyouneed  8 дней назад +1

      @@MichaelLaFrance1 thanks glad you enjoyed the video!
      Regarding your question, it’s important to stress this video only really covers an efficiency gain that reduces human labor in the training process. Their model also had other efficiency gains that change the amount of compute they needed which I don’t cover in this cost.
      But that being said, my expectation is nuanced:
      (a) GPUs and compute will continue to be an important resource and moat(doing all those generations is still taking a lot of GPU work). Another way of looking at it is that the threshold number of GPUs needed to apply an LLM to tasks that are already solved has probably gone down but we still need more compute for the unsolved tasks. GPUs are like money. There’s always bigger problems you can spend it on no matter how many or how much you have.
      (b) I expect a pre-Cambrian explosion of models using this technique given how much simpler it is (and within the reach of research orgs who didn’t have the budget for all that human annotation) but can’t promise they’ll keep pace with closed source.

  • @Razkoh
    @Razkoh 7 дней назад +1

    great video!!

  • @aslampervez2294
    @aslampervez2294 2 дня назад

    Thanks

  • @nitefure
    @nitefure 4 дня назад +1

    + Liked
    +Subscribed

  • @Spreadsheetsareallyouneed
    @Spreadsheetsareallyouneed  6 дней назад +1

    If you want to see some of the questions and the model trying to answer them, here's a link to the spreadsheet I showed in the video: docs.google.com/spreadsheets/d/1IdPdA6eOurRP6EFb2uwYpUh1HdkHCvjtZ50fB0gLHOs/edit?usp=sharing

  • @christianOver9000
    @christianOver9000 7 дней назад

    They made this video so hard to find!!

  • @ajt_2023
    @ajt_2023 6 дней назад

    Cool video! Is the Jupiter notebook you present around the 9th minute available somewhere? If I would like to play with training something similar, which hardware would I need?
    Thanks