this is the fastest AI chip in the world: Groq explained

Поделиться
HTML-код
  • Опубликовано: 22 янв 2025

Комментарии • 37

  • @remelin75
    @remelin75 11 месяцев назад +1

    Really good video. Looks very professional and the information was very easily understood. Great work!

  • @kylebehrend
    @kylebehrend 11 месяцев назад +6

    Quick is an understatement. I wonder if we get to a point where the systems actually record our inputs before we press enter. E.g. they are working in the background as we type in our input and when we press enter it may be instantaneous ?

  • @Shannenna-f7v4ggeu
    @Shannenna-f7v4ggeu 11 месяцев назад +4

    Current Groq configuration comes with limited stroage space of SRAM and it would not be sufficient for Training but Inference purpose. It 's comparable to Tesla's Dojo but with lesser computation power of Tesla Dojo. However when the potential buyers would have to ponder the limitation before making the purchase decision, Layers of KV Cache and Batch Size would make it only suitable for medium size computation. Not to mention that Complexity of Software should be taken into the consideration.

    • @waterbot
      @waterbot 11 месяцев назад +1

      i feeel like this first chip is more a proof of concept or a marketing effort from Groq more than anything, this video is a perfect exapmle of that.

    • @morethisdayinai
      @morethisdayinai  10 месяцев назад

      gotta make stuff for us hype boiz

    • @QH96
      @QH96 9 месяцев назад

      Hopefully their next chip fixes these shortcomings.

  • @sophiekandul6737
    @sophiekandul6737 11 месяцев назад +11

    Groq was before Grok

  • @RobertFletcherOBE
    @RobertFletcherOBE 8 месяцев назад +2

    1:37 That "Silicon Wafer" looks suspiciously like a circle of corrugated cardboard covered in tinfoil...

  • @dearlove88
    @dearlove88 11 месяцев назад +4

    It’s crazy though that it’s will cost about $2 million worth of cards to run a 70b model

  • @baheth3elmy16
    @baheth3elmy16 11 месяцев назад +8

    Thanks for the video! So is Groq going to be sold as a chip sometime? Like will it be installed on a motherboard like a GPU or RAM or NVME?

    • @razoraz
      @razoraz 10 месяцев назад

      I was thinking about this also - maybe as an add-on PCIe card. For the crowd that may believe that like "not your keys, not your coins" - "not your cloud, not your data". I could see Apple as a candidate for buying them to make their already-fast Neural Engine on the motherboard into something that puts OpenAI out of business.

  • @UselessHumanBeing
    @UselessHumanBeing 11 месяцев назад +3

    Obligatory comment. Hope to keep seeing more videos from your channel 🙏

  • @P__114
    @P__114 11 месяцев назад +10

    The branding confusion with Grok is going to be tough. One needs a rename.

    • @amykpop1
      @amykpop1 11 месяцев назад

      Nah, it's funnier this way.

    • @razoraz
      @razoraz 10 месяцев назад +1

      @@amykpop1 What would be even funnier, and differentiating, is if they decided to tell everyone to pronounce this one "groque", like "baroque" 😅

  • @backacheache
    @backacheache 11 месяцев назад

    I wonder if it is better with energy efficiency too?

  • @ngothianhquy7561
    @ngothianhquy7561 9 месяцев назад

    Does anyone know how to invest in this company

  • @WJohnson1043
    @WJohnson1043 9 месяцев назад +3

    Groq was never explained. You just quoted the obvious time and time again. The title is just click bait.

  • @byrnemeister2008
    @byrnemeister2008 11 месяцев назад +1

    So there is a one off gain in putting the algorithm / transformer. In a perfect world that one off gain should be in the range of 50 to 100 x. But this costs a lot of money and is a big gamble. If that algorithm is tweaked then you need new chips. So it’s a balance between flexibility and the resulting risk and cost. All the major service providers and semi companies are looking at this. Groq have a time to market advantage but it’s not going to be that big. 6-12 months.
    Big message. Short Nvidia stock.

  • @codediporpal
    @codediporpal 10 месяцев назад

    I imagine NVidia will have an LPU inference only offering soon.

  • @iritesh
    @iritesh 11 месяцев назад +1

    Didn't explain how it works but 👍

  • @g0d182
    @g0d182 10 месяцев назад

    barely any difference between the gpt3.5 and groq respond times at 0:20 and 1:00

  • @frankforrester42
    @frankforrester42 8 месяцев назад

    it cost the price of a new car

  • @toulasantha
    @toulasantha 11 месяцев назад

    Chamath is the Goat ❤

  • @user-ni3ti9cz5f
    @user-ni3ti9cz5f 10 месяцев назад

    Good content but I want to know who I’m getting my info from. Are you an engineer?

  • @johnpope1473
    @johnpope1473 11 месяцев назад

    (Can you guys stop mentioning simtheory on podcast - I've been on waitlist for 4 months ?? - I hit the page each week in hot anticpation only to be disappointed things are still closed.... sighh......(fix this please))

    • @morethisdayinai
      @morethisdayinai  11 месяцев назад

      Join the discord and send us a message with your email and we can let you in straight away: discord.gg/Rf7v6SZB

  • @Veneer22
    @Veneer22 11 месяцев назад

    Damn bro! Can see all the whites of your eyes! You on that high end addy or what?

  • @edwincloudusa
    @edwincloudusa 11 месяцев назад

    Why would you have that annoying background music that makes impossible to listen to what you are saying? I don't get it...