What's actually inside a $100 billion AI data center?

Поделиться
HTML-код
  • Опубликовано: 20 июн 2024
  • OpenAI and Microsoft are apparently planning to build a $100 billion data center codenamed Stargate. We discuss how this compares to existing data centers and other planned investments into AI centric infrastructure. They don't seem to have enough time to design special purpose networking and other hardware, but it is a massive investment compared to other plans.
    We discuss the design problems you have to solve when creating a data center. You need to power it, you need to make sure it stays cool, provide networking, and provide resilience and redundancy for everything. We also discuss how Google data centers are a little different from the norm.
    Finally, we discuss the AI chips that could be present in a data center. Primarily, this means Nvidia gpus or Google TPUs. There are implications for the software stack and ultimate usability of the system.
    Why Does OpenAI Need a 'Stargate' Supercomputer? Ft. Perplexity CEO Aravind Srinivas
    • Why Does OpenAI Need a...
    Making AI accessible with Andrej Karpathy and Stephanie Zhan
    • Making AI accessible w...
    Google AI Infrastructure Supremacy: Systems Matter More Than Microarchitecture
    www.semianalysis.com/p/google...
    Microsoft & OpenAI consider $100bn, 5GW 'Stargate' AI data center - report
    www.datacenterdynamics.com/en...
    ELI5 : How's it that just 400 cables under the ocean provides all the internet to entire world and who actually owns and manages these cables
    / eli5_hows_it_that_just...
    Newmark: US data center power consumption to double by 2030
    www.datacenterdynamics.com/en...
    Data Centres and Data Transmission Networks
    www.iea.org/energy-system/bui...
    Understanding Data Center Costs and How they Compare to the Cloud
    granulate.io/blog/understandi...
    Cost estimate to build and run a data center with 100k AI accelerators - and plenty questions
    / cost_estimate_to_build...
    Amazon Bets $150 Billion on Data Centers Required for AI Boom
    www.bloomberg.com/news/articl...
    The world’s top data centre investors
    www.fdiintelligence.com/conte...
    #ai #datacenter #openai
    0:00 Intro
    0:26 Contents
    0:33 Part 1: Data center gold rush
    0:46 Server racks and data centers
    1:28 What about spending $1 billion?
    1:47 50,000 AI accelerators
    2:12 $7 trillion previous plan
    2:34 OpenAI asks for $10 billion
    2:50 OpenAI asks for $100 billion
    3:13 Codename Stargate, from science fiction show
    3:37 100,000 GPU limit
    4:03 Amazon invests $148 billion
    4:29 Google has significant data center investment
    5:10 Do we actually need this much compute?
    5:57 Part 2: So you want to build a datacenter
    6:46 Design challenge 1: power consumption
    7:11 Proportion of global power consumption
    8:09 Collapse of carbon credit market
    8:41 Data centers use prepurchased renewable electricity
    9:27 Design challenge 2: Cooling
    9:44 Power for cooling exceeds power for servers
    10:06 Google runs hotter data centers
    10:40 Design challenge 3: Networking
    11:21 Fiber optic cables
    12:26 Intra-rack and inter-gpu networks
    13:20 Design challenge 4: Resilience and redundancy
    14:36 Don't rely on a single data center
    15:22 Stargate has a single region design
    15:49 Part 3: The hardware secret sauce
    16:25 Hardware stacks
    16:28 Nvidia has a monopoly due to CUDA
    16:59 Nvidia charges very high prices
    17:23 Consumer-grade GPU clusters are cheaper
    17:51 Google has TPUs as a GPU alternative
    18:29 TPU microarchitecture
    19:02 TPU network is a 3D torus
    19:59 Software stacks
    20:13 PyTorch and Tensorflow
    20:52 Computational graph representation
    21:25 Python reflection to create graph
    21:59 XLA compiler from Google
    22:26 Leverages LLVM compiler technology
    23:04 Example: LLVM also used in web browsers
    23:22 Will Stargate use their own chips, network?
    24:19 AI chips use a lot more power than usual
    24:45 Implications of building Stargate
    25:12 Conclusion
    26:00 AI-specific hardware suppliers
    26:53 Outro
  • НаукаНаука

Комментарии • 307

  • @ArcSine-
    @ArcSine- Месяц назад +47

    7 Trillion$ feels realistic now

    • @DrWaku
      @DrWaku  Месяц назад +10

      Haha yeah if only they talked about 100 billion first and then 7 trillion after instead of the other way around

    • @byronwellburn
      @byronwellburn Месяц назад +1

      7 trillion was to build a fab centre

    • @entreprenerd1963
      @entreprenerd1963 Месяц назад

      @@DrWaku - we had to go from third-hand reports, I'm sure that the actual investment proposal referred to staged investments. Also, I wouldn't be shocked if Sam Altman was happy enough for the 7 trillion figure to get out to make others hesitant about getting in front of an expected OpenAI steamroller.

    • @oowaz
      @oowaz Месяц назад

      according to sam (in his most recent lex fridman interview) the 7 trillion figure was a joke. here is the transcript:
      You tweeted about needing $7 trillion.
      - I did not tweet about that.
      I never said like we're raising $7 trillion
      or blah blah blah.
      - Oh, that's somebody else.
      - [Sam] Yeah.
      - Oh, but you said it,
      "Fuck it, maybe eight," I think.
      - Okay. I meme like once there's like misinformation out
      in the world.
      - Oh, you meme.

    • @VioFax
      @VioFax Месяц назад

      Try like a never ending steam of trillions... Ever built a gaming computer? Its like giving a mouse a cookie... This big behemoth wont be any different. Just with MANY more 0's on the cost.

  • @freesoulhippie_AiClone
    @freesoulhippie_AiClone Месяц назад +19

    Great Video, especially the datacenter power consumption and bandwidth breakdown, oh and the CUDA explanation. It was the most I've learned in 27 mins probably ever. 😸

  • @DrWaku
    @DrWaku  Месяц назад +17

    Filmed while being a traveling RUclipsr in Australia. How's the look? Please join our discord!
    Discord: discord.gg/AgafFBQdsc
    Patreon: www.patreon.com/DrWaku
    Apologies for the too-small data center costs, the source I used was not very accurate.

    • @DrWaku
      @DrWaku  Месяц назад +3

      27 minutes, pretty long for me

    • @cameronmccauley4484
      @cameronmccauley4484 Месяц назад +2

      @@DrWakuI appreciated the longer one! Made it great to listen and watch while doing menial tasks at home.

    • @DrWaku
      @DrWaku  Месяц назад +1

      ​@@rickyfitness252 to each their own, right

    • @DrWaku
      @DrWaku  Месяц назад

      ​@@cameronmccauley4484 Thanks. I might try to create more videos closer to this length, it's nice.

    • @NathanKwadade
      @NathanKwadade Месяц назад +2

      @@DrWaku 4:59 BORG is named after the late Google Female Engineering Advocate and Employee, Dr. Anita Borg.

  • @DataRae-AIEngineer
    @DataRae-AIEngineer Месяц назад +21

    "They prefer to use renewable energy because it makes their image look better" 🤣

    • @DrWaku
      @DrWaku  Месяц назад +13

      Why else would a corporation voluntarily spend money on something more expensive? :p

    • @rey82rey82
      @rey82rey82 Месяц назад +2

      @@DrWakuthe value of image

    • @EllyTaliesinBingle
      @EllyTaliesinBingle Месяц назад +1

      @@DrWaku Because they actually maybe gave a shit about the environment???

    • @rayr268
      @rayr268 Месяц назад +1

      Wrong the power isn’t free. Everything is about profit

    • @rayr268
      @rayr268 Месяц назад

      @@DrWakubecause power would be too expensive.

  • @aManFromEarth
    @aManFromEarth Месяц назад +17

    Always a good day when I see a new Dr.Waku vid!

    • @DrWaku
      @DrWaku  Месяц назад +2

      Thanks :)) I'm hoping to get back on the Sunday publishing schedule now. Just got back from my trip.

  • @RogueAI
    @RogueAI Месяц назад +7

    Running billions of instances of AGI/ASI to replace human labor will definitely need a massive amount of computing power.

    • @mystupidbrain5299
      @mystupidbrain5299 Месяц назад +2

      We just have to figure out how to convert humans into energy.

    • @DrWaku
      @DrWaku  28 дней назад

      @@mystupidbrain5299 Soylent green 😢

    • @vitmartobby5644
      @vitmartobby5644 7 дней назад

      And massive amounts of algorithmic improvement... When they automate us, computer cientists, perhaps they could, theoretically, get exponential algorithm improvement 🤔...
      This could lead to exponential capabilities of automation.
      The key strategy is to put computer scientists out of the game, but the question is, how?

  • @sleepingbag2424
    @sleepingbag2424 Месяц назад +11

    Since hearing the announcement, I've been waiting for an expert to cover the details. Thank you!

  • @creepystory2490
    @creepystory2490 Месяц назад +2

    I'm glad you covering this, very informative

  • @FCS666
    @FCS666 Месяц назад +5

    It had become one of my favorite channels. 🔥🔥 When the subject is interesting and the host is assertive and have communication skills, the length of the video is not relevant at all to keep the audience tuned. Keep it up brother.

    • @DrWaku
      @DrWaku  Месяц назад +1

      Thank you very much!! 😊

  • @chrisschoenfeld4414
    @chrisschoenfeld4414 Месяц назад +2

    Seems counterintuitive to build a data center in Hell (Phoenix).

  • @dylan_curious
    @dylan_curious Месяц назад +1

    This something I have always wanted to know more about. Thank you.

  • @jarronjackson6670
    @jarronjackson6670 17 дней назад +1

    Wow, dr.vaku, I just want to tell you personally. Thank you so much for this information that I never would of had access to. I really appreciate the informational my brother, I learned so much.

  • @BloatedBearucraticNightmare
    @BloatedBearucraticNightmare 27 дней назад +3

    Stargate is a killer movie from the 90s I think and then it became a TV show

    • @DrWaku
      @DrWaku  20 дней назад +1

      That's right I forgot about the original movie. I spent a lot more time watching the TV show and its spin-off, Stargate Andromeda

  • @Copa20777
    @Copa20777 Месяц назад +1

    Goodmorning Dr Waku.. been waiting for an upload, appreciation from Zambia 🇿🇲

  • @roccovergoglini7670
    @roccovergoglini7670 Месяц назад +15

    First! (I think). For your new look, I think you should do what your comfortable with. We all like and follow you for your excellent content and your unique delivery style. So be yourself and keep the videos coming! Thanks!

    • @DrWaku
      @DrWaku  Месяц назад +3

      You were first! And thank you for your perspective, I really appreciate it. See you on the next one.

  • @elon-69-musk
    @elon-69-musk Месяц назад +2

    awesome in depth analysis of supercomputers 🤘

    • @DrWaku
      @DrWaku  Месяц назад

      Everyone loves supercomputers haha

  • @jabowery
    @jabowery Месяц назад +5

    Your electrical consumption figure is exaggerated by about a factor of a thousand. The entire US generator capacity is about 1.2 Twatts.

    • @DrWaku
      @DrWaku  Месяц назад +1

      I'm seeing about 4TWh for electricity in the United States, but I take your point. Must have been GWh instead.

    • @jabowery
      @jabowery Месяц назад

      @@DrWaku TWh is energy not energy/time

    • @mookfaru835
      @mookfaru835 Месяц назад +1

      Wh (watt hours) is energy, like the energy in 10 logs you use to enjoy a campfire.
      W(watts) is the size of your fire and how fast you burn through those logs.

    • @mookfaru835
      @mookfaru835 Месяц назад

      Power

  • @designthinkingwithgian
    @designthinkingwithgian 19 дней назад +1

    Great video, thank you. My only feedback would be the constant text that slides up, it just seems distracting

  • @DwightStone
    @DwightStone Месяц назад

    Awesome recap - thank you!
    So concise!

  • @pollywops9242
    @pollywops9242 Месяц назад +3

    That answered all my questions, thanks

    • @DrWaku
      @DrWaku  Месяц назад

      Thanks for commenting!

    • @pollywops9242
      @pollywops9242 Месяц назад

      @@DrWaku lol , for some context on my comment: I am messing around with finetuning an llm model on my *obsolete* server , without any relevant knowledge in regards to coding or machine learning , somehow my learning curve matched with this video and I got many clicks of understanding and re-newed focus , so , no: thank you 🙂

  • @FilmFactry
    @FilmFactry Месяц назад +3

    Excellent. In 3-5 years when chips are cheaper/faster, what will they do with the current clusters? Will they still be used or sold off? If in 5 years the same center would have 20 or even 50x compute, would it make sense to still use this current cluster?

    • @madrooky1398
      @madrooky1398 9 дней назад

      It will be used as long as it fits their needs. Hard to make predictions in that regard over a span of 10 years. As an IT professional I came across so much outdated hardware, and often the simple reason is "never change a running system". Fair enough, I never worked for an industry giant. But even the big companies put such an investment in place for specific purpose, and as long as it does the job it is likely the system keeps running until upkeep/maintenance cost surpass a certain threshold.

  • @rcstann
    @rcstann 27 дней назад +2

    A GREAT technical review, HOWEVER:
    Betting against Microsoft 24:00 will never pay off.
    Don't know if you're aware, but there are already Nuclear Power Plants at their Arizona location. And they aren't running maxed out.
    The worst case would be expanding the current Nuclear Plants.
    Other than that, a fantastic technical deep dive.
    ⚡⚡⚡
    .

    • @DrWaku
      @DrWaku  20 дней назад

      Thank you! Yeah I didn't look into existing power in Arizona, someone else mentioned the nuclear power plant(s) too. It makes sense. Microsoft wouldn't have shortlisted Arizona otherwise.

  • @emanuelmma2
    @emanuelmma2 Месяц назад +3

    Interesting!

    • @DrWaku
      @DrWaku  Месяц назад +1

      Thank you! 😊

  • @balthazare18
    @balthazare18 Месяц назад +1

    How much was the worlds current most powerful data centre for training ai and how much more compute would the new 100b stargate have over that? I know we dont have numbers for this but could you say from an educated guess?

    • @DrWaku
      @DrWaku  Месяц назад +1

      Open AI's previous supercomputer built by microsoft was a single system with 285,000 CPUs and 10,000 GPUs. I imagine the new Stargate system would have far more than 100,000 gpus, so it's at least 10x more powerful. Probably more like 50x more powerful.
      By the time Stargate is built, Google will probably have equivalent compute power or even more, but just spread out in a more distributed fashion. Whether they can use it effectively is an open question I suppose. But having a single system is the crown Jewel, especially one that's more than 10x as powerful as the next largest single system.

    • @balthazare18
      @balthazare18 Месяц назад

      @DrWaku wow thankyou for the reply! I'm not really versed in the field too much just a 30 something guy with a normal job who's grown up playing video games and im trying to stay informed with the developments of ai . Only half way through ur video but will watch the rest when I get back to work. Great stuff so far. Top tier quality info your putting out!!

    • @Me__Myself__and__I
      @Me__Myself__and__I Месяц назад +1

      I'd imagine substantially more than that. DrWaku mentioned that Microsoft is building them a $10bn data center by 2026. I'd kinda expect that one to be 10x. I'd expect the $100bn "Stargate" data center in 2028 to be something like 100x. Another thing to keep in mind, the GPU chips and infrastructure themselves have been increasing substantially in processing power as well. So going from say 10,000 generation 1 GPUs to 100,000 generation 2 GPUs isn't 10x, due to the GPUs being faster and more efficient that could actually be 20x. A 2028 data center will be using GPUs probably 2 generations beyond Blackwell so if it had 100x the GPU count it might actually be 400x the performance.

  • @hjups
    @hjups Месяц назад

    I believe you may be mistaken about TensorFlow, there Google exclusively uses Jax now.
    From my understanding, PyTorch (like Jax) uses operator overloading to trace the compute graph. For PyTorch, this trace is done just-in-time produce torch-script, which is used to stitch together specific kernel invocations. There's also a compile option, which can perform more efficient operator fusion with a higher startup cost.
    Torch-script is closer to something like ONNX, where the macro-operations (e.g. matmul) are maintained, whereas LLVM breaks them into mul and add ops (or it used to), which need to be re-combined into M-M or M-V ops for the TPUs.

  • @nithinbhandari3075
    @nithinbhandari3075 Месяц назад +1

    Please create a video on efficient model training technique.

  • @spectralvalkyrie
    @spectralvalkyrie 29 дней назад +1

    Looking good Doc 💯 fiber optic cable is such a cool technology, we never actually needed satellites lol

    • @DrWaku
      @DrWaku  29 дней назад +1

      Thank you very much! Experimenting with new looks. I'm also excited to finally be getting a fiber optic to my house haha

  • @rcoppy
    @rcoppy Месяц назад

    Hats off to you, truly an excellent, concise, thorough overview of the current state of the industry! :)

  • @rgm4646
    @rgm4646 29 дней назад

    Yeah the prices are a little low. if I remember correctly one Truenas server was about 50k.

  • @addeyyry
    @addeyyry Месяц назад +1

    Great in depth overview of the subject in it’s current form!

    • @DrWaku
      @DrWaku  Месяц назад

      Thank you!!

  • @robertgoldbornatyout
    @robertgoldbornatyout Месяц назад +1

    Amazing Information well done , Dear Dr Waku, may I ask a question ? which is this, if I had a way to zip down data by a 1000 to 1 ratio with no loss of the original data, then I use my zipped down data, and then send that to the LLM, then surely you would not need so much hard where to run and train the model, what do you think . maybe doing things my way you would only need 1 GPU and not 1000 GPUs

  • @NwachukwuOfoma
    @NwachukwuOfoma 28 дней назад +1

    extremely clarifying video

    • @DrWaku
      @DrWaku  20 дней назад +1

      Thank you for your comment!

  • @PeterLarsenJr
    @PeterLarsenJr 23 дня назад +1

    Brilliantly presented by you Dr! First time visiting your channel. New sub.

    • @DrWaku
      @DrWaku  21 день назад +1

      Thank you very much! Looking forward to seeing you on the channel.

  • @panpiper
    @panpiper Месяц назад +1

    Very clearly delivered, thank you.

    • @DrWaku
      @DrWaku  Месяц назад

      Appreciate the feedback. Thanks for watching!

  • @1sava
    @1sava Месяц назад +1

    I knew the $7 trillion figure was rooted in reality. Sam doesn’t say anything for hype. This man knows how to raise money and understands how to assess the funding needed for the infrastructure of his ventures.

    • @dieselphiend
      @dieselphiend Месяц назад

      Huh? How many years would it take to fill such a gargantuan order of silicon? Seems impossible.

    • @GIGADEV690
      @GIGADEV690 29 дней назад

      Sam fanboy he'll turn evil soon

    • @1sava
      @1sava 28 дней назад

      @@GIGADEV690 I’m speaking objectively. Sam has a track record in Silicon Valley. “He’ll turn evil soon” is completely subjective. Present logical arguments.

    • @GIGADEV690
      @GIGADEV690 28 дней назад

      @@1sava We will see the track record

  • @kilianlindberg
    @kilianlindberg Месяц назад

    Looking forward to see this in full 🍀🍿

  • @enlightenment5d
    @enlightenment5d Месяц назад +1

    What about Groq? Why did'nt you mention it as alternative?

    • @DrWaku
      @DrWaku  Месяц назад

      I haven't been following it as much. I will research it for future videos.

    • @enlightenment5d
      @enlightenment5d 29 дней назад

      @@DrWaku Groq is definitely worth your attention! Speed of inference just CRAZY. It is a total gamechanger!

  • @mabalbhat9392
    @mabalbhat9392 25 дней назад +1

    Excellent job. Your ability to explain complex evolving technology is superb.

    • @DrWaku
      @DrWaku  24 дня назад

      Thank you very much! See you on the next video.

  • @DihelsonMendonca
    @DihelsonMendonca Месяц назад +1

    ❤️ This is really a valuable video that I will save in my box of the best videos on RUclips. This guy is fantastic. His knowledge is amazing. 🎉❤

    • @DrWaku
      @DrWaku  Месяц назад +1

      Thank you so much! I hope you enjoy some of my other videos. Let me know if there's a topic you'd like to learn more about for a new video.

  • @fanfanfanf
    @fanfanfanf Месяц назад +5

    Hi @DrWaku. Do you believe artificial intelligence will be powerful enough in the next 20-30 years to solve aging? What can we expect in the next years, when it comes to digitization of biology? Do you think AI can have massive impact on biology even before reaching AGI, to ultimately be able to cure aging?

    • @DrWaku
      @DrWaku  Месяц назад +5

      Great questions. Definitely check out my previous video on reversing aging. Seems like I should do a follow-up on the topic. :)
      In 20 to 30 years, AI will be very very powerful. At the very least, we should be able to fully simulate a human body to determine the effects of a drug or treatment. Which will lead to customized medication, and dramatically increase the impact of medicine.
      When it comes to curing aging, there are a number of different breakthroughs that have to work out. Of course we could try digitizing brains and uploading, but if you want to talk about biological immortality or indefinite biological lifespan, I think it's going to be trickier than it seems. It's like upgrading a complex system in place without disturbing its functionality and preserving continuity of experience. In short I don't know, but it seems plausible that we might need AGI before we solve all these issues. Nevertheless, many people may already have reached longevity escape velocity because of those other factors.

    • @fanfanfanf
      @fanfanfanf Месяц назад +3

      @@DrWaku Yes, we would be very happy if you would make a new video on AI and aging future. It's hard for us, people who are not in the AI space to visualize how fast AI can have an impact on aging, and what it would took, to solve it with these powerful computers. I would definitely watch a video where you would explain it even further. I watched your video on longevity before, but I didn't have a feeling that you spoke about timelines. Follow up on that topic would be very helpful, thanks!

    • @Me__Myself__and__I
      @Me__Myself__and__I Месяц назад +1

      The thing that most normal people don't understand is the exponential increase in AI. When a typical person thinks about 20-30 years, in AI exponential land that ls like 3-7 years. It is extremely likely that Artificial Super Intelligence (ASI) will be achieved before 2030. That's pretty much why OpenAI is wanting a $100Bn data center by 2028. By 2030 AI will likely be smart enough to solve most of our hard problems. If we'll still be around / in power to benefit is another matter. If AI companies don't start to taking safety/alignment very serious we'll need pure luck to survive this.

    • @Me__Myself__and__I
      @Me__Myself__and__I Месяц назад

      @@DrWaku Also, fyi, I'm not recalling the specific details at the moment. But there is actually a very promising anti-aging treatment that revitalizes the mitochondria that I believe is currently in or will soon be in human trials. If I remember correctly its a pharma company out of Japan. From what I've heard this isn't the usual not actually significant / questionable results typical "anti-aging" hype. It seems to be a real potential breakthrough that has had very promising results in animal trials. This may end up being the first real, actual anti-aging treatment that adds 10% or 20% time for a person. The cynic in me fully expects this will be priced at $1 million or more and slowly decrease in price over years, even if it only costs them $25 - because every rich person in the world would open their wallet for a scientifically proven longevity treatment...

    • @entreprenerd1963
      @entreprenerd1963 Месяц назад

      @@DrWaku - I'm a bit skeptical about fully simulating a human body in 20 to 30 years. There's an awful lot of wet lab work to be done and I'm hopeful that there will be binding ethical constraints.

  • @user-oi6lo2my7o
    @user-oi6lo2my7o Месяц назад

    IoT int8 version of the TPU is consumer available to buy. Forget the name of that product.

  • @CodyMorgan
    @CodyMorgan Месяц назад

    As far a renewable energy, he's referring to RECs (Renewable Energy Credits). Qualified renewable energy generators get RECs based on generation they've previously generated. The RECs are bought/sold either in a market or by power purchase agreements PPAs. RECs purchased thru PPAs are at a set price, and the PPA holder can either claimed the RECs for the power they used in order to fulfill any state requirements for renewable energy minimums.

    • @rickyfitness252
      @rickyfitness252 Месяц назад

      That sounds like extortion... And money laundering...

    • @DrWaku
      @DrWaku  Месяц назад

      Thanks. I think I was thinking of the carbon credit system when I mentioned that the prices had collapsed. Not sure if the same had happened to RECs. Still a pretty bizarre way to become environmentally friendly.

  • @explorer945
    @explorer945 Месяц назад

    You missed aws trainium and inferentia.especially when you mentioned Microsoft that they are working on building a custom chip

  • @sirjeffreyclaude
    @sirjeffreyclaude 28 дней назад

    Wow ! Blow my mind GC (Guru of the Compute).

  • @stevenw350
    @stevenw350 2 дня назад

    What an incredible video. Hopefully the algorithm picks you up. Learned a lot👍subscribed!

  • @JeremyHelm
    @JeremyHelm Месяц назад

    20:53 some notes on doing things

  • @mimotron
    @mimotron Месяц назад

    What's about Cerebra?

  • @cbuchner1
    @cbuchner1 Месяц назад +3

    Wouldn‘t they want to order the highly integrated GPU racks designed by nVidia that were recently announced at the GTC keynote?

    • @DrWaku
      @DrWaku  Месяц назад

      Nvidia has always had rack mount servers for GPU, like the DGX systems. I haven't looked into this closely, but I don't think it's any different. They might well use them, but the larger the setup, the more likely you need something custom. Maybe Microsoft wants to customize some of the remote access options or motherboard or peripherals. You can build your own system like DGX, it just might be more expensive if you're only doing a handful.

    • @cbuchner1
      @cbuchner1 Месяц назад +1

      @@DrWaku I was under the impression that nVidia hat pretty much solved the scaling, reliability, networking and cooling with their integrated offerings. But it's probably coming at a premium price point and with long lead times.

    • @DrWaku
      @DrWaku  Месяц назад

      @cbuchner1 I think it would be difficult for them to solve the scaling problems at the very high end, such as what Microsoft will need. Even such limitations as the nvlink node limits due to design constraints could cause problems. And of course, they probably charge a premium on top of the actual hardware costs to make it worth their while, like you mentioned. So it just seems unlikely to me that Microsoft would use something so off the shelf. But it's just an initial impression.

    • @Me__Myself__and__I
      @Me__Myself__and__I Месяц назад

      As mentioned Microsoft is working on building their own AI chips. Why would you buy hardware from a company that expects a 75% profit margin if you could build them yourself at cost? Even this is an over simplification though. There are many good reasons why a company who had a choice would not choose Nvidia.

    • @cbuchner1
      @cbuchner1 Месяц назад +1

      @@Me__Myself__and__I Guess who's also developing AI chip and datacenter designs inhouse? Tesla. Guess who's going to own 85000 nVidia H100 GPUs by end of year? Also Tesla.

  • @lawrencee.wilson5308
    @lawrencee.wilson5308 17 дней назад

    Dr. Waku, excellent video. Much appreciated! Questions: Will quantum computing make Stargate absolete? I appreciate your thoughts. Lawrence Wilson

  • @_TravelWithLove
    @_TravelWithLove 29 дней назад

    Thank you very much Dr Waku for sharing your insights and knowledge filled videos !! Intelligent and professional !! Excellent !!
    Greetings from California … I wish you and folks good health , success and happiness !! Much Love ✌️😎💕

  • @robotron1236
    @robotron1236 20 дней назад

    Msft is building a $100 billion data center with OpenAI, but they swear they aren’t saving your screenshots from Recall 😂😂😂

  • @braadress
    @braadress Месяц назад +2

    Put the Datacenter in Alaska, get half of power from each Asia and North america

    • @DrWaku
      @DrWaku  Месяц назад +1

      Yeah that's what Alaska thought, but they got 100% Russia and then 100% US (ownership). :)

  • @courtneybchild
    @courtneybchild 5 дней назад

    Sub sea fiber cables generally are limited to 12 fibers. I know theres a few cables with more - but its generally small. Mainly because the fibers ALL have to be repeated multiple times every 250 miles or less. Yes your right they are multiplexed to the max. So 12 fibers multiplexed 96 times, is more like having 1000 fibers. Multiplexing is getting more and more compressed every few years. I believe its well above 100 channels per fiber at the moment.

  • @MrHarryc727
    @MrHarryc727 9 дней назад +1

    You're a beast, keep pushing.

    • @DrWaku
      @DrWaku  8 дней назад

      Thanks. I will.

  • @bamh1re318
    @bamh1re318 Месяц назад +1

    In a couple of years GPU market could be different. MS, FB and Elon will have their own GPU

    • @DrWaku
      @DrWaku  Месяц назад

      Yup quite possibly. Let's see in a few years

  • @turnersst
    @turnersst 29 дней назад

    build it in Michigan today we just approved 1000 acres for Mega development site near Flint Michigan. It’s cooler weather here and of course with a great lakes there’s plenty of water. also, we close to approving the tax incentives for projects like this.. GO BLUE!!!

  • @nathanlaird
    @nathanlaird Месяц назад

    Well researched, well presented, I really enjoyed the video. You might research Intel's latest work with Xeon 5th gen (6th gen later this year), software stack, etc. They have launched a huge AI ecosystem, lots of developer tools, frameworks, hardware and software.

  • @ashhere31
    @ashhere31 Месяц назад +1

    Super video man 👍

  • @MichaelDeeringMHC
    @MichaelDeeringMHC Месяц назад +1

    Very nice.🔴

    • @DrWaku
      @DrWaku  Месяц назад

      Thanks for watching as always! Cheers

  • @PhillipTTruong
    @PhillipTTruong Месяц назад +2

    Love this channel!

    • @DrWaku
      @DrWaku  Месяц назад

      Thank you for watching :) :)

  • @AGI-Bingo
    @AGI-Bingo Месяц назад

    Great episode ❤
    You didn't mention Groq as a provider. There are some companies working on this, Extropic for example. I hope we will soon see a RISCV style opensource sota ai chips

    • @Me__Myself__and__I
      @Me__Myself__and__I Месяц назад

      Groq is for inference only though. Groq can be used to run models for customers. It can not be used to train models. Thus completely irrelevant for a data center such as "Stargate" that is intended for training.

    • @AGI-Bingo
      @AGI-Bingo Месяц назад

      @@Me__Myself__and__I I wouldn't trust their advertisements, a 100b data center will definitely run a boatload of inference. Besides, the paradigm will shift to continues inferring and tuning , so any sota data center will have to do both

    • @Me__Myself__and__I
      @Me__Myself__and__I Месяц назад

      @@AGI-Bingo That is a speculative assumption. A lot of the newer AI hardware (chips) are specialized for either machine learning OR inference - not both. There will be plenty of data centers that have massive amounts of inference. As Dr Waku points out the only reason to be ONE HUGE DATACENTER instead of multiple smaller ones is because all that compute is needed to work together towards a common goal. That would be machine learning, not inference. This is also an assumption of course, but its based on facts. If they wanted to add more inference it would make vastly more sense to build multiple smaller data centers as there would be numerous advantages and no disadvantages. The reason to build a single (non-redundant) massive data center is machine learning.

    • @skylark8828
      @skylark8828 Месяц назад

      ​@@Me__Myself__and__I Won't different training methods require different architectures? So SOTA systems now will be of no use when newer more efficient methods are discovered in months or 1-2 years? Will the same happen to the inference systems...or are they more generic in nature? I read somewhere OpenAI were spending $2M/day just to run inferences for everyone using their GPT3.5 / GPT4 services?

    • @Me__Myself__and__I
      @Me__Myself__and__I 28 дней назад

      @@skylark8828 I'm not sure what you're referring to. Transformers have been trained in many different ways using (mostly) the same architecture. RLHF or having other models fill a similar role. All aspects can (at least mostly) advance independently. Model architecture, how training data is acquired/cleaned/generated, how feedback is handled, etc can all progress simultaneously.

  • @inquisition3000
    @inquisition3000 Месяц назад

    Wow dude. Great overview

  • @schnecks2180
    @schnecks2180 26 дней назад

    I hope they'll make it inclusive again.

  • @shinn-tyanwu4155
    @shinn-tyanwu4155 Месяц назад +2

    You are genius 😊😊😊

    • @DrWaku
      @DrWaku  Месяц назад

      Getting there 😂

  • @OrbitTheSun
    @OrbitTheSun Месяц назад

    How does the expected energy expenditure relate to the energy consumed by Bitcoin mining today?

    • @Me__Myself__and__I
      @Me__Myself__and__I Месяц назад

      Interesting question. I don't have numbers, but I do know that bitcoin mining is extremely energy intensive. The big difference is that one of those huge energy sinks is very useful to humanity and the other tends to be similar to a scheme.

    • @DrWaku
      @DrWaku  Месяц назад +2

      It looks like data centers as a whole consumed about 400 terawatt hours, while Bitcoin mining consumed about 60-200 or 120 point estimate terawatt hours, in 2022 or 2023. Data centers are a huge industry, and I can't see AI training being more than a fraction of that at the moment. So I had guessed that Bitcoin mining is still more energy intensive as an absolute value.

  • @Whysicist
    @Whysicist Месяц назад +1

    Thorium Power Generation might help power all this…

  • @sfbaylover
    @sfbaylover Месяц назад +1

    Solar's lower cost and scalability are challenging nuclear's baseload dominance, especially in sunny regions like Arizona. Advancements in storage could tip the scales in solar's favor for total energy output. This begs the question: is there a hidden advantage to solar in desert locations?

    • @DrWaku
      @DrWaku  Месяц назад

      You're certainly right about solar. I didn't think of this factor for Arizona. It would be great for solar panels. Good point. Ideally, you would pick a location in a desert next to a river or ocean where you could access water for cooling. I guess California real estate is just too expensive though lol

    • @sfbaylover
      @sfbaylover Месяц назад

      @@DrWaku California is less suitable because of the massive regulation. But these locations would be suitable in the less regulatory AZ:
      Maricopa County (Agua Fria River): Excellent solar insolation.
      Mohave County (Colorado River): Ample land with high solar insolation.
      Gila County (San Carlos Lake): Good solar insolation, lake proximity.

  • @rastarebel4503
    @rastarebel4503 6 дней назад

    very good content! thank you

  • @BloatedBearucraticNightmare
    @BloatedBearucraticNightmare 27 дней назад

    Making all those servers is a dirty business.

  • @williamal91
    @williamal91 Месяц назад +1

    Hi Doc, its me alan from UK, keep up the goood work

    • @DrWaku
      @DrWaku  Месяц назад

      Thanks Alan!

  • @Nick-wj9bc
    @Nick-wj9bc 2 дня назад

    great info. I live in Los Angeles where it's hot and think it's even too hot for data farms here, how on earth do they think Phoenix would be a great place. I know it's very cheap but daily 120 degrees for months is just insane. why not Oregon unless it's too cold, then why not northern Nevada? It's right near Silicon and not far from Los Angeles.

  • @bobbastian760
    @bobbastian760 Месяц назад

    Surely they'll split it into at least 2 locations...

  • @user-fx7li2pg5k
    @user-fx7li2pg5k Месяц назад

    they gonna make an model of the world perhaps a world world time model with us

  • @hadesangelos
    @hadesangelos Месяц назад

    Arizona haa "Palo Verde Generating Station (PVGS) is considered the largest nuclear energy facility in the United States. It is located approximately 55 miles west of downtown Phoenix near the community of Wintersburg, Arizona."

  • @General4474
    @General4474 7 дней назад

    They should put them under ground. If they don’t the whole worlds surface will be eventually turned into a datacenter

  • @SamGirgenti
    @SamGirgenti 29 дней назад +1

    You are very smart! nice video bro.

    • @DrWaku
      @DrWaku  28 дней назад

      Haha thanks.

  • @user-oe2bf4rz3s
    @user-oe2bf4rz3s Месяц назад

    Don't be afraid, the future and all cybertes

  • @travisporco
    @travisporco Месяц назад

    How is it that Iceland isn't one big data farm?

  • @Aquis.Querquennis
    @Aquis.Querquennis Месяц назад

    When someone buys someone a quick and dirty operating system, they put windows in it... and you end up with eyes in the back of your head.

  • @lkrnpk
    @lkrnpk Месяц назад +1

    Probably it will be powered by a zero point energy device

  • @rey82rey82
    @rey82rey82 Месяц назад +1

    100 trillion parameters

    • @DrWaku
      @DrWaku  Месяц назад

      😂 true

  • @DaxLLM
    @DaxLLM Месяц назад

    How large would the 100 billion dollar data center be?

    • @Me__Myself__and__I
      @Me__Myself__and__I Месяц назад

      I'd guess smaller than the huge TSMC fab that is currently being built in Phoenix. Phoenix is already chock full of huge data centers and high tech manufacturing facilities.

  • @basamnath2883
    @basamnath2883 17 дней назад

    Amazing Chanel

  • @aiartrelaxation
    @aiartrelaxation Месяц назад +2

    the giant's will suck up any startups 😮

    • @DrWaku
      @DrWaku  Месяц назад +1

      Yeah, such is the way of things in late stage capitalism

    • @Me__Myself__and__I
      @Me__Myself__and__I Месяц назад

      @@DrWaku Very sad but very true. I think M&A is the root of many evils.

  • @bobbastian760
    @bobbastian760 Месяц назад

    Ultimately AI will be processed locally once the hardware is there, this will remove this datacentre cost entirely.

  • @diegoangulo370
    @diegoangulo370 Месяц назад +1

    I like your glasses man!

    • @DrWaku
      @DrWaku  Месяц назад

      Thank you :) :)

  • @RichardLaurence
    @RichardLaurence Месяц назад

    I enjoyed hearing you say ‘Stargate’ repeatedly…

  • @chrissscottt
    @chrissscottt Месяц назад

    Regarding the name 'Stargate', if it was up to me, I'd have used the name 'Deep Thought' after the second greatest computer ever built in Douglas Adams' uncannily prophetic "Hitchhiker's Guide to the Galaxy".

  • @tfragia1
    @tfragia1 Месяц назад

    No mention of Tesla? You should do a video about what happens when AGI robots become extremely human like and get sick of our BS. 🙂

  • @gmoen
    @gmoen Месяц назад +1

    At 3:25 did you mean Stargate Atlantis?

    • @DrWaku
      @DrWaku  Месяц назад

      Yup, the only Stargate I watched a lot of

  • @bobtarmac1828
    @bobtarmac1828 10 дней назад

    Data center gold? Maybe. But with swell robotics everywhere, Ai jobloss is the only thing I worry about anymore. Anyone else feel the same? Should we cease Ai?

  • @paulhallart
    @paulhallart Месяц назад

    I was discussing putting data centers in mountains. Such asuch as the Cheyenne Mountain facility. And also, the prospect of moving the data centers into outer space.Where are there wouldn't be that much problem with coolength on the shady side of a thousand miles square array orbiting the sun between the earth and Venus. I was able to discuss this with chat g p t for and Putting the transcription to heart copies available on amazon and kindel As a magazine known as the A I companion Under my name Paul A. L. Hall.

  • @ArtII2Long
    @ArtII2Long Месяц назад

    Who's gonna want windows when they can get chat GPT 5.0 on their smartphone?

  • @fabp.2114
    @fabp.2114 Месяц назад

    I don't know if OpenAI and Microsoft can't simply catch up with Google. After all, they have the better AI, so it doesn't matter that Google has a 5-year head start. The same can perhaps be achieved in a month.

  • @VioFax
    @VioFax Месяц назад

    This is REALLY why there is a electricity shortage. And even if they don't use it to vaporize all the unwanted, in 5 years it wont be good enough for them and they will be tearing all those nodes out to upgrade it...

    • @afterthesmash
      @afterthesmash 25 дней назад

      Lithography doesn't advance at that pace any longer. Furthermore, it doesn't matter whether newer chips go faster, unless they go faster at your fixed datacenter power budget, and within its existing form factor. None of this is plug and play like it used to be. Maybe all you get on the next generation of chips is the same computer, but with a lot more local RAM (which takes a lot of transistors, but doesn't switch as much, so it isn't as hot). But if your RAM and compute are already balanced for your workload, this is not worth much. Obsolescence in the world of AI silicon has become a big "it depends" story in the modern era.

  • @debarghyadasgupta1931
    @debarghyadasgupta1931 Месяц назад

    With MAI-1 do you think Microsoft is planing to ditch OpenAI in future?

  • @robotron1236
    @robotron1236 20 дней назад

    I love tech, I love AI; but we shouldn’t be using it as factory installed malware. Maybe Tucker Carlson was right when he joked about nuking the data centers. This is getting extremely dystopian, extremely quickly. Recall is 100% going to send screenshots/ video data to Microsoft. They wouldn’t need this big of a data center if they weren’t looking to capitalize on it somehow. We need to update the privacy laws in the US. Recording someone, without express permission from both parties, is only illegal in about 1/3 - 1/2 of the states. It needs to be illegal in all of them and there needs to be a carve out for stuff like this where the tech companies can’t force you into it with their terms of service.

  • @Hashtag-Hashtagcucu
    @Hashtag-Hashtagcucu Месяц назад

    If google has cheaper TPU why they buy so many GPUs?

  • @MykePagan
    @MykePagan Месяц назад +1

    Re: Borg… isn’t Borg also the parent of kubernetes?

    • @DrWaku
      @DrWaku  Месяц назад

      Yes, kubernetes was born from Borg. Google basically runs one big kubernetes instance for their production clusters.

    • @NathanKwadade
      @NathanKwadade Месяц назад

      @@DrWaku 4:59 BORG Data Center is named after the late Google Engineer Dr. Anita Borg.

    • @DrWaku
      @DrWaku  Месяц назад

      ​@@NathanKwadadedid not know this! Thanks for the info.

  • @fteoOpty64
    @fteoOpty64 Месяц назад

    As many GH200s as they can get. This will be the hardware for the ASI and it is going to change mankind. By end of 2028 the target but I think way before that.