Комментарии •

  • @JeffGeerling
    @JeffGeerling 2 года назад +52

    For that first shot I had some weird perspective thing going on; I thought the servers were on the floor behind you and were giant!

    • @ServeTheHomeVideo
      @ServeTheHomeVideo 2 года назад +7

      ha!

    • @OTechnology
      @OTechnology 2 года назад +7

      Did you see the wild raspberry pi at 16:42?

    • @CrazyLogic
      @CrazyLogic 2 года назад +2

      @@OTechnology i did :) i'm not surprised either - just a shame is wasn't a CM rather than cabled in.

    • @Nobe_Oddy
      @Nobe_Oddy 2 года назад

      @@OTechnology I THOUGHT that's what it was, but I didn't go back and check until I saw your comment HAHAHA!!!! To think a $10,000 server heat exchanger is actually running a pi4!! LMAO!!!

    • @aninditabasak7694
      @aninditabasak7694 2 года назад

      @@ServeTheHomeVideo Yeah and I thought someone was hiding in the servers with a gun trying to put a bullet in your head.

  • @wernerheil6697
    @wernerheil6697 2 года назад +16

    EXCELLENT VIDEO, Patrick. Basic thermodynamics I/we applied 20 years ago during my Ph.D. thesis, for instance. "High-chem" and QC on pumps is key here: longevity of components INCLUDING the coolant mix and mastery of its phase diagram behavior, because x(@10+ years)*365/24/7 shall not change - EVER !

    • @ServeTheHomeVideo
      @ServeTheHomeVideo 2 года назад +11

      Very much so. That is why we did another video while we were up there on how these are tested. Look for that soon.

    • @zachariah380
      @zachariah380 Год назад +1

      @@ServeTheHomeVideo I'd Love to see a video on creative solutions data centers have used for cool source water - like loops into a large body of water like an onsite pond or lake, or something like the new "space air conditioning" which basically converts the heat into infrared via special polymer tubing, and irradiates it up into and out of the atmosphere - almost like a reverse solar panel.

  • @bryanv.2365
    @bryanv.2365 2 года назад +8

    Yes! This is what I was talking about! Please do more of this kind of build content whenever you get a chance!

  • @JeffGeerling
    @JeffGeerling 2 года назад +6

    2:00 Raspberry Pi spotted!

  • @benjamintrathen6119
    @benjamintrathen6119 2 года назад +2

    Your enthusiasm is infectious. Thanks for this series.

  • @jeremybarber2837
    @jeremybarber2837 2 года назад +2

    This video is SO timely for my job, unreal. Looking into the AHx10 CDU to see if we can add 1 rack of liquid cooling. Thank you!!

  • @tad2021
    @tad2021 2 года назад +22

    You know Cool IT is serious when they have STH doing this and not LTT. /s
    Joking aside, Linus would have been trying to twirl the latest and most expensive system in the building.

    • @ServeTheHomeVideo
      @ServeTheHomeVideo 2 года назад +21

      Yea we tend to do these boxes before Linus does. E.g. A100 box, we did 3x and including liquid cooled ones a year ago. There is a MI250X node that makes a cameo in this video. That is faster at FP64 than the next-gen NVIDIA H100 GPUs. We just do more informative rather than entertainment aimed at gamers.

    • @nathanielmoore87
      @nathanielmoore87 2 года назад +3

      Well Linus kinda failed epically with his DIY whole room water cooling back in 2015. It's probably a good idea he didn't get anywhere near this equipment.

    • @YouTubeGlobalAdminstrator
      @YouTubeGlobalAdminstrator 2 года назад +4

      And Linus' humour is for kids...

    • @n0madfernan257
      @n0madfernan257 2 года назад +2

      pains me when i see linus building awesome systems just to run games. datacenters throwing and crunching gigs/tera of data to me is what these bad boys are for minus the linus... just ignore my ranting

    • @beauregardslim1914
      @beauregardslim1914 2 года назад

      Cool IT knows that Patrick is less likely to drop stuff, especially after months of having to lift hardware to show it in the new set.

  • @BillLambert
    @BillLambert 2 года назад +8

    I remember CoolIT from back when they made AIO liquid coolers, 15 years ago! I had a slightly modded Freezone Elite, an absolute beast of a unit, which could bring my CPU well below ambient if I let it. I later picked up a used 2U server that had a CoolIT system mounted where the front bays would have been, whisper quiet and perfect for my homelab.

    • @ilovehotdogs125790
      @ilovehotdogs125790 Год назад

      I thought you can't go below ambient with water cooling. Best case you are equal to ambient.

    • @ledoynier3694
      @ledoynier3694 Год назад +2

      @@ilovehotdogs125790 It's a Peltier unit actually, so it makes sense. With normal AIO y'a, can't go below ambient though :)

  • @JortKoopmans
    @JortKoopmans 2 года назад +5

    I just love how amazed Patrick is about the thermal capacity of water, indeed very effective! 😃
    But yes, taking the numbers given, the water should be about 38°C warmer than the input, if loaded with 80KW. That's easily handled by tubing/pumps etc.
    A lot of energy still, think about heating 1800 liters of water by 38°C every hour! 😛
    Water flow at 30L/min = 0.5L/s
    Water thermal capacity is 4186J/(L·°C)
    4186*0.5=2096J/(°C·s)
    80KW=80000J/s
    80000/2096=38.2°C temperature increase

    • @ankittayal8291
      @ankittayal8291 2 года назад

      Ahh, now i should use the hot water convert it into steam to generate electricity 🔥

  • @jolness1
    @jolness1 2 года назад +5

    Excited to see professional targeted water cooling.
    I love the benefits of water cooling but when the options are building a custom loop using consumer grade parts or air.. I choose air every time. To have something well tested as an option is exciting.

  • @sembutininverse
    @sembutininverse 2 года назад +1

    thank you Patrick for the video 🙏🏻

  • @Nobe_Oddy
    @Nobe_Oddy 2 года назад +2

    THAT IS MIND BLOWING!!! SOOOO QUIET!! You can even be in the SAME ROOM as a air cooled server, but with this you could build a crib out of liquid cooled servers and your baby would still turn out to be normal (well... as normal as it can be with parents that would make a crib out of running computers lol)

  • @nindaturtles613
    @nindaturtles613 2 года назад +5

    What’s the temp difference between both CPU’s under full load ?

  • @jurepecar9092
    @jurepecar9092 2 года назад +7

    Most exciting thing here: miniDP port on the back instead of VGA! Yes! Servers are finally entering 21st century.
    Jokes aside, next step must be single usb-c to carry display and keyboard/mouse signals, getting rid of usb-A slots too. Lets wait and see if we get this still in this decade ...

  • @ColdSphinX
    @ColdSphinX 2 года назад +4

    That Raspberry in the cooling unit, so lonely.

  • @balrajvishnu
    @balrajvishnu 11 месяцев назад +1

    Loved it, amazing job explaining it

  • @LordSesshomaru584
    @LordSesshomaru584 9 месяцев назад +1

    Very well put

  • @jfkastner
    @jfkastner 2 года назад +1

    Super cool video, thank you!

  • @zachariah380
    @zachariah380 Год назад +4

    I'd Love to see a video on creative solutions data centers have used for cool source water - like loops into a large body of water like an onsite pond or lake, or something like the new "space air conditioning" which basically converts the heat into infrared via special polymer tubing, and irradiates it up into and out of the atmosphere - almost like a reverse solar panel.

  • @nullify.
    @nullify. 2 года назад

    I like how there's a Raspberry Pi running the whole CDU. That's crazy.

  • @carmonben
    @carmonben 2 года назад +2

    I spy a raspberry pi in that CDU :)

  • @Noi5eB0mb
    @Noi5eB0mb 2 года назад +4

    I love videos like this- regular pc liquid cooling is interesting, and flashy - but this, this is proper high power stuff. My question is, in the OCP standard, have they also standardised liquid cooling? Or is it still open?

    • @ServeTheHomeVideo
      @ServeTheHomeVideo 2 года назад +4

      It is not standard yet, but the hyperscalers are looking at it or deploying it already

  • @typeer
    @typeer 2 года назад +1

    What's up, welcome to YYC 😎

  • @ewenchan1239
    @ewenchan1239 2 года назад +2

    From a mechanical engineering perspective, the chiller will be the next target for development (and I'm sure that there are already a lot of folks working on that) in order to make the chilling process more thermally, electrically, and mechanically more effiicent.
    Afterall, it's thermal/heat management. The heat has to all go somewhere.
    And the more efficient that you can make that process, the more efficient the entire cooling loop will be, even if you were running a data center in Phoenix, AZ, for example.

    • @aeonikus1
      @aeonikus1 2 года назад

      Heat from Datacenter should be used to heat homes. How cool would be home colocation of quiet, compact mining rigs during the winter? I know it would be challenging technicaly and perhaps legally (who'd be responsible for what in case of sth goes wrong etc) but just based on ecological and cost of electricity terms it would be a win-win. But realistically, unfortunately it's hard for me to imagine that. Unless doing some centralised heating, like locating many rigs in big boiler rooms that serves buildings or small communities etc.

    • @ewenchan1239
      @ewenchan1239 2 года назад

      @@aeonikus1
      "Heat from Datacenter should be used to heat homes."
      I agree. If there is an efficient way to store and transfer the heat.
      Case in point, older cities in the US used to have a municipal steam generation to provide heat to buildings. There's a reason why they don't do that anymore, and a LOT of that has to do with the fact that even with really heavily insulated pipes, a LOT of that heat is lost during transmission to the surrounding environment, before you even get to your intended building.
      So, you would have to "overdesign" the heat generation capacity to make up for those transmission losses so that you would actually get the heat that you need, in the building.
      "How cool would be home colocation of quiet, compact mining rigs during the winter?"
      I can tell you that with the mining rig that I have at home, it is literally more than enough to heat our ~1000 ft^2 house with only about 1.5 kW "space heater" running (read: mining) 24/7.
      It's not technically a cost efficient way of heating the house (turning electrical energy into thermal energy) - our natural gas furnance is a LOT more efficient for that purpose, but so long as proof of work crypto mining is still relatively profitable, (i.e. it is enough to cover the cost of electricity), then it is still a net profit PLUS heating the house.
      In fact, we didn't turn off our A/C until WELL into December, and we had to turn it back on I think as early as like February this year. (And the electrical costs that I am talking about is the cost of my electrical bill and NOT just the mining rig itself - so it's both mining and cooling/cost of running the A/C starting in February as well.)
      "Unless doing some centralised heating, like locating many rigs in big boiler rooms that serves buildings or small communities etc."
      So....yes and no.
      There are some data centers that actually don't bother with chilling the incoming air (or water if they're using liquid cooled servers), but the problem is that then in the data center, you end up with hot spots which lowers the computational efficiency of the systems themselves.
      So, the entire thermal management loop needs to be optimised so as to optimise the performance of the individual systems whilst allowing them to get at close to the "hot" temperature threshold without exceeding it so that you're also not spending a lot of energy cooling the incoming air/liquid supply (which just means that you need something else to be able to dump the heat somewhere else).
      As a computational infrastructure problem, it's actually quite challenging to hit that balance over a wide range of environmental and operating conditions.

  • @Traumatree
    @Traumatree 2 года назад +1

    The hottest components in this server you showed are the 20 x SFF HDDs running at 10k+ in the front, not the CPUs! But nice video anyway! Another thing, this is a big waste of fresh water if it is not reused...

  • @-MaXuS-
    @-MaXuS- 2 года назад +1

    Thanks for the really cool 😉 video!
    Really awesome content as usual with and by Patric! Would have been interesting to see the actual heat exchange element in CDS. If I understood it correctly the external water source cools the closed loop of the server system connected to the CDS. Seeing as the CDS don’t have fans pulling the heat away I’m really curious as to how that heat exchange works. Also how is the then heated water cooled?

    • @someguy4915
      @someguy4915 2 года назад +2

      The exchange is basically a normal radiator but instead of air flowing through it they flow water through it. That outside water (now warm) goes back to the datacenter cooling system, through the chillers (large AC equipment designed to cool the water) and back into a storage/buffer tank to then be pumped right around again and again.
      This sort of equipment is usually already mostly in place at most datacenters as they provide the cooling used by the air conditioning.

    • @nexxusty
      @nexxusty 2 года назад

      CoXuS.

  • @balrajvishnu
    @balrajvishnu 5 месяцев назад

    This is very helpful, can you make a video on how the RDHX works

  • @catulescumarian
    @catulescumarian 2 года назад +3

    Super cool!

    • @seylaw
      @seylaw 2 года назад +1

      It is actually super hot underneath. :D

  • @docjuanmd7397
    @docjuanmd7397 Месяц назад +1

    I cant imagine the magnitude of a disaster if a failure would happen with those garden hose!

  • @jacj2490
    @jacj2490 2 года назад +3

    Great Job, Very informative to honest my only concern regarding water cooling is safety. Imagine a water leak in "Top Of Rack Server", it will ruin the entire cabinet. I know their must be some safety mechanism like measuring waterflow or difference in return vs outlet "Like Current Earth Leakage Concept" but still the concern is there. I have experiences with servers failing constantly due to humidity how about water leak. But you are right it is efficient hence it is the future
    Thanks again

    • @ServeTheHomeVideo
      @ServeTheHomeVideo 2 года назад +3

      There are leak detectors people use. We also have another video with a tour of the lab coming to help show what is done to test and build reliable liquid cooling

    • @sirdeimos8968
      @sirdeimos8968 2 года назад +1

      You can also just use a non water cooling fluid like Galden HT200 in your primary loop and move your water cooled secondary loop away from your electrical equipment.

  • @-MaXuS-
    @-MaXuS- 2 года назад +5

    Would also been super cool to see how the fully water cooled system looks. Just saying. 🙏

    • @ServeTheHomeVideo
      @ServeTheHomeVideo 2 года назад +4

      Are you thinking the -ZL1 with the RAM and NIC? In the next video (and there was a cameo in this video) there is a MI250X node that is fully liquid cooled.

    • @-MaXuS-
      @-MaXuS- 2 года назад

      @@ServeTheHomeVideo Correct! Oh that’s awesome! Looking forward to the next video! To be fair I always look forward to your content. This because what you guys share with us is quite unique here on RUclips so immensely appreciated! 👊👌🤓

  • @careytschritter1108
    @careytschritter1108 2 года назад +2

    Had I known you were in Calgary I would have taken you to dinner! Maybe even gotten you a Calgary hat 🤠

    • @ServeTheHomeVideo
      @ServeTheHomeVideo 2 года назад +1

      I selfishly snuck out to Lake Louise the Saturday after we filmed the two videos in this series. I am a big fan

  • @DrivingWithJake
    @DrivingWithJake 2 года назад +1

    It's amazing. Only downside is so many data centers are not built to accept liquid cooling. Raised floors with power under is not so friendly should anything leak. :)

    • @ServeTheHomeVideo
      @ServeTheHomeVideo 2 года назад +2

      That is why the next video is going to be on the Liquid Lab we are in and how they test these systems. Starting later this year, not having liquid cooling is going to mean that the data center will only host lower performance/ lower density servers. We have been doing more on server liquid cooling over the past year to start pushing this concept so our readers/ viewers are ready for the transition.

  • @RaspyYeti
    @RaspyYeti 2 года назад

    How much of the daily heat load and hot water load for a surrounding campus can be recovered from a sweat spot sized data centre using water cooling?

  • @jlinwinter
    @jlinwinter 10 месяцев назад

    super cool!

  • @florianfaber9799
    @florianfaber9799 2 года назад +3

    Who have spotted the raspberry pi? 😉

  • @jeanmarcchartier6438
    @jeanmarcchartier6438 2 года назад

    Is that a Raspberry Pi controlling the CDU? How does that work and does it have any affect on redundancy in the unit?

  • @JasonsLabVideos
    @JasonsLabVideos 2 года назад +1

    OHH you are in Canada !! Come to Vancouver Island :) Bring me that Red ZIP up Jacket ! :)

    • @ServeTheHomeVideo
      @ServeTheHomeVideo 2 года назад +3

      Heck yes! Back in the US now but hopefully Canada again soon

  • @AlexandreAlonso
    @AlexandreAlonso 2 года назад

    I wait to see how to setup central water cooling solution on the rack

  • @owenness6146
    @owenness6146 2 года назад +2

    Normally I'd look up or call Performance PCs and put together a solution.

  • @frakman
    @frakman 2 года назад +1

    where does oil cooling (immersion) sit going forward? Is it being sidelined?

    • @ServeTheHomeVideo
      @ServeTheHomeVideo 2 года назад +1

      Immersion is still going forward. Just two different technologies. Using direct to chip liquid integrates easily into existing racks.

  • @mikeoxlong4043
    @mikeoxlong4043 2 года назад +1

    U shuld check out linuses data centre ibm video. Ur take would be interesting.

  • @zachariah380
    @zachariah380 Год назад

    Tech question on the actual gigabyte server - how do these 4 nodes access all of the storage drives up front? Are they equally distributed across the nodes? Or somehow pooled so that all of the nodes can access them?

    • @ServeTheHomeVideo
      @ServeTheHomeVideo Год назад +1

      Each node has access to one quarter of the front drive bays

    • @zachariah380
      @zachariah380 Год назад

      @@ServeTheHomeVideo thanks!

  • @ledoynier3694
    @ledoynier3694 Год назад +1

    even on the consumer market liquid cooling is now becoming a thing. Intel 13th gen and AMD 7000 will be the first to require liquid cooling for the higher tier units

  • @rRefuseToSelfCensor
    @rRefuseToSelfCensor 2 года назад +4

    So is it pronounced Cool IT (Eye-Tee) makes sense at least. I always hear their consumer products pronounced as CoolIT (Cool it) is that intentional as their consumer products wouldn’t emphasize the IT part? Currently have one of their pumps in my AIO right now.

    • @ServeTheHomeVideo
      @ServeTheHomeVideo 2 года назад +2

      I used to say it as well not I T.

    • @rRefuseToSelfCensor
      @rRefuseToSelfCensor 2 года назад +1

      Id imagine so many people have mispronounced it that both are acceptable to the company now.

  • @Phil-D83
    @Phil-D83 2 года назад

    Cooled with 3m Fluorinert tupe fluid?

  • @sugipoohtube
    @sugipoohtube 2 года назад

    ありがとうございます!

  • @zactron1997
    @zactron1997 2 года назад +1

    So this might be a stupid question, but since you'll have a datacenter producing many kilowatts (or maybe even megawatts) worth of hot water, why not hook that up to a steam generator and try to recover some of power consumed? It wouldn't be close to break-even, maybe less than 20% energy recovery, but still better than just flushing that perfectly function hot water away.

    • @gloop6589
      @gloop6589 2 года назад +2

      To make that work would require the water to be heated to well over 100C, which means that whatever is heating the water would also have to be well over 100C.

    • @zactron1997
      @zactron1997 2 года назад +1

      @@gloop6589 I think you're right, but surely you could use a heat pump to focus the heat? Things like sterling engines can sap energy out of much less heat so I'm sure you could do some energy recovery...

    • @ServeTheHomeVideo
      @ServeTheHomeVideo 2 года назад +1

      There are several projects looking at re-using the heat from servers for heating water and building heat applications. Steam less so.

    • @gloop6589
      @gloop6589 2 года назад +1

      I'd think that generating an appreciable amount of electricity would thermodynamically require a sufficient temperature differential to ambient temperature, which is the exact thing that we're trying to minimize when cooling a datacenter. Could be wrong though.

  • @SamGib
    @SamGib 2 года назад +2

    How high is the risk of water leaks?

    • @ServeTheHomeVideo
      @ServeTheHomeVideo 2 года назад +1

      We do not use the "L-word" in the lab. :-) But the next video that we recorded there is how they test in the lab to make sure no L-word events happen.

  • @someguy4915
    @someguy4915 2 года назад +4

    Kind of surprising that the CDU seems to be controlled by a Raspberry Pi, haven't seen those in enterprise equipment before (besides a cluster in a 4U box type of thing).
    You'd expect that some ARM based microcontroller or a simple SoC would be the preferred choice, it usually is right?
    Did Gigabyte tell why they chose for the Raspberry Pi, is it due to concerns on chip shortages with ARM SoCs, easier development or anything else?
    Does other enterprise equipment use Raspberry Pi computers? Perhaps another collab/competition with Jeff Geerling? :)

    • @ServeTheHomeVideo
      @ServeTheHomeVideo 2 года назад +7

      I think CoolIT was using RPi's because they were easy to source. I also think they are being phased out on newer products. This was also an older version of the CHx80 because we were running it on its side.

  • @Veptis
    @Veptis Год назад +2

    could you evaporate the water to dump more energy?

    • @ServeTheHomeVideo
      @ServeTheHomeVideo Год назад

      If you see our PhoenixNAP data center tour you can see those.

    • @Veptis
      @Veptis Год назад

      @@ServeTheHomeVideo went and watched that video and it reminds me of a lot of specialty installations that were shown on der8auers tour through a Hetzner data center with like the diesel generators and raised floor.

  • @allansh828
    @allansh828 2 года назад +1

    how do you cool the water?

    • @ServeTheHomeVideo
      @ServeTheHomeVideo 2 года назад +2

      Usually there are chillers outside the data center floor at the facility. You can see them in our PhoenixNAP data center tour video.

  • @__--JY-Moe--__
    @__--JY-Moe--__ 2 года назад +1

    pretty cool! nice cooling strategy! I thought it was cool in Canada, anyway! ha..ha..!!!

  • @francisphillipeck4272
    @francisphillipeck4272 2 года назад +1

    Cool....

  • @billymania11
    @billymania11 2 года назад

    A good video Patrick but I wonder about something. Shouldn't Intel and AMD being working on their horrendous power consumption? I mean at some point the power draw of their processors becomes scandalous doesn't it?

    • @Bexamous
      @Bexamous 2 года назад

      Energy effiency is the thing that matters, it is going up.

    • @ledoynier3694
      @ledoynier3694 Год назад

      transistor count will always go up exponentially. but efficiency limits the increase in power draw. They are as efficient as they get right now. you have to consider both aspects

  • @clausskovcaspersen5982
    @clausskovcaspersen5982 2 года назад +1

    u are my hero :-) hehe naa but u are cool
    good videos
    thanks

  • @chechitogmail
    @chechitogmail Год назад

    I suppose You have a very Big garden

  • @zactron1997
    @zactron1997 2 года назад +1

    13:40 so that's why I can't buy a Raspberry Pi right now 😉

  • @mamdouh-Tawadros
    @mamdouh-Tawadros 2 года назад

    Forgive me, no matter how liquid cooling becomes robust, they don’t match with servers environment i.e. 24/7 work.

    • @marcogenovesi8570
      @marcogenovesi8570 2 года назад +4

      liquid cooling is used in industrial applications (and vehicles) 24/7 with no problems. The point here is what grade is the equipment. Consumer liquid cooling, heck no. They are using industrial-grade stuff here so it is fine

  • @zippytechnologies
    @zippytechnologies Год назад

    so now all you need is a 120 ft deep water tank to dump the water into with valves and tubes on the bottom (up a little to not catch sediment) tied to another tank... dump the heated water in the top pull from the bottom as it cools to the other cold water tanks (small with exchangers to transfer as much heat out as possible and a pump that then pushes the cold recycled water back up to the massive server farm... waste not want not... swap in some better conditioners and corrosion prevention chemicals or better coolant and you have a pretty slick eco friendly solution... so tired of all the videos that talk about all the heat but never explain how they get rid of the heat other than massive water coolers on a building or giant server center hvac systems... I want to see giant cooling under ground instead of putting all that heat right back into the air around us... maybe I am nuts.

  • @brandonedwards7166
    @brandonedwards7166 Год назад

    Anyone want to pay me to host a rack? I need to heat my pool.

  • @Kingvoakahustla
    @Kingvoakahustla Год назад

    The Asus server has better liquid submerged cooling gigabyte sorry components do not last.

  • @KantilalMaheta-yo4ro
    @KantilalMaheta-yo4ro Год назад

    O

  • @g1981c
    @g1981c Год назад

    you don't need to explain what you're going to explain and how you're going to explain it - just show it