The Secret to NVIDIA Success
HTML-код
- Опубликовано: 21 сен 2024
- Invest in Blue-chip Art by signing up for Masterworks: www.masterwork...
Purchase shares in great masterpieces from Pablo Picasso, Banksy, Andy Warhol, and more.
See important Masterworks disclosures: www.masterwork...
Mentioned Videos:
Tesla DOJO Explained: • New Tesla DOJO superco...
Tesla DOJO Update: • Tesla AI Day 2: What’s...
Graphcore AI Chip: • The World’s First WoW ...
👉 Support me at Patreon ➜ / anastasiintech
📩 Sign up for my Deep In Tech Newsletter for free! ➜ anastasiintech...
Sign up for Masterworks www.masterworks.art/anastasi and purchase shares in great masterpieces from Pablo Picasso, Banksy, Andy Warhol, and more.
SC12+ FCSG< nb
*ur so fine woman*
My Wife in another Life.Detroit Michigan 48221. Thank you very much. I sub, liked and commented : Video paid for.
You look beautiful in black
No new video for so many days will put me into cardiac arrest
Gaming demand didn't cool so much as gamers are sick of the huge price increases.
I tend to agree
Exactly
That isn't going to change any time soon...unfortunately. They are selling GPUs hand over fist right now...regardless of gamers.
That's total bullshit. 7 million rtx40 GPUs sold in 2023, the 1st half of the year is ALWAYS slower than the second half. More than 3% of PC Gamers have an rtx40 gpu already. Gaming may have cooled relative to the pandemic, but gaming GPUs are still selling fine. The rtx4060 is CHEAPER than the rtx3060. Nvidia has REDUCED prices at the bottom of the stack! People are not generally using RTX gpus for non gaming purposes, there are better options for professional workloads.
@@Wobbothe3rdwhat's better than the RTX A6000 Ada?
Anastasi, your videos are amazing. I've only been watching for maybe 5 months, but have been going back and watching all your stuff. Sometimes I watch the same ones several times since there is always so much information presented. Please keep them coming.
deep dive! first video I watched which actually explained well why everyone is so exited about Nvidia and GPU rush.
Anastasi, the H100 is a monolithic design, not a chiplet design. Chiplet designs refer to designs that split the different components of a [traditionally singular] die across multiple dies on the same package, for example AMD's Epyc or MI300 processors having 'chiplets' of cores and a 'chiplet' for the memory controller. In datacenter products, HBM is often packaged with the die but this is not what a chiplet design implies. Also, there have been converters that can transform CUDA code to roughly corresponding OpenCL or HIP code for a while, but CUDA will never be able to run natively on AMD hardware as CUDA compiles to PTX, a proprietary virtual instruction set.
Sorry my mistake
absolutely, chiplet was initially designed by AMD for their cpus, and then applied it to their gpus. At first, it was simply MCM with 2d architecure, which was really good but had some drawbacks, like latency... Then they released chiplet with 3d Vcache, allowing for better performance, lower latency, better scalability and lower costs
I wonder if AMDs software stack is being underrated. I avoided their graphics cards for years because of the ROCm horror stories I had heard; and I think that was a smart choice at the time. But a month ago I bought a Radeon 6700 XT for my girlfriend's computer, and even without official ROCm support (after changing one line in a text file) it's been incredibly easy to use for stable diffusion and Leela Chess Zero. I'm just a hobbyist and don't have any commercial application for ANNs at the moment, so their are probably use case issues that I am completely unaware of. But I was absolutely amazed at how easy it was to get ROCm up and running on an AMD consumer GPU (it took like an hour, and it was only that long because of slow driver downloads). I ultimately found it less of a hassle than using CUDA on my Nvidia RTX 2060 Super because (at least in Ubuntu 22.04) the drivers that make it work for CUDA break it for Steam and vice versa. After setting up the 6700 XT with ROCm I was so sure I'd have problems getting Steam to run games, but everything went smoothly. I was simply mind blown. I hope AMD brings official ROCm support to their 7000 series cards ASAP. As a hobbyist I don't care that Nvidia cards might train a faceswap model slightly faster, I just care that everything runs smoothly.
China agrees and wants ROCm on RISC-V
Thank you for making things so easy to understand
There is no parallel between the two: gold retains its value, CPUs and GPUs become e-waste after 10 years or so
Indeed. One is a store of value, and the other is an investment. You can't really generate an income from gold, though, unless you're a pawnbroker buying at a massive discount.
@@klin1klinom actually Gold is both
e-waste in much lees than 10 years, maybe 5 years average and often much less than that.
What's the point of that nonsense? If you want to compare gold value, you compare it to investments in the ticker of those companies, make no sense to compare it to physical chips. Like yeah, obviously, nobody has ever bough a chip expecting it to go up in price ...
Chips can be used to make money, gold just sits there. Maybe gold increases in value, maybe not so much, if you're hoping gold is gonna make you rich if civilization suddenly goes headfirst into the crapbasket, you'll find you have much bigger problems than your portfolio. Value of any sort is ultimately faith based and situational, subject to unexpected change.
One of the most beautiful genuine geniuses that explains it easy for us to understand! Thank you so much always for your information and updates! You make something’s that can be boring very easy to want to watch!
SHE is always the goto source for credible AI & New Tech developments!
I cannot wait until Optical CPUs will be available over the counter working near the speed of light ❤ Quantum Computing is available for the enthusiastic people out there,
I remember the days of £500 CD writer reader & the 40MB Hard Drive thinking thats fantastic, now there so cheap, more & more solid state drives i purchased 200GB SSD for £6 in the UK 🇬🇧
Have a fantastic weekend ❤
Anastasia's videos are so great. I love listening to them.
The first computer I had as a teenager was a based on the MOS 6502. It's insane to realize how far these chip architectures and technologies have come since the mid 1980s.
A lot of useful devices were built around Chuck Peddle's 6502 "RISC" architecture... now it probably couldn't even serve as a supervisor/side processor for today's multi-core beasts. 😁
Forget about 4MHz becoming 4GHz... As the complexity of the chip architecture and software also grows exponentially... what is it that is looming to take over the burden of design and planning? Our brainchild, which goes by many names, but all call it AI.
Once let loose into this world, and our sattelites and probes... it will live forever and ever. Ever Growing... Ever Learning...
- Fascinating!
- Thx for keeping us updated.
- Oh and, great presentation. Keep up the great effort/content...
Turned on notifications! Can't miss these vids. Thank you for teaching us and keeping us updated! 🤖🌹
Some amazing stuff coming down the line. I'm looking forward to Graphcore's Good computer next year.
Can you explain the differences between x86, arm, risc-V and the GPU architecture that I don't know what is called :D
Thx for the great videos. It's an andidote to the pessimism that the consumer side industry has
Thank you for another excellent video which in a nutshell allows me to stay updated about the GPU/ CPU « universe », the upcoming technical hurdles and challenges, as well as where the future may be going.
Your videos are the best! And you look and sound great.
Another terrific video Anastasi! Well done.👍
These hardware companies don't actually need a parity with CUDA. They only need to implement custom backends for TensorFlow and PyTorch libraries. This is much easier than replacing CUDA. I even did similar stuff myself, on top of Direct3D 11 compute shaders.
When do you think companies like Meta and Google can use Intel's or AMD's GPU's for their TF or PT?
@@maxjames00077 I think they already can, it just involves more friction to setup and use.
I believe AMD is better than Intel for that. According to PyTorch devs, support for ROCm became stable in version 1.12, June 2022. According to AMD, both PyTorch and TensorFlow work on AMD GPUs on top of ROCm. Both require a Linux OS for that, Windows is unsupported. However, Internet companies like Meta and Google don’t care about Windows support because they run their AI stuff on servers, Linux is better for that use case.
excellent video. very well researched. lets hope Cerebras (that actually just closed a deal with AMD for a supercomputer to help with ML) and AMD are able to pull ahead and actually compete. the fact nvidia has such a big piece of the pie with proprietary software is a real problem imo... or else no AI for everyone cause its a company focused only on profits and extremely greedy at that.
i pray Rocm has a chance to shine but amd needs to actually focus on it massively.
AI is in its infancy so theres a chance for others to join the fight.
Amd would have done same if they pulled up before Nvidia did.
You guys shouldn't behave as if AMD is a saint and will not overprice
🎯 Key Takeaways for quick navigation:
00:00 📈 The demand for GPUs, particularly for AI applications, is surging, causing supply struggles.
03:02 🚀 Nvidia dominates the GPU market due to its high-performance hardware and proprietary software, like CUDA.
05:03 💹 The new Nvidia H100 GPU offers significantly better performance for large language model training compared to CPUs.
09:51 💼 Alternatives to Nvidia GPUs for AI applications include Tesla's Dojo, Google's TPUs, Cerebras' Wafer Scale Engine, and Intel/AMD GPUs.
16:34 💡 The future of AI hardware is promising with intense competition, and diversifying investments is recommended, including alternative investments like Fine Art through Masterworks.
Made with HARPA AI
Super work Anastasi 👍
CUDA is way more than a parallel computation framework, it's a complete series of libraries for solving problems on the GPU with C/C++ including linear algebra, deep learning, sparse matrix, solver, FFT, nvJPEG, profilers, etc.
There are higher level abstractions like pytorch and tensorflow keras
@@hdthor Yeah that add layers of performance killing kruft on top of CUDA.
Love your videos, but justed wanted to say: I love your watch, the Santos is one of my all time favourites
Best Market advisors are those who knows the most about the company and the technology.
Hey, I saw my name as a supporter! Thanks, Anastasi! 😍
Very informative, and exciting.
Me encantaria un video sobre el impacto que puede tener el descubrimiento de LK99 en el universo de los microchips
You are extremely educational. Thank you.🎉
Hi Anastasiia your English is fantastic you are explaining reasons for the shortfall and high prices so clearly and concisely keep up the great work gb
Alright will watch this after work
Thanks again for another great informative vlog. Until next time thank you and keep up the great work
Great video Anastasi.
Crazy how much the other chip makes slept.
Cartier watch?? looks pretty. Will be interesting to see the role GPU chips and quantum chips may play soon...
You like? Cartier tanks are pretty popular among buyers. That one should be a 36mm Cartier tank they're expensive tho.
you have a great ability to explain . would be a great teacher
Great work Anastasi video always very interesting I believe that NVIDIA will be increasingly the protagonist in the realization of supercomputers given the billion-dollar investments I foresee exciting solutions in a short time !!!”
Another insightful Pro-Video - Thanks !
'seems a dedicated AI to help streamline and optimize the software layers of MUX/DeMUX (pls excuse the hardware terminology) for all these parallel processors will be another 'Cutting Edge' tech
Awesome growth ahead and happening now.
:) strix point apu is the only way for me.
This from Model 3 manual, May also apple to model Y ( I would think): To experience the same amount of deceleration whenever you release the accelerator pedal, regardless of the state of the Battery, you can choose to have the regular braking system automatically engage whenever regenerative braking is limited. Touch Controls > Pedals & Steering > Apply Brakes When Regenerative Braking is Limited.
When even the auto-generated Google subtitles can't quite understand you, you know that you should provide a transcript of your speech in subtitle form.
Ive been a big proponent of Open Source since desktops first came into the market.
As a gamer I'm hoping there would be some GPU dribble to us. For example failed chips getting new life as gamer GPUs.
4090 is a failed 6000 Ada chip.
Between AI replacing the financial interest of anyone capable of producing gpus and the economies of scale for mobile and dedicated devices, I think those are slowly replacing the pc gamer market, and with it, the gamer gpu market. Wont be long before your playing your old pc games on an emulator running on your phone. Just give it a few years for mobile to catch up while investment in the desktop drops off.
Almost all my gaming is on a home made vr treadmill. Theres already one company, Virtuix, releasing a dedicated all in one device for this purpose. Even Kat-VR is working on a device, the kat nexus, to use their slidemill with quest2 and psvr2 so theirs won't need a pc either. Dont know if thats diversification or they also see the market headed in that direction.
Of course, some new development could turn everything upside down or I could be just wrong about this. I'm into video games, but mostly for exercise, and only a small part of that. I suspect us casual gamers are really the bigger market.
Wasn't it Clive Sinclair and associates back in the 1990s who came up with the idea of using a single wafer ( at that time likely a 200 mm wafer) as a single large RAM device, similar in general idea to the Cerebras full wafer TPU?
_What?_ Covering my hobbies, now, too? Gee, thanks!
More seriously, you made a good case for Cerebras. Taken together, these trends may spell deep trouble for Intel in particular.
Intel is the only one who isn't in deep trouble in the long run when they will be manufacturing the chips with TSMC. AMD on the other hand...
GPUs are currency now!
Well right now with NVDA people are assuming massive sales in the future even if the market isn't there for it and the market can't supply what NVDA needs. Good luck with that.
plz bring videos on quantum computing
thanks for the video Anastasi !
The flip side of this is that AMD is making serious market shares in the private sector which may have massive implications later in this trend.
Thank you for this video and all of previously you did. As i know, if we could make a digital chip implement Non Volatile elements such as Memristor or MTJ device, it will be actually useful 👌
Thank you!
enjoyable and informative as always thank you
How do you explain TSMC earnings report and guidance of 10% reduction in sales and Ai only accounting for 5% of those sales?
i came back from work awsome thanks for info!
First Comment And First Like.......Love your videos.. I'll watch your Videos and become more intelligent..
Thank you.
The watts to do the same ISOwork are the key factor. Cost of the rack much less so. Heat and software difficulty a distant third and fourth place. H100 GPU on TSMC N4 and better tensor cores, chiplet memory, CoWoS is impressive but... When a RISC-V solution on DUV with AMD's parallel franework comes out, itll be far cheaper i exoect.
VM limited on single chip versus the Dubai Wafer?
Misspelling out what you are saying on screen really helps our comprehension.
Thanks!
Thank you!
I feel like Cerebras can benefit if these bottleneck by selling the GPU's that NVIDIA can't supply yet.
Nicely explain , I was born 1965 now i am 57 years old ten years old than bill gate. when i was at University i am as lecture asst System Information , and Electronic Data Processing
Has Apple some in-house super-computer for ai training? The M series arm processors show good performance for that.
Спасибо за информацию, очень интересно.
very exciting! The real world kinda sucks but the computer world is so promising! cheaper, better, faster... exciting opportunities in tech, beautifully, and organically presented! I don't think she can be replaced with a bot!
Why is the celebrus chip square and not round? Seems to he loosing some of the silicon area.
Talk about AI designing itself, how much of that happens, chipwise?
FYI Grace Hopper: en.wikipedia.org/wiki/Grace_Hopper
Grace was a consultant to DEC (I'm a former employee)
I feel that Cerebras direction sounds promising and really hope and wish that AMD catches up with nVidia and ROCm succeeds because nVidia is a monopoly as of now and it hurts of all us
Could you pls explain how much extra energy ( e.g. MG=Megawatt's) is aprox. in future needed to fuel AI ? Elon Musk made a warning about a lack of power plants in US (equal to europe)
What do think about Tiny Corp and their future Tiny Box ?
Will AMD stocks fly too, what do you think?
Since NV can not deliver to the demand, the next best will catch up.
I just dont see both of them as very stable and predictable.
Even more at NV where news been spread way to slow and therefore too late to counter react in case.
Anastasi, the hissing of your S sounds are killing me. Do you use a de-esser for your mic. I don't know, it might just be me.
Thanks for the information.
Hey, supremely tech lady maybe in the future they use some sort of messenger RNA coupled with special types of fluid dynamics to impart AI with some sort of human thought! Maybe it takes a decade or 2 or...!
Towelie of South Park has a TPU
Can you make a video on the new room temperature superconductor paper?
Excellent presentation across the board. I would love to see Cerebras compared to GH200. That would be more apples-to-apples. And the NVLink vs Cerebras in build memory - should give a good comparison.
hi from Belgium, just one question, isn't it precisely at this time that China is imposing restrictions and sanctions on rare earth exports to the USA;;;; very useful for making all the technology you said on your video?? have a nice day ( or night)
Which sounded fine to me
Wish someone would just make AGI and solve all our problems and answer all our questions.
why are you not talking about other AI brands AMD for 1???
Yes, yes it costs 40k. But what is it's production cost?
It's annoying cuz it means the next-gen gaming chips will be super-expensive
there is already big difference between gaming and AI GPUs. They just keeps the term the same (i.e. GPU) for marketing purposes. however, at some point of time it won't make sense anymore
16:25 Overall, proprietary code can be relevant to AI development, including LLM and autonomous AI. However, it is crucial to strike a balance between proprietary and open-source components in the AI ecosystem to promote innovation, collaboration, and transparency while protecting intellectual property rights.
Why do the GPUs have to be so expensive?
Is this supposed to be an inclusive pricing model?
How do you justify the price vs the cost to make them?
Tensores y transformadores. Los chips se sofistican especializándose en lo que funciona. La función hace al órgano o viceversa.
I'm hoping intel arc GPUs brings in some real competition. Other than that some sort of low power setup for those that don't mind waiting a day or two.
i see many happy Arc a770 owners already!
You should do a video about multimodal models and see if/when we'll have proper and accurate 3D STEP files imagined from AI inputs. STEP is the only real engineering format for 3D IMO. It means you can machine the parts and build assemblies to 100% proper engineering specifications. When we get there, we'll see a step change in everything.
Room temp superconductor patent was released Ana. Maybe we could have mini super computers oneday. Could you explain how superconductors would help computing?
99.99% of the human race has little to no idea what they're in for over the next ten years.
The things we think are so important right now are going to seem trivial in ten short years.
Sit back and enjoy the ride, you have no other choice.
I need updates on running Cuda on rocm
The Goddess of Technology
Software development's the hard part.
Insightful and clear. Thanks.
I thought it was the NPU's that's doing the machine learning.