AI Just Solved a 53-Year-Old Problem! | AlphaTensor, Explained

Поделиться
HTML-код
  • Опубликовано: 23 дек 2024

Комментарии • 491

  • @tane_ma
    @tane_ma 2 года назад +330

    I really admire the amount of work put into this video: the research, editing and everything else. Kudos.

    • @underfitted
      @underfitted  2 года назад +13

      Thanks, Rafael! Yeah, it’s a lot of work, but we’ll worth it if the videos help more people.

    • @mistycloud4455
      @mistycloud4455 2 года назад +1

      A.G.I Will be man's last invention

    • @metaparcel
      @metaparcel 10 месяцев назад

      I always think Kudos sounds like what they call Cheetos in Greece for some reason.

  • @absence9443
    @absence9443 2 года назад +135

    Without having considered the counts themselves, I would have expected it to be a larger channel given the way the video is structured and how precise some edits are. But that aside, it's interesting how we went from finding neural patterns to solve tasks to finding neural patterns that optimize finding neural patterns to solve tasks. Essentially analogous to a factory creating better factories. I'm aware that in public topics such as AI are usually only mentioned in a shallow manner, referencing typical generic developments of poor face or speech neural networks, even providing merely mediocre samples of them additionally, whereas there are so many more intricate creations from the past 10 yrs. While the media is often seen as immediate, it depicts new types of technology as concepts of the future or reduces their depth, even though they already are defining the present. Maybe too metaphorical, I just think very few people are aware of the state of models and have an insight on what is to come soon.

    • @underfitted
      @underfitted  2 года назад +10

      Thanks for the comment! Really appreciate it! Yeah, we are making progress at a breakneck speed. I can’t even imagine where we will be by the end of 2023.

  • @harrytsang1501
    @harrytsang1501 2 года назад +45

    It used to be a manual process. When transistor count increased, one of the first thing computer engineers did was to put 10x more logic gates into the multiplication part of the ALU by using many predicates like filling in a truth table.
    Having a known software method to optimize them further is nice, but be reminded that hardware made for the task is already super optimized with memory bandwidth and latency being the limit

    • @gljames24
      @gljames24 2 года назад

      Even with 3d stacked cache and large matrices?

    • @n4rzul
      @n4rzul 2 года назад +2

      I thought the point of this was that it could optimize on any hardware. Those optimizations can then be built into the next iteration of the hardware. I don't see this as a purely software optimization. Wait till the AI is capable of implementing those changes in hardware by itself...

    • @abulkhaiyrtalapov3038
      @abulkhaiyrtalapov3038 2 года назад

      When using truth tables as part of the digital design process, we get ridiculous logic gates as we add bits to the data buses we operate on. As far as I know, we do increase the speed of the custom circuitry, but differently, by adding more gates to the designs, but still I am not sure about perfect, being an engineer is realizing there is never a roof you can reach, I think.

    • @asdfasdf71865
      @asdfasdf71865 2 года назад +1

      it is not all about performance. if you optimize at the wrong place you get vulnerabilities which could ruin your whole business

  • @Axacqk
    @Axacqk 2 года назад +12

    The real breakthrough here is AI being able to improve the design of a crucial building block of its own implementation.

  • @paulallen1597
    @paulallen1597 2 года назад +90

    Always high quality, high entropy content. DeepMind's application of Reinforcement Learning (RL) to solve problems is fascinating, and I agree that it will be interesting to see how they use RL moving forward: what other problems will be "gamified" in the future ...

    • @underfitted
      @underfitted  2 года назад +8

      Right?! I'm tempted to say "the next decade will be interesting" but who am I kidding: 2023 will be really interesting!
      Thanks for the comment! Appreciate your words!

    • @JorgetePanete
      @JorgetePanete 2 года назад +1

      ​@@underfittedmaybe an AI learns how to create AIs given a set of task-network

    • @marilynjensen476
      @marilynjensen476 2 года назад

      A.I. could (likely already is) gamify the "human farming game," what neither Marx or Rothschild, nor any other man has totally figured out to perfection.
      The last 75+ years of Technocrats and Cyberneticians have indeed made tremendous documented progress in innovating improvements to the human farm; yet, just imagine what machine learning from and for us (and our digital twins) could bring forth from the human stock.
      Then maybe, by the time we get elderly the brain to silicon tech will be ready for perfect virtual eternity with our tremendous A.I. pals. No belief in God needed or expected.

    • @locowachipanga561
      @locowachipanga561 2 года назад

      Your compliment was sufficient. Let's see Paul Allen's comment.

    • @Corpsecreate
      @Corpsecreate 2 года назад +1

      I want DeepMind to build an AI that takes as input a training set and labelled dataset, and as output gives you the best hyperparams to learn that data with the lowest loss and best generalisation error.

  • @jackrdye
    @jackrdye 2 года назад +6

    Jesus ai advancements in the last year have been insane. This is phenomenal

    • @underfitted
      @underfitted  2 года назад +1

      Yeah. When I look back at the list of things they happened this year alone, it’s crazy!

  • @ozzyphantom
    @ozzyphantom 2 года назад +48

    This is the video I’ve been looking for describing AlphaTensor. Concise and to the point about why this is important. Subbed.

  • @andrewmalcolm79
    @andrewmalcolm79 2 года назад +9

    Can AlphaTensor solve x265 decode on Playstation 3? There are no good compilers, even IBM had separate compilers for the PPE and the SPEs in their own CellBE CPU and nVidia kept the architecture details of the G70 confidential. Better yet, can AlphaTensor produce an LLVM backend for Playstation 3? Or optimise LLVM backends more generally?

  • @daniellim1212
    @daniellim1212 2 года назад +4

    For non square matrix multiplications:
    (axb)(bxc), the number of multiplications will be = a*b*c

  • @michaelnurse9089
    @michaelnurse9089 2 года назад +4

    Its one of those things that are great in theory but mildly useful in practice. Something like 40% speedup was achieved by the paper authors.

  • @jaykaku7646
    @jaykaku7646 2 года назад +13

    I loved your hardware interpretation of the results because today IBM and basically everyone is trying to make better chips for matrix mul tasks from ASICs to Accelerators, and a new algorithm eliminates a huge compute cost 💯💯. Thanks a ton for the video, I too am working with matrices this moment.

  • @SinanAkkoyun
    @SinanAkkoyun 2 года назад +27

    I am stunned, now subbed and shocked by how juicy and fun to watch && also insightful you made this video, you brought us through the whole journey in minutes and explained every obstacle.
    On top of that, your visuals and audio is fantastic, extremely high quality content you produce.
    Please more :)

    • @underfitted
      @underfitted  2 года назад +1

      Glad you enjoyed it! Thanks for taking the time and writing such a thoughtful comment. More to come!

    • @SinanAkkoyun
      @SinanAkkoyun 2 года назад

      @@underfitted Of course, this is the least I can give back! Hope to see more awesome ML content and also the awesome color grading ;)

  • @DeathSugar
    @DeathSugar 2 года назад

    So can we expect generalized version of optimal matrix multiplication?

    • @DF-ss5ep
      @DF-ss5ep 2 года назад +1

      Yes. The generalized version will eventually be: make the AI optimize it on the fly

  • @mmuschalik
    @mmuschalik 2 года назад

    Thank you. I heard about the article when it came out, but didn't realise the implications at a practical level. The machine improving its own hardware and software to get better. That's the path to singularity

  • @richardwargo5201
    @richardwargo5201 2 года назад

    Did it take into account finite precision and truncation issues?
    Now, how about matrix inversion problem?

  • @mmenendezg
    @mmenendezg 2 года назад +7

    Your channel and your tweets are really helpful in my ML learning process.
    The quality of your videos is really high and the content is incredible useful to better understand ML.
    Vamos Santiago!

    • @underfitted
      @underfitted  2 года назад

      Marlon, thanks so much for your comments! Really appreciate them!

    • @hiankun
      @hiankun 2 года назад

      Totally agreed! And I always wondering why the traffic in YT seems much less than Twitter. These videos deserve more views!

  • @karelvanderwalt3625
    @karelvanderwalt3625 2 года назад

    Is stability of the algorithm not neglected for the big O ?

  • @williamseipp9691
    @williamseipp9691 2 года назад +1

    I think I understand the implication? We're talking about an algorithms ability to adapt, right?
    We can see that in some neural networks that novel behavior appears after millions of millions of repetitions, which I guess is similar to adaptation. But a AI's ability to use different techniques on different problems as a fundamental feature versus a result of extensive training might be a big deal.
    I really enjoyed the video. You kept me interested and whisked me along your train of thought with very little effort on my part. Perfect story telling, great explanation of all the necessary topics and you made it engaging. Awesome work.

  • @didles123
    @didles123 2 года назад +1

    I wonder if these multiplication optimizations are useable in GPGPU. Since GPGPU can calculate each element independently in parallel, keeping the operations independent is important. I don't know as much about these multiplication algorithms, but maybe they can be parallelized in their own way.

    • @GRAYgauss
      @GRAYgauss 2 года назад +2

      We already use GPUs (or TPUs) for all this. Alphatensor is computed with billions of f16 mat4 calcs on your gpu, and its results could be used to optimize the next iteration of hardware pipelines, of course?...(or perhaps, less usefully, emulated on the same device.) Why couldn't we bootstrap math and theory, its how these devices make improvements in the first place? Gpgpu, yeah. I don't really get what you're trying to say tbh.

  • @JasonAndrewsUK
    @JasonAndrewsUK 2 года назад +12

    Dude! Fantastic video. Full marks on making an exciting, clear, attention-grabbing and informative video! Honestly, the production was faultless.

    • @underfitted
      @underfitted  2 года назад +1

      Thanks, Jason! Much appreciated! Still learning a lot, but I'm glad the videos are coming out better.

  • @johnpayne7873
    @johnpayne7873 2 года назад

    Provocative presentation - instant subscriber.
    Things you said that caught my attention:
    1. Concisely stated the problem, i.e. computation time for matrix multiplication
    2. Nice historical summary of approaches to solve the stated problem
    3. Clear understanding of "small gains" may yield large rewards. Examples of applying this "algorithm optimization method": solving complex problems found in general relativity and fluid mechanics (non-linear differential tensor and vector systems) or standard quantum mechanics (or string/membrane theory)
    4. Your enthusiastic imagination; unafraid to jump ahead to what might be next
    Looking forward to what else grabs your attention

    • @underfitted
      @underfitted  2 года назад

      Thanks! Appreciate such a thoughtful comment!

  • @vinhkhangtrankhuong7203
    @vinhkhangtrankhuong7203 2 года назад

    In quantum computing, when we introduce noise to the system, the calculation of the density matrix for the state of the quantum system grow exponential. Maybe this will help a lot in that field.

  • @jairoc.peralta
    @jairoc.peralta 2 года назад

    and, can you share the algorithm? or should i keep multiplying matrices the naive way?

    • @underfitted
      @underfitted  2 года назад

      The paper shows the results for 4x4 by 4x4 and 4x5 by 5x5

  • @juanmadg1
    @juanmadg1 2 года назад

    Did the paper specify which method was found by Alphatensor for each matrix dimension? Did they make that information public or just the minimum number of multiplications found?

    • @underfitted
      @underfitted  2 года назад +2

      Their paper includes the specific steps found by AlphaTensor to multiply 2 4x4 matrices and 4x5 by 5x5. www.nature.com/articles/s41586-022-05172-4

  • @MiccaPhone
    @MiccaPhone 2 года назад +2

    0:27 : you cannot "solve this equation" because it IS NO EQUATION! lol 😂🤣😂🤣

    • @underfitted
      @underfitted  2 года назад +1

      You are right. Thanks for taking the time!

  • @pedrorebelo6671
    @pedrorebelo6671 2 года назад +8

    What a great video. You really know how to capture your viewers attention.

    • @underfitted
      @underfitted  2 года назад

      I appreciate that, Pedro! Thank you!

  • @andresgoens
    @andresgoens 2 года назад

    Great video, thanks for making it! I have a small nitpick, which you do consistently: x^2 - y^2 is not an equation, where's the equals sign? You can call it a polynomial, or say you want to compute that expression, but there is no equation there to solve.

    • @underfitted
      @underfitted  2 года назад +1

      💯 I should have said expression.

  • @ominollo
    @ominollo 2 года назад

    Where is the link to the paper?

  • @MrHaggyy
    @MrHaggyy 2 года назад

    There is also some neat work out there to make multiplication and things like sin, cos, tan much faster. Autonomous driving is a lot of 8x8 matrix multiplication and we have algorithm/logic to do this very specific thing. But it's a really hard engineering challange and tradeoff between big-complex-slower chips that offer matrix multiply and smaller-simpler-faster chips that only offer really basic math but stupidly fast. Which one is more efficient really comes down to the job of the hardware and the algorithm at hand.

  • @RuairiODonnellFOTO
    @RuairiODonnellFOTO 2 года назад +3

    It's limited by the fact that each algorithm it finds differs depending on the matric dimensions?
    So I don't see a wide application of this for math, unless we stop each algorithm for each matric shape.
    Interesting, yes but not ground breaking

    • @underfitted
      @underfitted  2 года назад +4

      For machine learning is very important. Finding a different algorithm depending on the dimensions of the matrices is more a feature than a bug. The goal here is speed, not generalizability.
      The other aspect of it that I really like is the application of AlphaZero on a different domain than games. It opens many possibilities for us!

  • @waldtrautwald8499
    @waldtrautwald8499 2 года назад +1

    4:17 This is not quite true. There are algorithms with a better time complexity: O(n^2.4) vs. Strassen's O(n^2.8). The problem is that these theoretically faster algorithm only outperform Strassen for truly gigantic matrices, which is why they aren't used in practice.

    • @underfitted
      @underfitted  2 года назад

      What time stamp are you referring too? 4:17 is not the one, I think.

    • @waldtrautwald8499
      @waldtrautwald8499 2 года назад +1

      @@underfitted I'm referring to your quote "53 years later and we still don't know if there is a better way to multiply matrices." at roughly that moment in the video. I guess it depends on your definition of "better" but asymptotically faster matrix multiplication algorithms than Strassen have been discovered.

  • @OlatundeAdegbola
    @OlatundeAdegbola 2 года назад +1

    Quickly saw this is an awesome channel and quickly subscribed.
    But I'm shocked it only has 12k subscribers.
    You deserve a million.
    Thanks for sharing.

    • @underfitted
      @underfitted  2 года назад +1

      I’m just starting. Thanks for the support!

  • @fludeo1307
    @fludeo1307 2 года назад

    i have read that AI hardware is going to move to analog signals, because you can do multiplication just passing a current through a resistance. The other problem of matrix multiplication is the accumulated error provoked by floating point representation in binary (digital) system.

  • @juwonkim1782
    @juwonkim1782 2 года назад +1

    I just subscribed. Keep on your fascinating works.

  • @famnyblom6321
    @famnyblom6321 2 года назад

    Haven't read the paper but I am curious how well the new algorithms are numerically. Some algorithms that are fast are also not practical to use due to their poor numerical stability.

    • @kazedcat
      @kazedcat 2 года назад

      Machine learning does not care about stability. They are even using 8bit int instead of 32bit or 16bit float to do the inference calculation to reduce the hardware needed. Going from 16bit training to 8bit inference have very poor effect on numerical stability.

  • @mathepunk
    @mathepunk 3 месяца назад

    An essential part of an equation is an equal sign which is missing in all of the items referred here as "equation".

  • @cbunix23
    @cbunix23 2 года назад

    To what extent can Alpha Tensor help with very large matrices? On very large matrices I got a 40% speed up simply using cache coherent operation ordering.

  • @MikeBTek
    @MikeBTek 2 года назад

    Gee, back in the mid 1970's (high school), before we all had home computers, I gave myself a project of reducing the number of math operations required to solve large multiplication problems. Came up with some reductions, but did not have much use for it at the time. Now I'm trying to remember just what I did. Question: In general, does AI come up with solutions because it tries all the possibilities, eliminating those processes that apparently don't work, while optimizing the procedures that show some promise? For example if the problem was to come up with sets of 3 positive biquadrates that sum to another biquadrate, would AI just test combinations of numbers or would it actually develop a parametric equation if possible?

    • @underfitted
      @underfitted  2 года назад

      Very similar process to what you described, yes.

  • @parthsavyasachi9348
    @parthsavyasachi9348 2 года назад +4

    And no, this hasn't made any significant dent on the problem of multiplication. Even in their own hand picked selected few problem the gain is what like 6 to 8 percent. Thats not ground breaking. One gets that speed up or down just in the cache access.

    • @underfitted
      @underfitted  2 года назад

      We’ll have to disagree on that one. 😊

    • @parthsavyasachi9348
      @parthsavyasachi9348 2 года назад +1

      @@underfitted so how much was the speed up?

    • @underfitted
      @underfitted  2 года назад

      You should read the paper.

    • @parthsavyasachi9348
      @parthsavyasachi9348 2 года назад

      @@underfitted i read it thatd why i said. The max improvement that is reported is like 18 percent.
      Thats not ground breaking. Not even close.
      I have a patent with my company where i take search in 1 dimension array and have it searched in o(3)to o(4) time regardless of data size.(once the data structure has to be constructed). The same thing usually is done with binary search that is size dependent.
      This is called improvement. Not 10 to 20 percent cost improvement.
      I would not even count this as anything if it doesn't really give good advantage.

    • @underfitted
      @underfitted  2 года назад

      Language is such a funny thing.

  • @davidzhang4825
    @davidzhang4825 2 года назад +5

    I like how the video title rhymes with "Attention is All You Need" paper on transformer

    • @underfitted
      @underfitted  2 года назад +3

      Yeah, I did that on purpose :)

  • @40NoNameFound-100-years-ago
    @40NoNameFound-100-years-ago 2 года назад +1

    I read that paper, a while ago....and Inwas pretty amazed with what is written in it.

  • @ronaldlau9363
    @ronaldlau9363 2 года назад

    Aside from the (mis)use of "equation" which you must have got for the millionth time by now, can you check your expression for M6 in the case of the 2 by 2 matrix? I think it should be a negative sign in the first pair of brackets in M6 just like M7, otherwise you don't get to calculate the ab_22 element in the resulting matrix.
    Also, showing how to combine M1 to M7 to get the 4 elements might be good (as much as one who is interested can simply look it up and verify from Wikipedia).

    • @underfitted
      @underfitted  2 года назад

      Yup, 2,000,000 comments mentioning that I said "equation" instead of "expression." :)

  • @johnpayne7873
    @johnpayne7873 2 года назад

    Another thought ...
    Accuracy is improved with fewer calculation steps as well as speed; in some instances this gain might be even more important

  • @JustATempest
    @JustATempest 2 года назад +1

    You deserve to be on the level of 3blue2brown or computerphile. You're amazing!

  • @priyamgupta5524
    @priyamgupta5524 2 года назад

    I wonder how much improvement we can get if we combine alpha tensor with analog computer chip like one created by Mystic AI.

  • @AleX_028
    @AleX_028 2 года назад +8

    Lemme just say...
    Sir you are amazing🤩
    The hardwork and information with atmost detail with creative editing makes this video very very engaging and connecting.
    Hats 👑⛑👒🎩 off to your effort.
    Subscribed right away. 😀😄

    • @underfitted
      @underfitted  2 года назад

      Thank you so much 😀! Really appreciate your comment!

  • @3bdo3id
    @3bdo3id 2 года назад

    Waited such video but did google published the results they got?

    • @underfitted
      @underfitted  2 года назад +1

      They publish their paper on Nature.

  • @fg-zm2yu
    @fg-zm2yu 2 года назад

    The conclusion seems very similar to what Stephen Wolfram says in a New Kind of science: start searching in the whole set of simple programs for the one that fits the behavior you are looking for. Boring task -> give it to a program.

  • @Stl71
    @Stl71 2 года назад +4

    This is terrifying in some way. A machine that is so 'clever', that can write code itself.

    • @underfitted
      @underfitted  2 года назад

      It'll happen.

    • @DiamondSane
      @DiamondSane 2 года назад +2

      The cell is so clever it can evolve to an embryo. I'd call it amazing.

    • @MrHaggyy
      @MrHaggyy 2 года назад

      We actually have this with generic algorithm and some strategies of system identification in engineering already.
      It's just finding the right question and initial data as well as proving your results is often times harder than solving the problem in the first place. But this approach really helps with discovering math and physics as well as applying it.

  • @willyouwright
    @willyouwright 2 года назад

    Lookup tables are much faster but require memory. Things can be done to hybridise calculations. I.e process easy calls and lookup hard ones.. I.e. you don't need to process 2x3 AND 3x2 they are the same. . You also need to only calculate what is necessary and not everything at high resolution.. the trick is know when to round off and when to be more accurate..

  • @fuzzy-02
    @fuzzy-02 2 года назад

    His hands are shaking from excitment.
    Truly a maths genuis.

  • @ajayparwani2403
    @ajayparwani2403 2 года назад

    I read somewhere after finding these new algorithms Alpha tensor applied these to itself and got faster!!
    I believe this is the next step in the world of AI where systems are not only able to figure out better algorithm but infact use them to become faster overtime

  • @DrWrapperband
    @DrWrapperband 2 года назад +1

    Sounds like the A.I. computer operating system that learns all operations from scratch, i.e. reading disk, loading memory, running programs, is almost here.

  • @FresGamerMan
    @FresGamerMan 2 года назад

    Would the system be able to optimize its own algorithms? Would it be able to find better algorithms this way to improve itself recursively?

    • @underfitted
      @underfitted  2 года назад

      Not in its current version, no

  • @datascienceproject
    @datascienceproject 2 года назад

    Just subscribed and started to binge-watch all the videos. Channel is brand new, I hope it grows fast :)

    • @underfitted
      @underfitted  2 года назад

      Thanks for the support! Yeah, channel is 114 days old 😋

  • @maboesanman
    @maboesanman 2 года назад

    Maybe what’s next is an alphatensor powered assembly optimizer. Seems like it would be the same problem but way bigger. Given that people treat the optimizers like black boxes anyway it would be pretty incredible to have a significantly smarter optimizer just show up. Maybe it has different versions for different instruction sets and cache sizes. Maybe it has a pre build model for every consumer cpu. Squeezing every drop from your hardware with AI optimization is a pretty cool idea.

  • @אביבשמואל-מ2צ
    @אביבשמואל-מ2צ 2 года назад

    You can multiple matrix by recursions.its the same operation time but you can try loan results from smaller dimensions matrix's.

    • @underfitted
      @underfitted  2 года назад

      There are many ways to skin a cat.

  • @Supramonk
    @Supramonk 2 года назад

    Do you think P vs NP can be solved?

    • @underfitted
      @underfitted  2 года назад

      I don’t think P = NP. The other question is whether we can prove that. I’m not sure about that one.

  • @NicholasRenotte
    @NicholasRenotte 2 года назад

    This was brilliant, props to you!

  • @abdelrahmanaymen382
    @abdelrahmanaymen382 2 года назад

    Dude! Fantastic video. I enjoyed watching it 😮 You are awesome 🌟

  • @locowachipanga561
    @locowachipanga561 2 года назад

    That is a scary concept. AIs finding better ways to code AIs. Great video.

  • @khaledhamza1805
    @khaledhamza1805 2 года назад +2

    Always brilliant Santiago and you choose your topics verl well
    your content is very helpful and the quality of your videos is astonishing

  • @arbitrandomuser
    @arbitrandomuser 2 года назад +1

    its not an equation , its an expression , equation is when there is an "="

  • @fixfaxerify
    @fixfaxerify 2 года назад +1

    0:30 A simple "equation" indeed, it even lacks an equality! 😄

    • @underfitted
      @underfitted  2 года назад +1

      Yup. I should have said "expression."

  • @lgmuk
    @lgmuk 2 года назад +2

    Very good video! And I had already read the paper and blog post!

    • @underfitted
      @underfitted  2 года назад +1

      Thanks a lot for the comment! Really appreciate it!

  • @persiancarpet5234
    @persiancarpet5234 2 года назад +1

    Great video, I got the chills thinking about how crazy this actually is

  • @VanillaSunadae
    @VanillaSunadae 2 года назад

    So a bunch of matrices finds a new way to mutiply a bunch of matrices. Very Cool.

  • @ameermirawdeli4148
    @ameermirawdeli4148 2 года назад +4

    Thank you for this informative video

  • @jamesjones2212
    @jamesjones2212 2 года назад +1

    What comes next is "Physics Diffusion" where we can input all the elements and have AI create exotic material that exhibit negative space. Then boom flying cars 😀

  • @thoward0
    @thoward0 2 года назад

    Literally the machines are training us to make them smarter.

  • @crosswalker45
    @crosswalker45 2 года назад

    U deserve more subs... Your research and editing is fabulous.

  • @mrknarf4438
    @mrknarf4438 2 года назад

    Wonderful video: short enough, clearly explained, enthusiastic.
    You've earned a subscriber!

  • @wasifnabi4553
    @wasifnabi4553 2 года назад +1

    Appreciate you for compiling such informational content. 😊

  • @fluidmodels3796
    @fluidmodels3796 2 года назад

    This means new possibilities. I think it Will also optimize memory management. Because a lot of time can be spent only on waiting for the data. Thanks

  • @einzelsorten5783
    @einzelsorten5783 2 года назад

    Thanks. Great info! Absolutely dislike the continuous dramatization, though. It's like, "OMG, OMG, the zombie is catching up" adrenaline rush and grabbing the webcam every few seconds. Completely unnecessary. Sorry if this hurts your feelings. Not the intention. Really want to improve your channel. I subscribed!

    • @underfitted
      @underfitted  2 года назад

      No feelings hurt. I'm learning, and improving with every video. Everyone's feedback is valuable because it's my only way to make progress. Thanks for reaching out!

  • @muzaffarnissar1978
    @muzaffarnissar1978 2 года назад

    First time came to ur channel, salute to you sir, the efforts, research etc etc has made made this video awesome.
    Looking for your guidance.
    Thanks and regards

  • @EngineerNick
    @EngineerNick 2 года назад

    Next it also designs and optimises the hardware. Bet there is a way to drop a heap of transistors from every operation too

    • @underfitted
      @underfitted  2 года назад

      I don't doubt our ability to do anything, to be honest.

  • @sinfinite7516
    @sinfinite7516 2 года назад

    Amazing video, great editing, interesting topic. You got a new subscriber!

    • @underfitted
      @underfitted  2 года назад +1

      Awesome, thank you! Really appreciate it!

  • @Jo_Wick
    @Jo_Wick 2 года назад

    This man actually replies to (almost) *every* comment. Now that's dedication.

    • @underfitted
      @underfitted  2 года назад

      I try to. As long as RUclips shows them to me 😁

  • @vinitvsankhe
    @vinitvsankhe 2 года назад +2

    I am confused. Is an AI (whose DNA is matrices and their multiplication) now multiplying matrices as a problem to make them multiply faster? Does that mean it makes itself faster with time?
    Shouldn't that fall into an inception loop?
    By the way did you see the video of "analogue chips" by veritasium? It intends to speed up multiplication through voltage, current and resistance relationship sacrificing unnecessary precision that digital chips bring along (and sometimes not necessary to the final outcome in AI's decision making)

  • @nonothik
    @nonothik 2 года назад

    2:22 Indeed, instead of going 2 times 4, we can do 1+1+1... 8 times, that's one less multiplication needed 😂

  • @pedromoya9127
    @pedromoya9127 2 года назад +2

    thank you, great video. Perhaps the next step is something more and more exciting, like a 'living' tool that creates and replace itself in more powerful and general tools each iteration, like living RL and Goodfellow's GANs derivatives, or like our brains... persistent focus is succesfull in nature, and birth-death cycles even more!

  • @siddheshb.kukade4685
    @siddheshb.kukade4685 2 года назад +1

    Great Video great Graphics and Information Please continue making such videos

    • @underfitted
      @underfitted  2 года назад

      Thanks! Really appreciate the comment!

  • @NoNTr1v1aL
    @NoNTr1v1aL 2 года назад +1

    Absolutely amazing video!

  • @mohegyux4072
    @mohegyux4072 2 года назад

    instant sub, love the content, just hoped the energy is 10% lower

    • @underfitted
      @underfitted  2 года назад

      Man… I wish I had a regulator 😋 thanks for the sub and the comment! I’ll keep improving!

  • @bayzed
    @bayzed 2 года назад

    Wow, this was really interesting and the way you delivered it made it even more fascinating! You've got a new subscriber :)

  • @tamasgal_com
    @tamasgal_com 2 года назад

    0:34 this is not an "equation"

  • @maxprofane
    @maxprofane 2 года назад

    Very interesting topic and very good presentation. Kudos

  • @naveens6114
    @naveens6114 2 года назад

    Hi, The Content and Editing is in Top Notch. After ML Concepts, Continue with Deep Learning & Like this Breakdown too.

    • @underfitted
      @underfitted  2 года назад +1

      Thanks a ton! Really appreciate your comment!

  • @ErikS-
    @ErikS- Год назад

    Pretty good editing skills.

  • @koendos3
    @koendos3 2 года назад +1

    I fucking love this video, a cool mix between math and computer science! Well done

  • @joshpoe9339
    @joshpoe9339 2 года назад

    The tick sounds effects during the work switches became very distracting.

    • @underfitted
      @underfitted  2 года назад

      You are right. I got rid of it on my last few videos. Thanks for the feedback!

  • @zemanntill
    @zemanntill 2 года назад +1

    Imo the ending (“What comes next?“) nails it. There is some juicy potential for papers if you figure out how to turn other hard fundamental CS problems into games and apply Alphazero.

  • @paulsalele3844
    @paulsalele3844 2 года назад

    Awesome content!!

  • @pnachtwey
    @pnachtwey 2 года назад

    But multiples and adds take only one clock cycle on many CPUs now.

  • @charlinhos0824
    @charlinhos0824 2 года назад

    What a wonderful an engaging way to explain the complexity in simple terms, thank you for all the hours and hearth you put in your content.

  • @zemanntill
    @zemanntill 2 года назад +2

    Hey, i think the editing style doesn’t fit for this kind of content (too fast/ dizzy) - i know the other comments liked it and it borrows from large youtubers but i feel like a bit less would be better (even if retention might drop a little).

    • @underfitted
      @underfitted  2 года назад +2

      Really appreciate your comment! Thanks for letting me know!
      I've been trying to find a good balance, but overall, I'm trying to make the type of content that I like to watch. For example, although RUclips is full of "academic" videos, I find them boring and don't learn too much from them.
      But I hear you! Everyone has a different style, and the secret is to find a good balance! I'll keep trying.

    • @josephthehansen
      @josephthehansen 2 года назад

      @@underfitted I completely agree with you. Your content reminded me a little bit of Jonny Harris's videos - very engaging, very creative and keeps the story alive.
      I'm very interested in learning more about science topics and AI, but everyone I come across on RUclips is so academic and monotone about it that it becomes impossible to watch at times.
      Two Minute Papers for instance is great, and I adore Károly Zsolnai-Fehér, but his monotonous voice does sometimes become a bit hard to listen to, and he also loses audience when he starts talking about technical data.
      Your content was a breath of fresh air, and you've earned yourself a sub! I look forward to seeing more content from you in the future :)
      My one recommendation for now would be to either soften or remove the volume of the clicks whenever the word changes on the screen - I found it to be slightly detracting from what you were saying.
      But otherwise it is really good stuff!

    • @underfitted
      @underfitted  2 года назад +1

      Joseph, this "... but everyone I come across on RUclips is so academic and monotone about it that it becomes impossible to watch at times" describes exactly the reason I started making these videos.
      There are fantastic creators on RUclips talking about AI/DS/ML, but I need something different.
      A few people have commented here because they find this style "weird" and "not serious" for ML content, and I agree :) But that's a feature, not a bug.
      I want to build something different.

  • @DiegoDelagos
    @DiegoDelagos 2 года назад +1

    First of all, x²-y² is not an equation.

    • @underfitted
      @underfitted  2 года назад

      You are right. What’s second of all?

  • @---zg7ex
    @---zg7ex 2 года назад

    so clearly explained, please do more on science, math related topics

  • @PritishMishra
    @PritishMishra 2 года назад +9

    The Editing is on fire 🔥
    I also upload videos on ML on my channel and you are a true motivation for me to get up and make a new one. Your consistency inspires me.

    • @underfitted
      @underfitted  2 года назад +1

      Thanks, Pritish! I'm so happy you liked the editing. It took quite a bit of work, but I'm happy how it turned out.
      Keep going at it!