Gradient descent, how neural networks learn | DL2

Поделиться
HTML-код
  • Опубликовано: 17 дек 2024

Комментарии • 3,2 тыс.

  • @3blue1brown
    @3blue1brown  7 лет назад +2882

    Part 3 will be on backpropagation. I had originally planned to include it here, but the more I wanted to dig into a proper walk-through for what it's really doing, the more deserving it became of its own video. Stay tuned!

    • @HeyItsSahilSoni
      @HeyItsSahilSoni 7 лет назад +11

      Can you provide some link to the training set?
      I'm quite new and I'm trying to learn this "Hello World" of NN,

    • @pinguin1009
      @pinguin1009 7 лет назад +4

      Did you consider to do a part about phase functioned neural networks? Would be awesome!

    • @nyroysa
      @nyroysa 7 лет назад +2

      As the part progresses,
      we're getting closer to seeing that "lena.jpg" picture

    • @akhileshgangwar394
      @akhileshgangwar394 7 лет назад +10

      You are doing very good job , lots of hard work behind this video , i salute your hard work thanks

    • @mynameisZhenyaArt_
      @mynameisZhenyaArt_ 7 лет назад +9

      So have you decided to do more of these videos? There is a line with CNNs and LSTMs in video series...

  • @Shrooblord
    @Shrooblord 7 лет назад +2367

    I'm only 12 minutes into this video right now, but I just wanted to say how much I appreciate the time and spacing you give to explaining a concept. You add pauses, you repeat things with slightly different wording, and you give examples and zoom in and out, linking to relevant thought processes that might help trigger an "a-ha" moment in the viewer. Many of these "hooks" actually make me understand concepts I've had trouble grasping in Maths, all because of your videos and the way you choose to explain things. So thanks! You're helping me a lot to become a smarter person. :)

  • @snookerbg
    @snookerbg 7 лет назад +2452

    One of youtube's highest quality content channels! Chapeau

  • @colonelmustard7078
    @colonelmustard7078 Год назад +226

    Not only the videos themselves are great on this channel but the lists of the supporting materials are amazing too! Drives me down a breathtaking rabbit hole every time! Thank you!

  • @DubstepCherry
    @DubstepCherry 4 года назад +1865

    I'm an IT student, and we have an Assignment on exactly this topic. We even have to use the MNIST data set. I have to say, this is absolutely lifesaving and I can not thank you enough Grant. What you do here is something that only a handful of people on this planet can do, explain and visualize rather complicated topics beautifully and simple. So from me and A LOT of students all around the globe, thank you so so much

    • @vsiegel
      @vsiegel 3 года назад +44

      Yes, it is just extremely good, in an objective way. He is brilliant at it, and spends a lot of time on each video. If there is an explanation of something by 3blue1brown, you will not find anything explaining it nearly as good.

    • @johndough510
      @johndough510 2 года назад +6

      @@vsiegel bro you guys are so much smarter than i am im jealous

    • @vsiegel
      @vsiegel 2 года назад +22

      @@johndough510 If you are thinking about how smart you are, you are probably smarter than you think. No worries.

    • @johndough510
      @johndough510 2 года назад +7

      @@vsiegel thanks for being so cool about it man, hope you have a good one

    • @alfredoaguilar2076
      @alfredoaguilar2076 2 года назад

      @@johndough510 I'm kill

  • @plekkchand
    @plekkchand 4 года назад +736

    Unlike most teachers of subjects like this, this gentleman seems to be genuinely concerned that his audience understands him, and he makes a concerted and highly successful effort to convey the ideas in a cogent, digestible and stimulating form.

    • @hamidbluri3135
      @hamidbluri3135 2 года назад +5

      TOTALLY agreed

    • @redflipper992
      @redflipper992 Год назад +1

      concerted with whom? I don't think you understand how to use that word.

    • @werwinn
      @werwinn Год назад

      he is a true proffesor!

    • @macchiato_1881
      @macchiato_1881 Год назад

      @@redflipper992 I don't think you understand what concerted means. Stop trying to act smart and think you're better than everyone here. Be humble. You are irrelevant in the big picture.

    • @JeffCaplan313
      @JeffCaplan313 Год назад +2

      ​@@redflipper992I read that as "concerned"

  • @hangilkim245
    @hangilkim245 5 лет назад +132

    "But we can do better! Growth mindset!" at 5:18 .... a wholesome intellectual i love to see it

  • @melkerper
    @melkerper 5 лет назад +9431

    Dissapointed you did not animate a 13000-dimensional graph. Would make things easier

    • @havewissmart9602
      @havewissmart9602 5 лет назад +181

      No.... No it would not....

    • @mrwalter1049
      @mrwalter1049 5 лет назад +492

      A 2-dimensional projection of a 13000-dimensional graph would probably look like a pile of garbage.

    • @cchulinn
      @cchulinn 5 лет назад +389

      If 3Blue1Brown cannot animate a 13000-dimensional graph, then noone can.

    • @tehbonehead
      @tehbonehead 5 лет назад +142

      @@mrwalter1049 You're not thinking fourth-dimensionally!!

    • @mrwalter1049
      @mrwalter1049 5 лет назад +33

      @@tehbonehead No-one is. We're trapped in three dimensions. That's why you could never imagine what a 4-dimensional cube looks like.
      Making a 4-dimensional projection of a 13000-dimensional object isn't significantly better than 3 dimensions.
      If you meant to be humorous I hope someone gets a chuckle, because I didn't. Then your effort won't be in vain.
      Have a nice day 🙂

  • @jasonzhang6534
    @jasonzhang6534 10 месяцев назад +16

    my professor has explained this in 3 lectures for about 6-7 hours. 3B1B explained it in 30 mins and it is much more clearer. I can now visualize and understand the what/why/how behind the basic deep learning algorithms. Really appreciate it!!!

  • @bikkikumarsha
    @bikkikumarsha 6 лет назад +129

    You are changing the world, shaping humanity. I wish you and your team, happy and peaceful life. This is a noble profession, god bless you guys.

  • @seC00kiel0rd
    @seC00kiel0rd 7 лет назад +3576

    My math career is over. Once I learned about gradient descent, it was all downhill from there.

    • @rlf4160
      @rlf4160 7 лет назад +302

      I had a similar fate, except mine went negatively uphill.

    • @yepyep266
      @yepyep266 6 лет назад +215

      just remember there are people in an even lower minima than you are.

    • @jomen112
      @jomen112 6 лет назад +128

      Yea, but making random choices makes you eventually reach the bottom.

    • @thetinfoiltricorn7797
      @thetinfoiltricorn7797 6 лет назад +36

      It's all planar vectors from here.

    • @nateschultz8973
      @nateschultz8973 5 лет назад +42

      You just need to take a few steps back and turn your life around.

  • @frenchycman
    @frenchycman Год назад +1

    The visualizations are so great. Appreciate the thoughtfulness of these videos

  • @kraneclaims
    @kraneclaims 4 года назад +259

    I just sat through a 3 day ML accelerator class and you series did a far better job at explaining them with 4 twenty minute videos. Well done mate. Really appreciate it. Thank you

    • @NoobJang
      @NoobJang Год назад +10

      this youtuber is the best in maths and engineering in general
      i have never been so astounded for how easy learning machine learning can be, without having to take in bunch of complex topics that doesnt add to the discussion.
      Like most of the courses try to make you understand various different complex topics and by the time you finished it, you will ahve forgotten mostly about machine learning. Why dont you just explain the catch for each concept then allow us learn it in depth afterwards
      like these channels only explaining the concepts with both ease of learning and depths are the best.

    • @b2782
      @b2782 14 дней назад

      could you pin which videos are you referring to?

  • @hikaruyoroi
    @hikaruyoroi 7 лет назад +37

    I love you so much. I'm taking multivariate calculus and I'm doing some neural network work right now, and none of my teachers have the passion nor the capability to teach as well as you. You help me keep my passion for learning alive

    • @ssa3101
      @ssa3101 6 лет назад

      Incompetence galore.

  • @dsmogor
    @dsmogor 7 месяцев назад +13

    I think what puts this material apart from the competition is the authors intuition of the focal points where the audience might loose the plot. Then he takes a patient and systematic turn to reiterate what have been learned so far to reinforce the basics to decrease the cognitive leap needed to grasp the next step. This ability is in my experience pretty unique.

  • @Skydmig
    @Skydmig 7 лет назад +60

    That end comment with Lisha Li really points out how important it is to put a lot of effort into gathering and creating good and structured data sets. I know it's cliché to state "garbage in, garbage out", but these findings put very precise context and weight to this particular issue.

    • @atlas7425
      @atlas7425 7 лет назад +3

      Haha, "weight".....get it?

    • @theespatier4456
      @theespatier4456 6 лет назад +2

      StiffWood True. This also becomes ethically important in medical applications of AI, where poor input can create racist AI and the like.

  • @souvikroy7570
    @souvikroy7570 4 года назад +83

    Hands down, I have never seen anyone explain mathematics so beautifully the way he does. Kudos!

  • @welcome2bangkok-d1x
    @welcome2bangkok-d1x 2 года назад +64

    i have no words to describe how thankful i am. thank you so much for such great content.

  • @pseudo_goose
    @pseudo_goose 4 года назад +26

    Some thoughts on the results:
    1. 14:01 The weights for the first layer _seem_ to be meaningless patterns when viewed individually, but combined, they do encode some kind of sophisticated pattern detection. That particular pattern detection isn't uniquely specified or constrained by this particular set of weights on the first layer; rather, there are infinitely many ways that the pattern detection scheme can be encoded in the weights of this single layer.
    These infinite other solutions can be thought of as the set of matrices that are row-equivalent to the 16x700ish matrix where each row is the set of weights for each of the neurons on this layer. You can rewrite each of the rows as a linear combination of the set of current rows, while possibly still preserving the behavior of the whole NN by performing a related transformation to the weights of the next layer. In this way, you can rewrite the patterns of the first layer to try and find an arrangement that tells you something about the reasoning. Row reduction in particular might produce interesting results!
    2. 15:10 I think I understand the reason why your NN produces a confident result - it's because it was never trained to understand what a number _doesn't_ look like. All of the training data, from what I can tell, is numbers associated with 100% confident outputs. You'd want to train it on illegible handwriting, noise, whatever you expect to feed it later, with a result vector that can be interpreted as 0% confidence, by having small equal weights, having all weights to zero, or maybe an additional neuron that the NN uses to report "no number".

    • @watchm4ker
      @watchm4ker 2 года назад +2

      2 is a painfully easy mistake to make, because it requires the human assembling the programming data to think outside the box.

    • @josboersema1352
      @josboersema1352 Год назад

      Quite hard to read your comment, but it seems that we have the same idea: the neural network _is_ detecting smaller elements like "edges and loops" (as video author puts it), assuming those pictures 14:01 are of the actual results. The next layer then starts combining these elements, and it seems that if you stare at it long enough you can almost start guessing what it might be doing, like combining row 1 column 2 = strong + row 3 column 1 = strong + row 3 column 4 = strong + row 2 column 4 = weak + row 1 column 4 = weak, you might be going toward an 8 on those counts, and depending on some other combination of pattern strengths it might be a 6 or 9 if there is absence of signal upper/right or lower/left given by some of those patterns which are matched with the input. This is almost certainly not accurate as an example, but it seems to be the theme of how it works. 16:05 _"... picking up on edges and ... not at all what the network is doing."_ This statement in the video seems wrong.
      P.S. If this first part above is true, then the neural network might be capable of drawing a 5 (15:39). You just have to extract that answer in the way that it is in there, which is a bit more involved that following it's normal operation for which it is built. If you look into what combination of patterns from the first layer output, in what strengths, leads to a number (5 for example), than you could superimpose those patterns unto each other, and that would be what this neural network thinks is that number. It shouldn't be too hard to write a function to the already trained network, to draw this out.

  • @JockyJazz
    @JockyJazz 3 года назад +349

    3:38 you missed the chance of using the meme *"AI: I've found an output, but at what cost?"*

  • @darshita1270
    @darshita1270 3 года назад +21

    Math courses in my college are basically trash compared to your videos , finally now I understand how math is being applied in computer science . Thank you so much for teaching in such an illustrative way .

  • @Nyhilo
    @Nyhilo 7 лет назад +414

    After watching your first video, I ended up drawing a "mock" neural network up on paper that would work on a 3x3 grid (after all what else are you supposed to do during a boring lecture class?). It was supposed to recognize boxes, x's, sevens, simple shapes, and I defined the 7 or so neurons that I thought it might need by hand. I did all the weighted sums and sigmoid functions on paper with calculator in hand. It took maybe an hour and a half to get everything straight but once I did, it worked. It guessed with fairly good accuracy that the little seven I "inputted" was a little seven. All that excitement because of your video.
    Later that evening and the next one, I tried to program the same function taking PNGs as inputs and definitions of the neurons and it honestly was only a little more rewarding. But now that I see what the hidden neurons *actually* look like, I only want to learn so much more. I expected the patterns to be messy, but I was really surprised to see that it really does almost look like just noise.
    Thank you for making these videos. I find myself suddenly motivated to go back to calculus class tomorrow and continue our less on gradients. There's just so much out there to learn and it's educators like you that are making it easier for curious individuals like me to get there.

    • @3blue1brown
      @3blue1brown  7 лет назад +155

      That's so cool, thanks for sharing! I didn't expect anyone to actually go an play with it by hand, but simplifying down to a 3x3 grid seems really smart. Stay curious!

    • @Dom-nn1kg
      @Dom-nn1kg 7 лет назад

      Nyhilo +

    • @Dom-nn1kg
      @Dom-nn1kg 7 лет назад

      3Blue1Brown +

    • @jayeshsawant6734
      @jayeshsawant6734 6 лет назад +5

      Were you able to do all of that by watching this video series alone? Please can you add other resources you referred? Thanks!

    • @michaelhesterberg702
      @michaelhesterberg702 6 лет назад

      .....go anD play... AND AN+D

  • @andrasiani
    @andrasiani 7 лет назад +27

    how can anyone dislike these videos? very detailed, accurate explanations and cool animations. Keep up the good work!!

  • @imad_uddin
    @imad_uddin 3 года назад +17

    Cant believe you explained this so easily. I thought it would take me ages to wrap my head around what neural networks basically are. This is truly amazing explanation!

  • @obsidianblade4228
    @obsidianblade4228 7 лет назад +1547

    Did anybody else feel bad for the network after he called the output utter trash?😢

    • @sarahmchugh4169
      @sarahmchugh4169 5 лет назад +143

      I know, especially with those sad computer eyes. Tragic

    • @MarioRodriguez-or9fn
      @MarioRodriguez-or9fn 4 года назад +64

      Yes, specially when he called it bad computer :(

    • @DavidLee8981
      @DavidLee8981 4 года назад +7

      we are all utter trash for future robots

    • @theshermantanker7043
      @theshermantanker7043 4 года назад +50

      Bruh there's literally Reinforcement Learning where the Network is tortured by the researchers when it gets a wrong answer and the torture continues until it gets the right answer lol

    • @David5005ful
      @David5005ful 4 года назад +1

      Lmao.

  • @cody._.--._.--.
    @cody._.--._.--. 5 лет назад +274

    I wish someone would have introduced this to me at a young age back in the 90s. I had no idea neural network have existed for so long

    • @lopezb
      @lopezb 4 года назад +44

      Now it's easier to explain. He couldn't have made a video like this back then, both because RUclips didn't exist, and all the relevant stuff would be in technical papers...

    • @Djorgal
      @Djorgal 4 года назад +25

      @@lopezb Also, it was a really niche field that didn't show that much promise.

    • @damienivan8946
      @damienivan8946 4 года назад +7

      Also, from my understanding, modern neural networks are very different from the one in the 90s

    • @bubblelyte401
      @bubblelyte401 4 года назад

      It's a college graduate course.

    • @georgalem3310
      @georgalem3310 4 года назад

      In the 90s NN fell into disfavor.

  • @bishalthapaliya4069
    @bishalthapaliya4069 4 года назад +15

    Probably, even a 5 year old would master deep learning when taught in this way. What a video man ! Awesomeeeeeeeee

  • @SPYTHandle
    @SPYTHandle 5 лет назад +127

    How confident I feel in my current knowledge of neural networks:
    15:41 - *"Uh...I'm really more of a multiple choice kinda guy."*

    • @natchu96
      @natchu96 3 года назад +2

      The neural networks themselves generally feel the same, so at least we won't be alone in that sentiment.
      Assuming thinking rocks and metal count as good companionship at any rate.

  • @GAment_11
    @GAment_11 6 лет назад +8

    When I watch your videos, all I want to do is keep going. Thanks for motivating me, as well as others, with your amazing content. I really appreciate it.

  • @Shubhi021
    @Shubhi021 Год назад +11

    This video is truly a work of art. The animations are mesmerizing. Thank you for all your work, Grant!

  • @tonraqkorr230
    @tonraqkorr230 6 лет назад +750

    We need AI to recognise what the doctors write

    • @frankchen4229
      @frankchen4229 3 года назад +55

      whoever designs the algorithm and engineers the software deserves a nobel peace prize

    • @flyinglack
      @flyinglack 3 года назад +1

      @@frankchen4229 LOL

    • @johnbarbuto5387
      @johnbarbuto5387 3 года назад +4

      Who writes any more??? That horse left the barn a long time ago. Besides, we are no longer doctors. Courtesy of insurance companies we are "providers". (The same strategy of devalued identities has long been used by invading armies to anonymize those being conquered, an apropos metaphor.)

    • @MrWite1
      @MrWite1 3 года назад +6

      @@johnbarbuto5387 why so mad

    • @centerfield6339
      @centerfield6339 2 года назад

      @@johnbarbuto5387 not courtesy of insurance companies; courtesy of the fact that healthcare needs to be paid for. State systems are also payer systems.

  • @musthavechannel5262
    @musthavechannel5262 6 лет назад +487

    "I'm more of a multiple choice guy" LOL

  • @Heisenberg355
    @Heisenberg355 2 года назад +1

    This man is a living legend. I really sincerely believe he's one of the best "explainers" for many complex mathematical topics. I found your channel because of linear algebra, and now I'm relieved whenever I search for a topic and see one of your videos. You truly are the master of your league

  • @superj1e2z6
    @superj1e2z6 7 лет назад +777

    Watching 3b1b
    Step 3b. Drop Everything
    Step 1b. Watch religiously.

    • @jonasvanderschaaf
      @jonasvanderschaaf 7 лет назад +23

      oh the accuracy of this comment

    • @spiderforrest7816
      @spiderforrest7816 7 лет назад +6

      My god I relate

    • @Petch85
      @Petch85 7 лет назад +31

      for me it is.
      step 3b: make sure you are ready. you need to be 100% focused.
      step 1b: Watch it critically, be sure not to strengthen your miss believes. If it seems simpel and obvious I am probably misunderstanding it.

    • @fossilfighters101
      @fossilfighters101 7 лет назад +1

      +

    • @Cosine_Wave
      @Cosine_Wave 7 лет назад +9

      counting level: Parker

  • @sweepy84
    @sweepy84 5 лет назад +328

    You sir deserve a noble, or an oscar...what an incredibly effective method of teaching. thank you so very much!!!
    NO! BAD COMPUTER! made me crack up! lol

    • @shakhaoathossain5032
      @shakhaoathossain5032 3 года назад +5

      A balon di or too

    • @sohambhattacharjee951
      @sohambhattacharjee951 2 года назад +1

      @@shakhaoathossain5032 XD good one.

    • @Phoenix-nh9kt
      @Phoenix-nh9kt 2 года назад +3

      @@shakhaoathossain5032 add a grammy in there too hahaha

    • @ChemEDan
      @ChemEDan 2 года назад

      The fact it was recorded digitally meant he said that to a real computer.😭
      AND SO DID YOU!!! 😠

  • @user-rw6iw8jg2t
    @user-rw6iw8jg2t Месяц назад

    Perfect Gradient Descent explaiantion Best one indeed ! 12:30 Training the Neural Nets with 13,000 weights, MNIST data set Image classification case study.

  • @Iextrimator
    @Iextrimator 7 лет назад +4

    Absolutely love your videos! I'm trying to show this video to my friends who doesn't know English so well, and I decided to make subtitles. Hope you approve them, I really want to spread word about your work.

  • @skintaker1949
    @skintaker1949 4 года назад +1946

    So uhhhh, did you just say that this was the "Hello World!" of neural networking.....

    • @GabrielCarvv
      @GabrielCarvv 4 года назад +71

      ​@Winston Mcgee "p r e t t y m u c h i t"

    • @rudigerbrightheart7304
      @rudigerbrightheart7304 4 года назад +36

      Well, the data are the hello world, because it is the first image set that people take to test or learn about an algorithm.

    • @junkailiao
      @junkailiao 4 года назад +18

      Because you don't need all those knowledge to build a network that can read digits. It's easy with Keras even my grandmother can do it

    • @domizianostingi9504
      @domizianostingi9504 4 года назад +15

      Yes it is, because the dataset is veerry clean and CNN through Keras is very easy to implement, though you need to have huge background in math and code-writing (I'm a statistician so I have a little bit of both) :)

    • @namlehai2737
      @namlehai2737 4 года назад +12

      Use a package. People already did the hard stuff, you just have to call their function / use their models

  • @matteo7861
    @matteo7861 Месяц назад +1

    I’m a grad student in physics and i wanted to thank you. It is insane to find such good videos on such advanced subjects !

  • @olesyabondar4826
    @olesyabondar4826 5 лет назад +6

    The graphics of this video is absolutely stunning! Thank you for your work ♡

  • @Tri_3st
    @Tri_3st 7 лет назад +4

    Hi 3B1B, as a technical physics student, beeing interested into this topic for quite a while now, and also enjoying your content for quit a while, i really wanna thank you for not only going into this topic particulary, but also for educating a relatively large audience with your informative videos and improving the interest into mathematical sciences for a lot of people including me, which is pretty important in my opinion! Keep it up!!

  • @-beee-
    @-beee- Год назад +1

    Keep coming back to this series and sharing it with so many people. This whole channel is truly a gift. Thank you so much for making these!

  • @blunderbus2695
    @blunderbus2695 6 лет назад +793

    "It's actually just calculus."
    "Even worse!"
    i'm dead

    • @fitokay
      @fitokay 5 лет назад +5

      Actually, AI just lie to people of the world

    • @fitokay
      @fitokay 5 лет назад +1

      so far

    • @the.abhiram.r
      @the.abhiram.r 3 года назад +2

      calculus is the easiest form of math

    • @ahmedezat1353
      @ahmedezat1353 3 года назад +1

      @@the.abhiram.r
      I wish you are joking

  • @genoir-itsmusicart9169
    @genoir-itsmusicart9169 6 лет назад +16

    This is mindblowingly interesting and extremely well explained. Thank you!

  • @VincentKun
    @VincentKun 2 года назад +2

    I saw this video when i know nothing and i had a lots of intuitions, I'm rewatching after studied a lot more and I'm still learning a lot. You're a great teacher

  • @jabug_1144
    @jabug_1144 5 лет назад +7

    Once I graduate and start working, I’m gonna send you the money I owe you for watching all these videos. I’m doing BSEE for control systems so hopefully it works out.

  • @jamesluc007
    @jamesluc007 7 лет назад +9

    You explained in less than 4 minutes something that took me several days to understand from other sources. You are awesome!

  • @super266
    @super266 2 года назад +2

    You are the best science teacher I have every seen. If anyone upstairs is serious about our education system they should use your videos as baseline for how to teach properly; you never use a term that wasn't clearly defined prior, you use analogies perfectly, and you tie new technical info back to the original concept thereby making sense how the new info fits together in the larger picture. If my high-school and college teachers were like you I would have done infinitely better at school.

  • @amagicpotato5511
    @amagicpotato5511 7 лет назад +60

    Hi 3b1b i love your vids and they are one of the reasons why I know so much of how the universe works. Your channel inspires me to know more and you show the beauty of all of it. Please dont ever stop making these videos, you are making so many lives greater.

    • @3blue1brown
      @3blue1brown  7 лет назад +18

      Thanks so much Amagic, I'll do my best.

    • @charliedexter3202
      @charliedexter3202 7 лет назад

      Hello....I would like to learn how to make these animations....I teach math, particularly statistics at the graduate level, and I find the way you make numbers illustrate the idea actually helps understand the flow of parameters in question in a much better manner . Do give me certain leads so that I can pick this up. I shouldn't have problems programming once I know which platform to work with

    • @neerajtiwari5365
      @neerajtiwari5365 7 лет назад

      @Amagic potato, For the world of me, I cannot possibly figure out how watching 3blue1brown's videos helped you attain that enlightenment about How The Universe Works...

    • @flumsenumse
      @flumsenumse 7 лет назад +5

      +charlie dexter
      He has all of the code for his animations (written in python) here: github.com/3b1b/manim

  • @cgmiguel
    @cgmiguel 5 лет назад +5

    Your videos with such wonderful LaTeX animations are just as high level as a BBC awarded documentaries. Very impressive to say the least.

    • @soliduscode
      @soliduscode 2 года назад

      I agree. I need to learn more about this LaTex animations

  • @DavidMauas-j6t
    @DavidMauas-j6t Месяц назад

    you know something? your channel is the best math teaching I have ever laid eyes upon.
    it is brilliant. beyond amazing.
    everything is meticulously choreographed to perfection.
    I WISH I could have learned math from someone like you when I was younger, and I am so happy I get to have the occasional brush with your videos. they are sublime.

  • @n00urddine
    @n00urddine 7 лет назад +27

    Thanks for the part 2, I was waiting for it impatiently.

  • @siddheshmisale3904
    @siddheshmisale3904 4 года назад +20

    Would just take a moment here to appreciate the sheer brilliance of Grant on this series. I would not have reached a decent level of NN w/o these explanations and so would so many other people. Single best series on NN / Math out there in general.

  • @mashab9129
    @mashab9129 5 месяцев назад

    after looking though many udemy, orelilly and other youtube videos finally found this one - beginner friendly but on a profound enough level, explained in a comprehensible way, that does not lose you in the middle because it jumps from abc to hard concept - this channel is a gem. thank you!

  • @jacoblund8289
    @jacoblund8289 7 лет назад +6

    I love how he has people like Desmos and Markus Persson supporting him on patreon

  • @HaouasLeDocteur
    @HaouasLeDocteur 7 лет назад +52

    WOO BEEN WAITING FOR THIS

  • @imtanuki4106
    @imtanuki4106 11 месяцев назад

    Possibly one of the best mini-courses on ML anywhere. Clearly explained concepts, beautiful post-production. kudos

  • @shwetamayekar1863
    @shwetamayekar1863 5 лет назад +12

    Love the eye/ pi animations! :) Gets me smiling amidst all the complexities of Neural Networks 😲

  • @shaylempert9994
    @shaylempert9994 6 лет назад +21

    Pause and ponder?! Every 10 seconds I stop for a minute of thinking! And on all of your videos!
    This time I had a time I thought for like half an hour.

  • @navidutube
    @navidutube 6 месяцев назад +1

    This is simply the best channel on RUclips

  • @ThinkTwiceLtu
    @ThinkTwiceLtu 7 лет назад +82

    great explanation, thank you:)

    • @UltraRik
      @UltraRik 7 лет назад

      did you honestly understand any of this did this video honestly help you comprehend something

    • @chibrax54
      @chibrax54 7 лет назад +1

      +Patrik Banek it did help me ! But I was already familiar with these concepts. If you don't understand, watch the video again and look for different sources of explanation it will help you :)

    • @UltraRik
      @UltraRik 7 лет назад

      Okay thanks for the advice

    • @chibrax54
      @chibrax54 7 лет назад

      +Patrik Banek You're welcome :) If you specifically don't get how the gradient can help reduce the error, you should learn what is the point of a derivative in a simple variable function and dig into multivariable calculus and optimization !

  • @Treegrower
    @Treegrower 7 лет назад +148

    @ 3:39
    Wow... I didn't realize 3B1B likes to bully neural networks. That was ruthless.

    • @Brian.001
      @Brian.001 5 лет назад

      Yes, it's a jungle in there.

    • @oskarjung6738
      @oskarjung6738 5 лет назад

      @- RedBlazerFlame -
      ' Oversimplified' reference

    • @theshermantanker7043
      @theshermantanker7043 4 года назад

      There's a training method called Reinforcement Learning where you literally torture the Network when it gets the wrong output lol

  • @GhostyOcean
    @GhostyOcean 4 года назад

    I can't describe the moment of pure joy I had when I noticed you put the label of "tease" on Fermat at 18:05

  • @danielamurphy8560
    @danielamurphy8560 Год назад +6

    I'm doing my Masters in applied Econ right now and we briefly went over Neural Networks in my advanced econometrics class. Some of the terminology was a bit different and I felt like I could understand it decently in office hours with my professor, but this was still a great resource to solidify my understanding of the concept. (Also we looked at the MNIST dataset in class too) :D

  • @geregeorge1589
    @geregeorge1589 7 лет назад +262

    At the 16 minute mark, I got sucker punched. After having gone through this and the previous video on machine learning and just loving how an art student like myself is enjoying math such as this and feeling like I'm making some progress..... You tell me that this is all stuff that was figured out in the 80s and I'm like...... Oh Come On! Lol!

    • @apuapustaja2047
      @apuapustaja2047 6 лет назад +89

      Honestly, the 80s is actually very recent compared to other stuff. In math undergrad I was learning concepts from the 1800s lmao

    • @SimberLayek
      @SimberLayek 5 лет назад +17

      @@apuapustaja2047 yup! Math is older than all of us... it's our discoveries that are "new"~

    • @DiegoGonzalez-vn3qx
      @DiegoGonzalez-vn3qx 5 лет назад +45

      Honestly, don’t feel discouraged. General Relativity was formulated almost a century ago, but that doesn’t mean it is easier to understand.

    • @SimberLayek
      @SimberLayek 5 лет назад +2

      @Dark Aether some definitely could say that~

    • @DiegoGonzalez-vn3qx
      @DiegoGonzalez-vn3qx 5 лет назад +18

      @Dark Aether What do you even mean by that? Right now, we are living in a moment in which scientific knowledge is being acquired at the fastest rate we have ever seen. The number of active scientists right now, as you might expect, is the largest in history. Now, if you are talking about "raw" intelligence... well, I'm pretty sure evolving into creatures with a noticeable higher intelligence is going to take a long, long, long time.

  • @somag6810
    @somag6810 11 месяцев назад

    "Our growth mindset is reflected when we think always if we can do better!" You are always awesome. Thanks for all the informative videos that imparts a lot of fundamental knowledge to people like me.

  • @minerawesome28
    @minerawesome28 7 лет назад +6

    I was looking forward to this video all week.

  • @theflaminglionhotlionfox2140
    @theflaminglionhotlionfox2140 3 года назад +123

    Me in part 1: Ah I think I'm starting to understand this whole thing.
    Me in part 2: Nevermind...

    • @saicharansigiri2964
      @saicharansigiri2964 3 года назад +4

      excatly

    • @osmanyasar9602
      @osmanyasar9602 3 года назад +8

      Once you learn more math it will be meaningful. I guess if you dont understand this video then something is missing in your calculus and/or linear algebra

    • @FivosTheophylactou
      @FivosTheophylactou 5 месяцев назад

      Rewatch it 3 times. I did

  • @mtrifiro
    @mtrifiro 9 месяцев назад

    I am so happy I discovered this today. I ignored all (well, most) of the math, and I still came away with a pretty solid understanding of how it works. Your explanations are ridiculously clear; you have a gift.

  • @YaLTeRz
    @YaLTeRz 7 лет назад +72

    Pretty sure at 11:03 the weights should either start at w1 or end at w13,001.

    • @3blue1brown
      @3blue1brown  7 лет назад +50

      Gah! Good catch.

    • @thesecondislander
      @thesecondislander 7 лет назад +19

      This just goes to show that off-by-one errors really do happen to the best ;)

    • @nikoerforderlich7108
      @nikoerforderlich7108 7 лет назад +4

      +thesecondislander
      Well, next to cache invalidation and naming things it's one of the two big problems in computer science :P

    • @columbus8myhw
      @columbus8myhw 7 лет назад +13

      It gained weight.

    • @williamwilliams1000
      @williamwilliams1000 7 лет назад

      So my little pony, whats the appeal?

  • @hakimr7986
    @hakimr7986 5 лет назад +66

    15:19 seems interessting, just like you have to train your own (biological) NN to draw a human face, although you saw millions of them

    • @TrendyContent
      @TrendyContent 5 лет назад +8

      Hakim R thats a very good analogy

    • @sohampatil6539
      @sohampatil6539 4 года назад

      This might be relevant: look up general adversarial networks

  • @johannes523
    @johannes523 2 месяца назад +1

    This is, like, literally the most important video on the internet.

  • @ajnelson1431
    @ajnelson1431 7 лет назад +204

    3:40 "NO! Bad computer!"

    • @stydras3380
      @stydras3380 7 лет назад

      AJ Nelson Bad boy!

    • @NF30
      @NF30 7 лет назад +41

      I felt so sorry for the computer...

    • @johnchessant3012
      @johnchessant3012 7 лет назад +17

      "To say that more mathematically..."

    • @Shockszzbyyous
      @Shockszzbyyous 7 лет назад +1

      i heard eric cartman say it.

    • @Dom-nn1kg
      @Dom-nn1kg 7 лет назад

      +

  • @Maffoo
    @Maffoo 7 лет назад +4

    This series is fantastic and just the right level of being complex but understandable. Thanks!

  • @gersonribeirogoulart9895
    @gersonribeirogoulart9895 Год назад

    When something is amazing, it will look like
    with your work. Even your bg voice is totally understandable, legit and direct

  • @sashimanu
    @sashimanu 4 года назад +21

    10:10 biological neurons are continuous-valued as well: their firing frequency varies.

    • @MOHANKUMARAPGPBatch
      @MOHANKUMARAPGPBatch 4 года назад +2

      still, the frequency cannot be decimal right ? so its still discrete input where calculus cannot be applied.....

    • @nullbeyondo
      @nullbeyondo 3 года назад +1

      @@MOHANKUMARAPGPBatch No. Calclus can always be applied and your idea of a frequency is horrible since it can easily be represented by many other methods like time or transforming it. And anyway, that's not how a biological machine works. The "decimals" in math serve no real purpose in reality cause everything in our universe is quantumized.

    • @MOHANKUMARAPGPBatch
      @MOHANKUMARAPGPBatch 3 года назад +1

      @@nullbeyondo still the time representation will not be continuous since the irrational values will not be included in the domain. I think you should read more about it. A lot more.

  • @rubyjohn
    @rubyjohn 6 лет назад +5

    BEST VISUAL THERAPY IN MY LIFE

  • @AarshWankarIITGN
    @AarshWankarIITGN 8 месяцев назад

    Thanks a lot, at 2:45 AM in the morning, sitting peacefully in in the hostel of my institute, you actually cleared a lot of things up in the first two videos. This is the first time I understood to some extent what gradient descents and weights and cost functions were all about. Looking forward to continuing this journey of learning on your awesome channel 😃

  • @VikasYadav1369
    @VikasYadav1369 6 лет назад +73

    Draw a 5 for me.
    "I am a more of a multiple choice guy"

  • @bytenommer
    @bytenommer 7 лет назад +318

    Could you please just drop everything else you are doing and do these videos full time for the rest of your life.

    • @iLoveTurtlesHaha
      @iLoveTurtlesHaha 6 лет назад +43

      But his videos are a result of his other interests. XD

    • @raycharlestothebs
      @raycharlestothebs 5 лет назад +2

      @@iLoveTurtlesHaha Just like you saying 'XD' is......

    • @thibauldnuyten2891
      @thibauldnuyten2891 5 лет назад +5

      Sadly people still need to work to fricking live.

    • @sgracem2863
      @sgracem2863 3 года назад +1

      @@thibauldnuyten2891 Wouldn't he be rich off these videos though? I mean they all have millions of views and he almost has 4m subs

  • @anjanit2006
    @anjanit2006 Год назад

    Suprised u pulled this off real well. I am 26 years old and working in google for a 1.3 crore job in IT. I am about to be a millionaire all because of u . Like seriously u are the most helpfull person in my life.

  • @emberdrops3892
    @emberdrops3892 6 лет назад +24

    3:41 Oh that poor little network... Say something good to it so it's happy again!

  • @AnshulGuptaAG
    @AnshulGuptaAG 7 лет назад +39

    That XKCD comic is how a lot of people consider neural networks to work :P
    Great video again, 3B1b!
    Edit: Waiting eagerly for your ConvNets and LSTMs :D

    • @somedude4122
      @somedude4122 7 лет назад +2

      And it certainly isn't wrong btw

  • @swealf-nonofficial
    @swealf-nonofficial Год назад +1

    You do such a great job at explaining complex topics in a "simple" way, great job

  • @nigeljohnson9820
    @nigeljohnson9820 7 лет назад +115

    Humans have a habit of seeing images in random data, such as clouds, craters on the Moon or Mars or hearing voices in random radio static. Is this similar to identifying a 5 in a field of random data?

    • @5up3rp3rs0n
      @5up3rp3rs0n 7 лет назад +12

      well for human you see things from the shape or outline that looks like a particular object, kinda like the "See a digit by the loops and lines it has" ideal for this system. So it's all notthe same as that of picking a number and being very confident about it from a static.

    • @seditt5146
      @seditt5146 6 лет назад +17

      But is that what our brain is doing? Is it looking at a static, or are our neurons going .... ok, straight line... then round edge... another round edge..... hmmm that looks like the other 5s I seen... than triggering memory banks to look for other 5s. and again compare.

    • @jomen112
      @jomen112 6 лет назад +39

      No. As explained in the video the network been (more) punished for providing multiple answers than single output wrong answers. That means a multiple answer does not exists as an option for the trained network, i.e. the set of output pattern it has been trains to respond with does not contain multiple choices. That is to say, the alternative answer "I dont know" or "maybe this or that" does not exists for the networks as an answer.
      Regarding clouds or craters, this is not "random data", the shapes we recognize are real and can be agreed upon to exists. This is not the case with noise, i.e. random data. Per definition random data does not not contains pattern and that is why noise carries no meaning to our brains.
      Regarding hearing voices in random static, I would suspect you only would hear voices if there is a pattern (signal) of some form which the brain pics up on and tries to make sense of. How prone you are to hear an actual voices might depend on how your brain be trained, i.e. biased, to detect voices (for instance if you believe one can communicate with ghosts you might be more prone to hear voices were others hears none). Because in the end, detecting meaning, i.e. label stuff, is all about being biased towards a certain interpretation of reality.
      So to conclude, the "reality" for the neural network in the video is biased, or limited, towards a singe neuron output and anything it "perceives" will get a response as such. However, human brains are a little bit more complex and biased differently, i.e. wired up in unique ways, which makes up for the diversity in believes and reasoning among people.

    • @parthshrivastava6325
      @parthshrivastava6325 6 лет назад +2

      That falls under the imagination bracket,it's more like changing the value of the pixels instead of the weights or biases to get a desired output.

    • @ottrovgeisha2150
      @ottrovgeisha2150 6 лет назад +3

      @@seditt5146 Not likely. Nobody knows. Brain s truly bizarre and the connections between cells are actually differently wired. No ask yourself: how does a brain know it exist, how are feelings developed etc. Brain is still a mystery.

  • @PV10008
    @PV10008 5 лет назад +4

    This is the best educational channel on RUclips by a long mile.

  • @terence0till
    @terence0till 9 месяцев назад

    This is so so much better Information visualization than any of my teachers eher had! Plus your calm Voice and humour. I just Like it!

  • @MikhailFederov
    @MikhailFederov 7 лет назад +4

    I wish I had these video when I was first learning. Damn you Tom Mitchell and your formal explanations.

  • @vivekd296
    @vivekd296 7 лет назад +4

    i found your video on jacobians on khan academy at first i was like i don't know this new person he's not sal and then i read the comments and found out it was you !! it was a pleasant surprise

  • @sagig8881
    @sagig8881 2 года назад +1

    Dude you did in 20 minutes what my professor couldn't in 3 hours, Thanks a lot !

    • @klajdiklb1520
      @klajdiklb1520 2 года назад +1

      That's not true. He probably spent hours and hours to design the animations , edit the video and to write the script as smoothly as possible.

  • @ItsCugel
    @ItsCugel 7 лет назад +43

    What about given it just noice and make it learn that it is no digit? Could that improve the output? Maybe also other images that is no digits. It is not only what things are that we can learn from but also that they are not.

    • @3blue1brown
      @3blue1brown  7 лет назад +27

      Try it!

    • @primarkbandit8535
      @primarkbandit8535 7 лет назад +6

      Rofl falafel, this is why we're watching it in the first place. If we could we would, honey. I'll give this a go once I learn more, right now I'm mooching off the really smart kids in the year above until I learn a bit more. This channel is really fantastic and seems to be some kind of frequency illusion phenomenon. Your essence of calculus series came right as we started actually building up intuitions for what is going on in calculus, and now this!

    • @redjr242
      @redjr242 7 лет назад +7

      Very interesting! I agree you should try it.
      The vast majority of all possible 26x26 images look random or noisy so my thoughts are that it might learn to properly classify only a small subset of all possible noise (the kind used in the training data) and the rest will still activate random neurons.

    • @NilesBlackX
      @NilesBlackX 7 лет назад +8

      The problem is with the problem itself, not the cost function etc.
      If you give it random input, the *actual* 'correct' answer, like you would get from a human, would be 'none of the above/indecipherable', not 'well I can see this much of these aspects of these different numbers in here'. So you need an additional output to match the additional input you give it.
      Otherwise it's like when you ask someone "okay, is Jerry's favorite color black or white?" They make a guess, and then you're like "no, it was splork, a color I just made up you big dumb idiot".
      The problem isn't with the guesser (or in this case, the mechanics of the network and cost function) but with the fact that the available outputs do not match the possible inputs. So one of the two requires adjustment, not something else.

    • @ganondorfchampin
      @ganondorfchampin 7 лет назад +7

      +Niles Black
      The problem is not with the problem, it's with the training data, because it COULD output "utter garbage", it just never got rewarded for doing so.

  • @willw4096
    @willw4096 Год назад +4

    3:25 3:36 5:21
    5:36 why gradient descent is necessary (b/c closed-form solution isn't always available)
    5:51 gradient descent
    6:24 different starting points (initially random as weights are initially assigned randomly) could lead to different local minima
    6:42 making stepsize proportional to slope helps prevents overshooting (stepsize = learning rate * derivation, so it's already there)
    10:58 a correction here: a negative derivative indicates we should increase to minimize loss function, instead of decrease.
    15:23

  • @WiredWizardsRealm-et5pp
    @WiredWizardsRealm-et5pp 5 месяцев назад

    Man , it feel so good to learn everything in zero shot now.. the neural networks , gradient decent , backpropagation . I used to get frustrated with lot of challenging concepts.. cuz I did not know maths , and AI terms.. but now after learning it for year it feels worth learning. Thanks to 3Blue guy.. whatever course he touched is worth all lectures combined i can't say. Its just pure core concept with animation. Quality at par

  • @sagjounkani
    @sagjounkani 7 лет назад +7

    Your videos are awesome in the truest sense of the word. Hence, I thought it worthy to mention something which was slightly confusing to me from the videos.
    At 18:05, I think that the concept of deep neural networks generalizing to a randomly labelled image set can be explained better. When I first saw this part of the video, I thought that shuffling of labels for images maintained the 1:1 mapping of the image classes to labels, albeit the false ones. In such a case, the learning would definitely not result to be any different than if the 1:1 mapping was correct. The part of the video in question should bring out the fact (as intended in the concerned paper) that, after random shuffling of labels, the image class to label mapping was 1:Many i.e. two dog images can have labels 'cow' and 'lion' respectively and the network would still be able to achieve very good training accuracy, showing that it is failing to learn features general to dogs but memorizing the image to labels mapping (which does pose a very potent argument to re-think generalization)
    Thanks for making these. :)

    • @3blue1brown
      @3blue1brown  7 лет назад +4

      +Sagar Jounkani I think you are right.

    • @_noirja
      @_noirja 7 лет назад +1

      Thank you for pointing this out! It's fairly confusing. Quite obviously the simple act of calling things by different names would not change anything. I get the jist of what you mean but I still don't fully understand, I should probably read the paper haha!

    • @rulzzzzzz
      @rulzzzzzz 6 лет назад +3

      Hey, I know this comment is one year old but you are the only one to point out the same confusion I had! I still don't quite get it, when you say 'the network would still be able to achieve very good training accuracy' do you mean it still recognizes a picture of a dog as a dog even though the labels are randomized?

    • @glowish1993
      @glowish1993 6 лет назад +1

      I have the same question regarding the recognition of the dog picture even though the labels can be randomized to be other things like 'cow' and 'lion' in a 1 to many mapping, does it mean that the majority of the dog pictures were labelled correctly, with the exception of some labelled incorrectly? Otherwise, wouldn't the network not learn anything if most of the labels are incorrect?

  • @Jkirek_
    @Jkirek_ 5 лет назад +52

    "Is it... **Learning**?!" (How terrifying)
    -"It's just calculus"
    "Even worse!"
    I love the realism, where people would rather have a potential robot uprising than have calculus

    • @sagadegeminis9722
      @sagadegeminis9722 4 года назад +1

      Change the gradient based optimization to genetic algorithm and its "evolving"

    • @codinghub3759
      @codinghub3759 4 года назад

      @@sagadegeminis9722 english please?

  • @santircastillo
    @santircastillo 2 года назад

    This is the first video I found about AI that actually explains the how, instead of just dumbing everything down and leaving just as confused as before watching. Thank you!

  • @OrigamiCreeper
    @OrigamiCreeper 5 лет назад +5

    I am so happy that vsauce recommended this channel because it is amazing!

  • @wolfboyft
    @wolfboyft 4 года назад +3

    4:33 "And that's a complicated thing!"
    Blues: *shuffle nervously*

  • @rob651
    @rob651 2 года назад

    I've watched many videos and done some reading on how neural networks work (learn), but I couldn't find a satisfactory explanation until I watched this video. Your examples, analogies, visuals... were just perfect. Thank you so much.

  • @abhijain8295
    @abhijain8295 5 лет назад +4

    Thank you for providing us with such nice content. Keep it up.