Why Transformers Are So Powerful

Поделиться
HTML-код
  • Опубликовано: 9 июн 2024
  • I find most explanations get lost in the details so i challenged myself to come up with a one sentence description. It's a new kind of layer capable of adapting its connection weights based on input context. This allows one layer to do what would have taken many. I hope this helps you!

Комментарии • 166

  • @filovirus1
    @filovirus1 9 месяцев назад +60

    finally! this is most succinct description I found about Transformers that actually increased my understanding. thank you for this. can't wait for the longer version.

  • @AtulShukla1
    @AtulShukla1 6 месяцев назад +11

    True, majority of the explanations in the internet are just blabbering same stuff without any insights. Thank you for the effort and simple explanation you provided.

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад +2

      thank you, and yes there is so much garbage out there.

  • @barts5040
    @barts5040 6 месяцев назад +11

    Don't stop. I literally have discovered your channel yesterday and really, really love your style, deep preparation and the energy you're putting into those videos! Thanks!!

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад +1

      So glad you found it! I definitely will be posting more, i'm just slow. I'm thinking of re-editing some of my old stuff to help build momentum for next series...

  • @yahiiia9269
    @yahiiia9269 9 месяцев назад +7

    I can't believe I found this hidden gem of a channel just through typing in "RSA encryption explaination". You're like... perfect for me.

    • @ArtOfTheProblem
      @ArtOfTheProblem  9 месяцев назад +2

      yay welcome to the family, i know it's hard to find my channel as YT isn't friends with it anymore

    • @yahiiia9269
      @yahiiia9269 9 месяцев назад +1

      @@ArtOfTheProblem Use a long list of keywords in your description. It will help.

    • @ArtOfTheProblem
      @ArtOfTheProblem  9 месяцев назад

      I didn't know that still worked, I'll try it. If you have other thoughts let me know!@@yahiiia9269

  • @JJ-eb4tx
    @JJ-eb4tx 9 месяцев назад +26

    Easy, "more than meets the eye"

  • @VoldiWay
    @VoldiWay 6 месяцев назад +8

    Whaaaat????! I have watched many hours of videos about transformers, trying desperately to understand how they work, and this is the first video that explained how attention heads are actually dynamic hidden layers that are adjusted on the fly. The other videos made it seem like attention heads are simple score values, but I had no idea that they are in fact large matrices (rivaling the size of static hidden layers) with weights that are adjusted in real-time as each new token is generated! 🤯 Thank you for sharing this huge insight! 🥰

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад +1

      wooo!!! i know i'm so happy to have had this insight and shared it, and it's working for some :)))

    • @traenkler
      @traenkler 2 месяца назад +1

      @@ArtOfTheProblem❤

    • @ArtOfTheProblem
      @ArtOfTheProblem  2 месяца назад

      thanks!@@traenkler

    • @adamcole918
      @adamcole918 12 дней назад +1

      Yes! Exactly what @VoldiWay said. I too suffered from the same gap in understanding. And @ArtOfTheProblem, you are the 1000 watt epiphany light bulb 💡

    • @ArtOfTheProblem
      @ArtOfTheProblem  12 дней назад

      woooo!!!@@adamcole918

  • @Tubeytime
    @Tubeytime 9 месяцев назад +6

    This is akin to changing my focus while giving feedback to artists. The input gives clues about how advanced they are, which informs what kind of feedback they would find most relevant.

    • @ArtOfTheProblem
      @ArtOfTheProblem  9 месяцев назад +1

      Interesting analogy, it also explain in context learning too

  • @BrianSalamone
    @BrianSalamone 6 месяцев назад +2

    Glad to see you've been posting videos. I'll grab some popcorn for them. :)

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад +1

      welcome back! really just one new video, let me know what you think

  • @ArtOfTheProblem
    @ArtOfTheProblem  9 месяцев назад +9

    Let me know if this insight helps you at all or spurs other ideas. The next video will be the conclusion to my AI series which is currently here: ruclips.net/video/YulgDAaHBKw/видео.html&ab_channel=ArtoftheProblem

    • @NoahElRhandour
      @NoahElRhandour 9 месяцев назад +2

      do you have a patreon or another efficient way to get a donation? not sure if membership through youtube is efficient (twitch for example takes half of it, and then you ofc still have to tax your half)

    • @ArtOfTheProblem
      @ArtOfTheProblem  9 месяцев назад +2

      thanks for asking I do have one here: www.patreon.com/artoftheproblem@@NoahElRhandour I really appreciate it

  • @BulentErdemir
    @BulentErdemir 6 месяцев назад +3

    Finally. I went through the same hurdles, too. Couldn't grasp the main concept behind Transformers. This video captures the essence of it very successfully. I wonder why no one has expressed it this way before, too.. Thank you very much for your efforts.

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад +1

      appreciate this, I know....i took me a long time. But i think i've seen others start to land on this too...just once in a talk I was watching last week

  • @Financeification
    @Financeification 9 месяцев назад +3

    Often can't intuitively grasp subjects until I watch your videos. I'm eager for your next release.

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад

      Took 2 years to finish this one, finally live would love your feedback: ruclips.net/video/OFS90-FX6pg/видео.html

  • @pcfreak1992
    @pcfreak1992 9 месяцев назад +6

    Can’t wait for the new video!

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад

      new vid! ruclips.net/video/OFS90-FX6pg/видео.html

  • @deadalnix
    @deadalnix 9 месяцев назад +2

    I read a lot about tranformers myself and came to the same conclusion. Look forward to see your video.

    • @ArtOfTheProblem
      @ArtOfTheProblem  9 месяцев назад

      would love to hear how you were thinking about it.

    • @deadalnix
      @deadalnix 9 месяцев назад

      @@ArtOfTheProblem My mental model was similar to the visualization you had with the musical piece.
      But I think the way you look at it is probably better.

  • @Grateful.For.Everything
    @Grateful.For.Everything 2 месяца назад +1

    Very cool visualization!

  • @nezv71
    @nezv71 9 месяцев назад +2

    Excellent! Very excited for the full video. Always great work man!

  • @mattmatty1234
    @mattmatty1234 9 месяцев назад +5

    I am glad that your back to making content

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад

      new vid! ruclips.net/video/OFS90-FX6pg/видео.html

    • @mattmatty1234
      @mattmatty1234 6 месяцев назад

      Already watched it 2x man
      Glad you’re back.

  • @71sephiroth
    @71sephiroth 9 месяцев назад +2

    Just, WOW! Can't wait for the final video!

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад

      it's done! ruclips.net/video/OFS90-FX6pg/видео.html

  • @signalch1025
    @signalch1025 4 месяца назад

    You are having a great skill to explain complex topics with simple words without using advanced graphics.

    • @ArtOfTheProblem
      @ArtOfTheProblem  4 месяца назад +1

      thank you! i did this one very rough and I'm glad it still works well

    • @signalch1025
      @signalch1025 4 месяца назад +1

      @@ArtOfTheProblemI saw may videos on this topic with lot of math and graphics but I don’t pic exact anaswar for my level of understanding but all your videos on deep learning & generative AI gives excellent explanation on complex topics like this. We are greateful for having you on RUclips .

    • @ArtOfTheProblem
      @ArtOfTheProblem  4 месяца назад +1

      thank you, stay tuned for more!@@signalch1025

  • @keghnfeem4154
    @keghnfeem4154 5 месяцев назад +1

    Brilliant work sage!

  • @benoitleger-derville6986
    @benoitleger-derville6986 6 месяцев назад +1

    So happy to hear from you again :)

  • @pi5549
    @pi5549 8 месяцев назад +1

    I'm amazed I haven't seen this insight before. As you observe, every explainer simply goes down the pre-existing path-of-least-resistance. Would be great if you had a Discord.

    • @ArtOfTheProblem
      @ArtOfTheProblem  8 месяцев назад

      thanks! happy to hear this, i don't have a discord but I do stay active in comments

  • @alderiatefanclub1576
    @alderiatefanclub1576 7 месяцев назад

    Thank you so much ! Your videos always provide crystal clear explanations.

  • @vitabranks8775
    @vitabranks8775 5 месяцев назад +1

    Hey this is a late discovery of this video. I always love your work. (Haven't seen your next video yet) Another way I want to think about transformers based on my limited understanding is that they're just like CNNs (more so than RNNs despite the frequent association between them) but the structures that makes these networks smaller and trainable isn't a *convolution filter* but an *attention head*. Then, just like CNNs, upon passing them through a filter of some sort, they pass these into a smaller fully-connected NN.

    • @ArtOfTheProblem
      @ArtOfTheProblem  5 месяцев назад

      great to have you back, let me know what you think of the next video please. And yes your intuition is in the right direction.

  • @dianes6245
    @dianes6245 6 месяцев назад +1

    This video demystifies Transformers. It moves beyond explanations that get lost in query, key, value matrices and positional encoding. Instead, we explain that a unique layer adapts its connection weights based on input context. This expands the Transformer's efficiency and prowess.
    We Compare this dynamic nature with static layers in traditional networks. And see why Transformers excel at complex tasks with fewer layers.
    Get a visual grasp of how mini networks within layers, known as attention heads, act as information filters. They adjust to input and enhance the model's learning capability. [At training time?] This explanation sheds light on what makes Transformers a game-changer in deep learning.
    [so it works at training time and inference time - those get confused]

  • @ramanshariati5738
    @ramanshariati5738 Месяц назад +1

    amazing work!

  • @bg-mq5hz
    @bg-mq5hz 4 месяца назад

    Adapt its connection weights based on its inputs! Dynamic connection weight!! Now it all is starting to make sense 🙏🙏🙏

    • @ArtOfTheProblem
      @ArtOfTheProblem  4 месяца назад +1

      yes! was excited when I had this insight and i'm glad it's helpful

  • @thouys9069
    @thouys9069 6 месяцев назад +1

    Really good summary. For me it first clicked when I saw Karpathy's transformer series. The data "describes itself", resulting in the Keys. Data can also ask for other kinds of data's keys using Queries. And finally Values are simply modified versions of the data. Each head modifies data differently, and learns different ways of summarizing/describing the data. So what we learn during training is this meta-learning process of describing data, literally the "transformers".

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад +1

      thanks for sharing, i'm curious did you find this video after seeing the history one, or was it recommended? i'm just trying to figure out what's happening as i'm finally getting a ton of views

    • @thouys9069
      @thouys9069 6 месяцев назад +1

      A friend sent me the history video, which struck me by its quality. Then I looked at other videos on your channel. Keep it up! @@ArtOfTheProblem

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад +1

      word of mouth, awesome! welcome to the family

  • @scottekim
    @scottekim 8 месяцев назад

    This is the type of explanation i want. I want insight into how and why these mechanisms are able to act in an intelligent way. I eagerly await your longer video.

  • @belibem
    @belibem 9 месяцев назад +5

    Very interesting. Subscribing to the comments! Also I had to ask GPT4 for some definitions: In the context of Transformer models in machine learning, an "information head" usually refers to a specific part of the model that focuses on capturing some aspect of the data. Think of it like a mini-detector within the model that's good at picking up a certain kind of information, like understanding the meaning of a word in relation to other words.
    An "information filter," on the other hand, is a mechanism that decides what information should be passed on and what should be ignored. Imagine it as a sieve that only lets important bits of information flow through, helping the model focus on what really matters for the task it's trying to accomplish.

    • @belibem
      @belibem 9 месяцев назад +1

      also GPT4 seems to agree with the insight!
      Q:Does it make sense to explain a transformer model as a combination of several smaller models with just one hidden layer (i.e. the attention heads) and a filtering layer that feeds the input in parallel to the many attention heads?
      Answer: Yes, that's a reasonable way to simplify the complex architecture of a Transformer model. You could think of each attention head as a mini-model focusing on a specific part of the input data. These attention heads operate in parallel, like multiple workers each doing their own specialized task.
      The "filtering layer" you mentioned could be thought of as the mechanism that decides what information is most relevant for each attention head to look at. It essentially prepares the input data so each attention head can do its job more effectively.
      After all attention heads have done their work, their outputs are usually combined and passed through additional layers to make the final prediction or interpretation. So, in simple terms, yes, you can describe a Transformer as a combination of several smaller models (attention heads) with a filtering layer that guides them.

    • @ArtOfTheProblem
      @ArtOfTheProblem  9 месяцев назад

      interesting, I'm still mulling over this@@belibem

  • @bhaumikpatel4902
    @bhaumikpatel4902 5 месяцев назад

    2:18 There’s something profoundly beautiful in that visual of a transformer processing music

  • @JavierSalcedoC
    @JavierSalcedoC 9 месяцев назад +1

    AotP uploads are always on time

  • @griggiorouge
    @griggiorouge 9 месяцев назад +1

    I want to quit my career and do something with computers just because your videos make it so amazing.

    • @ArtOfTheProblem
      @ArtOfTheProblem  9 месяцев назад +1

      :) what do you do for a living? I find myself wanting to get away from computers practically speaking, but I do love thinking about them and GPT really was a spark that got me dreaming again....

    • @griggiorouge
      @griggiorouge 8 месяцев назад +2

      @@ArtOfTheProblem studied medicine in south america... mostly paperwork in a failing system centering on benefiting clinics and big pharma. computer people could change medical education with their knowledge... it sucks life and money away and its not worth it unless you manage to get into residency which is statistically unlikely here, and even then most specialists say its too much work and little benefit. technology could change that.

  • @michal261
    @michal261 8 месяцев назад

    Can't wait for the full video

  • @m-pana
    @m-pana 9 месяцев назад +10

    I think the idea of "network that adapts to the input" can be useful to get a good intuition on how attention works, though I don't think that's really the key idea here. I mean, every neural network technically adapts to its input. Intermediate activations change based on the input. Self-attention weighs the linear combination of input embeddings based on... the input. So, as you have argued yourself, they are not so different from a black-box perspective - they just crunch numbers differently, and you could argue they would be functionally equivalent given enough memory and training time (it would be vary fancy to mention universal approximation theorems here, but I'm definitely not qualified to do that).
    I would say that, while there are no definitive answers to why transformers actually work and, most importantly, scale so well, IMHO the key insight is that attention is an incredibly effective inductive bias in the architecture - it just makes it more suitable to process sequential data, since it somehow injects in the architecture the concept that "some elements are more important than others, depending on the context". Mind that this is by no means exclusive to transformers, attention was around at least a couple of years prior: this is a nice reconstruction of its history, though you have probably read it already: lilianweng.github.io/posts/2018-06-24-attention/
    Regardless of where it comes from, I think this inductive bias of attention is what you refer to as "squishing the network": the architecture effectively has sequence processing built inside of it. I would instead avoid the term "adaptive network" because I would argue that every network is somehow "adaptive".
    Having said that, I'm a big fan of your work and I can't wait to see the finished product!

    • @ArtOfTheProblem
      @ArtOfTheProblem  9 месяцев назад +1

      Great comment thanks for sharing. Yes definitely functionally equivalent in theory, and yes every network 'adapts' to input using fixed weights - will think more

    • @ArtOfTheProblem
      @ArtOfTheProblem  9 месяцев назад +1

      I'm focused on the weightings between word pairs being adaptive, which is as you say "a different way of crunching numbers" - when you say adaptive you are talking about 'activations' , right? That makes sense to me, but those adaptive activations are a result of fixed weights and dynamic inputs interacting. what about dynamic and dynamic interactions?
      I find your comment "sequence processing built inside of it" interesting, can you say more about this? because in regular fully connected networks you can think of as each layer as a "step in a processing sequence" (networks "learn a program"). I don't see sequential data being key as Transformers (attention) is good at non sequential data too

    • @belibem
      @belibem 9 месяцев назад

      Also the terms adaptive/dynamic can be misleading because someone may think that the weights of the model are dynamically changing, which is not true. Once the model is trained the weights are static if I get it right.

    • @ArtOfTheProblem
      @ArtOfTheProblem  9 месяцев назад +2

      It's a bit trickier because in transformers they are static (the weights of the QKV matricies) BUT it's designed so that the weighting between word pairs is a result of the interaction between the two matricies they are passing information through - and that result will depend on the words themselves. That's what I mean by adaptive weights. One issue with my analogy is i'm trying to make a very very high level claim about low level details @@belibem

    • @ArtOfTheProblem
      @ArtOfTheProblem  9 месяцев назад

      @m-pana

  • @EllyTaliesinBingle
    @EllyTaliesinBingle 6 месяцев назад

    Your vids are fantastic!

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад

      thank you! let me know what you'd like to see more of

  • @swfsql
    @swfsql 8 месяцев назад +1

    Thanks! I'm trying to study the field, and it's difficult to try to reach for the essence like you did.

  • @B1_66ER
    @B1_66ER 9 месяцев назад

    This is a cool explanation.

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад

      Took 2 years to finish this one, finally live would love your feedback: ruclips.net/video/OFS90-FX6pg/видео.html

  • @nillchen
    @nillchen Месяц назад

    thanks for boiling it down to the essential, which is a sign of great intelligence!

    • @KainniaK
      @KainniaK 10 часов назад

      That's the key insight. Intelligence and compression might be the same thing!

  • @chrisanderson687
    @chrisanderson687 Месяц назад

    Aha! Thank you! My goodness I never got this insgiht despite also reading a ton about these things. And it makes sense: our neural connections are very dynamic, so it's no surprise really that adding dynamism to simulated neural networks would make them "smarter" and more efficient. I hope in the future people figure out how to add even more dynamism, so that these networks can learn properly over time like we do, instead of needing massive training runs that are so very static and expensive (and not personal to a single AI agent's experience).

  • @ytpah9823
    @ytpah9823 7 месяцев назад

    🎯 Key Takeaways for quick navigation:
    00:00 🤖 Transformers' core insight is self-attention: a network with dynamic connection weights that adapt based on input context.
    00:28 💡 Self-attention layers are context-sensitive, making them more efficient at processing information in one layer than traditional networks.
    01:10 🧠 Self-attention allows for shorter networks while maintaining efficiency, unlike fully connected networks that are impractical due to excessive parameters.
    01:37 🔄 Attention heads within the layers dynamically adjust connections between input words, acting as information filters.
    02:32 🎓 Mini networks in the attention heads are learned by the network during training, defining the dynamic interaction of input words.

  • @arashputata
    @arashputata 9 месяцев назад +3

    Thanks

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад

      new vid! ruclips.net/video/OFS90-FX6pg/видео.html

  • @mauroabidalcarrer4083
    @mauroabidalcarrer4083 9 месяцев назад

    I can't wait for that vidéo to drop i still remember thé vidéo you made about neural networks years ago.

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад

      Took 2 years to finish this one, finally live would love your feedback: ruclips.net/video/OFS90-FX6pg/видео.html

  • @QualityControla
    @QualityControla 9 месяцев назад

    Can't wait!

  • @guyindisguise
    @guyindisguise 9 месяцев назад +5

    I am so looking foward to the next video!
    As you said, for some reason transformer explanations are all so similar...
    The worst tutorials are the ones that start with talking about the downsides of RNNs and LSTMs and why transformers fix that. I want to know how the thing works, not why it fixes flaws of other things that I don't fully understand!
    Ideally I would love a video with a simple example that one could follow by hand with pen and paper, similar to the explanation by Brandon Rohrer on deep neural networks back in the day: ruclips.net/video/ILsA4nyG7I0/видео.htmlsi=bycRDWN6o0y12ju6 (the first ~ 12 minutes are really original and insightful and what makes that explanation special, the 2nd half is more a standard explanation of backpropagation).
    But I fear that it's more difficult to understand the buildup of concepts over multiple layers in a transformer.
    I already like the shift in mental models to understand how transformers work.
    Can't wait for the next video. :)

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад +1

      Took 2 years to finish this one, finally live would love your feedback: ruclips.net/video/OFS90-FX6pg/видео.html

  • @nicolastelfyr6792
    @nicolastelfyr6792 9 месяцев назад

    thanks))

  • @AbdennacerAyeb
    @AbdennacerAyeb 3 месяца назад

    This is what I need. Adaptive Learning

    • @ArtOfTheProblem
      @ArtOfTheProblem  3 месяца назад

      there you go!

    • @ArtOfTheProblem
      @ArtOfTheProblem  2 месяца назад

      would love if you could help share my newest video: ruclips.net/video/5EcQ1IcEMFQ/видео.html

  • @ikoino
    @ikoino 6 месяцев назад +1

    I saw you later video first, so am back-tracking. I wonder if you could think of each Attention Head as a Huffman tree? Where the branches can span a variable length number of time intervals?

  • @georgiepentch
    @georgiepentch 9 месяцев назад +1

    HES BACK!!!!!

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад

      new vid! ruclips.net/video/OFS90-FX6pg/видео.html

  • @MrMehrd
    @MrMehrd 8 месяцев назад +1

    Thx, waiting for new video on transformers, yea

  • @amir-ali8850
    @amir-ali8850 5 месяцев назад +1

    Excellent insight! Where can I get the full video of transformer music generation?

  • @numannebuni
    @numannebuni 9 месяцев назад

    Thank you! I'm looking forward to your long video. Btw, do you have a twitter account?

  • @Jack-cm5ch
    @Jack-cm5ch 4 месяца назад +1

    you are the GOAT!!!

  • @btnt5209
    @btnt5209 8 месяцев назад +3

    This is a great explanation and also the one that most people who truly understand the architecture ascribe to -- although few explain it that way, curiously. Transformers perform KNN in a latent space -- the difference is that they calculate a latent space specific to your input context, and are able to modify that latent space as new information comes in.

  • @ioannischristou2362
    @ioannischristou2362 6 месяцев назад +1

    you say that an attention layer can dynamically adapt its weights during run-time (as opposed to training time) ? this is something new to me, as I did not realize that the LLM can be changing its own weights during inference time, and I would love to hear some clear (mathematical if possible) explanation about it. Also, you mention that the attention mechanism is equivalent to transforming an otherwise deep network with many hidden layers to a shallow network with one big hidden layer that can be trained easier? isn't this completely in contrast to the whole deep learning "movement", whereby the deep network usually wins because it can "transform" the otherwise local minima in the search-space into "saddle-points" from which with enough training you can escape? Thank you for your videos.

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад

      thanks!
      1) It's still "deep" just not "so deep as to be untrainable"
      2) the dynamic connections I'm talking about are defined by the Key/Value matricies which interact on a word level at each layer
      lmk if you have more Questions

  • @YoussefMohamed-er6zy
    @YoussefMohamed-er6zy 2 месяца назад

    hi there, thanks for the content, btw what is the source of the transformer processing music animation ?!

    • @ArtOfTheProblem
      @ArtOfTheProblem  2 месяца назад +1

      hey this should be in the description or in the video watermark i found it on twitter

  • @notdian
    @notdian 9 месяцев назад +2

    Can you please add more tiers between second and last in your patreon ❤

  • @marcfinns4302
    @marcfinns4302 5 месяцев назад

    Brilliant. Is the longer video coming out at some point?

    • @ArtOfTheProblem
      @ArtOfTheProblem  5 месяцев назад

      it's out! ruclips.net/video/OFS90-FX6pg/видео.html

  • @yorailevi6747
    @yorailevi6747 5 месяцев назад

    yes!

  • @samarthtandale9121
    @samarthtandale9121 Месяц назад +1

    Bro has police running behind him 😂 🚨🚓

  • @EliathahBoda
    @EliathahBoda 6 месяцев назад

    Any update on the longer video that this is part of?

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад

      I'm actually just finished the script and hope to post in a week or so, sorry for delay! if interested my script is here as i'm looking for notes: docs.google.com/document/d/1s7FNPoKPW9y3EhvzNgexJaEG2pP4Fx_rmI4askoKZPA/edit

  • @georgiepentch
    @georgiepentch 9 месяцев назад

    What's the source of the music-processing video at 1:57?

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад

      I have the source in this new video: ruclips.net/video/OFS90-FX6pg/видео.html

  • @ttthttpd
    @ttthttpd 9 месяцев назад +1

    Giant robot movies with bayhem and big 'splosins.

  • @Libertariun
    @Libertariun 6 месяцев назад

    0.40 ... a new kind of layer, which can adapt its connection weights based on the context of the input...

  • @rogerscott529
    @rogerscott529 6 месяцев назад +1

    It would be useful if you could be more specific about what you mean by "more efficient". "Efficient" in what respect(s)? "More efficient" usually means "doing more (of something X) with the same, or less (, of something else Y). So what are the something X and the something else Y here?
    A really non-obvious (to me, at least) thing about attention is why does having multiple attention heads (as all real transformer layers do) do anything more than what one head would do? The answer is that they're all randomly initialized to they have at least the opportunity to all learn separate ways to contribute (in practice, this doesn't work perfectly -- some do the same thing as others, and some contribute little or nothing).

    • @ArtOfTheProblem
      @ArtOfTheProblem  6 месяцев назад

      In this case efficient means fewer layers of depth in the network. and ultimately fewer connections.
      the attention heads each learn different aspects, it turns out that adding them helps, up until a certain point - this is active research

  • @sethstewart1103
    @sethstewart1103 4 месяца назад +1

    Correct!!! The genius of transformers isn't even attention. It's the connectivity pattern. The ability to treat data as virtual neurons.

    • @ArtOfTheProblem
      @ArtOfTheProblem  4 месяца назад +1

      yes! i was originally thinking data creating 'synthetic connections'

    • @sethstewart1103
      @sethstewart1103 4 месяца назад +1

      @@ArtOfTheProblem It bears considerable resemblance to metric learning in that it is flexible, extensible and retrieval-based. When explaining transformers to others I often consolidate the query-key product to a general relationship I refer to as the "relevance function" of that data point to this data point in the current context. One term fewer and it simplifies the discussion into something we already think about natively

    • @ArtOfTheProblem
      @ArtOfTheProblem  4 месяца назад

      yes, perfect@@sethstewart1103

  • @dogecoinx3093
    @dogecoinx3093 5 месяцев назад

    SHA-256 can compress ngrams of any size into a 64 chars.

  • @kaitlyn__L
    @kaitlyn__L 9 месяцев назад +1

    Is it annoying when progress is happening so quickly that you (presumably) have to keep updating your voiceover? I understand you focus on the fundamentals and those are the same, but even just references to the current chatGPT version etc gotta be updated right?

    • @ArtOfTheProblem
      @ArtOfTheProblem  9 месяцев назад +4

      tell me about it, I started my next video (on how networks process sequential data) before chatGPT even came out, threw me for a looooong loop. but at this point I'm ready to make the next video and hit on all the high level Ideas which will stay relevant regardless of GPT 5 etc.. I'm excited about it.

  • @buh357
    @buh357 3 месяца назад

    this is it. :))))))

  • @NeoShameMan
    @NeoShameMan 9 месяцев назад

    Me: those are AND logics,
    but the real power isn't the transformer, it's the embeddings, AND I don't want to spill the bean just yet.

    • @ArtOfTheProblem
      @ArtOfTheProblem  9 месяцев назад

      say more!

    • @NeoShameMan
      @NeoShameMan 9 месяцев назад

      "input self similarity pre learned"@@ArtOfTheProblem

    • @ArtOfTheProblem
      @ArtOfTheProblem  9 месяцев назад

      interesting, so are you saying that the real power isn't the transformer but the importance of embeddings and how they might already encapsulate some notion of 'input self similarity'. @@NeoShameMan

    • @NeoShameMan
      @NeoShameMan 9 месяцев назад

      They prevent the network from learning an arbitrarily sparse input, ie the data is self sorted, the self similarity encode the distribution bias that define the inherent semantic of the input. LLM are "special" in machine learning because the input distribution and the output distribution is the same.
      You can basically bog down neural network at large into two main function, sorting input data together, then tagging them with class.
      The transformer is merely a clever acceleration structure, but if you look at the evolution of power of neural network, limiting the input field (attention mechanism) as been a way to accelerate learning and increase power.
      Most big LLM are very sparse with neuron only reacting to specific Ngram (aka bag of word), I encourage you to look at how chatbot using chatscript are program too (suzette and rosette).
      The reason is that you tokenized the input allowing better composition (ie higher level token), reproducing an adhoc parsing tree. This pose the question, what limit a dag impose on that structure.
      LLm goes one step beyond and dissociate teh learning of the input (embedding training) and the output token generation) ie it's virtually deeper than the "physical" model because really it's two model chained together. @@ArtOfTheProblem
      But I would go much deeper, I would ask you, what function are the neuron ACTUALLY are doing, I don't mean logistic regression and the kind, I mean *concretely*, try to separate each mechanism of the neuron (beyond the bs brain analogy or obscure math) and question yourself what do they do (the mul, the add, the bias, the activation).
      After all the math is merely the same as groceries store ticket you look at to see if you got over budget or not...
      IMHO LLM will go the way of NeRF vs Gaussian splatting ... YAGNNNI! (You Aren't Going to Need Neural Network for Intelligence)

  • @joanyadhatan
    @joanyadhatan 9 месяцев назад

    "Transformers is a series of science fiction action films based on the Transformers franchise of the 1980s." (wikipedia)
    calling a new technology by the same name is more idiotic than the film/series itself. 🤦🏼‍♂
    it is more efficient to name it XxX btw... :(

    • @AB-wf8ek
      @AB-wf8ek 8 месяцев назад

      "A transformer is a deep learning architecture that relies on the parallel multi-head attention mechanism. The modern transformer was proposed in the 2017 paper titled 'Attention Is All You Need' by Ashish Vaswani et al., Google Brain team." (Wikipedia)
      Not being able to add "AI" to your search in order to find it, says more about your search skills than anything.

  • @user-jw4kv6zq6b
    @user-jw4kv6zq6b 8 месяцев назад

    imagine having a brain that understands what he is saying..lol..

  • @joanyadhatan
    @joanyadhatan 9 месяцев назад

    ok, but... can you explain in one sentence what the fuck Transformers is so I can google it?

  • @macethorns1168
    @macethorns1168 9 месяцев назад

    Energon makes them powerful.

  • @GregMatoga
    @GregMatoga 5 месяцев назад

    Thanks

    • @ArtOfTheProblem
      @ArtOfTheProblem  5 месяцев назад

      WOW thank you so much for this. I'd love to know if you were struggling with this concept? what else are you looking for help with. right now i'm working on an RL video