What OVERFITTING Teaches You About Life | Machine Learning and Statistics

Поделиться
HTML-код
  • Опубликовано: 4 фев 2025

Комментарии • 53

  • @Flourish38
    @Flourish38 День назад +58

    You got SO CLOSE to mentioning the bias-variance tradeoff, you even described it quite well, but you didn’t. It’s the name of the rule underlying this principle, and it runs much deeper than machine learning, even describing some aspects of information theory! Very worth understanding, and I highly recommend anyone interested looks it up.

  • @seancheong922
    @seancheong922 День назад +44

    life is more like differential equations rather than polynomials

    • @gotri7
      @gotri7 День назад +16

      Life is more like life rather than mathematics.

    • @sonicmaths8285
      @sonicmaths8285 День назад +6

      @@gotri7hmmm pretty obvious since all things that are things are at least similar but actually exactly equivalent to themselves
      That observation is more recurrent than helpful LOL

    • @trevorallen3212
      @trevorallen3212 День назад +6

      Hmm idk about that i think life is about differintengral equations.

    • @LuckbyNumbers
      @LuckbyNumbers  День назад +5

      Especially a system of non-linear PDEs - life is chaotic and there are no general solutions!

    • @positivenozy6065
      @positivenozy6065 23 часа назад +1

      bro how about normal distribution or statistical models?

  • @waterbird2686
    @waterbird2686 День назад +23

    what bro sends me after I tell him he should study for the final instead of gaming all day

    • @LuckbyNumbers
      @LuckbyNumbers  День назад +6

      Haha! But more people underfit by barely studying at all!

  • @philosolog
    @philosolog День назад +13

    Your channel is CRIMINALLY underrated. This was a brilliant watch.

  • @obamos7870
    @obamos7870 День назад +7

    Never saw it that way but it's true. I definitely overfit when studying for exams. Thank you for the insightful video!!

  • @Luxof_
    @Luxof_ 4 часа назад

    I take advantage of this ALL the time on tests and exams. Instead of memorizing specific things word-for-word (which i find painfully slow and just painful in general the few times i do it) i just moderately memorize the few key points and talk out my ass.

  • @rej1skyspell686
    @rej1skyspell686 19 часов назад

    This also remembers me kind of the "Einstellung" effekt which makes worked examples more useful at least if you are interleaving enough because like that you better learn when f.e. use the product rule.

  • @rej1skyspell686
    @rej1skyspell686 19 часов назад

    Never thought about the Barnum effect as overfitting thanks for the insight :)

  • @KayKay-ob6tz
    @KayKay-ob6tz 7 часов назад

    When i was a teen i thought i could make a machine learning that is made by bayesian logic but also works like decision tree to classify(but instead of branching deeeply, it uses and and or logic to classify) even tho i dont have deep understanding of programming even by now

    • @Luxof_
      @Luxof_ 4 часа назад

      Wild educated guess: you're still a teen or below and you're not from the U.S.
      - Cover up ("when i was a teen")
      - Grammar ("have deep understanding")
      - Lack of capitalization
      Farewell and git gud at covering things up lmao

  • @joebucket1471
    @joebucket1471 День назад +4

    Have you heard about grokking though (generalization beyond overfitting in machine learning) ?

    • @LuckbyNumbers
      @LuckbyNumbers  День назад +1

      Never heard of that. Let me look that up!

    • @pytguyid
      @pytguyid День назад

      Was coming here to mention grokking.

  • @GermanSeverin
    @GermanSeverin День назад +2

    When I watched this vid I thought you had at least 100k subscribers and years of youtube expirience. Great video! keep it up!

  • @thomasmcghee360
    @thomasmcghee360 21 час назад

    Amazing content, you deserve more recognition for the quality you produce!

  • @vadimzhdanov3246
    @vadimzhdanov3246 День назад

    Hey, thank you for posting this video.
    I wasn't in a mood to dive into this specific topic deeply, but just watching segments of this video and hearing this nice voice with this nice music, contemplating graphs and formulas were so relaxing and kind of inspiring, I feel better now and going back to my learning routine.
    Great impact on curious minds, dear Author.

  • @M1szS
    @M1szS 21 час назад +1

    I avoid overfitting by not studying at all

  • @askebe3
    @askebe3 21 час назад +2

    Maybe a polynomial for a temperature evolution is just completely wrong in general

    • @yurisich
      @yurisich 19 часов назад

      I agree. Any curve can be described by combining a group of sine waves using a Fourier transform. It seems much more efficient to catalog these frequencies when describing the fit of a function, and would respond better to additional data when tuning against validation.

  • @aarash6982
    @aarash6982 16 часов назад

    Instant sub

  • @linezich8680
    @linezich8680 18 часов назад

    Pretty underrated content ❤

  • @brycehabel207
    @brycehabel207 2 дня назад +1

    great video man! just subscribed

  • @abderrahm3n
    @abderrahm3n День назад

    The video is extremely insightful and creative ,love it ,keep it up

  • @ripper5941
    @ripper5941 23 часа назад

    This is some great content, keep it up

  • @neuralnetwork-nahigotnoclue
    @neuralnetwork-nahigotnoclue 23 часа назад

    But in overparameterized regime, it starts to generalized after the interpolation threshold in the second descent. Does that mean we study even more 😅?

  • @leoliang1
    @leoliang1 День назад

    Very insightful. Lovely!

  • @barbozafernandezclaudioema1585
    @barbozafernandezclaudioema1585 21 час назад

    Mythical pull, nice vid

  • @duongpham-zj6hv
    @duongpham-zj6hv 2 дня назад +1

    great vid

  • @yacinebel-hadj6559
    @yacinebel-hadj6559 20 часов назад

    maybe is you study even more you experience the double descent phenomena

  • @mj9665
    @mj9665 День назад +1

    interesting

  • @colinpitrat8639
    @colinpitrat8639 День назад

    I didn't watch the whole video but the opening analogy is clearly wrong. Overfitting is when your model is too powerful for the amount of data you have and your problem. Or seen from the opposite perspective, you don't have enough data for your model.
    When you study, you're (supposedly) not touching the complexity of your model (your brain, number of neurons and connexions it has) but you're adding more data. How could this lead to overfitting?
    The problem could be that you're choosing the wrong exercises to practice, focusing on a small part of your curriculum. But then, it's not overfitting, it's biased data.

    • @LuckbyNumbers
      @LuckbyNumbers  День назад +1

      Thanks for the comment! When you do practice problems for an exam, you are training your brain's problem-solving model to solve the exam questions. Overfitting occurs when people memorise the answers to the practice problems instead of understanding the reasoning behind (or in some scenarios, overcomplicating things).
      Towards the end of the video, I touched upon this and explained how we can avoid overfitting by training our brain using practice problems, but then use a timed mock exam as validation to ensure that our brain is not just memorising the specific answers to the practice questions but actually understanding them and applying the knowledge to unseen questions.
      So, overfitting is not a result of adding more data points (as you have corrrectly mentioned), but a result of focusing too much on the specific examples or patterns in the training set without generalising the underlying principles.

  • @joseleperez8742
    @joseleperez8742 День назад

    Well, hyperfitting and the double U want to enter the chat…