Machine Learning-Bias And Variance In Depth Intuition| Overfitting Underfitting

Поделиться
HTML-код
  • Опубликовано: 26 июн 2024
  • In statistics and machine learning, the bias-variance tradeoff is the property of a set of predictive models whereby models with a lower bias in parameter estimation have a higher variance of the parameter estimates across samples, and vice versa
    Please join as a member in my channel to get additional benefits like materials in Data Science, live streaming for Members and many more
    / @krishnaik06
    Please do subscribe my other channel too
    / @krishnaikhindi
    If you want to Give donation to support my channel, below is the Gpay id
    GPay: krishnaik06@okicici
    Connect with me here:
    Twitter: / krishnaik06
    Facebook: / krishnaik06
    instagram: / krishnaik06

Комментарии • 455

  • @heartsbrohi9394
    @heartsbrohi9394 3 года назад +185

    Good turorial. My thoughts below (hope it adds to someone's understanding):
    We perform cross validation (to make sure that model has good accuracy rate and it can be used for prediction using unseen/new or test data). To do so, we use train and test data by properly splitting our dataset for example 80% for training, 20% for testing the model. This can be performed using train_test, train_test_split or K-fold (K-fold is mostly used to avoid under and overfiting problems).
    A model is considered as a good model when it gives high accuracy using training as well as testing data. Good accuracy on test data means, model will have good accuracy when it is trying to make predictions on new or unseen data for example, using the data which is not included in the training set.
    Good accuracy also means that the value predicted by the model will be very much close to the actual value.
    Bias will be low and variance will be high when model performs well on the training data but performs bad or poorly on the test data. High variance means the model cannot generalize to new or unseen data. (This is the case of overfiting)
    If the model performs poorly (means less accurate and cannot generalize) on both training data and test data, it means it has high bias and high variance. (This is the case of underfiting)
    If model performs well on both test and training data. Performs well meaning, predictions are close to actual values for unseens data so accuracy will be high. In this case, bias will be low and variance will also be low.
    The best model must have low bias (low error rate on training data) and low variance (can generalize and has low error rate on new or test data).
    (This is the case for best fit model) so always have low bias and low variance for your models.

    • @karthikparanthaman634
      @karthikparanthaman634 3 года назад +2

      Wonderful summary!

    • @farhathabduljabbar9879
      @farhathabduljabbar9879 3 года назад +8

      You should probably create articles coz you are good at summarising concepts!
      If you have one please do share!

    • @adegboyemoshood809
      @adegboyemoshood809 3 года назад

      Great

    • @roshanid6523
      @roshanid6523 3 года назад +1

      Very well written 👍🏻
      Thanks for sharing
      👍🏻 Consider writing blogs

    • @AnandKumar-to6ez
      @AnandKumar-to6ez 3 года назад +1

      Really very nice and well written. After watching video, if we go through your summery, its a stamp on our brains. Thanks to both for your efforts.

  • @sandipansarkar9211
    @sandipansarkar9211 3 года назад +217

    This video need to be watched again and again.Machine learning is nothing but proper understanding of ovrfitting and underfitting..Watching the second time.Thanks Krish

    • @adipurnomo5683
      @adipurnomo5683 2 года назад +4

      Ageeed!

    • @batman9937
      @batman9937 2 года назад +7

      This is what they asked me in OLA interview. And the interviewer covered great depth on this topic only. It's pretty fundamental to ML. Sad to report they rejected me though.

    • @ashishbomble8547
      @ashishbomble8547 2 года назад

      @@batman9937 hi man plz help to know what other questions they asked .

    • @carti8778
      @carti8778 Год назад +2

      @@ashishbomble8547 buy the book :: ace the data science interview by Kevin Huo and nick singh .

  • @shashankverma4044
    @shashankverma4044 4 года назад +15

    This was my biggest doubt and you clarified it in so easy terms. Thank you so much Krish.

  • @westrahman
    @westrahman 4 года назад +22

    XGBoost, the answer cant be simple, but what happens is when dealing with high bias, do better feature engineering n decrease regularization, so in XGBoost we increase depth of each tree and other techniques to handle it to minimize the loss...so you can come to conclusion that if proper parameters are defined (including regularization etc) it ll yield low bias and low variance

  • @rafibasha1840
    @rafibasha1840 2 года назад +34

    Hi Krish,thanks for the explanation ..6:02 it should be high bias and low variance in case of under fitting

    • @mohitzen
      @mohitzen 2 года назад +3

      Yes exactly i was looking for this comment

    • @singhvianuj
      @singhvianuj 2 года назад

      Amazing video by Krish. Thanks for pointing out this. @Krish Naik please make a note of this

    • @shailzasharma5619
      @shailzasharma5619 2 года назад

      yess!!!

    • @rohitkumar-gi8bo
      @rohitkumar-gi8bo Год назад

      yess

    • @ramyabc
      @ramyabc Год назад

      Exactly! I searched for this comment :)

  • @ellentuane4068
    @ellentuane4068 2 года назад +8

    Can't express my gratitude enough ! Thank you for explaining it so well

  • @MrAli2200
    @MrAli2200 3 года назад +1

    You can't get a clearer explanation than this, hats off mate

  • @donkeshwarkavyasree8632
    @donkeshwarkavyasree8632 2 года назад

    The best explanation among the whole youtube channels 👏. I love the way how you always keep things simple. Glad to find out about your channel, sir.

  • @dishashaktawat2219
    @dishashaktawat2219 3 года назад

    providing these info makes you a great teacher... the way you explain everything going to brain.....

  • @khursiahzainalmokhtar225
    @khursiahzainalmokhtar225 Год назад +1

    What an excellent explanation on bias and variance. I finally understood both terms. Thank you so much for the video and keep up the good work!

  • @sidduhedaginal
    @sidduhedaginal 4 года назад +1

    This guy is really great...Thank you so much for effort you put for us.

  • @randomdyno
    @randomdyno 3 года назад +4

    krish sir i hope God bless you with whole heart you are doing great job and thanks for the INEURON it made my life easy.

  • @kaushalpatwardhan
    @kaushalpatwardhan 2 года назад +2

    I have been trying to understand this concept since long ... But never knew its this simple 😀 thank u Krish for this amazingly simple explanation to understand.

  • @manishsharma2211
    @manishsharma2211 4 года назад +3

    Very good. Revised my concepts perfectly 🔥🔥

  • @teegnas
    @teegnas 4 года назад +2

    I really love his in-depth intuition videos ... compared to his plethora of videos!

  • @vatsal_gamit
    @vatsal_gamit 3 года назад +18

    at 6:10 you made it all clear to me in just 2 lines!! Thank you for this video :)

  • @cheeku5568
    @cheeku5568 4 года назад +2

    One video all clear content... thanks bro it was really a nice session.. u really belong to low bias n low variance human. Keep posting such clear ML videos..

  • @yashkhandelwal6558
    @yashkhandelwal6558 3 года назад +1

    Great I learnt by watching your entire playlist.

  • @nandinisarker6123
    @nandinisarker6123 3 года назад

    I really was in great need of such an excellent explanation of Bias and variance. great help!

  • @VigneshVicky-cn8ek
    @VigneshVicky-cn8ek 2 года назад

    You nailed it man ! Great work ! Respect your time and effort!

  • @vinayakbachal8134
    @vinayakbachal8134 3 года назад +2

    bhai, tu bahot sahi hai, 2.80 lacs fees bharke jo baat nahi samzi easily wo tumne 16 minutes me bata di..kudos..amazing word dear, all the very best

  • @jerrylu1452
    @jerrylu1452 3 года назад

    Thank you for a detailed explanation of bias and variance. Great teaching!!

  • @nevillesantamaria7487
    @nevillesantamaria7487 2 года назад

    Beautifully explained. My concept are now clear on Over fitting and Under fitting models. 👍 Thanks 🍻

  • @chandrachalla3466
    @chandrachalla3466 Год назад +1

    This is an awesome video - was fully confused earlier - this video made it all clear !! Thanks a lot sir !!

  • @YashSharma-es3lr
    @YashSharma-es3lr 3 года назад

    sir after watching this video , mera confusion ek baar mein clear ho gya between bias and variance , awsome explaination

  • @kanhataak1269
    @kanhataak1269 4 года назад +1

    After watching this video doubt is clear really helping this. And Thanks given ur precious time...

  • @adiMallya
    @adiMallya 4 года назад +1

    Thanks for revising these important concepts

  • @06madhav
    @06madhav 4 года назад +2

    Thanks for this. Amazing explanation.

  • @satishbanka
    @satishbanka 3 года назад

    One of the best explanations of Bias and Varianace w.r.t Overitting and underfitting...

  • @delwarhossain43
    @delwarhossain43 4 года назад +1

    Very important discussion on important words in ML. Thanks. Easy explanation on hard words.

  • @nissankarki917
    @nissankarki917 3 года назад

    Awesome video. It explained many concepts significantly.

  • @eswaranthanabslan2607
    @eswaranthanabslan2607 3 года назад

    Krish, you are a master in statistics and machine learning

  • @ameysawant2301
    @ameysawant2301 4 года назад

    You make one of the best tech videos on youtube !!!!

  • @gurjeetkaur1605
    @gurjeetkaur1605 3 года назад

    Thank you , This video is really helpful to understand the Bias and Variance concepts

  • @tarunkumar-gg5ks
    @tarunkumar-gg5ks 3 года назад

    Thank you for good explanation of bias & variance..❤️

  • @Cat_Sterling
    @Cat_Sterling 2 года назад

    Awesome video, thank you so much for these wonderful explanations, they are much needed!

  • @72akshayvasala59
    @72akshayvasala59 3 года назад

    U are Reallly great sir ... ur explanation is very much Crystal Clear

  • @agarwalneha1986
    @agarwalneha1986 3 года назад

    You explained it really well!! Thank you!

  • @avinashankar
    @avinashankar 2 года назад

    Thanks Krish, had scourged the net, but this understanding was great. Good memory hook! Thanks for this.

  • @milanbhor2327
    @milanbhor2327 23 дня назад

    The most clear and precise information 🎉 thank you sir❤

  • @rinksist
    @rinksist 3 года назад

    Thank you sooo much for making it so easy to understand.

  • @shekharpandey9776
    @shekharpandey9776 4 года назад +3

    Please give a video on some mathematical terminology like gradient descent etc. You are really doing a great job.

  • @rahul1994negi
    @rahul1994negi 3 года назад

    Thank you, this video cleared all my doubts :)

  • @shreyasb.s3819
    @shreyasb.s3819 3 года назад

    Superbbb explained..it connected my dots. Thank u

  • @aseemjain007
    @aseemjain007 11 дней назад

    Brilliantly explained !! Thank you !!

  • @gunjalvis
    @gunjalvis Год назад

    ultimate discussion and person who discussed

  • @bauyrjanjyenis3045
    @bauyrjanjyenis3045 2 года назад

    Very succinct explanation of the very fundamental ML concept. Thank you for the video!

  • @anaghaparadkar0511
    @anaghaparadkar0511 3 года назад

    Very well explained. Thank you so much sir.

  • @hemdatascience369
    @hemdatascience369 2 года назад

    Today, I got clarity about this Topic, Tq u sir

  • @pavithrad9543
    @pavithrad9543 3 года назад

    Krish thankyou so much., this is the best channel for data science that I ever seen. Great efforts Krish. Thanks again.

  • @shantanupagare7141
    @shantanupagare7141 4 года назад

    Great Explanation Sir, thanks a lot for the video

  • @mansishah8151
    @mansishah8151 3 года назад

    Love watching your video’s..You explain very well.

  • @kamaleshp6154
    @kamaleshp6154 Год назад

    Best Explanation on Bias and Variance!

  • @alexsepenu
    @alexsepenu 3 года назад

    you made my work easy by this explanation. thanks.

  • @sampurngyaan2867
    @sampurngyaan2867 2 года назад

    It was really good video and it clears all the doubts I have.

  • @vladimirkirichenko1972
    @vladimirkirichenko1972 Год назад

    Insanely good video. Also this has amazing energy!

  • @kiran082
    @kiran082 4 года назад

    Awesome explanation.Thanks a lot for the video

  • @khmercuber5570
    @khmercuber5570 2 года назад

    Thanks Mr. Krish for your best explanation, now I can clearly understand about Bias and Variance :D

  • @tagoreji2143
    @tagoreji2143 Год назад

    Tqsm Sir.Very Valuable Information

  • @yashkhilavdiya5693
    @yashkhilavdiya5693 Год назад

    Thank You so much Krish Sir..!!

  • @sandipansarkar9211
    @sandipansarkar9211 3 года назад

    watched this again,Each time I feind something new

  • @robertselemani3775
    @robertselemani3775 2 года назад

    Well articulated, thank you Krish

  • @CamiloDiazBogotaDC
    @CamiloDiazBogotaDC 2 года назад

    tbh, best video on youtube about Bias And Variance.

  • @emamulmursalin9181
    @emamulmursalin9181 3 года назад +30

    At 06:08 it is said that the underfitted data, the model has high bias and high variability. To my understanding, the information is not correct.
    Variance is the complexity of a model that can capture the internal distribution of the data points in the training set. When variance is high, the model will be fitted to most (even all) of the traiining data points. It will result in high training accruacy and low test accuracy.
    So in summary :
    When the model is overfitted : Low bias and high variance
    When the model is underfitted :High bias and Low variance
    Bias : The INABILITY of the model to be fit on the training data
    Variance : The complexity of the model which helps the model to fit with the training data.

    • @tejanarora1221
      @tejanarora1221 2 года назад +2

      yes bro, you are correct

    • @rohitkumarchoudhary
      @rohitkumarchoudhary 2 года назад

      I also have same doubt. @Krish Naik sir , please have a look on it.

    • @swapnilparle1391
      @swapnilparle1391 2 года назад

      But under fitting suppose to have low accuracy of training data know ? Confusing !!

    • @prachinainawa3055
      @prachinainawa3055 2 года назад +1

      Have I learned the wrong definition of bias and variance by krish sir's explanation? Now I am confused😑

    • @swapnilparle1391
      @swapnilparle1391 2 года назад

      @prachi... not at all concept is at the end same

  • @hiteshikajadhav5540
    @hiteshikajadhav5540 3 года назад

    you explain everything so well :)

  • @kajalkush1210
    @kajalkush1210 2 года назад +1

    Way of explanation is woww.

  • @nahidzeinali1991
    @nahidzeinali1991 Год назад

    your are so awesome, I love your teaching

  • @lekshmigs123
    @lekshmigs123 2 года назад

    Thank you very much for the simple and proper explanation...

  • @sauravb007
    @sauravb007 4 года назад +40

    XGBoost should have low bias & low variance !

    • @brahmihassane8499
      @brahmihassane8499 3 года назад

      Not really it will depend how do you tune the hyperparameters of the model, for this reason it is important to tune a model in order to find a compromise that ensure a low biais (capacity of the model to fit a theoritical function) and low variance (capacity of model to generalisation)

  • @shreyagupta222
    @shreyagupta222 2 года назад

    Thanks a lot for the wonderful explanation

  • @saydainsheikh3123
    @saydainsheikh3123 2 года назад

    brilliant video!!!!! explained everything to the point.

  • @basantmounir
    @basantmounir 3 года назад

    Unbelievably amazing 👏

  • @vennapoosasurendranathredd1027
    @vennapoosasurendranathredd1027 3 года назад

    Mam even though I am studying AI in my clg probably this is easy to understand thanks man..

  • @abdelwahabfaroug9656
    @abdelwahabfaroug9656 Год назад

    very useful lecture , it helps me much to understand this topic in a simple and easy way please keep going

  • @BusinessDexterity
    @BusinessDexterity 3 года назад

    You are amazing, thanks a lot for your wisdom

  • @brahimferjani3147
    @brahimferjani3147 3 года назад

    Good pedagogy and easy explanation. Thanks a lot

  • @bakerkar
    @bakerkar 11 месяцев назад

    excellent tutorial. better than IIT professors who r teaching machine learning.

  • @yourtube92
    @yourtube92 11 месяцев назад

    Very good video, easiest video for understanding logic of bias & variance.

  • @basharb5215
    @basharb5215 2 года назад

    Great explanation. Thank you so much!

  • @benyaminem.9385
    @benyaminem.9385 7 месяцев назад

    Thank you so much bro ! So clear !!!

  • @alihamdan425
    @alihamdan425 2 года назад

    GREAT SIR I GOT IT, THANKS FOR YOUR EFFORT.

  • @dadoll1660
    @dadoll1660 3 года назад

    I just love this guy, - from PH.

  • @anirudhsrivastava3530
    @anirudhsrivastava3530 2 года назад

    very niceee, filled the gap in my knowledge

  • @sharawyabdul6222
    @sharawyabdul6222 2 года назад

    Very well explained. Thanks

  • @santhoshkumar-dd6xq
    @santhoshkumar-dd6xq 4 года назад

    Simple and crisp explaination.

  • @janakiraam1
    @janakiraam1 3 года назад

    Clear explanation. @krish sir thanks for making this video

  • @amitjajoo9510
    @amitjajoo9510 4 года назад

    Thanks for updating this video

  • @kuthaardana
    @kuthaardana 3 года назад

    Simple, easy to understand. Thanks

  • @sabaamin6442
    @sabaamin6442 Год назад

    woow awesome, great work done in one single video. insightful

  • @malikumarhassan1
    @malikumarhassan1 Год назад

    Perfectly explain sir

  • @MukeshYadav-wb5uo
    @MukeshYadav-wb5uo Год назад

    Thanks for such a nice explanation

  • @souravbiswas6892
    @souravbiswas6892 Год назад

    XGBoost has the property of low bias and high variance, however it can be regularised and turned into low bias and low variance. Useful video indeed.

  • @Raja-tt4ll
    @Raja-tt4ll 2 года назад

    Perfect Lecture!!! Thanks Krish :)

  • @kumarssss5
    @kumarssss5 4 года назад +2

    Excellent teaching

  • @shirish3008
    @shirish3008 3 года назад

    This is the beautiful explanation👏

  • @RAHULPANCHALMUSIC
    @RAHULPANCHALMUSIC Год назад

    Sir superb explanation 🙏🙏

  • @supriyapatil5025
    @supriyapatil5025 2 года назад

    Thanks..very useful..perfectly understood sir.

  • @RinkiSingh-ph6oo
    @RinkiSingh-ph6oo 2 года назад

    well-done sir ....keep it up

  • @bhargavsai8181
    @bhargavsai8181 3 года назад

    Good video sir, its a great help for learners like me.

  • @Agrima_Art_World
    @Agrima_Art_World 4 года назад +48

    Underfitting : High Bias and Low Variance
    OverFitting : Low Bias and High Variance
    and Generalized Model : Low Bias & Low Variance.
    Bias : Error from Training Data
    Variance : Error from Testing Data
    @Krish Please confirm

    • @videoinfluencers3415
      @videoinfluencers3415 4 года назад +1

      I am confused ...
      It means that underfitted model has high accuracy on testing data?

    • @akashhajare223
      @akashhajare223 4 года назад +8

      Underfitting : High Bias and HIGH Variance

    • @kiran082
      @kiran082 4 года назад +4

      @@videoinfluencers3415 I mean under fitting model has low accuracy on Testing and Training Data and the difference between the Training accuracy and test accuracy is very less, that's why we get low variance and high biased in Under fitting models.

    • @hiteshchandra156
      @hiteshchandra156 4 года назад +1

      You are correct bro I checked on Wikipedia also..and in some different sources too.
      @Krish Please Confirm.

    • @sindhuorigins
      @sindhuorigins 4 года назад +2

      If it makes it any clear for other learners, here's my explanation...
      BIAS is the simplifying assumptions made by a model to make the target function (the underlying function that the ML model is trying to learn) easier to learn.
      VARIANCE refers to the changes to the estimate of the target function that occur if the dataset is changed when implementing the model.
      Considering the linear model in the example, it makes an assumption that the input and output are related linearly causing the target function to underfit and hence giving HIGH BIAS ERROR.
      But the same model when used with similar test data, will give quite similar results and hence giving LOW VARIANCE ERROR.
      I hope this clears the doubt.