Это видео недоступно.
Сожалеем об этом.

Linear Regression and Multiple Regression

Поделиться
HTML-код
  • Опубликовано: 12 окт 2017
  • In this video, I will be talking about a parametric regression method called “Linear Regression” and it's extension for multiple features/ covariates, "Multiple Regression". You will gain an understanding of how to estimate coefficients using the least squares approach (scalar and matrix form) - fundamental for many other statistical learning methods.
    ⭐ Coursera Plus: $100 off until September 29th, 2022 for access to 7000+ courses: imp.i384100.net/Coursera-Plus
    MATH COURSES (7 day free trial)
    📕 Mathematics for Machine Learning: imp.i384100.net/MathML
    📕 Calculus: imp.i384100.net/Calculus
    📕 Statistics for Data Science: imp.i384100.net/AdvancedStati...
    📕 Bayesian Statistics: imp.i384100.net/BayesianStati...
    📕 Linear Algebra: imp.i384100.net/LinearAlgebra
    📕 Probability: imp.i384100.net/Probability
    OTHER RELATED COURSES (7 day free trial)
    📕 ⭐ Deep Learning Specialization: imp.i384100.net/Deep-Learning
    📕 Python for Everybody: imp.i384100.net/python
    📕 MLOps Course: imp.i384100.net/MLOps
    📕 Natural Language Processing (NLP): imp.i384100.net/NLP
    📕 Machine Learning in Production: imp.i384100.net/MLProduction
    📕 Data Science Specialization: imp.i384100.net/DataScience
    📕 Tensorflow: imp.i384100.net/Tensorflow
    INVESTING
    [1] Webull (You can get 3 free stocks setting up a webull account today): a.webull.com/8XVa1znjYxio6ESdff
    More on Matrix Calculus: atmos.washington.edu/~dennis/...

Комментарии • 189

  • @xavierfournat8264
    @xavierfournat8264 3 года назад +6

    Fantastic work. Usually all tutorial videos about linear regression or multiple regression are simply giving the formulas out of nowhere, without explaining the rational in the background. Thanks for taking the time for diving through the underlying maths :)

  • @arpitbharadwaj8799
    @arpitbharadwaj8799 4 года назад +23

    after multiplying and opening the brackets at 9:00 third term of the resultant should have transpose of B hat and not just B hat

  • @martinsahmed9107
    @martinsahmed9107 5 лет назад +2

    This exposition is timely. I have battled over the disappearance of Y transpose Y in the matrix approach of a least squares for months until I came across this video. This is awesome. I am speechless.

  • @sanathdas4071
    @sanathdas4071 4 года назад +3

    very well explained. I have been searching such video for many days. Now, the concept is crystal clear.

  • @bluejays440
    @bluejays440 2 года назад +3

    Excellent video, highly illuminating to finally see a comprehensive explanation of things that are too often left unexplained. I wish far more people, books, and videos explained statistics in similar detail.

  • @yuxiongzhu4249
    @yuxiongzhu4249 4 года назад +4

    Wow! This is the best video to quickly understand the derivation of linear regression formulas!

  • @sidharthramanan6780
    @sidharthramanan6780 6 лет назад +59

    This is a great video - I was looking for the math behind calculating the co-efficients in multiple linear regression and this explains it perfectly. Thank you!

    • @CodeEmporium
      @CodeEmporium  6 лет назад +2

      Thanks Sidharth! Glad it helped! Mind sharing the video to help others like you? :P

    • @sidharthramanan6780
      @sidharthramanan6780 6 лет назад +3

      Thank you for the video! And I'd love to share it with others :)
      Also, you just got a subscriber! Let's see you get to 1K soon !

    • @CodeEmporium
      @CodeEmporium  6 лет назад +2

      Thank you! Much Appreciated! I'm trying to upload more regularly than I have done in the past. There should be a lot more where that came from very soon.

    • @sidharthramanan6780
      @sidharthramanan6780 6 лет назад

      Yup! Any platform I can network with you on by the way? Quora for example?

    • @CodeEmporium
      @CodeEmporium  6 лет назад +1

      Sidharth Ramanan Quora is good. I'm under the name "Ajay Halthor".

  • @ledzeppelin1212
    @ledzeppelin1212 3 года назад +2

    Needed some refresher on a math class from grad school, and this really hit the spot. Thank you!

  • @Anonymous-ho1mt
    @Anonymous-ho1mt 4 года назад +1

    I have tried many ways to find a decent derivation for multiple regression, I found the key term is understanding matrix derivation rules which I was missing all those times. this is first time I got the clear understanding of the formula. Thanks a lot.

  • @anujlahoty8022
    @anujlahoty8022 5 лет назад +2

    This video just made my day.
    Absolutely loved it...

  • @FindMultiBagger
    @FindMultiBagger 5 лет назад +1

    Hats of for your efforts ! Really Fun way to learn algorithms, Please post more videos of other machine learning algo.

  • @kerolesmonsef4179
    @kerolesmonsef4179 4 года назад +8

    after week of searching . finally i found you . Thank you so much
    great explanation . keep going on

    • @CodeEmporium
      @CodeEmporium  4 года назад +1

      The search is over. Join me in turning this world into -- nah just kidding. Glad you Finally found me. Hope you stick around

  • @elishayu8002
    @elishayu8002 5 лет назад +1

    Super helpful and very clear! Thank you so so much!

  • @speakers159
    @speakers159 2 года назад +8

    Pretty amazing, especially since nobody really covers the mathematics behind ML, really appreciate the math based content.

    • @CodeEmporium
      @CodeEmporium  2 года назад +1

      Yesss! Math is underappreciated

    • @joelwillis2043
      @joelwillis2043 6 месяцев назад

      this is a very well made video but this is always covered in statistics

  • @MrStudent1978
    @MrStudent1978 4 года назад +1

    Very nice explanation! Very clear! I was looking for exactly the same.

  • @CK-vy2qv
    @CK-vy2qv 5 лет назад +1

    Excellent!
    Very nice see the scalar and matrix approach :)

  • @anandachetanelikapati6388
    @anandachetanelikapati6388 5 лет назад +2

    Excellent explanation with precise terminology!

  • @satyendrayadav3123
    @satyendrayadav3123 4 года назад +1

    Hands down my dawg❤️❤️ Very well explained

  • @ahmedmarzouki3991
    @ahmedmarzouki3991 5 лет назад +1

    thank you very much for this amazing video, it was really helpful
    do you have any other videos about : polynomial regression and non linear regression ?

  • @trackmyactivity
    @trackmyactivity 6 лет назад +8

    Amazing, thanks to the map you just drew I feel confident to learn the deeper concepts!

  • @ashlee8140
    @ashlee8140 5 лет назад +2

    This is a great video and explained things so clearly! Thanks!

  • @kuntsbro4856
    @kuntsbro4856 6 лет назад

    Thanks a lot. This is the most comprehensive regression video on RUclips.

    • @CodeEmporium
      @CodeEmporium  6 лет назад +1

      Kunt's Bro Thanks! Regression is an important topic, thought I'd take time explaining it

  • @capnic
    @capnic 4 года назад +1

    Wonderful video! very useful and clear!

  • @mike_o7874
    @mike_o7874 4 года назад +1

    Great video, exactly what i was searching for,
    how did they get that matrix equation was exactly what i needed!
    thanks a lot man!

  • @lucaslopes9907
    @lucaslopes9907 5 лет назад +1

    Thank you! It helped me a lot.

  • @tomarkhelpalma138
    @tomarkhelpalma138 4 года назад

    Way too cool!!! I am enjoying this video!

  • @andrewzacharakis8583
    @andrewzacharakis8583 Год назад

    great job man !

  • @souravdey1227
    @souravdey1227 3 года назад

    Extremely clear. Bang on!

  • @MOTIVAO
    @MOTIVAO Год назад

    Amazing work

  • @donaldngwira
    @donaldngwira Год назад

    One of the best explanations on this topic. And the presentation is superb

  • @nilankalord5974
    @nilankalord5974 5 лет назад

    Great video, thank you!

  • @hyphenpointhyphen
    @hyphenpointhyphen 8 месяцев назад

    I am binging the concepts and might forget to like - great channel.

  • @krishnachaitanyakr1237
    @krishnachaitanyakr1237 4 года назад

    Very well explained. Thank you.

  • @snehashishpaul2740
    @snehashishpaul2740 6 лет назад

    Very nicely explained

  • @kastenkarsten
    @kastenkarsten 2 года назад

    Incredible video for the derivation!

  • @carolyneatieno1927
    @carolyneatieno1927 4 года назад

    This was really helpful. I'm taking a unit on Data mining with no statistics background. Thank for sharing your knowledge 👊

  • @user-gn7op1nq3d
    @user-gn7op1nq3d 6 лет назад +3

    Thanks! You just saved my life!

    • @CodeEmporium
      @CodeEmporium  6 лет назад

      Raquel Morales Anytime. Saving lives is what I do.

  • @derrickmwanja2020
    @derrickmwanja2020 2 года назад

    Thank you very much it was very helpful

  • @DiegoAToala
    @DiegoAToala 4 года назад

    Great math , thank you

  • @NEOBRR
    @NEOBRR 2 года назад

    Thanks, this is an amazing video. It was very helpful.

  • @JAmes-BoNDOO7
    @JAmes-BoNDOO7 4 года назад

    Finally a video which makes perfect sense. Thanks a lot bro.

  • @gcumauma3319
    @gcumauma3319 2 года назад

    Excellently explained. Very lucid

    • @CodeEmporium
      @CodeEmporium  2 года назад

      Glad this is useful. Thank you :)

  • @yunqianggan2906
    @yunqianggan2906 6 лет назад

    Thank you so much, this video helps a lot :)

    • @CodeEmporium
      @CodeEmporium  6 лет назад +1

      Yunqiang Gan Thanks! Glad you liked it!

  • @tmrmbx5496
    @tmrmbx5496 4 года назад

    Good job, thanks

  • @ashokpalivela311
    @ashokpalivela311 4 года назад

    Very well explained!! Tq❤

  • @benisbuff
    @benisbuff 4 года назад

    Great explanation

  • @juanitogabriel
    @juanitogabriel 4 года назад +1

    Nice video. What software do you use for writing that math expressions? I mean, is it editor equations from ms word? Thank you.

  • @rahuldeora5815
    @rahuldeora5815 6 лет назад

    Hey great video. Can you suggest some mathematical test that explains ML with some debt as in this video?

  • @tsehayenegash8394
    @tsehayenegash8394 Год назад

    really nice explanation you have deep knowledge. hoa can we minimize the error term?

  • @amineelfarssi3902
    @amineelfarssi3902 5 лет назад

    Very clear explanation … better than doing it by considering the projection on the model space and using the projection formula (t(X)X)^-1t(X)Y

  • @alphatakes
    @alphatakes Год назад

    this is the best video on multiple linear regression

  • @christophereicher3928
    @christophereicher3928 5 лет назад

    Thank you!

  • @MrKaryerist
    @MrKaryerist 6 лет назад +4

    Great video! Probably the best in explanation of math behind linear regression. Is there a way to do multiple non-linear regression?

  • @k_anu7
    @k_anu7 4 года назад

    1 question.
    The method that you described above is of normal equation as of andrew ng machine learning course. The other way to find coeff. are gradient descent, BFGS, L-BFGS etc.
    Correct me if I am wrong.

  • @leochang3185
    @leochang3185 9 дней назад

    At 11:10, the quadratic form of matrix differentiation should be x^T(A^T + A). Under the condition of A being symmetric could the derivative be 2 x^T A (as being used in the last term of d(RSS)/dx).

  • @mustafizurrahman5699
    @mustafizurrahman5699 2 года назад

    Splendid and now words are sufficiently enough for such lucid explanation

  • @jean-michelgonet9483
    @jean-michelgonet9483 3 года назад +7

    In minute 10:27: X is mx1 and A in mxn. The 3rd differentiation rule is about y = X*A. But, given the sizes of the matrices, how can you multiply X*A?

    • @noelcastillo3369
      @noelcastillo3369 Год назад

      I have the same question. Were you able to clear it up?

    • @xiaodafei4738
      @xiaodafei4738 4 месяца назад

      here the differentiation rule should be: let scalar y=x^T A, then dy/dx = A^T
      It's nice that the video shows some matrix differentiation rules, but I recommend the more serious propositions in: atmos.washington.edu/~dennis/MatrixCalculus.pdf

  • @amogh2101
    @amogh2101 3 года назад +2

    Is there something wrong at @8:58? Shouldn't B(hat) be B(hat)(Transpose)?

  • @MrDiggerLP
    @MrDiggerLP 5 лет назад

    My friend, you Saved my Bachelorpresentation.

  • @runethorsen6230
    @runethorsen6230 Год назад

    Is it possible that there is a little error at 12.54 min. 3'rd term of RSS: beta should be transposed?

  • @naveengabriel9368
    @naveengabriel9368 5 лет назад

    I did not understand the part where it explains "for n samples number of operations.." Can anyone explain that to me.

  • @deeptigupta518
    @deeptigupta518 4 года назад

    How did you get y=xA as A transpose ? As both x A doesnt have the dimensions to get multiplied?

  • @Wisam_Saleem
    @Wisam_Saleem 5 лет назад +1

    GREAT!

  • @PetaZire
    @PetaZire 3 года назад

    thank you, you also saved my life :)

  • @rajeshsoma143
    @rajeshsoma143 Год назад

    9:40 How does the third formula work?
    Here, the dimensions of xA do not satisfy the condition for matrix multiplication

  • @akzork
    @akzork 2 года назад

    Can you upload a pdf of these formulae you shows in this video?

  • @ellenma3523
    @ellenma3523 3 года назад +2

    beginning from 8:54 the RSS should have the third term as -(β_hat)^T X^T y instead of -(β_hat) X^T y, the transpose sign is missing here

    • @nguyenkuro3827
      @nguyenkuro3827 3 года назад +1

      You're Right. And I think it should be:" y=x^T.A => dy/dx = A" .

  • @baa3124
    @baa3124 3 года назад

    thank you so much from korea

  • @zachc5261
    @zachc5261 11 месяцев назад

    Great video, thanks for your effort 😁
    I just have two questions:
    1. in the last RSS equation, why is T removed from beta_hat in the third term
    2. how is y = xA feasible given x has dimension (m x 1) and A has dim (n x m)
    Appreciate your help please. Thanks!

  • @sukanyaweangnon5717
    @sukanyaweangnon5717 4 года назад

    Excuse me , i would like to learn about bayesian linear regression ,Can you help me plsss
    Thank you for this video .It excellent .

  • @jugglingisgreat
    @jugglingisgreat 6 лет назад +1

    You explained this 1000000000000000000000000000x better than my professor. Thank you!

    • @CodeEmporium
      @CodeEmporium  6 лет назад

      Ryan Smith Thanks! So glad it was useful!

  • @sfundomabaso3200
    @sfundomabaso3200 4 года назад

    where's the rest.
    please make a video on normal linear models (I'm using a book called An Introduction to Generalized Linear Models by Dobson). the book is so confusing please help

  • @techchit6797
    @techchit6797 4 года назад

    Does the inverse( X*transpose(X)) always exists in the formula? Why?

  • @1UniverseGames
    @1UniverseGames 3 года назад

    How can we obtain intercept and slope of B0 and B1 after shifting line l to l'

  • @aigaurav5024
    @aigaurav5024 5 лет назад

    Thanku so much

  • @deeptigupta518
    @deeptigupta518 4 года назад

    Can you give us the reference for the matrix differentiation used here?

  • @eewitht3190
    @eewitht3190 3 года назад

    amaizing. thanks

  • @youcefyahiaoui1465
    @youcefyahiaoui1465 2 года назад

    In your logistic regression, I am not sure how you came up with the two exponents when you formed the two product of the product of p(x) and 1-p(x)

  • @dellazhang5785
    @dellazhang5785 2 года назад +1

    Super good

  • @vishu8770
    @vishu8770 Год назад

    Thanks a lottt

  • @surajjoshi3433
    @surajjoshi3433 3 года назад

    9:44 Actually it should be 2AX if A is a symmetric matrix, am i correct ??,help me anyone please

  • @troxy1935
    @troxy1935 4 года назад

    why for calculate b_1, the 1/n, becomes n?

  • @Dra60oN
    @Dra60oN 6 лет назад +3

    Hey, in 2nd example, you got y = xA, how can you even multiply those two when dimensions don't match? (m x 1) * ( n x m) , thus 1 != n
    Similar for 4th example where you got y = transpose(x) Ax ... I think A should be square matrix in this case (mxm).

    • @CodeEmporium
      @CodeEmporium  6 лет назад +2

      2nd example: y = Ax, not xA.
      4th example: You're right here. x^T A x has shape (1 x m) *(n * m)*(m*1). This is true if n = m i.e. A is a square matrix. Good catch! Should have mentioned that. In the derivation, we use it with X^T X -- which is square.

    • @Dra60oN
      @Dra60oN 6 лет назад +1

      Hey,
      sorry my typo,I was referring to the 3rd example, y = xA, not the 2nd one.
      And also are you sure that the last term B^T * X^T * X * B is the case of your 4th example. Because you can rewrite that expression as (X*B)^T * (X * B) and then it's a norm squared of matrix, and you say g(X) = X * B, and then you can apply derivative with respect to beta given by this formula: 2 * g(X)^T * d(X*B) / dX, which in this case would yield the same result, so after all you might be correct as well.
      All the best.

  • @rajinfootonchuriquen
    @rajinfootonchuriquen 2 года назад

    Something i don't understand is this: in simple linear regresión, you take the mean of square of error, but in múltiple regresión, what happend with taking the mean?
    X and y in the result fórmula have components with the mean?

  • @swapnilborse3150
    @swapnilborse3150 4 года назад +1

    Thanks....put more videos on regression analysis

    • @CodeEmporium
      @CodeEmporium  4 года назад +1

      Glad you enjoyed it! Will think of more Regression based videos in the future

  • @chaosido19
    @chaosido19 Год назад

    I come from Psychology and am following data science courses rn. The completely different way of approaching regression was a mystery to me, but this video helped me a lot. I do feel like I should practise stuff like this myself too, do you have any suggestions for places where to find exercises.

    • @CodeEmporium
      @CodeEmporium  Год назад

      Thanks for commenting and watching! Maybe a textbook might be good for establishing a foundation. You can check out the “Introduction to Statistical Learning”. Aside from that I have I lol playlist on linear regression, though I admit it hops around some concepts. It still might be worth your watch.

  • @raihanahmimi8233
    @raihanahmimi8233 4 года назад +2

    27.6K Subscriber on 13 July 2020... is that close enough from your prediction?

  • @xeeeshaunali
    @xeeeshaunali 3 года назад

    Which book you consulted??

  • @purvanyatyagi2494
    @purvanyatyagi2494 3 года назад

    Pls tell me that RSS is same as mean squared error

  • @BHuman2024
    @BHuman2024 Год назад

    8.54 min: last line 3rd term, I cannot match, could anybody clear me, please?

  • @ananthakomanduri1567
    @ananthakomanduri1567 5 лет назад

    THANKS

  • @P3R5I3dark
    @P3R5I3dark 6 лет назад +2

    At 8 56 last line, 3rd term shouldn t it be BTXTy instead BXTY? It also doesn t make sense bcs matrix sizes dont fit

    • @CodeEmporium
      @CodeEmporium  6 лет назад +3

      Nevermore You're right. Someone mentioned that in the comments as well. When you take the derivative wrt Beta, you get the same result. So that part of the video is incorrect, but the rest is fine.

    • @P3R5I3dark
      @P3R5I3dark 6 лет назад +1

      Yea the thing is -yTXB=BTXTy so you obtain deriv(RSS)=deriv(-2BTXTy+BTXTXB)=2XTXB-2XTY=0. Wich lead us to b=XTY*(XTX)^(-1) .
      Ty anyways for video, it helped to understand better this method. Tomorrow i have exam from this method's aplication in System's models.

    • @CodeEmporium
      @CodeEmporium  6 лет назад +1

      Death Xeris yup. You're right again. Still, glad the video was helpful. Good luck with your exam!

    • @anmoltariqbutt8987
      @anmoltariqbutt8987 5 лет назад +1

      @@P3R5I3dark -yTXB = BTXTy? How? If we use this equality then the whole term will be eliminated

  • @cghale3692
    @cghale3692 4 года назад +1

    when u removed the bracket...what happened to B transposition and X transposition while multiplying with Y? B transposition is not there just B is there ...the last line of simplification ?

    • @xofjo6239
      @xofjo6239 4 года назад

      why is that, I am very confused

  • @mohannadbarakat5885
    @mohannadbarakat5885 4 года назад +1

    why is sum(sqr(e)) = e^T * e

  • @peacego624
    @peacego624 2 года назад

    Thanks

  • @BlackJar72
    @BlackJar72 5 лет назад

    Good, but I think I need to review some general math and sit down to work it out -- solving its not hard, but its good to know why it works.

    • @CodeEmporium
      @CodeEmporium  5 лет назад

      For sure. One can always use builtin libraries to code it in a single line. But understanding why it works the way it does will help understand when to use it.

  • @shyamyadav6842
    @shyamyadav6842 4 года назад +1

    can anyone tell why (p-1)^2 operation instead of 2*(p-1) at 6:16

    • @tuscarorastreet1564
      @tuscarorastreet1564 4 года назад +1

      and shouldn't it be p instead of p-1 also?? since the B goes from B0 to Bp... which is p+1 number of beta's

  • @allensrampickal1997
    @allensrampickal1997 4 года назад

    How to find 'ε' ?

  • @Rigou27
    @Rigou27 3 года назад +1

    There's a mistake at minute 9:00, the third term of the expanded version of RSS is -(beta' x' y)

  • @marinpostma8958
    @marinpostma8958 4 года назад

    why is XTX invertible?