How to check assumptions of linear regression in Python | How to check linear regression assumptions

Поделиться
HTML-код
  • Опубликовано: 2 дек 2024

Комментарии • 40

  • @gowthamsr3545
    @gowthamsr3545 2 года назад +3

    There are so many videos explaining the assumptions of linear regression but no one was explaining how to do it............I was searching for this from last 3months,,, thank you sir.....
    Thank you 🙏

  • @nikhilgupta4859
    @nikhilgupta4859 2 года назад +2

    Heyy Aman, I am your subscriber from past 1.5 year and I feel honoured to tell you, after following you I finally got a job transition as a senior data scientist at an MNC 6 month back. Now I have understood the datascience project ecosystem in my company. You are one of the contributors for my success.
    Thanks a Ton!!!!!
    Also I would like to open my hands for helping learners. So learners you can tag me asking any doubts. I would be more than happy helping you.

    • @UnfoldDataScience
      @UnfoldDataScience  2 года назад +1

      Thanks Nikhil, your comments are precious.

    • @nikhilgupta4859
      @nikhilgupta4859 2 года назад

      @@UnfoldDataScience Thank you Aman!!

    • @UnfoldDataScience
      @UnfoldDataScience  2 года назад

      Join here for free to share your experience live with me
      www.lighthall.co/class/6fdae050-85a4-48c4-9ecc-8dd2fa6ff175

  • @spicytuna08
    @spicytuna08 Год назад

    among all uTubers, i would rank this guy #1 for insightfulness. what a gift of teaching!!!

  • @commonNigerian
    @commonNigerian 2 года назад +1

    You're the best bro! I understood this explanation of yours more than I understood anybody else's. I've also saved a copy of the notebook (from Google Drive) and imported it into my DataSpell IDE so I can easily refer to it whenever I want to check assumptions - I've heard that memorising syntax is not important as long as one understands the logic :). Much love from Nigeria. Subscribed and liked. Cheers!

    • @UnfoldDataScience
      @UnfoldDataScience  2 года назад

      Thanks, Goriola. Your words mean a lot to me. Keep learning.

  • @mayankmehta8480
    @mayankmehta8480 2 года назад +1

    Well explained. I was very useful. Pls continue uploading lectures.

  • @shekharkumar1902
    @shekharkumar1902 2 года назад

    The title of your Channel could be " unfolding the untold data science" . Aman ji you reach and teach what no one dare to teach or explain . Amazing job!

  • @naveenkrish4037
    @naveenkrish4037 2 года назад

    When do we need to check these assumptions? After we do the Train Test Split and prediction, or before the start of Train Test Split?

  • @asitnayak636
    @asitnayak636 2 года назад +1

    Clean !!!

  • @anmolanreja1227
    @anmolanreja1227 Год назад

    Can you please tell me if we have large data means we have more than 30 columns in our data so linear regression will be good for training that data or not?

  • @sixlife.official
    @sixlife.official 2 года назад

    Great video! I wanted to ask, in order to find whether there is linearity in the first 4 scatterplots, shouldn't one plot the line of best fit? Also, regardless of whether that's true or not, how would one plot the line of best fit?
    "a, b = np.polyfit(X,Y,1)
    plt.plot(X,a*X+b)" doesn't seem to do anything and obviously im doing something wrong I'm just not sure what

  • @developerboy8341
    @developerboy8341 2 года назад

    Very intuitive video.

  • @andrewhenderson593
    @andrewhenderson593 Год назад

    How can we make a Residuals vs. Leverage plot which displays the Cook's distance?

  • @kalam_indian
    @kalam_indian 2 года назад

    you are excellent always sir

    • @UnfoldDataScience
      @UnfoldDataScience  2 года назад +1

      Thanks for your positive feedback. Please share with others as well who could be benefited from such content.

  • @beautyisinmind2163
    @beautyisinmind2163 2 года назад

    Sir, does it mean relation between predictor variable and target variable isn't linear means it does non follow Normality condition(Non normally distributed) and vice versa?

    • @UnfoldDataScience
      @UnfoldDataScience  2 года назад

      No, it does not mean it. Neither way. It's quite possible that a variable is not normally distributed however has a linear relationship with target variable and vice versa.

  • @pradeept328
    @pradeept328 Год назад

    Actually what should be normal in linear regression? The training data(target variables) or the residuals only?

    • @UnfoldDataScience
      @UnfoldDataScience  Год назад

      Both ideally

    • @pradeept328
      @pradeept328 Год назад

      But after we create a OLS regression model only we get the residuals and residuals normal distribution checks right?
      But we are saying normality of residuals should be confirmed before making the regression model.
      How the both cases satisfies each other?

  • @ajaykushwaha4233
    @ajaykushwaha4233 2 года назад

    Hi Aman, if we have encoded any categorical variables to numerical variables by count or frequency, or by onehotencoding or by top categories by any mean so that columns also need to be converted to Gaussian distribution form ?

    • @UnfoldDataScience
      @UnfoldDataScience  2 года назад +2

      Good question. No, not needed because gaussian is primarily defined for continious only.

  • @sanyamsingh4907
    @sanyamsingh4907 2 года назад

    Hi Aman can you Please advise us on "How to keep the relations with different Data Science Managers long lasting on Linked In?

  • @karthebans2420
    @karthebans2420 2 года назад

    Can you please explain about errors and residuals, I am not able to get the concept clearly in websites

    • @UnfoldDataScience
      @UnfoldDataScience  2 года назад

      error is the difference between actual and predicted.

  • @chalmerilexus2072
    @chalmerilexus2072 2 года назад

    One question though not related to this video. At what point we do train - test split. BEFORE preprocessing like normalization, imputation etc or AFTER preprocessing?