Great video, but you should credit the book you took the method almost verbatim from : introduction to statistical learning, chapter 6, algorithm 6.2 - but again, really helpful, thank you !!! EDIT : IQmates actually credited several sources in an introduction video, my mistake.
Hi Dr. Lizard. In my introduction video, I do credit the sources I use. Most structure came from ISLR and I talk of sources such as Towards Data Science blog, Analytics Vidhya, ELSR and so on. Thank you for the feedback though. I will edit the caption.
Really cool explanation. Please keep producing more videos. If I had some investment money, I would have put it right on your channel just to make it more fancier, content is superb!
Sai Kishore Subramaniam thank you for the positive feedback. I am back to recording now. About to finish the machine learning course and I’ll start on deep learning 😊
Thanks for the wonderful explanation.. I have a doubt hope you get sometime to clarify. For example when we do RSS or R^2 to find the best model what should be the value of parameter(Beta) we need to substitute with? Normally we find the parameters value like that best fit our data and keep adjusting until our RSS is reduced. But in the above case where in each loop we check for many models (i.e., by adding additional argument) how is our Beta calculated?
First we create a model. This model gives estimate of parameter beta. Then we check the RSS(or R2). If RSS is low among other models that model is selected and hence its beta estimate is for that model is selected.
Wahyu Wisnu Wardana thank you for the feedback. I will try use fewer features for the explanations. My aim was to show how the method can be difficult computationally as we have more features 😊
The principle is there is no defined stopping criteria for this. You test all models and compare them to get the best one. For example, when you are doing k = 2, you create all models that have two variables and then you test them to get the best one of those models, then you do the same for k = 3 (all models with 3 variables). You keep doing that, getting the best model for each k. After you are done getting the best models for each k value, you compare those best models using AIC or something to get the best of the best. So there is no stopping criteria. You look at all possible models.
nice explanation, get to totally understand forward&backward method finally.
Most helpfull things in my study🙏🏿🙏🏿
very good explanation !
Great Explanation.
Thanks Vikas Bhardwaj. I’m glad it made sense. Stay tuned for more videos!
Great video, but you should credit the book you took the method almost verbatim from : introduction to statistical learning, chapter 6, algorithm 6.2 - but again, really helpful, thank you !!! EDIT : IQmates actually credited several sources in an introduction video, my mistake.
Hi Dr. Lizard. In my introduction video, I do credit the sources I use. Most structure came from ISLR and I talk of sources such as Towards Data Science blog, Analytics Vidhya, ELSR and so on. Thank you for the feedback though. I will edit the caption.
@@IQmates My apologies, I had not seen the introduction video, I will edit my comment. Thank you for your reply
Really cool explanation. Please keep producing more videos. If I had some investment money, I would have put it right on your channel just to make it more fancier, content is superb!
Sai Kishore Subramaniam thank you for the positive feedback. I am back to recording now. About to finish the machine learning course and I’ll start on deep learning 😊
I feel the same too. Great work
This is an amazing site! Thank you pal. If you have a patreon account you should post it. The other thing - are you sure its B1X2 - or is it B2X2?
Thanks for the wonderful explanation.. I have a doubt hope you get sometime to clarify. For example when we do RSS or R^2 to find the best model what should be the value of parameter(Beta) we need to substitute with? Normally we find the parameters value like that best fit our data and keep adjusting until our RSS is reduced. But in the above case where in each loop we check for many models (i.e., by adding additional argument) how is our Beta calculated?
First we create a model. This model gives estimate of parameter beta. Then we check the RSS(or R2). If RSS is low among other models that model is selected and hence its beta estimate is for that model is selected.
Good explanation! But, just wanna give a suggestion: why don't you provide less than 10 features only? Lets say 3 or for is enough. :)
Wahyu Wisnu Wardana thank you for the feedback. I will try use fewer features for the explanations. My aim was to show how the method can be difficult computationally as we have more features 😊
Its a good explain. But I still have confuse. When the process will stop?
The principle is there is no defined stopping criteria for this. You test all models and compare them to get the best one. For example, when you are doing k = 2, you create all models that have two variables and then you test them to get the best one of those models, then you do the same for k = 3 (all models with 3 variables). You keep doing that, getting the best model for each k. After you are done getting the best models for each k value, you compare those best models using AIC or something to get the best of the best. So there is no stopping criteria. You look at all possible models.
I think it stops when the addition of any variable doesn't augment the model or the additional variables are insignificant
hi i m farhan i cannot understand your main step when you 3rd model
Which part exactly muhammad farhan ? Can you put the time stamp of where you are questioning.
plzz expiain it