Vector Projections : Data Science Basics

Поделиться
HTML-код
  • Опубликовано: 19 сен 2019
  • What does it mean to project one vector onto another one?
    ---
    Like, Subscribe, and Hit that Bell to get all the latest videos from ritvikmath ~
    ---
    Check out my Medium:
    / ritvikmathematics

Комментарии • 80

  • @jirasakburanathawornsom1911
    @jirasakburanathawornsom1911 2 года назад +33

    Just want to say, you are really gifted teacher. I have seen so many videos from this channel and all of it hit the right spot, clear, deep enough, but concise. and with only one whiteboard, the animation is not even needed to understand your teaching. very impressive. Thank you so much. I would love to support and looking forward to all your contents in the future.

    • @larsaugustat7029
      @larsaugustat7029 21 день назад

      Agree partly, but would not use the word gifted. I guess he just put in the hours and learned to be good

  • @datasciencewithshreyas1806
    @datasciencewithshreyas1806 3 года назад +21

    Finally someone took this up !!!
    Thank you so so much Ritvik, you are Amazing.

  • @arongil
    @arongil 2 года назад +7

    Hey Ritvik, I just discovered your channel and wanted to say it is super helpful for me.
    Quick sidenote: at first I agreed with you that we need 9 data points to store the projected vectors, but now I think we only need 8. This is because the unit vector u needs only one number to store, namely the angle -- we already know its length is 1.
    Thank you very much for making these videos. I am learning so much!

  • @knp4356
    @knp4356 4 года назад +6

    Thank you!! Finally videos that connect LA to DS! Very good explanation and really like how you provide application/use-case before starting the lecture.

  • @omaraflak
    @omaraflak 3 года назад +4

    Side Note: The variable we were looking for was k. The answer you gave, by identification of the last result, is dot(x, v) / norm(v), which is totally correct. However, there is another way to get there pretty quickly.
    The dot product between two vectors **is** the product between the norm of the projected vector and the norm of the vector on which we project (the dot product is commutative). From there we can easily see that, if we only need the norm of the projected vector (k), all we have to do is divide the dot product by the norm of the vector on which we project. Hence: dot(x, v) / norm(v).

  • @spiritunchecked2281
    @spiritunchecked2281 4 года назад +2

    Awesome explanation. Thanks a lot!!

  • @jonathanfarache5218
    @jonathanfarache5218 4 года назад +2

    this is an excellent explanation - thanks!

  • @iOSGamingDynasties
    @iOSGamingDynasties 3 года назад +7

    Crystal clear, thank you for an amazing lecture.

  • @drumsFreeman95
    @drumsFreeman95 2 года назад

    Thank you for providing the actual derivation and intuition behind the equation.

  • @MZmakesgames
    @MZmakesgames 2 года назад +3

    For those who are checking videos for their ML exams. This is the best course for university students: concise and profound. Not so shallow as most top ML videos, but not so purely math-driven that it might be hard to understand.

  • @flloyd
    @flloyd 2 года назад

    so far my understanding
    is getting better. The best explaination of how and why.

  • @devika2250
    @devika2250 2 года назад

    Very good lecture ..People like u gift us more knowledge through these kind of videos..Thank you so much

  • @adrianfletcher8963
    @adrianfletcher8963 7 месяцев назад

    If you are watching this and you don't follow: I watched this when it first came out and I was in the same boat. I am coming back now to review some concepts and it all makes so much more sense!

  • @MrMoore0312
    @MrMoore0312 3 года назад +2

    I'm very new to the actual math behind these concepts, so first off want to give big thanks to this channel for your absolutely dedicated work.
    I want to leave this comment just in case anyone else runs into the, admittedly novice, problem i experienced and how i solved the problem. Sorry to all u big wigs who are going to find this blatantly obvious, but I've never taken a linear algebra or calculus class before; totally self study.
    From vector projection video we have the formulation: the projection of X onto vector V is called p and is equivalent to = (X dot V / magnitude(V) ) matrix multiply (V / magnitude(V) )
    = ((X • V)/||V||)*(V/||V||)
    When I saw the formulation here, as :
    U1transpose * Xi * U,
    p = U1T * Xi * U, I was totally lost and didn't see where that came from. Again I know this is an incredibly stupid question now but i first had to rewatch vector projection to make sure i didn't miss anything obvious, then break down the question into what I did and didn't understand, and finally arrived at how to make that jive.
    Solution:
    V/||V|| = U, directly from the vector projection video, and finally after some time staring at it
    U1 • Xi = U1T*Xi
    So we have
    ((Xi • U1)/||U1||)*(U1/||U1||) =
    U1T * Xi * U,
    Or to make your math come out exactly:
    (U1T * Xi) * U = (Xi • U1) * U
    And
    ((U1T * Xi) * U)/||U1|| = ((Xi • U1)/||U1||) * U
    Sorry, again I know this is obvious now but just wanted to leave it in case some other newbie stumbles across the same "issue".
    Thanks Ritvik!!!

    • @ritvikmath
      @ritvikmath  3 года назад +2

      First off, I really appreciate the kind words. And second, I am extremely grateful to you for writing all this out. People learn in different ways and neither I nor any other person/textbook/resource out there can account for all of them. I undoubtedly think your comment will help others.

    • @MrMoore0312
      @MrMoore0312 3 года назад +1

      @@ritvikmath to be honest, I should also make it clear that
      U1T *Xi * U works perfectly if U1 = U = unit vector with ||U||=1, which you stated explicitly in your video was necessary or the case in your example. Therefore, my formulations only work if you are too laze to turn U1 into a normalized unit vector U (for example i just used my vector V from projections video without realizing that it needed to be/ should have been initialized by normalization first z lol straight over my head). Turns out your answer needed no explanation from me, but I really am thankful that you provide this content; i love it!
      My favorite challenge so far has been proving that the "derivative" of XT * X is 2X, from the Lagrange video, which like you said I would be able to handle on my own without you needing to prove it in the video, AND I DID! You are amazing sir, thanks

  • @gracia9775
    @gracia9775 3 года назад

    God bless u mann !!! going through Data Science Basics series and it is wonderful !!

  • @pulkitmehta1795
    @pulkitmehta1795 7 месяцев назад

    Awesome ...this made vector projection very clear..I think the idea of deriving it in algebra makes it so much more clear

  • @alphar85
    @alphar85 3 года назад

    I do not know how to thank you man. things get easier when I watch your videos. I really want to see this channel picking up with more subscribers and views.
    I suggest you start charing for data science courses like 10 streamed courses or key concepts in DS. I wish you well mate

  • @user-su5gu4jm4e
    @user-su5gu4jm4e 10 месяцев назад

    It helped me make more sense of linear regression!!!

  • @user-mt4nz6tv1d
    @user-mt4nz6tv1d 4 года назад

    Thanks a lot for kind explanation. these lecture should be big help for my automatic appraisal business.

  • @tobydunbar1153
    @tobydunbar1153 3 года назад +2

    Beautiful and coherent explanation! Thank you! 💯

  • @yinstube
    @yinstube 4 года назад +2

    Love this!

  • @kristenchow7859
    @kristenchow7859 10 месяцев назад

    Thank you so much for your clear explanation

  • @redherring0077
    @redherring0077 2 года назад +4

    Hi. Awesome video. If you could maybe suggest some books/ papers for each individual concept like PCA or SVM that would be wonderful for all the novice researchers like me out there.

  • @siliconvalleyguy
    @siliconvalleyguy 4 года назад +4

    Light example is quite intuitive!!! I just landed to watch your Eigen vector video but ended up watching all your videos...

    • @sofiyavyshnya6723
      @sofiyavyshnya6723 4 года назад

      Same! Thank you so much for the great content!

    • @terrypan6168
      @terrypan6168 4 года назад

      same here, just ended up watching all. It's so magic

    • @gesuchter
      @gesuchter 3 года назад

      Same here hahah :D

  • @paulbrown5839
    @paulbrown5839 3 года назад

    This guy is one of the hero's of the internet

  • @AlexHop1
    @AlexHop1 2 года назад

    Thank you!

  • @hariharannair3281
    @hariharannair3281 4 года назад +3

    brother u r doing a great job. its a commentary on the system why y u have so less views.

  • @mustafaelmukadem9017
    @mustafaelmukadem9017 2 года назад

    amazing my hero professor

  • @MohamedMahmoud-jl1kh
    @MohamedMahmoud-jl1kh Год назад

    Big thanks

  • @goutamchatterjee1196
    @goutamchatterjee1196 3 года назад +2

    Can someone pls explain how projecting X on the unit vector helped us??

  • @adityaghosh8601
    @adityaghosh8601 4 года назад

    thank you please make video for partial derivatives as optimisation require partial derivatives

  • @SoumyajitGanguly_ALKZ
    @SoumyajitGanguly_ALKZ 3 года назад +4

    hey! really like all your awesome content. PCA brought me here... but please fix the camera focus :P

  • @MrChaluliss
    @MrChaluliss Год назад +2

    I can't help but wonder if there are some limitations to vector projection. Its working to compress data, but isn't some data lost in the process? Can we recover the X1 by X2 table from the vector projection method of storage if we wish to? That is to ask: how are the directions of the other vectors being stored within the vector projection?

    • @GoToMan
      @GoToMan Год назад +1

      It is the same as fitting a histogram of population onto a Gaussian normal distribution, the particularsof the data are lost but its features are retained, it allows us to bypass privacy restrictions too.

  • @n.m.c.5851
    @n.m.c.5851 2 года назад

    Amazing content im into watching the 20th video now i guess

  • @thomasstiglich3484
    @thomasstiglich3484 2 года назад

    Great video. Confused regarding the memory reduction in the first half of the video. When projecting a data point (vector) onto another vector (v), there is no way to get the original data point back. Correct? I believe that in order to do so, one would also need to know its projection onto the vector normal to v.

  • @paulntalo1425
    @paulntalo1425 3 года назад

    Thank you for the video. MSC data science student

  • @bosepukur
    @bosepukur 4 года назад

    great explanation

  • @ravipasumarthy8732
    @ravipasumarthy8732 6 месяцев назад

    Thanks Rithvik, your explanations are really good. If I may request you, how about writing a book possibly for math behind ML. Or if any books are there to suggest. Kindly let me know. I would like to keep refering to them regularly. On another note, people like me, who have spent a lot of their career in IT, some times, just one more lower level would be helpful. A little bit of more explanation or depth is required. For example in your covariance video, I found it difficult to understand how apple banana covariance is applied for vectors. Esp multi variant formula is difficult to understand. If you can emphasize on deriving univariate to multivariate in videos or blogs, that would super helpful. I wish you good luck in your journey and happy to even discuss more if you have some free time.

  • @tracyyi4478
    @tracyyi4478 4 года назад +2

    why is p+d=x? Suppose to be Pythagorean theorem , p square + d square = x square

  • @n1984ster
    @n1984ster 2 года назад

    at 11:16 does it mean that if the target vector is shorter or longer such that the projection is unable to fall at the right angle, then projection cannot happen?

  • @shrantunimoon3600
    @shrantunimoon3600 Год назад

    thank you

  • @felipe28g
    @felipe28g Год назад

    Just to clarify, could we say at the end there that the unit vector of v squared times x is the projection vector?

  • @cliffkwok
    @cliffkwok 4 года назад

    Want to learn more on auto regression, can u do more..?

    • @ritvikmath
      @ritvikmath  4 года назад

      more videos coming up soon!

  • @kingat8056
    @kingat8056 2 года назад +1

    i love u

  • @gerardsagliocca6292
    @gerardsagliocca6292 6 месяцев назад

    Why didn't you add a bar or some way to show p is a vector and d is a vector and u is a vector by using some common way of indicating it is a vector ? Just curious.

  • @srinivaspachika1996
    @srinivaspachika1996 4 года назад

    Your videos are the best videos among all the videos I have watched. Could you please make a video on the math behind SVM

    • @ritvikmath
      @ritvikmath  4 года назад

      good idea! I'll look into it

  • @osamaalsheikh7093
    @osamaalsheikh7093 3 года назад

    went from hero to zero real quick hahahha

  • @vishalvishwakarma7603
    @vishalvishwakarma7603 4 года назад

    projection of a vector on an another vector is what??scalar or vector??

  • @edvogel56
    @edvogel56 2 года назад

    Thank you so much I have been flunking my Coursera LA quiz on this and was getting frustrated. Vector projection just seemed so arbitrary. Now with understanding from your vid that it is part of larger schemes to compress data nd much more I am much better motivated to continue studying LA.
    Extra Special Easter Egg - I am watching season 3 of Silicon Valley. Right after your vid I watched the episode where the fired Nucleon programmers discover the missing link in their understanding of the "middle out" algorithm.

  • @MRIalyze
    @MRIalyze 18 дней назад

    My name is Rithvik!!!!

  • @Unremarkabler
    @Unremarkabler 2 года назад

    it seems not make sense to me, you do P * d = k*x*u - k*u*u = 0 -> u*u equals 1, so you get P*d=k*x - k, but later you got P = (x*u)u -> x * v/||v|| * u, the later formular also seems indicate P = x according with your P*d theory?

  • @mohammadbinmahbub9160
    @mohammadbinmahbub9160 3 года назад

    (13:19) If p=(x.u).u then it follows that p= x.(u.u) == x
    So is p=x? Doesn't look right!!

    • @curdyco
      @curdyco 7 месяцев назад

      Yeah he made a mistake in the end.

  • @ibrahemnasser2744
    @ibrahemnasser2744 Год назад

    9:33 p+d=x? that can't be true! maybe you mean pythagorean theorem?

  • @TheGibberingGoblin
    @TheGibberingGoblin 3 месяца назад

    bump

  • @jsridhar72
    @jsridhar72 4 года назад

    More confused...Cant get anything..I am not mathematically strong.

    • @ritvikmath
      @ritvikmath  4 года назад

      It's ok if you are not mathematically strong, we can all learn! Do you have any specific questions?

  • @tsunningwah3471
    @tsunningwah3471 6 месяцев назад

    nn

  • @ericklestrange6255
    @ericklestrange6255 4 года назад

    could you explain IN AN INTUITIVE WAY why am i using dot products in the perceptron? does it have to do with projections?
    def net_input(self, X):
    """Calculate net input"""
    return np.dot(X, self.w_[1:]) + self.w_[0]
    mllog.github.io/2016/11/04/Python-ML-Chapter-2/#bellow-is-implemetation-of-the-perceptron-learning-algorithm-in-python

  • @gpulogic1260
    @gpulogic1260 4 года назад +2

    YOUR TITLE IS WRONG. Vector Projection. When you are doing Scalar Projection. OOPS. it means you are doing PARROT TEACHING - NOT UNDERSTANDING.
    Also start of video, you are failing to emphasize the angel of projection is always 90 degrees to the "onto vector", at the start of the lesson (later it is clear). As you assume the viewers knows this but, you are ambiguous in the first diagram with the little x each side. The focus is lost because you did not bridge the gap between "Scalar Projection" and "Vector Projection".