SVM Dual : Data Science Concepts

Поделиться
HTML-код
  • Опубликовано: 29 ноя 2024

Комментарии • 100

  • @robertcanberkozturk7725
    @robertcanberkozturk7725 Год назад +18

    The only man who actually explains the concepts on the internet. God bless your soul Ritvik

  • @AlexandraWu-t8g
    @AlexandraWu-t8g Месяц назад +1

    I know this is an old video but I just want to express my appreciation for doing this! I have suffered for an entire semester of my professor jumping into complicated math without ever explaining the reason behind this. I binged watched all your SVM video and is moving on to the Kernel one right now. This is the first time everything about SVM suddenly becomes clear and even interesting for me! Thank you so much!

  • @mental4508
    @mental4508 7 месяцев назад +5

    Man I swear you are one of the very few that actually understands what it takes to make these concepts understandable.

  • @jjabrahamzjjabrhamaz1568
    @jjabrahamzjjabrhamaz1568 3 года назад +47

    Ritvik you my man are a godsend. Thank you for sharing your extremely
    technical expertise on youtube for free. You teach this material better
    than any platform (university or online).

  • @teegnas
    @teegnas 3 года назад +30

    Important video for those who want to understand SVM properly ... thanks for uploading!

  • @SS-xh4wu
    @SS-xh4wu 2 года назад +5

    Thank you so much for your videos. As a stats major person, I still learned so much from your channel. In many cases I learned the math in school, but no one ever talked about the intuition behind. Really appreciate your work here

  • @SHREEANSHMOHANTY
    @SHREEANSHMOHANTY 11 месяцев назад +2

    I have watched 6 times now, still can't wrap my head around it. Enjoy your views!!!

  • @alaham2590
    @alaham2590 7 месяцев назад +1

    You just give the exact amount of details needed to grasp these concept ! very nice !! thx

  • @pranjalpatil9659
    @pranjalpatil9659 2 года назад +1

    althogh this dosen't include all the maths in depth but this is enough for most of us to understand SVM formulation very well

  • @atrus3823
    @atrus3823 Год назад

    Along with StatQuest, these are by far the best ML videos on the RUclips. Thank you!

    • @ritvikmath
      @ritvikmath  Год назад

      Wow, thanks!

    • @atrus3823
      @atrus3823 Год назад

      @@ritvikmath No problem! I took SVM's in school, and have read about them many times, and I have only ever seen 1 of 2 levels of explanation: 1) just stating the results without question, or 2) diving deeply into measure theory and other crazy math that though sounds interesting, but I don't really have time for. This is the first source I've found that explains the gist of how the kernel trick works without diving super deeply into the math.

  • @ThangNguyenVan-tm3qm
    @ThangNguyenVan-tm3qm Месяц назад

    This guy is so underrated! period!

  • @Trubripes
    @Trubripes 6 месяцев назад

    Amazing that SVM is derived from KKT conditions. Never noticed that until watching this.

  • @adam_long_unique_name
    @adam_long_unique_name Год назад +1

    Thanks for these great videos.
    I would suggest to reorder the playlists to single topics - for example SVM,
    and add in the description links to the playlist or all other videos on the topic

  • @mohamedgaal5340
    @mohamedgaal5340 Год назад

    Thank you so much for explaining why we need the dual formation. This video made me understand some of the concepts I jotted down from other videos without understanding.

  • @CodeEmporium
    @CodeEmporium 3 года назад +4

    I love the way you throw the marker and hope to catch it. And if it's too far, you don't snap and point at the camera. Ah. And Nice videos. It comforts my soul

    • @ritvikmath
      @ritvikmath  3 года назад +1

      hahaha, thanks for noticing something I wasn't even aware of :)

  • @yitang7577
    @yitang7577 3 года назад

    Thanks for the wonderful video!!! It's really benefit for me who don't know the Lang and Dual problem.

  • @user-xi5by4gr7k
    @user-xi5by4gr7k 3 года назад +4

    Fantastic video. I have been binging your content. Any chance you will make a series on stochastic calculus?

    • @ritvikmath
      @ritvikmath  3 года назад +2

      thanks for the suggestion!

  • @e555t66
    @e555t66 Год назад

    Really explained well. If you want to get the theoretical concepts one could try doing the MIT micromasters. It’s rigorous and demands 10 to 15 hours a week.

  • @amrdel2730
    @amrdel2730 8 месяцев назад

    Thanks for the explanation of crucial points in this topic , thank you for the effort

    • @ritvikmath
      @ritvikmath  8 месяцев назад

      Glad it was helpful!

  • @bryankelly8820
    @bryankelly8820 2 года назад +2

    Thank you for posting this informative video. You mention that alphas only need to be calculated for support vectors. That does simplify things considerably; however, how can one in practice determine which vectors are support vectors without doing the minimisation?

  • @kally3432
    @kally3432 3 года назад +2

    Great explanation, I enjoyed it a lot. Thanks

  • @annakotova3508
    @annakotova3508 3 года назад

    Very-very clear explanation! Thanks!

  • @marisa4942
    @marisa4942 2 года назад

    Thank you so much!! Very clear and intuitive explanation!!

  • @TheKingLikesArt
    @TheKingLikesArt 2 года назад

    You are a legend my man

  • @TheJuankpunk
    @TheJuankpunk 10 месяцев назад

    Hi @ritvkikmath, thank you for these videos. In what type of degree do you usually cover these subjects? I would like to enroll in one.

  • @bhailog_gaming_2000
    @bhailog_gaming_2000 3 года назад +2

    Please make more videos on SVM!!
    Thanks a lot

  • @sejmou
    @sejmou Год назад +1

    Thanks for another wonderful video!
    However, there's one thing I really want to understand: in the formulation of the hard margin SVM problem (the 'SVM Math' video) you indeed stated that we want to minimize ||w||, not 1/2 ||w||^2. where does this difference come from, and why are both approaches equivalent? can anybody shed some light on this?

    • @k_1_1_2_3_5
      @k_1_1_2_3_5 10 месяцев назад

      I am in the same boat.. really would like to understand that

  • @lumos309
    @lumos309 3 года назад +2

    Hello, regarding the efficiency of the two forms of the problem: What about the inner product gives the dual form a complexity of N^2 rather than NP? It seems that the inner product operation would have complexity O(P) since it is dependent on the number of terms in the input vectors x(i) and x(j); similarly W(T)x(i) in the primal form also has complexity O(P). So this O(P) term shows up in the same place in both forms, which would mean they are both dependent on P. Is there some other place in which only the primal form complexity scales with P? Or is complexity not even the right way to analyse this?

  • @abhishekchandrashukla3814
    @abhishekchandrashukla3814 2 года назад

    apostle of machine learning !!

  • @liq3395
    @liq3395 3 года назад +5

    Thanks for the great teaching!!! Just have 1 question that why it was Max{Σα_i - 1/2(Σ.....)} after substituting w, but not Max{Σα_i + 1/2(Σ.....)}? I tried for several times and still got "+" not "-", might I ask for illustration on this calculation?

    • @mateuszserocki4026
      @mateuszserocki4026 2 года назад +3

      Hi man, I stuck with the same problem but we did it wrong! Resolved it finaly, first element w^T*w is reduced to 0! The secound element - sum(alpha_i * y_i * w^T*x_i +... you have to extend that w^T here then you have double sum that is not reduced to zero and here you go with this minus. Hope it helps you

    • @adefajemisin
      @adefajemisin Год назад

      When you substitute, the first two terms are the same, except the first one in multiplied by 1/2. 0.5x - x = -0.5x.

  • @SwapravaNath
    @SwapravaNath 8 месяцев назад

    11:30 but while solving the optimization, we don't know what are the support vectors. so, we need to solve the convex program for all cross terms but we'll find \lambda to be zero for most of them.

  • @nimeesha0550
    @nimeesha0550 3 года назад

    Keep going buddy!! Amazing work. Really helpful :) Thank You.

  • @emptygirl296
    @emptygirl296 2 года назад

    You are a legend really

  • @lilianaaa98
    @lilianaaa98 6 месяцев назад

    The math this video covers is kind of sophisticated :(

  • @winstonong9593
    @winstonong9593 2 месяца назад

    How are the support vectors determined in the first place, such that only pairs of support vectors need to be considered for the final minimisation problem?

  • @alenscaria1516
    @alenscaria1516 11 дней назад

    Thanks mahn! U just saved me!!

  • @muhammadaliyu3076
    @muhammadaliyu3076 3 года назад +12

    I think you are one of the great teachers on youtube. But I think the only reason I understood this video is because I already have knowledge about Optimization, linear algebra and Multivariable Calculus. And also because I already understood dual SVM problem. To be honest, some of your videos are totally not for beginners. I think you should try to be proving the math behind the algorithms you are explaining from the first principal. Its better for beginners.

    • @ritvikmath
      @ritvikmath  3 года назад +15

      thanks for the feedback! It's important for me to strike a balance between making the videos accessible to everyone and covering complex topics.

    • @jjabrahamzjjabrhamaz1568
      @jjabrahamzjjabrhamaz1568 3 года назад +6

      @@ritvikmath I think CrimaCode might not be fair here, because if you even come to this video for SVM's you will have atleast some knowledge about the realm of linear algebra and multivariate calculus. Both are basic things taught at most highschool levels. I think ritvik is doing a great job and video is dope for beginners.

    • @kroth5810
      @kroth5810 3 года назад +2

      I found the video very helpful despite not being well versed in svm before watching it. People learn in different ways :)

    • @quant-prep2843
      @quant-prep2843 3 года назад

      @@kroth5810 okay dicc

    • @thirumurthym7980
      @thirumurthym7980 3 года назад

      @@ritvikmath agree. May be you can try to cover basics of those basics separately for the pure beginners. But all your videos are awesome.

  • @johnstephen8041
    @johnstephen8041 8 месяцев назад

    Bro - Please discuss on the VC dimension

  • @TheInevitable360
    @TheInevitable360 Год назад

    Thank you very much for this great video. I have a question: you said that for non-support vectors the alpha values would be zero due to the fact that they have no contribution to the final solution. But here we have considered the Hard Margin version of the SVM. how about the soft margin. because in the soft margin, all the points have a contribution to the solution and we can't ignore non-support vectors. are they still 0 or not?

  • @sagarpadhiyar3666
    @sagarpadhiyar3666 2 года назад +1

    Hello ritvik,
    superb video of dual SVM. nicely explained. but I have one doubt in final equation of dual form of SVM. Where is that xj term coming from?? actually, what is that term?? if xi is our training data then what is xj??

    • @SwapravaNath
      @SwapravaNath 8 месяцев назад

      a different training instance.

  • @EW-mb1ih
    @EW-mb1ih 3 года назад

    At 6:30 Why do we take the alpha that MAXIMIZE the solution to the inner minimization?

  • @AmithAdiraju1994
    @AmithAdiraju1994 2 года назад

    _/\_ for this.
    I do have on question though, regarding non-support vectors requiring alpha(i)'s to be zeros.
    Intuitively, that would mean only support vector data points ( which are very few ) would contribute to optimal weights of the model, wouldn't that be bad for model weights ? i.e., only few contributions ?
    I'm wondering if model would not generalize well with only few contributions from few data points ?

  • @kanishkgarg423
    @kanishkgarg423 2 года назад

    thanks, this was very helpful 😀😀

  • @LingkarPeduli
    @LingkarPeduli 2 года назад

    can you make video explain about twin support vector machine. thank's in advances

  • @johnspivack6520
    @johnspivack6520 11 месяцев назад

    Wonderful video thank you

  • @zhangeluo3947
    @zhangeluo3947 Год назад

    Hello, sir I have 2 doubts: 1. Once you found the optimal values for alpha, how to determine the optimal value for b(bias)?
    2. You said that for non-SVs, the alpha value is just 0, but how to determine those non-SVs in the real data set? I know those non-SVs are outside the bottom lines for -1 and 1, but in practice, for given alpha values for each data point, you can get the optimal values for w or b , then you can get the decision boundary given w and b based on that set of alpha values, so to use that decision boundary to evaluate the support vectors and non-SVs? Is that true?
    Thank you!

  • @nuamaaniqbal6373
    @nuamaaniqbal6373 2 года назад

    You are the boss!

  • @anthonyburn6505
    @anthonyburn6505 9 месяцев назад

    Excellent videos

    • @ritvikmath
      @ritvikmath  9 месяцев назад

      Glad you like them!

  • @GanadoFO
    @GanadoFO 2 года назад

    Once we get the point where we are trying to minimize the equation on the right of the screen, I don't understand how you actually do the minimization.
    Let's say we have 3 support vectors, so we will have a function of three variables, alpha_1, alpha_2, alpha_3. How do you minimize a multi-variable function? What does that even mean? I've only ever done minimization of one variable.

  • @arminhejazian5306
    @arminhejazian5306 Год назад

    great explanation

  • @alexz3346
    @alexz3346 2 года назад

    Cool jacket!

  • @nieeel9112
    @nieeel9112 3 года назад

    super clear! really helpful:)

  • @ziadtarek633
    @ziadtarek633 3 года назад

    you're brilliant

  • @xinking2644
    @xinking2644 2 года назад

    good video with clearly explain!!!!

  • @shaktijain8560
    @shaktijain8560 2 года назад

    Thanks for the video 😄😄

  • @tryfonmichalopoulos5656
    @tryfonmichalopoulos5656 2 года назад

    Using stationarity we can get rid of 'w' but how did we get rid of the 'b' that does not appear in the dual formulation of the problem on the top right?

    • @gautamchopra9939
      @gautamchopra9939 2 года назад

      I think this is due to the dL/db = Σα_i*y_i. When we plug these value we have b*Σα_i*y_i such that the term containing b becomes 0. We don't have an explicit statement that tells us that b = 0.

  • @hari8568
    @hari8568 3 года назад

    Hey when you code this out how exactly would he choose the alphas like what are the upper bounds of alpha?

  • @prysrek8858
    @prysrek8858 3 года назад

    You're great

  • @ayushtiwari6932
    @ayushtiwari6932 3 года назад

    The term |w| is squared only for mathematical convenience ?

    • @SwapravaNath
      @SwapravaNath 8 месяцев назад

      yes, and makes the objective function differentiable as well.

  • @PF-vn4qz
    @PF-vn4qz 3 года назад

    how did you get rid of b?

  • @prajwalsatannavar4576
    @prajwalsatannavar4576 Год назад

    Thank u...

  • @RahulDable
    @RahulDable 3 года назад

    thanks

  • @johnspivack6520
    @johnspivack6520 11 месяцев назад

  • @mwave3388
    @mwave3388 2 года назад

    Thank you. But from scratch, it is too difficult.

  • @ryanleyland5565
    @ryanleyland5565 Год назад

    Kumar?

  • @johnspivack6520
    @johnspivack6520 11 месяцев назад

    Wonderful video thank you

  • @johnspivack6520
    @johnspivack6520 11 месяцев назад

    Wonderful video thank you

  • @johnspivack6520
    @johnspivack6520 11 месяцев назад

    Wonderful video thank you

  • @johnspivack6520
    @johnspivack6520 11 месяцев назад

    Wonderful video thank you