The Kernel Trick in Support Vector Machine (SVM)

Поделиться
HTML-код
  • Опубликовано: 12 дек 2024

Комментарии • 152

  • @nicoleta-vw3ql
    @nicoleta-vw3ql Год назад +89

    I listened to my lecturer and I was convinced that not even she understood something of her lecture...You clarified a lot for me only in three minutes...

    • @technosapien330
      @technosapien330 10 месяцев назад +6

      More common than you'd like to think, I'm also convinced my lecturers don't understand it either.

  • @TheKrasTel
    @TheKrasTel 2 года назад +56

    Your vidoes are absolutely amazing. Please keep making these, eventually the serious view numbers will come!

  • @srinayan7839
    @srinayan7839 9 месяцев назад +3

    Please don't stop doing videos you are helping a lot of people

  • @jeffguo490
    @jeffguo490 2 года назад +29

    This video is perfectly clear. I learnt SVM in class while I was confused by the lecture, and it is much clearer now.👍

  • @think-tank6658
    @think-tank6658 Год назад +4

    I went through class video for 1 hour didnt understand a thing..thank god you thought me in 3 min ..you are a legend bro

  • @7embersVeryOwn
    @7embersVeryOwn 2 месяца назад +2

    Fantastic job at making a not so simple concept easily understandable, the video was perfect, nothing can be removed from it.

  • @Baldvamp
    @Baldvamp 8 месяцев назад +5

    Incredible video, no messing around with long introductions, not patronising and easy to follow. It should be used as a guide for other people making educational videos!

  • @marounsader318
    @marounsader318 4 месяца назад +2

    thank you i dont know why uni profs wont explain stuff this easy

  • @maibster
    @maibster 2 месяца назад +2

    Dude this video is so awesome. Teaching in perfection. Thank you for your service to humanit

  • @peaceandlove5855
    @peaceandlove5855 2 года назад +10

    seriously !!!! simplest explanation of Kernel in SVM ever seen , just wow
    thank you so so much bro for the hard work you are doing to make such great videos ;*

  • @PowerhouseCell
    @PowerhouseCell 2 года назад +25

    Wow! I can't believe I didn't find this channel until now- your videos are amazing! As a creator myself, I understand how much work must go into this, so HUGE props!! Liked and subscribed 💛

  • @satyamverma4726
    @satyamverma4726 2 года назад +2

    Please keep making these contents. They are so intuitive. I don't understand why such channel don't grow on the other hand shitty contents are growing exponentially.

  • @LH-et7of
    @LH-et7of 2 года назад +1

    The best and most concise tutorial on Kernel tricks and SVM.

  • @sunritpal9596
    @sunritpal9596 2 года назад +12

    Absolutely incredible explanation 👏

  • @RAyLV17
    @RAyLV17 6 месяцев назад +4

    Man, I just checked that you haven't uploaded any new videos since 2 years!
    Hope you're doing well and come back with these amazing videos

  • @tomashorych394
    @tomashorych394 2 года назад +6

    Thanks a lot! I struggled to understand the difference between conventional dimension lifting and the "trick". Now it's crystal clear! Great explanation.

  • @gunduqrkdxhdrkd94
    @gunduqrkdxhdrkd94 Год назад

    I am korean so i am not good at english, but your teaching is very clear and easy to understand. Thank you teacher!

  • @angelo6082
    @angelo6082 6 месяцев назад +1

    You saved me for my Data mining exam tomorrow
    🙏

  • @vil9386
    @vil9386 10 месяцев назад

    understood fully. thank you. giving a code sample is like a bonus. Awesome explanation.

  • @JI77469
    @JI77469 Год назад +6

    Fantastic video, but I think you should mention at the end why the "kernel trick" isn't practical with lots of data (i.e. why deep learning is used much more than the "kernel trick" in this age of big data): given a kernel, you need to store the entire matrix of data inputs into the kernel. There are ways to mitigate this a bit (for example Random Fourier Features and Nystrom method) but still this is a huge issue that no one has seemed to figure out how to fix.
    On the other hand, if you have a small amount of complicated data then the kernel trick is very useful! For example a medical researcher might only have access to the horrifically complicated genomic data of patients at their hospital.

  • @judgelaloicmoi
    @judgelaloicmoi 2 года назад +3

    so amazingly simple and clear explanation, thank you so much !

  • @pakkateluguikkada
    @pakkateluguikkada Месяц назад

    You are amazing bro, come back and make more stuff like this

  • @amardeepkumar1267
    @amardeepkumar1267 2 года назад

    Great Visuals and explanation. Got it in One go. Thanks

  • @archer9056
    @archer9056 8 месяцев назад

    Amazing videos by explaining different concepts in simple words.
    Please have long vides as well..

  • @lucaspecht9105
    @lucaspecht9105 2 года назад +2

    excellent video, short but informative!

  • @AshishOmGourav
    @AshishOmGourav 2 года назад

    Wow...this is for free.
    Amazing visuals!

  • @mukul-kr
    @mukul-kr 2 года назад

    just subscribed after watching this video only. Hoping to find more good content as these in your channel

  • @5MassaM5
    @5MassaM5 2 года назад +1

    Simply AMAZING! Thanks a lot!!!

  • @ishsharma6529
    @ishsharma6529 Месяц назад

    the simplest lecture on this to exist

  • @jonahturner2969
    @jonahturner2969 2 года назад

    Wow, I am sharing this everywhere bro. Fantastic videos, we will grow together !!

  • @Sameer-jv5vx
    @Sameer-jv5vx 2 года назад

    thankyou so much you have nailed it it is crystal clear about kernel after watching your video thanks again

  • @CrispySmiths
    @CrispySmiths Год назад +4

    What software do you use to make these videos? and why have you stopped making videos!? and why did you start?

  • @negarmahmoudi-wt5bg
    @negarmahmoudi-wt5bg 7 месяцев назад

    Thank you for this clear explanation.

  • @cryptovolts
    @cryptovolts 3 месяца назад

    Holy sh*, my lecturer recommend us a reading but I was lost in the math terms and formulas, and I didn't even understood what was the purpose of kernel, incredible what you did in just three mins, thank you !

  • @0xabaki
    @0xabaki 10 месяцев назад

    super smooth explanation. Thanks!

  • @dialgos7574
    @dialgos7574 Год назад +3

    At 2:11 the shown Kernel and Transformation function do not match. The transformation function is missing the element ,1 as its last component and needs a scaling factor of sqrt(2) innfront of first 3 elements

  • @Bibhu_Nanda
    @Bibhu_Nanda 2 года назад +2

    Keep doing the Good work 👏

  • @shubhamgattani5357
    @shubhamgattani5357 10 дней назад

    Boy am I feeling lucky that I watched this video

  • @HeduAI
    @HeduAI 2 года назад

    Such an amazing explanation!

  • @kingnetwork-8519
    @kingnetwork-8519 24 дня назад

    i swear that's better than 2h lecture

  • @real7876
    @real7876 2 года назад

    to the point, concise, easy to understand, and even with code sample
    thanks!

  • @marishakrishna
    @marishakrishna Год назад

    Great.....Great presentation...This is what I mean when I say use Visuals graphics to explain a concepts

  • @imalive404
    @imalive404 Год назад

    i have never understood kernel trick in SVM better.

  • @0-sumequilibrium421
    @0-sumequilibrium421 2 года назад

    Very good video! Nice visualization! :)

  • @abdulhameedmalik4299
    @abdulhameedmalik4299 5 месяцев назад

    in given time stamp it is a good explanation

  • @jamesr141
    @jamesr141 2 года назад

    Gonna try the Kernel Trick!

  • @user-wr4yl7tx3w
    @user-wr4yl7tx3w Год назад +1

    why did you have "1 + " term in your polynomial kernel?

  • @batikanboraormanci
    @batikanboraormanci Год назад

    Wow, very nicely explained

  • @hantiop
    @hantiop 6 месяцев назад

    Quick question: How do we choose the gamma parameter in the RBF kernel at 3:00? By, say, cross validation?

  • @chzpan
    @chzpan Месяц назад

    The feature map f(x) shown at 2:11 is incorrect, please double check.

  • @abdulelahaljeffery6234
    @abdulelahaljeffery6234 2 года назад

    Hmmmm!
    Grant Sanderson is getting a serious contender right here

  • @footballistaedit25
    @footballistaedit25 2 года назад +2

    Thanks, Sir. Your explanation is incredibly amazing

    • @VisuallyExplained
      @VisuallyExplained  2 года назад

      So nice of you! You are most welcome!

    • @footballistaedit25
      @footballistaedit25 2 года назад

      @Visually Explained It helps so much. I'm waiting for the next video

  • @letsplay0711
    @letsplay0711 2 года назад

    Amazing Man.. 😍
    Just Awesome...

  • @TalGalili
    @TalGalili 2 года назад

    I wish I could like this video twice.

  • @photogyulai
    @photogyulai 9 месяцев назад

    Bravo! Great lecture! How the hell do you do this interactive function animations?

  • @chiragvaghela07
    @chiragvaghela07 4 дня назад

    Damn, liked and subscribed! :) Thanks!

  • @胡天翊
    @胡天翊 Год назад

    incredible explanation!

  • @lucasw8032
    @lucasw8032 11 месяцев назад

    Thank you so much, excellent content.

  • @תומרנדיבואחרים
    @תומרנדיבואחרים 2 года назад

    Wow! Great video, thanks :)

  • @gama3181
    @gama3181 Год назад

    i love it ! thanks for the explanation

  • @michele23
    @michele23 8 месяцев назад

    Can you make a video on unique game conjecture / unique label cover ? That would be very helpful

  • @harithm8172
    @harithm8172 Год назад

    Amazing explanation

  • @lip6792
    @lip6792 8 месяцев назад

    Super good explanation

  • @cleverclover7
    @cleverclover7 Год назад

    fantastic contribution

  • @vaishnavi4354
    @vaishnavi4354 Год назад

    that's so powerful to understand

  • @SosoKoko21
    @SosoKoko21 5 месяцев назад

    That was amazing, thanks 😊

  • @Kirill-xp9jq
    @Kirill-xp9jq 13 дней назад

    Are you computing the kernel value for each pair of points?

  • @enicay7562
    @enicay7562 3 месяца назад

    Thank you sir
    Loved it

  • @tlchannel2359
    @tlchannel2359 Год назад

    i hope u come back , i really like this content , pls

  • @sophiaxiao5071
    @sophiaxiao5071 Год назад

    I really like this video. Thanks!~

  • @Infinium
    @Infinium 2 года назад +1

    Really great video, thanks for sharing! Out of interest, what do you use to produce your videos?

  • @man49231
    @man49231 Год назад

    awsm video please make a video explaining K-Nearest Neighbors Algorithm also

  • @raven5165
    @raven5165 2 года назад +1

    Amazing video! How do you animate your videos?

  • @GamerAvi33
    @GamerAvi33 Год назад

    Thank you!! Nice video :)

  • @rokibulhasan2005
    @rokibulhasan2005 Год назад

    Great Video Sir

  • @ishahzaibkhan
    @ishahzaibkhan Месяц назад

    kindly continue your content

  • @julienblanchon6082
    @julienblanchon6082 2 года назад +2

    What did you use for visualization ?

  • @sisialg1
    @sisialg1 2 года назад

    you are amazing, you saved me !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!

  • @Mutual_Information
    @Mutual_Information 2 года назад

    Holy shit these are good videos!

  • @fizipcfx
    @fizipcfx 2 года назад +1

    did you use the Unity to produce this video?

  • @rohanjamakhandi5461
    @rohanjamakhandi5461 Год назад

    How to I use this when my decision boundary needs to be spiral?

  • @fortherest8490
    @fortherest8490 Год назад

    Im confused, how to display the hyperplane with polynomial kernel? Help me please?

  • @JulianHarris
    @JulianHarris Год назад

    Thanks! How do you know which gamma to use?

  • @chaddoomslayer4721
    @chaddoomslayer4721 2 года назад

    What kind of software did you use to create such beautiful illustrations?

  • @MamourBa-dc3fv
    @MamourBa-dc3fv 8 месяцев назад

    Fantastic video

  • @rouanemouaad
    @rouanemouaad Год назад

    Just fantastic explanation, i was wondering how much time takes to make such a high quality video, and what software he is using to do it. ? ? anyone knows ?

  • @saxonharries9033
    @saxonharries9033 2 года назад

    Fantastic video, thankyou

  • @ingrid1156
    @ingrid1156 5 месяцев назад

    very informative video!!

  • @fortunebezetshali7468
    @fortunebezetshali7468 2 года назад

    really a superbe video ! Thank you.

  • @rishisingh6111
    @rishisingh6111 2 года назад

    Awesome! Thanks a ton!

  • @benaoudamamchaoui8384
    @benaoudamamchaoui8384 Год назад

    amazing !! subscribed

  • @cy-ti8ln
    @cy-ti8ln 7 месяцев назад

    As I see, kernel-based regression is a type of symbolic regression

  • @mjwschuringa5464
    @mjwschuringa5464 Год назад

    Why did you stop making these video's?

  • @leogacitua1926
    @leogacitua1926 Год назад

    Great video! How does this differ from kernal ridge regression?

    • @JI77469
      @JI77469 Год назад

      kernel ridge regression is just that: regression using the kernel trick. Namely, instead of a hyperplane of best fit, you do the kernel trick to implicitly nonlinearly map the data into a high dimensional (sometimes infinite dimensional) space. But like with SVM, we don't need this map, and just need the kernel matrix at all the data points to practically perform algorithms.
      "Ridge" just refers to adding an l2 penalty to avoid overfitting. "Lasso" refers to l1 penalty, and I think in practice people even use l1+l2 penalties.

  • @-________9732
    @-________9732 7 месяцев назад

    In witch programmulka can doing this

  • @bengkeldepan6638
    @bengkeldepan6638 Месяц назад

    Label array must be binery?

  • @donharris8846
    @donharris8846 2 года назад +1

    Why isn’t the use of kernels considered overfitting?

    • @MrNoremac99
      @MrNoremac99 2 года назад +1

      You can control the flexibility of the kernel to avoid overfitting. For example, in the case of the radial basis function kernel you could use a lower value of gamma to avoid overfitting, as is shown in the video.

  • @lancelotdsouza4705
    @lancelotdsouza4705 2 года назад +1

    Loved your video..can you Mentor me?

  • @frederickaltrock5018
    @frederickaltrock5018 2 года назад

    I really would have liked to know why you claim that we only need to computer inner products. Does it arise from the dual problem? If i remember correctly that problem features such scalar products. And why is that better?

    • @VisuallyExplained
      @VisuallyExplained  2 года назад +2

      If you write the dual problem (e.g., page 4 of www.robots.ox.ac.uk/~az/lectures/ml/lect3.pdf) you can see indeed that it only depends on inner products of the training points. There is, however, a more "intuitive" way to see why SVM only cares about the inner products without actually writing the dual problem. If I give you the matrix X = (xi^Txj) of pair-wise inner products of n points, you can recover the coordinates of the n points "up to rotation" by computing the matrix-square root of X for example. In other words, all sets of n points that have the inner product matrix X are obtained by rotating the n points whose coordinates are the columns of sqrtm(X). Now, if you want to separate the n points with a hyperplane, your problem doesn't fundamentally change if all the points are rotated in some arbitrary way (because you can just rotate the separating hyperplane in the same way). So the knowledge of the matrix of inner products X is sufficient for SVM to do its job.
      As to why that's helpful, let's say have 100 points, and each point has a million features (which can easily happen if you "lift" to the data). That's a 100 million numbers you need to store. However, the matrix of inner products will be 100x100 only, which is a huge saving!

    • @crush3dices
      @crush3dices 2 года назад

      @@VisuallyExplained this is my other Account. The first one is for work, didn't realize i used that one.
      Thanks for taking the time to answer :). Interesting, not quite trivial, that the matrix contains all relevant information.
      I have already read some more sources on it now. Since you seem to understand involved math, maybe u can help me with another Question. The Question is, why exactly we use the kernel trick instead of simply using a usual transformation into another vectorspace and then use usual linear svms. This seems like it would work so there has to be a motivation for the kernel trick. I already read that this has better performance. But even the book "hands on machine learning" only says that it "makes the whole process more efficient" which says practically nothing about the motivation. One thing one can easily notice is that since the dual problem optimizes only for the lagrange multipliers,, we have to calculate the kernel only once before training. This also seems to be the reason why the kernel trick only works for the dual problem. But i was wondering weither this is the whole Motivation or if there is some more magic that i missed here?

    • @VisuallyExplained
      @VisuallyExplained  2 года назад +2

      @@crush3dices There are basically two main reasons. The first that you already alluded to, is for performance reasons. It's more efficient to compute k(x,x') than the transformation f(x) if f is very high dimensional (or worse, infinite dimensional). The second reasons is practical: sometimes, it is easier to choose a kernel than a transformation f

    • @crush3dices
      @crush3dices 2 года назад

      @@VisuallyExplained alright thanks.

  • @ttiigogo5897
    @ttiigogo5897 2 года назад

    Still don't really understand, but I'm closer, thanks!