Inverse Transform Sampling : Data Science Concepts

Поделиться
HTML-код
  • Опубликовано: 8 фев 2025
  • Let's take a look at how to transform one distribution into another in data science!
    Note: I should have included a lambda in front of the exponential PDF. I mistakenly forgot it. I appreciate the comments which helped me realize this mistake.
    ---
    Like, Subscribe, and Hit that Bell to get all the latest videos from ritvikmath ~
    ---
    Check out my Medium:
    / ritvikmathematics

Комментарии • 145

  • @gomolemartifex
    @gomolemartifex 5 лет назад +34

    This video just transformed my life

  • @shivamathghara2870
    @shivamathghara2870 5 лет назад +97

    pdf of exponential is (lambda)*e^(-lambda x)

  • @bhaskarroy8753
    @bhaskarroy8753 2 года назад +5

    Great video. It made the underlying concept crystal clear. Thanks a lot, Ritvik.

  • @thiagobarreto9056
    @thiagobarreto9056 3 года назад +6

    Just saw two 20 minutes long videos before this, none made me understand this at all. Then, I saw this 10 minutes long video of yours and it made this subject so much clearer than before. Amazing professor, congratulations!

  • @mjf1422
    @mjf1422 5 лет назад +24

    Thank you so much for doing these videos.

    • @fyaa23
      @fyaa23 5 лет назад

      I can't agree with you more.

  • @sepidet6970
    @sepidet6970 5 лет назад +23

    That was a great intuitive explanation of inverse Transform Sampling. It seems so easy to me after watching this video,. Thanks a lot.

    • @Arriyad1
      @Arriyad1 5 месяцев назад

      It only seems easy. Inverting the cdf is difficult. The exponential distribution is kind enough to let itself invert, but many other ones are mean.

  • @whyyflo
    @whyyflo 2 года назад +2

    omg you saved my stats degree, much thanks

  • @Bksemsem
    @Bksemsem Год назад +1

    I really want to thank you because your clear explanation helped me get an A in my statistical programming exam. You are a hero.

  • @shueibsharif9955
    @shueibsharif9955 2 года назад +2

    I can't thank you enough. You have been of help in many subjects from time series analysis to this. I would like to see EM algorithm, latent class models, and hidden Markov models in the future.

  • @EubenM
    @EubenM 4 года назад +5

    You solved a big curiosity I had. I learned about the power of MonteCarlo analysis and how easy it is to get a uniform distribution from Excel, but knew I would always need more specific distributions. So the question was how to get any distribution from a set of randomly generated numbers from the usual Excel Rand() generator. Thanks for the brilliant and easy demonstration! Congrats for your terrific work!

  • @elias043011
    @elias043011 Год назад +1

    You have brilliantly and simply explained a topic that I have been struggling with for a whole semester. Thank you so much! :)

  • @ellyakawecki422
    @ellyakawecki422 Месяц назад

    Excellent video, super clear! One thing on the maths side, when you do the final derivation step at around 7:28. I think you should do the following, since you have F_X(x) = T^-1(x) = u, you then apply the inverse of F_X to the left and right hand side to get x = F_X^-1(u), i.e. we take the CDF inverse of the uniform sample.

  • @therockbottom2539
    @therockbottom2539 Год назад

    Love how calm you are. I'm shitting myself when I have to explain topics like these to someone.

  • @phuongdinh3769
    @phuongdinh3769 Год назад

    Trying to wrap my head around this in class but to no avail. Thank you so much for your amazing explanation

  • @joaopedroxavier8474
    @joaopedroxavier8474 4 года назад +1

    Thanks for the video! I was struggling to understand the motivation behind it, but your explanation has made it much easier for me :)

    • @ritvikmath
      @ritvikmath  4 года назад

      Glad it helped!

    • @EubenM
      @EubenM 4 года назад

      João, veja meu comentário acima para um exemplo de aplicação.

  • @rmiliming
    @rmiliming 2 года назад

    Thanks a lot! your videos on DS and Stats is the best!

  • @lilmoesk899
    @lilmoesk899 5 лет назад +3

    Thanks for the video! I'm still struggling with this, but your explanation definitely helped!

  • @roayadiamond
    @roayadiamond 5 лет назад +2

    He is going to be a fabulous professor

    • @ritvikmath
      @ritvikmath  5 лет назад

      Haha I appreciate the kind words :)

  • @samersheichessa4331
    @samersheichessa4331 4 года назад +1

    Just fantastic ! keep it up man great videos and great explanation

  • @hahahaYL-h3x
    @hahahaYL-h3x 10 месяцев назад

    Brilliant teacher! I guess it is a sort of gift.

  • @MegaNightdude
    @MegaNightdude 4 года назад +2

    Brilliant!!!!

  • @Maikpoint11
    @Maikpoint11 4 года назад +1

    Super helpful, thank you very much!

  • @andyak93
    @andyak93 3 года назад

    nice! Thanks for the work. Like the way you explained concepts in a straightforward and smooth way. Please keep it up ! :)

  • @liamobrien8610
    @liamobrien8610 5 лет назад +8

    Great video! Your exponential density is missing it's normalizing constant, though. Since your CDF is correct, no harm , no foul, but it might confuse some people.

    • @algrant33
      @algrant33 4 года назад +2

      Yep, I'm looking for the lambda*e^ -(lambda*x).

  • @ramn9071
    @ramn9071 2 года назад

    Well explained .. thanks. One minor suggestion .. if there is a way you can make the video screen capture friendly or leave a screen capture slides to the video, that would be super helpful. Thanks for the clear presentation.

  • @dedecage7465
    @dedecage7465 2 года назад

    This was super pedagogical, thank you very much.

  • @bartomiejpotaman6973
    @bartomiejpotaman6973 2 месяца назад

    Captivating!

  • @deepanshu7714
    @deepanshu7714 Год назад

    u r best teacher ever

  • @emilioalfaro4365
    @emilioalfaro4365 Год назад

    very clear explanation, thanks for sharing!

  • @presbytes
    @presbytes 3 года назад

    Quite an amazing explanation. Well done!!

  • @aytekin8669
    @aytekin8669 3 года назад +1

    thanks for good explanation about Inverse Transform sampling !

  • @emre2889
    @emre2889 Месяц назад

    finally i understood, thank you

  • @malhajed
    @malhajed 3 года назад

    I love your explanation always produce the best please don’t stop

  • @fredericoamigo
    @fredericoamigo 2 года назад

    Excellent explanation! Keep ut the good work!

  • @markusnascimento210
    @markusnascimento210 Год назад

    Greatly explained! Thanks!

  • @caiyunwurslin2468
    @caiyunwurslin2468 3 года назад

    Thank you. Our instructor did not explain it and just gave the theorem. I was confused like I have three heads.

  • @jindai5850
    @jindai5850 4 года назад

    Yo Ritvik not sure if you still remember me we talked during orientation (I was the guy work with Tasty). We had a class last week about MCMC and I was confused about certain parts and RUclips directed me to this video lol. Great job man keep it up. Hope we can catch up when things get back to normal after the pandemic

  • @berkayyucel1538
    @berkayyucel1538 5 лет назад +1

    That was awesome. Thank you !!!!!

  • @mostafaalkady6556
    @mostafaalkady6556 10 месяцев назад

    Great explanation! Thanks.

    • @ritvikmath
      @ritvikmath  10 месяцев назад

      Glad you enjoyed it!

  • @aryang5511
    @aryang5511 2 года назад +1

    Great video, it really helps me out a lot. One thing I still dont really understand is why we might do this. As in, why would we use the inverse transformation method to find the exponential random variable instead of just using the exponential PDF directly if we have lamda?

  • @gavinresch1144
    @gavinresch1144 2 года назад

    Hey - great video! I think you might have forgotten the lambda in front of the exponential for the exponential PDF. If you calculate the CDF from what you have written you will get a 1/lambda factor.

    • @ritvikmath
      @ritvikmath  Год назад

      Yup you’re definitely right !

  • @Busterblade20
    @Busterblade20 4 года назад

    Thank you so much. You help me a lot with a homework I have.

  • @fionnmcglacken35
    @fionnmcglacken35 4 года назад

    Brilliant, thank you so much.

  • @ahmetkarakartal9563
    @ahmetkarakartal9563 2 года назад

    you saved my life

  • @adityasaini491
    @adityasaini491 4 года назад +1

    That subtle pen flip at 5:49.. Damnn

    • @EubenM
      @EubenM 4 года назад

      LOL

    • @adityasaini491
      @adityasaini491 4 года назад

      @@EubenM You replied :DD Great videos man! Your channel is awesome :DD

  • @sebastianmathalikunn
    @sebastianmathalikunn 2 года назад

    Hi Ritvik, great videos! would be interested to have a set of videos explaining variational bayes, ELBO etc. in order to perform bayesian optimisation on hyper-parameters

  • @katiedunn7369
    @katiedunn7369 3 года назад

    very helpful, thanks for this video!

  • @dwightsablan3571
    @dwightsablan3571 4 года назад +1

    Thank you, this helped a ton! :)

  • @realimaginary5328
    @realimaginary5328 3 года назад

    Excellent. !

  • @shubhamthakur3461
    @shubhamthakur3461 3 года назад

    Great Explaination! Thanks so much :)

  • @Juan_Genaro
    @Juan_Genaro 4 года назад

    Thank you for this, it has helped me a lot! :)

  • @rishikalodha1236
    @rishikalodha1236 7 месяцев назад

    Thank you for this

  • @trollingenstrae2207
    @trollingenstrae2207 4 года назад

    great explanation, thanks a lot!

  • @phalanxz11_
    @phalanxz11_ 5 лет назад +3

    Can you please do a video about Copulas? For example in a (credit) risk management context

  • @HadiAhmed7546
    @HadiAhmed7546 2 года назад

    Thanks a lot bro, so helpful

  • @praburocking2777
    @praburocking2777 2 года назад

    great explanation

  • @adishumely
    @adishumely 3 года назад

    great video! thanks!

  • @teojunwei2000
    @teojunwei2000 3 года назад +10

    hi, is there an error with the PDF function? f(x) = lambda * exp^(-lambda)(x)? thank you for this video!

  • @anastasialeonhart
    @anastasialeonhart 5 лет назад +2

    Hi! I loved the video, but I've got a question. What are the cases when the CDF is not invertible? And what are the strategies then? Should we try to make the CDF invertable by interpolating it or should we use another random variate generation technique? Thank you in advance! Happy New Year.

    • @ritvikmath
      @ritvikmath  5 лет назад +3

      Happy new year! And great question, indeed this technique is good only if you can find the inverse of the CDF, so if that is not possible, interpolation is a great idea as long as the fit is "good enough"

    • @jonatangarcia9285
      @jonatangarcia9285 Год назад

      You can use the generalized inverse of the function. This is a function g such that g(y) is the infimum of the x such that F_X(x) >= y. Since F_X is a continuous function from the right this is always a minimum. So this function is such that F(g(y)) =y, it works like the inverse and the difference is that if there are other values with the same image you take the least of them and you can always do that. This is the same function to calculate quaintiles, so Q_{0.5} = g(0.5). Take in account that g(0) = -infinity and g(1) = infinity, to get the values right. More information here en.wikipedia.org/wiki/Probability_integral_transform

  • @emanelsheikh6344
    @emanelsheikh6344 2 года назад

    Thank you 🙏

  • @airesearch2024
    @airesearch2024 5 лет назад +2

    could you make more advance time series tutorial? really like your videos and i'm struggling in grad level time series course

    • @ritvikmath
      @ritvikmath  5 лет назад

      More time series vids coming up soon!

  • @musondakatongo5478
    @musondakatongo5478 4 года назад

    Well explained. Thanks a mil

  • @OscarBedford
    @OscarBedford Год назад

    What is the role of lambda? I've seen other videos that don't include it, so now I'm curious. Amazing explanation btw!

  • @TheiPodMac1234
    @TheiPodMac1234 4 года назад

    Very helpful thank you !

  • @ec-wc1sq
    @ec-wc1sq 4 года назад

    thanks, this is a great video!

  • @lm58142
    @lm58142 2 года назад

    Thanks for sharing. Just one small comment....pdf of the exponential is lambda*e^(-lambda*x).

  • @lancelofjohn6995
    @lancelofjohn6995 3 года назад

    Nice lecture!

  • @riaddjaid7428
    @riaddjaid7428 10 месяцев назад

    thank you so much sir, I would like to know which probability distributions commonly used that we use inverse method with.

  • @tianjoshua4079
    @tianjoshua4079 4 года назад +1

    Great video. Quick question: at the end of the video, you said we could swap 1- u for u. That means 1 - u = u, which translates into u = 1/2. Yet u is a random variable, it is not necessarily 1/2, right? What am I missing?

    • @ritvikmath
      @ritvikmath  4 года назад +5

      Good question! We are not swapping 1-u for u in an algebraic sense (in which case you would be absolutely correct). Rather, we note that u is a uniform random variable between 0 and 1. Therefore 1-u is also a uniform random variable between 0 and 1. Thus, it does not matter (in terms of probability) whether we use 1-u or u. And using just u makes the formula look a bit nicer.

    • @tianjoshua4079
      @tianjoshua4079 4 года назад

      @@ritvikmath Oh. I understand. RVs are not really variables. When it comes to RVs, what matters is not the specific value of the RV, yet it is the distribution of the RV that matters. Since u and 1-u are both RVs with the same distribution, they are interchangeable.

  • @learnphysics6455
    @learnphysics6455 3 года назад

    Gem level bhau

  • @nishitshukla4139
    @nishitshukla4139 4 года назад +3

    Lets say u = 0.25. Then 1 - u = 0.75, right? Could someone explain how 1- u = u in the uniform distribution?

  • @sheeta2726
    @sheeta2726 2 года назад

    Thank you!!!!!!!!!!!

  • @stipepavic843
    @stipepavic843 2 года назад

    this guy is epic!!!

  • @annabelseah920
    @annabelseah920 3 года назад

    perfect!

  • @maximegrossman2146
    @maximegrossman2146 4 года назад

    excellent

  • @yelnady
    @yelnady 4 года назад

    Thank you man

  • @kobi981
    @kobi981 9 месяцев назад

    Very nice video! thank you!
    The uniform should be (0,1] without 0 right? so the ln will be defined.

  • @zelousfoxtrot3390
    @zelousfoxtrot3390 4 года назад

    Thanks!

  • @tj9796
    @tj9796 3 года назад

    Great video. Could you do one on copulas, building on this one?

  • @grjesus9979
    @grjesus9979 3 года назад +1

    Then, why is important the uniform pdf?. I mean you could sample directly from one distribution to another just by putting the value returned from the CDF of the first pdf as input to the inverse CDF of pdf you want to arrive at. Am I wrong?

  • @geoffreyanderson4719
    @geoffreyanderson4719 3 года назад

    I have question about the math, on how to derive other inverse transformations especially for datasets that predict number of clicks on a web page for example. Some of them are tricky and might even need estimation by iterative numerical methods or ML, because the Poisson is simple to find the inverse function for. And then how do you put the inverse transform into an sklearn pipeline exactly? Here's why I ask this: Sometimes I am using a Generalized Linear Model which provides a convenient link function already built-in, but we are not always going to just use a linear model as we might need to use for example the large feature vectors that an NLM model is producing to describe some text. GLM is not necessarily the only tool to consider. Besides for random sampling, Transforms are also good for ML preprocessing and postprocessing pipelines to help your model learn easier. The log(Y) and e(Y) are the Poisson distributions transformations when your response Y is a count. Quasipoisson and Negative Binomial are good for count data when the mean and variance are not staying equal as the Poisson requires, but instead are showing some overdispersion or underdispersion. There's also zero inflation model which combines a logistic model and a Poisson model together in sort of an ensemble to help pre-predict the count = 0 case when 0 appears a lot more often than plain old Poisson can account for alone.

  • @rachidwatcher5860
    @rachidwatcher5860 5 лет назад +1

    Thx body u the best

  • @kerguule
    @kerguule Год назад

    I don't get it why the exponential distribution is called memoryless? Yes, I know that that lambda or hazard rate is constant but isn't that just the speed or rate of the probability (not the actual probability because the lambda can be more than 1). From the exponential PDF, you can clearly see that the chances in the early phase are bigger than in the later phases so why is it called memoryless? If I sampled time to failures, should I get more numbers early than later because of that decreasing curve?

  • @nicnicco
    @nicnicco 10 месяцев назад

    Are there any resources I can look at to understand why it's valid to assume that p(T(U)

  • @Reftsquabble
    @Reftsquabble 11 месяцев назад

    TLDR: The distribution of the CDF of _any_ PDF is uniform, so if you want to sample from a PDF that has an invertible CDF, you can sample from the uniform distribution and convert it to the desired distribution with the inverse of the CDF.

  • @BlueSkyGoldSun
    @BlueSkyGoldSun 3 года назад

    In data science can we transform weibull distribution into Gamma or poison distribution?

  • @Ivan_e9812
    @Ivan_e9812 4 года назад

    Not very proficient in statistics, but in sum, if I do the transformation and have the final function, given a number u that is randomly generated from a uniform distribution, I will get an equivalent randomly generated number that falls under an exponential distribution?
    great video, I will subscribe and continue to watch them!

  • @annali9577
    @annali9577 3 года назад

    this is super clear and I can go to bed very happy

  • @chonglizhao2699
    @chonglizhao2699 4 года назад

    If I understand correctly, the reason why uniform distribution is used because its output range from 0 to 1. Just out of curiosity, can we use beta distribution to replace uniform distribution?

  • @PavelSTL
    @PavelSTL 5 лет назад

    Was hoping to hear more about motivations for WHY i need to know this method for DS. "that's how computer gives you random samples from a distribution" is not enough to care about it. What about cases where maybe I don't have a pdf or it cannot be integrated or I get only proportionality of pdf (like in Bayesian model) so I can't just plug in the variable into the proportional pdf and get accurate samples..... maybe that's when I need to use this method.....

    • @ritvikmath
      @ritvikmath  5 лет назад

      I appreciate the feedback!

  • @庭瑞赵
    @庭瑞赵 5 лет назад

    Having a hard time understanding EWMA and GARCH model ,can you make some videos introducing them?thx

    • @ritvikmath
      @ritvikmath  5 лет назад +1

      GARCH is coming up soon!

  • @7ignatios
    @7ignatios 5 лет назад

    Can you do (or recommend) a video on Granger Causality?

    • @ritvikmath
      @ritvikmath  5 лет назад

      Thanks for the suggestion! I'll look into it

  • @konstantinkulagin
    @konstantinkulagin Год назад

    I probably missed this moment: why transformation to CDF actually gives you desired distribution?

  • @piyushsinha3344
    @piyushsinha3344 3 года назад

    in order to find the inverse of CDF, we just find the value of x..why? in other word, how come x is the inverse of CDF?

  • @scarlettwang2643
    @scarlettwang2643 4 года назад

    if the distribution we want is not the exponential distribution, are the steps are still the same?

  • @cozyASMRing
    @cozyASMRing 3 года назад

    So if I had some other distributions apart from exponential one, I just need to derive its inverse, and set the number of simulations I will like to do with a U that is unif from 0 to 1? I just need clarification in that part.

  • @unnikrishnanadoor
    @unnikrishnanadoor 5 лет назад

    I have a question if we graph the inverse function of that exponential function how it will looks like? whether it looks similar to graph of uniform distribution? otherwise how this can be equal?

  • @martinschulze5399
    @martinschulze5399 5 лет назад

    its not a ''datascience'' method (which sounds like it comes from modern era). it is known as smirnov method who lived around 1900 and likely known before

  • @zhoucyrus5797
    @zhoucyrus5797 Год назад

    there is an error for the pdf of the exponential distribution, the lambda is missing.

  • @AhmedMohamed-dd4ef
    @AhmedMohamed-dd4ef Год назад

    Question : Hi, i have rainfall data as a 2d matix/frame of the UK every 5 minutes so the data is spatially and temporarily correlated. The data has severely positive skewness. Around 90% of pixels or points are less than 10 and 10% between 10-128. When i train a cnn, it is only predict rainfall of low values because of the data imbalance. I would like to transform to uniform distribution. I tried log transformation which compressed the data but still there is imbalance. Do you know how to convert to a uniform distribution so all of the values have the same chance to be predicted? It is a regressio task to predict the next 12 frames of rainfall. The data is represented by only one continuous variable, rainfall intensity. Many thanks