Machine Learning with 10 Data Points - Or an Intro to PyMC3

Поделиться
HTML-код
  • Опубликовано: 11 сен 2024

Комментарии • 56

  • @ravink
    @ravink Год назад +9

    I'm a core dev on PyMC. This is a great video. For newer watchers know that PyMC3 has been superseded by PyMC v5 and its got so many cool new things

    • @lashlarue7924
      @lashlarue7924 Год назад

      Thank you, Sir! 🫡

    • @stanleynwanekezie5355
      @stanleynwanekezie5355 Год назад

      Hi, I am working on a model update involving pymc. I have a function, say func, which an given array, produces a float or -inf. Func used to be wrapped within a pymc.stochastic decorator before it was deprecated. I understand that func serves to confirm whether a draw during sampling of a tensorvariable satisfies requirements so that only draws that do are retained in the posterior. Now, I am trying to use the pymc.DensityDist or pymc.Potential but because a TensorVariable (not an array) is passed to func, it is unable to perform inequality checks. I have used the eval method of the tensorvariable but that has also failed to work. Please help

    • @ravink
      @ravink Год назад

      @@stanleynwanekezie5355 The best place to ask is the pymc discourse. All the core devs are there and your question I believe has been asked already so there may already be a solution!

  • @cameronwebb9851
    @cameronwebb9851 Год назад +1

    Dude your video is amazing. So my clarity and simplicity around complex topic.
    I really like how you keep re explaining basic terms as you cover them because it really helps following through into more advanced areas

  • @brycedavis5674
    @brycedavis5674 3 года назад +2

    Wow I really hope you dive into pymc3, I always had difficulty on understanding programming the priors. You're the best! This video is great. Sending love from South Korea :)

  • @xxshogunflames
    @xxshogunflames 3 года назад +2

    SUPER good video, I have been practicing on jupyter notebooks for data science and it has taken time but these videos make me feel like im standing on the shoulders of giants. THANKS

  • @sunilmathew2914
    @sunilmathew2914 3 года назад +4

    You teach so well! Please keep making videos!

  • @matakos22
    @matakos22 2 года назад +1

    Very nice video, well-organized and neatly explained concepts. One thing I would like to see at the end is some discussion on how would you use the posterior distributions, given that the true values are not obvious to infer from those at all :)

  • @pgbpro20
    @pgbpro20 3 года назад +1

    Amazing as always! Just enough to wet the appetite for more PyMC3 learning. It looks like I may be using this library really soon.

  • @111dogger
    @111dogger 3 года назад +5

    Very informative. Thanks!

    • @ritvikmath
      @ritvikmath  3 года назад +3

      Of course! Thanks for watching

  • @user-or7ji5hv8y
    @user-or7ji5hv8y 3 года назад +7

    More video like this on Bayesian approaches.

    • @ritvikmath
      @ritvikmath  3 года назад +7

      Haha you'll probably be happy with Wednesday's video :) stay tuned

  • @jiaqint961
    @jiaqint961 10 месяцев назад

    This tutorial really ties everything together. Thank you.

  • @EdoardoMarcora
    @EdoardoMarcora 3 года назад +3

    Nicely done! Very clear, concise yet informative. One thing missing from intro tutorials like this one is a real world example, with a a bigger dataset and more variables. Can pymc3 run on gpus, or computer clusters, etc or should I look elsewhere (pyro, tensor flow.probability)? Just giving an idea for future videos! Keep up the good work

    • @ritvikmath
      @ritvikmath  3 года назад

      Valid point and great suggestion! Thank you

    • @ravink
      @ravink Год назад

      Yes. PyMC can run on GPU TPU backend using Jax. Source I'm core dev of the library

  • @MeshRoun
    @MeshRoun 2 года назад +1

    I'm subscribing, your explanation was on point!

  • @umamiplaygroundnyc7331
    @umamiplaygroundnyc7331 9 месяцев назад

    Amazing job! So clear & easy to understand

  • @ResilientFighter
    @ResilientFighter 3 года назад +2

    Very nicely done!

  • @dwivedys
    @dwivedys 10 месяцев назад

    Brilliant - I loved it

  • @MrMoore0312
    @MrMoore0312 3 года назад +1

    Great video, very well explained!! I would love to see you 💪 on another equation and distribution, something just a little harder like
    Ytrue = X1^3 - X2^2 - X3
    Y = Ytrue + binomial error distribution
    Or something weird like that lol
    Thanks for another great lesson!

    • @ritvikmath
      @ritvikmath  3 года назад +1

      Hey great suggestion! Thanks

  • @ppybmjc
    @ppybmjc 3 года назад +7

    What are the advantages of this approach over use of the confidence intervals calculated in a linear model (e.g. as output by statsmodels?)

    • @axscs1178
      @axscs1178 2 года назад +2

      Confidence intervals don't actually give you the probability of a parameter being inside them. They instead tell you that, over repeated sampling, say 95% of the time the true value will be contained in those intervals. Bayesian approach gives you intervals with the probability of a parameter being inside them

  • @jfndfiunskj5299
    @jfndfiunskj5299 2 года назад

    Great stuff. You've won a new subscriber.

  • @FabulusIdiomas
    @FabulusIdiomas 2 года назад

    In other words, the MC in PyMC3 is for Markov Chain eh?

  • @renaspersonal9854
    @renaspersonal9854 Год назад

    Ur videos r awesome thanks for adding in theory!

  • @robertc6343
    @robertc6343 3 года назад

    Fantastic material. Thanks.

  • @komuna5984
    @komuna5984 Год назад

    Thanks a lot for this video and the corresponding codes in GitHub! May Allah bless you!!!

  • @zsoltczinege3014
    @zsoltczinege3014 3 года назад +3

    What's the advantage of Bayesian analysis compared to calculating confidence intervals with linear regression and bootstrap?

    • @ritvikmath
      @ritvikmath  3 года назад +3

      This is a great question and I'm going to have to give the unsatisfying answer that I'll probably address this in a future video

    • @zsoltczinege3014
      @zsoltczinege3014 3 года назад

      @@ritvikmath That's totally satisfying, as long as we get that video. :) Thank you!

  • @qiguosun129
    @qiguosun129 2 года назад

    So cool!

  • @jiayiwu4101
    @jiayiwu4101 9 месяцев назад

    2:45 question - would linear regression give you distribution too? consider the confidence intervals of the parameter estimates.

  • @ArgumentumAdHominem
    @ArgumentumAdHominem 10 месяцев назад

    Great video. I'm just shocked about the runtime. 2 whole minutes. So, the mcmc needs to run a few thousand iterations, where for each iteration it should sample a few random numbers from standard distributions. I would have expected this to be done in milliseconds. What am I missing?

  • @lulzimi
    @lulzimi Год назад

    @ritvikmath Do you have any book recommendations to learn PyMC3?

  • @cbasile22
    @cbasile22 2 года назад

    Hi ritvikmath, great videos, this one , the gibbs, metrolopis, the ridge/lasso!! Do you have any suggestions to understand how to best do a prediction in the bayesian way? And to get a credible set of the prediction, we have credible set for each of the parameters, if I use their means I can make predictions , if I were to use a bit lower than their mean I can produce a sightly different value , so there many possible prediction ranges for a given x input. I would appreciate any suggestions, thanks !!!! I could not find a good resource for that.

  • @user-or7ji5hv8y
    @user-or7ji5hv8y 3 года назад

    Could we not simply use p-value to gauge our confidence of our point estimate? Or is there additional benefit from being able to see the full distribution?

  • @thenayancat8802
    @thenayancat8802 3 года назад

    Posterior median would probably give a better estimate of the mode of the distribution of sigma

  • @user-or7ji5hv8y
    @user-or7ji5hv8y 3 года назад

    how come you did not plot a histogram of the posterior, given that you drew samples from it?

  • @rajns8643
    @rajns8643 Год назад

    Can somebody pls tell where did we use the observed values of y in determining the posterior distribution?
    Is it perhaps used in determining just the normalizing factor in the Bayes theorem...?

  • @stanleynwanekezie5355
    @stanleynwanekezie5355 Год назад

    Hi, I am working on a model update involving pymc. I have a function, say func, which an given array, produces a float or -inf. Func used to be wrapped within a pymc.stochastic decorator before it was deprecated. I understand that func serves to confirm whether a draw during sampling of a tensorvariable satisfies requirements so that only draws that do are retained in the posterior. Now, I am trying to use the pymc.DensityDist or pymc.Potential but because a TensorVariable (not an array) is passed to func, it is unable to perform inequality checks. I have used the eval method of the tensorvariable but that has also failed to work. Please help.

  • @nb9797
    @nb9797 3 года назад

    good teacher

  • @robertwest6244
    @robertwest6244 3 года назад

    Does Pymc3 do the same type of thing that RJags does? Does it “send” the model to another language? Or is it doing all of the mcmc sampling itself?

    • @cornagojar
      @cornagojar 3 года назад

      pymc3 is a builtin library of python, not an interface like rjags. Moreover rjags only allows gibbs sampling (I think) while pymc3 allows hmc

  • @user-or7ji5hv8y
    @user-or7ji5hv8y 3 года назад

    Do you prefer PyMC3 over TensorFlow Probability?

  • @ryan_chew97
    @ryan_chew97 3 года назад

    whats the reasoning why mcmc is better for smaller datasets?

  • @musiknation7218
    @musiknation7218 11 месяцев назад

    Make some videos for Bayesian prior selection

  • @junkbingo4482
    @junkbingo4482 3 года назад

    i'm not sure it's a good idea to model 10 data with a model; the test theory is really different ( and just says ' watch out')