For those of you who are absolute beginners, yes this course might be quite hard to follow. However for people like me, who have not took statistics or probability theory courses during their undergraduate studies but have some background whatsoever, this course is very insightful and beautifully brought together.
I'm a data scientist trying to review the basics, and I often have a hard time just figuring out what he's trying to say. I don't think he put enough thought into his delivery. Lecture notes would be helpful instead of just slides. It's often easier to convey your ideas in writing if you're a novice teacher.
If you take this class my advice is first dont let this be your first stat class, you will be in trouble. This professor might be super clear if you already know what he is explaining but if you dont, he is a terrible professor. The only ones that do good are the ones that already took other courses. He knows his stuff of course, but he is not an efficient communicator, he cant clearly think in the student perspective, and his TAs are just like him too. You will be dazed and confused, just a fair warning.. Plus adding to the above he also makes several mistakes in what he writes, and has the annoying habit of speeding up and talking fast as the content gets more complicated, instead of doing the exact opposite like a professor should. Case in point, at 47:04 professor changes variables to x, should be y. Its a small detail right? But this in newly introduced concepts can cause confusion, and put this together with a fast pace, and you will be struggling a lot of things that are quite easy, but because his communication skills are lacking, the easy becomes confusing and difficult fast.
Let us tell you our tale of woe... it's of materials captured by instructors, outside camera crews, and distance learning systems. Problems of dead batteries, badly patched audio, camera audio from the back of the room, and of AV crews that are not actively monitoring their feeds. We try to do the best we can with the materials we are given. 🤷
at time 1:12:36, the instructor stated the expectation to be P. What is P representing here? In the case of the statistical model, P represents the Family of Distributions......
Right after saying P, he said "the probability of turning your head right". He alluded to the example of the "Kissing Problem" (or whatever the name was). In that problem, the estimator for P( the parameter of the Bernoulli distribution) was simply the expectation of the random variable X and P was the parameter of the model. But, in general, think of the output of an estimator to be the value of the parameter of the model and you can denote it using some letter like P in this case
I've watched the first 3 videos so far and I find it's not easy to follow the professor. It looks like I should read through all the slides before coming back to the videos..
The white noise in the background is the most irritating part of these lectures. Does anyone know a way to denoise these videos? It's so difficult to focus because of this static noise.
Because given infinite data you can only pin down the ratio of mu/sigma. This fails to uniquely determine a vector (mu, sigma), since any constant multiplication of mu/sigma is also mu/sigma. It follows that (mu, sigma) is not injective. However, mu/sigma is injective. This makes sense because mu/sigma is intuitively a smaller class than (mu, sigma), though they ofc have the same cardinality
It means the square of the difference between the limiting distribution and those converging to it, (f_l-f)^2, has integral epsilon, converging to zero. Roughly it implies that the in the limit things agree up to a set of measure zero, so that as probability distributions there is no difference.
Here is what we have for statistics and probability: ocw.mit.edu/courses/find-by-topic/#cat=mathematics&subcat=probabilityandstatistics. Best wishes on your studies!
love the tranquil waterfall in these MIT classrooms
Your comment made my day man
For those of you who are absolute beginners, yes this course might be quite hard to follow. However for people like me, who have not took statistics or probability theory courses during their undergraduate studies but have some background whatsoever, this course is very insightful and beautifully brought together.
yep i started with 18.600 for this reason
I'm a data scientist trying to review the basics, and I often have a hard time just figuring out what he's trying to say.
I don't think he put enough thought into his delivery.
Lecture notes would be helpful instead of just slides. It's often easier to convey your ideas in writing if you're a novice teacher.
Agreed, you will need some background in statistics and calculus to follow. But if you have the pre-requisites this is excellent
I agree , Philippe Rigollet is very good besides being refreshing and entertaining lecturer. Enjoyed thoroughly the lectures so far.
I really recommend this course! there are a lot of in-depth comments that refresh your understanding about statistic.
Is there a technology at MIT to effectively remove chalk from the chalkboard?
Good job! Such a refreshing lecture and easy to follow if you have background in Statistics.
MIT loves their thick chalk, which is great, except when the buildings people don't use the squeegee in between your class and the previous
Hi guys! Can you please reupload this with the background static removed?
Excellent Lecture ! Thank u MIT !
MIT can't figure out how to apply a filter to the audio to remove the hiss?
Why is the volume so low on this video? I'm pumped up to full and I can barely hear him.
1:00:00 shouldn't it be 1-P(Z
Yes. Or equivalently Phi(mu / sigma).
I agree with you. there is a mistake..
How do we calculate the mode if the highest frequency is same two times.
@@sumitkumarsingh5290 Then both of those are modes. That’s the thing about mode, there need not be just one.
@@CathyZhang No they are the same. The normal distribution is symmetric. phi(x)=phi(-x)
If you take this class my advice is first dont let this be your first stat class, you will be in trouble. This professor might be super clear if you already know what he is explaining but if you dont, he is a terrible professor. The only ones that do good are the ones that already took other courses. He knows his stuff of course, but he is not an efficient communicator, he cant clearly think in the student perspective, and his TAs are just like him too. You will be dazed and confused, just a fair warning.. Plus adding to the above he also makes several mistakes in what he writes, and has the annoying habit of speeding up and talking fast as the content gets more complicated, instead of doing the exact opposite like a professor should. Case in point, at 47:04 professor changes variables to x, should be y. Its a small detail right? But this in newly introduced concepts can cause confusion, and put this together with a fast pace, and you will be struggling a lot of things that are quite easy, but because his communication skills are lacking, the easy becomes confusing and difficult fast.
愛我了個去~頭一次看見騎自行車上課的,可以的這很MIT,抱拳了。
教授脚受伤了,带伤上课。那个“自行车” 是单腿的轮椅
All models are wrong but some of them are useful!
Is my data iid? For each individual party - maybe yes!
Parametric models vs non parametric
Did MIT clone Joe Blitzstein?
2:00 dang I should go to MIT
I don't undertand why MIT publish such a good material with so terrible the audio
Let us tell you our tale of woe... it's of materials captured by instructors, outside camera crews, and distance learning systems. Problems of dead batteries, badly patched audio, camera audio from the back of the room, and of AV crews that are not actively monitoring their feeds. We try to do the best we can with the materials we are given. 🤷
at time 1:12:36, the instructor stated the expectation to be P. What is P representing here? In the case of the statistical model, P represents the Family of Distributions......
Right after saying P, he said "the probability of turning your head right". He alluded to the example of the "Kissing Problem" (or whatever the name was). In that problem, the estimator for P( the parameter of the Bernoulli distribution) was simply the expectation of the random variable X and P was the parameter of the model. But, in general, think of the output of an estimator to be the value of the parameter of the model and you can denote it using some letter like P in this case
because the parameter of bernoullis is generally denoted by p. Note that this p is small cap and P when denoting a distribution is big cap
I've watched the first 3 videos so far and I find it's not easy to follow the professor. It looks like I should read through all the slides before coming back to the videos..
You're supposed to read or at least skim the textbook before lecture, and then again after lecture with the professor's outlines.
@Neeraj Sahu ae19m023 leme know if you find out
@@abcdxx1059 All of statistics - wasserman
The white noise in the background is the most irritating part of these lectures. Does anyone know a way to denoise these videos? It's so difficult to focus because of this static noise.
use the slutsky theorem
@1:08:40 - "it is weakly convergent if it converges in probability, and strongly convergent if it converges al naturel?"
strongly if it converges almost surely.
isnt theta hat an estimator not a statistic?
Nice
what is the text book for this subject
No textbook is listed for this course. See the course on MIT OpenCourseWare for more info at: ocw.mit.edu/18-650F16. Best wishes on your studies!
Why are the parameters of one of the normal distributions he describes identifiable and the others aren't?
Because given infinite data you can only pin down the ratio of mu/sigma. This fails to uniquely determine a vector (mu, sigma), since any constant multiplication of mu/sigma is also mu/sigma. It follows that (mu, sigma) is not injective. However, mu/sigma is injective. This makes sense because mu/sigma is intuitively a smaller class than (mu, sigma), though they ofc have the same cardinality
@1:16:00 what is convergence in l2, why does it imply convergence in probability
It means the square of the difference between the limiting distribution and those converging to it, (f_l-f)^2, has integral epsilon, converging to zero.
Roughly it implies that the in the limit things agree up to a set of measure zero, so that as probability distributions there is no difference.
watch the last lecture
Is there any other statistics course from MIT
Here is what we have for statistics and probability: ocw.mit.edu/courses/find-by-topic/#cat=mathematics&subcat=probabilityandstatistics. Best wishes on your studies!
參數式
Excellent instructor but poor hand writing.
This Lecture seems to be somewhat lacking in examples and illustrations.
@ 32:16 here atm
Prof seems a bit disorganized otherwise a good lecture series
estiMAYter
Every time this guy says "Poisson" I think of Carl saying croissant
ruclips.net/video/dg2S08WAIAo/видео.html