What are Maximum Likelihood (ML) and Maximum a posteriori (MAP)? ("Best explanation on YouTube")

Поделиться
HTML-код
  • Опубликовано: 24 дек 2024

Комментарии •

  • @arvindp551
    @arvindp551 2 года назад +2

    What you just fabulously explained in 15 lines, takes 4+ blackboards to many Indian teachers to explain even less than that. Thank you so much for sharing your knowledge here.

  • @ayushkumarrai1117
    @ayushkumarrai1117 4 года назад +44

    You're the professor I wished I had in my college! Thankyou!!

  • @rahulrustagi7029
    @rahulrustagi7029 11 месяцев назад +2

    Amazing. Make a series of Probabilistic ML Models!

  • @burakkurt3027
    @burakkurt3027 3 года назад +2

    You explain the concept not only very concise way but also in saving paper. I appreciate you for both the topic and the saved paper.

    • @iain_explains
      @iain_explains  3 года назад +4

      Thanks. I hadn't realised how environmentally responsible I was being. 😀 I think it really helps to fit everything onto a single sheet of paper so that the whole explanation is visible all at once, so the viewer can easily refer back at any point.

    • @burakkurt3027
      @burakkurt3027 3 года назад

      @@iain_explains This is very logical :D

  • @junchiliu7350
    @junchiliu7350 2 года назад +3

    The best explanation on ML and MAP! I finally understood them. Thank you!

    • @iain_explains
      @iain_explains  2 года назад

      I'm so glad the video helped, and that you liked the explanation.

  • @davebordenkircher8131
    @davebordenkircher8131 2 года назад

    I see now that the MAP estimator is like a weighted version of the ML estimator, where the weights come from prior knowledge of the measurement target. The different conditional distributions fy(y|xi) are “pushed up” or “pushed down” based on the value of the corresponding fx(xi). Of course, provided that all fx(xi) are equiprobable, the MAP estimator reduces to the ML estimator which we commonly see in optimal communications system analysis.
    I have a question for you, why is it that the equiprobable symbol scheme is considered most optimal? I am inclined to assume that it is because it yields the highest entropy. Also, I would like to know how it is that we ensure equiprobable signaling?
    Thank you

    • @iain_explains
      @iain_explains  2 года назад +1

      Excellent question. Yes, when a random source is compressed to its minimal representation (using an entropy achieving codebook) it results in a binary sequence that has equally likely ones and zeros. This video provides more insights: "What is Entropy? and its relation to Compression" ruclips.net/video/FlaJPxP8sd8/видео.html

  • @maclee3841
    @maclee3841 3 года назад +1

    Thanks a lot! One of the most simplest explanations on RUclips

  • @Amine1z
    @Amine1z 3 года назад +1

    Dear prof, you're the best

  • @NoahElRhandour
    @NoahElRhandour 3 года назад

    holy what a clear explanation. it ended my 2 day struggle of not getting it in 18 minutes!!!! thank you

  • @Peolorios
    @Peolorios 3 года назад +12

    I found your videos at the right moment, they cover a lot of the basics of my 1st semester master courses. Thank you. A nice topic you could cover that comes up a lot in detection and estimation is the Cramer-Lao Lower Bound

    • @iain_explains
      @iain_explains  3 года назад +2

      Thanks for your comment. Glad the videos are helpful. And thanks for the C-R suggestion, I'll add it to my "to do" list.

  • @jizong234
    @jizong234 2 года назад

    This is the best explanation in the world, thank you !

    • @iain_explains
      @iain_explains  2 года назад +1

      I'm so glad to hear that you liked the video.

  • @_Sam_-zh7sw
    @_Sam_-zh7sw 2 года назад

    may be i am missing some pre-requisite knowledge because i am confused a little bit. have we inverted the graph of the function here? f(x) is plotted horizontally and x is plotted vertically. But how can there be a different distribution function of ax1,ax2...ax(n) if there is just one input and output?

    • @iain_explains
      @iain_explains  2 года назад

      Um no. x is not plotted vertically. f_Y(y|x) is plotted vertically. This is the density of the random variable Y, given a specific value of the random variable X. This is a different function for each different realisation (value) of X (ie. x_1, x_2, ...).

  • @ronaldmtenga8093
    @ronaldmtenga8093 Год назад

    are the differenct x values you are checking for maximum likelihood each a possible input signal or are we searching on a bit by bit basis?

    • @iain_explains
      @iain_explains  Год назад

      They are a possible realisation of the random variable X. If X represents binary data, then it would be "searching on a bit by bit basis", but it X represented higher order modulation then it would be on a "symbol" basis.

  • @NeilDeshpande313
    @NeilDeshpande313 3 года назад +2

    I finally get the difference between the two! Thank you!

  • @usmanshabbir4902
    @usmanshabbir4902 2 года назад

    very precisely explained.

  • @aidankennedy6973
    @aidankennedy6973 3 года назад +7

    This is a fantastic video that answered so many questions I had while working through my academic coursework. Thank you so much for uploading!

  • @tuongnguyen9391
    @tuongnguyen9391 3 года назад

    Do you ever have to do a rehearsal beforehand ? I see the explanation is quite smooth.

    • @iain_explains
      @iain_explains  3 года назад +1

      Thanks, I put quite a bit of thought into how to explain things in the best way.

    • @tuongnguyen9391
      @tuongnguyen9391 3 года назад

      Do you follow a systematic procedure to construct the explanation process. If so I really hope that you could share this procedure :). Although everything is short I find that the information is delivered clearly with many subtle points and detail carefully summarized. Thank you for your inspiring lecture.

  • @JoaoVitorBRgomes
    @JoaoVitorBRgomes 3 года назад +1

    Beautiful explanation. Very helpful.

  • @nikkoNine
    @nikkoNine Год назад

    Very intuitive explanation! 🙏

  • @osamasaeed7743
    @osamasaeed7743 Месяц назад

    woow so good to understand the MAP and ML

  • @AlexColeAldems
    @AlexColeAldems 2 года назад

    Does demodulating using ML require channel state information? (i.e. an estimation of the AWGN noise variance)

    • @iain_explains
      @iain_explains  2 года назад +1

      Yes. Good point. It's almost never mentioned. It's not too hard to get an estimate of the receiver noise - by taking measurements when nothing is being sent (of course you need to be able to work out when nothing is being sent!) It's harder to estimate other parameters, such as channel gain. And there's lots of things that are done to make that possible. See eg: "Channel Estimation for Mobile Communications" ruclips.net/video/ZsLh01nlRzY/видео.html

  • @BlueSkyGoldSun
    @BlueSkyGoldSun Год назад

    Can this be applied in marketing?

  • @Dutta1605
    @Dutta1605 Год назад

    i beleive the title of the video is genuinely true.

    • @iain_explains
      @iain_explains  Год назад

      Thanks. It was a comment someone else had made about the video, so it's good to know that you also agree.

  • @heroszeros
    @heroszeros Год назад

    Very good explanation with right amount of details and relevant examples. Thanks a million.

  • @ivanrodionov9724
    @ivanrodionov9724 Год назад +1

    Outstanding video! You sir have saved the day, again!

    • @iain_explains
      @iain_explains  Год назад +1

      I'm so glad the video helped. It's great to read these comments, and know that my videos are making a difference for people. Thanks.

  • @peterkimemiah9669
    @peterkimemiah9669 2 года назад

    Thanks for your super simple explanation. I now understand how to apply it.

  • @abderrahmanetayebcherif7325
    @abderrahmanetayebcherif7325 3 года назад

    best explanation of the ML and MAP on youtube
    thank you

  • @passionatechristianmaker
    @passionatechristianmaker 9 месяцев назад

    great explanation!

  • @prietessutube
    @prietessutube 2 года назад

    Thanks so much for this video, explained it much better with my lecturer!!!

  • @AleeEnt863
    @AleeEnt863 Год назад

    Great explanation...

  • @sayantankundu973
    @sayantankundu973 2 года назад

    Definitely "Best explanation on RUclips" !! ❤ Thanks a lot Sir.

    • @iain_explains
      @iain_explains  2 года назад

      Thanks. I'm glad you think so. And I'm glad it was helpful.

  • @cartercheng5846
    @cartercheng5846 2 года назад

    Very helpful video. I have a question there. The MAP is explained as MLE weighed by the probability of the parameter x, and the parameter follows a certain distribution. If X is a continuous random variable, what is the mathematical meaning for f_y(y|x)f_x(x)?

    • @iain_explains
      @iain_explains  2 года назад

      It equals the joint pdf f_{X,Y}(x,y)

  • @ks.4494
    @ks.4494 2 года назад

    Thanks for the Video, is there any reference ( book, ...) for that, particulary for numerical solution?

    • @iain_explains
      @iain_explains  2 года назад

      I'm not sure if this is what you're looking for exactly (eg. I'm not sure it has the numerical examples you might be looking for), but I like this book: H. Vincent Poor, “An Introduction to Signal Detection and Estimation”

  • @Balance-fl1zc
    @Balance-fl1zc Год назад

    Beautiful explanation sir, thank you!

  • @SouthEsVideos
    @SouthEsVideos 4 года назад +3

    I'm currently learning about autoencoders and it's based on this topic! very helpful and intuitive. Thank you!

  • @ravindratomar9916
    @ravindratomar9916 3 года назад

    sir, how to estimate channel in case of correlated rayleigh fading channel. for example y1=hx_1 + hx_2 +n_1, y2=hx_1 + hx_2 +n_2.

    • @ravindratomar9916
      @ravindratomar9916 3 года назад

      n_1 and n_2 are white gaussian noise with different variances.

  • @mangomilkshakelol
    @mangomilkshakelol 11 месяцев назад

    Thank you so much! That's clear. One question: for MAP, what's f_X(x)?

    • @iain_explains
      @iain_explains  11 месяцев назад +1

      It's the probability density function for the variable X.

  • @akinsolaidris9292
    @akinsolaidris9292 3 месяца назад

    Is AwGN the same as normally distributed err or bias?

    • @iain_explains
      @iain_explains  3 месяца назад

      This video should help: "What is White Gaussian Noise (WGN)?" ruclips.net/video/QfUQMzHfbxs/видео.html

  • @saipavan7703
    @saipavan7703 2 года назад

    Hi Iain. Loved your explanation. I wanted to ask a question about MLE. In the plots of x1,x2,xn, When each x1/x2 give a single value for the function, Why does plot exist for x1 when the function takes a single value for x1. Thank You

    • @iain_explains
      @iain_explains  2 года назад

      Sorry, I'm not sure what you're asking.

    • @rishabhsrivastava1241
      @rishabhsrivastava1241 2 года назад

      I think I got what you are saying and there seems to be some gap in your understanding. Let me try to fill that although Ian mentioned it in this video.
      What you are saying is that for a given value of x, there is only one single value of y through it's distribution f(y/x) but that is not true. Actually, there are SEVERAL different distributions of y depending on the SEVERAL values of x's. So, when Ian says that for a given x, the distribution's center shifts, it is actually a new distribution centered around that given x value. Then comes the concept of a single value from these distributions, now that is y(bar), this is an observation of all the f(y/x) pdf value among all the distributions of y's for those SEVERAL x's. That is the single value that you are thinking of.
      Hope I was able to answer your question to some degree. :)

    • @mrr7115
      @mrr7115 2 года назад

      The single value is a result of having measured/fixed y (the y bar hat equation). The plot is for all y (ie it’s a function of y not of x1). x1 is just a guess of the true parameter of the Gaussian distribution (proportional to mean). The horizontal axis (independent variable) is y. Also the function, which is a Gaussian, takes more than just a*x1, it also takes in the variance from the noise. To prove to yourself that the function takes y, look at the form of the Gaussian equation, see the y in there?

  • @alexanderpaskal1387
    @alexanderpaskal1387 2 года назад

    Liked and subbed, very clear and accessible explanation of a concept that made no sense to me as it was presented in my class

  • @robertpollock8617
    @robertpollock8617 Год назад

    So it is L(x|y)? - we want to maximize the likelihood of x given the data values y? . So we are in a sense trying to say that we have high likelihood that this data observed could come from or be predicted by this model of x? Where the probability is P(y|x). Maybe you are saying that and I am not picking up on this. I think you might be but I might not be understanding your notation.

  • @shahabansari5201
    @shahabansari5201 3 года назад +1

    Wow! Amazing way of explaining these complex ideas.

  • @gastonparra2971
    @gastonparra2971 Год назад

    Gran explicación.. Gracias por subir el video.

  • @gamingandmusic9217
    @gamingandmusic9217 3 года назад

    at a particular point in the density function, the probability is zero right? I'm a little confused.

    • @iain_explains
      @iain_explains  3 года назад

      I'm not exactly sure what you're asking. The density function is a "density" (as the name indicates). This means you need to integrate it over some range of values, in order to find the probability. The probability of any _exact_ value is zero (since the base has zero width, for a single _exact_ value). See: "What is a Probability Density Function (pdf)?" ruclips.net/video/jUFbY5u-DMs/видео.html

    • @gamingandmusic9217
      @gamingandmusic9217 3 года назад

      @@iain_explains oh sorry, im wrong. Thank you so much sir.

  • @zhipenglin
    @zhipenglin 5 месяцев назад +1

    excellent video!!

  • @ashokmahipathi8133
    @ashokmahipathi8133 Год назад

    Thank you so much Sir Iain. You made my day. Great explanation regrading MAP and ML. Hats off Iain

  • @KCODivyanshuKatyan
    @KCODivyanshuKatyan Год назад +1

    best in the game 🙌🙌

  •  2 года назад +1

    MAP starts at 10:35

  • @kalyansunkara9464
    @kalyansunkara9464 2 года назад

    Sir which text book should we follow for detection and estimation theory?

    • @iain_explains
      @iain_explains  2 года назад +1

      I like the book: H. Vincent Poor, "An Introduction to Signal Detection and Estimation", Springer.

  • @saraho5338
    @saraho5338 2 года назад

    Thank you for the great video

  • @ashleyhart1907
    @ashleyhart1907 Год назад

    Incredibly helpful. Thank you!

  • @gofaonef.mogobe1306
    @gofaonef.mogobe1306 3 года назад +1

    Hi..very helpful video. Kindly assist me understand how I can factor in the concept of consistency of MLE with respect to the graph illustrations?
    Particularly, I've learned that as n gets large, mean turns to zero as MLE becomes an even more consistent estimate.

  • @林泓均-q4j
    @林泓均-q4j 4 года назад

    Really big thanks for your video!!
    May you take another video for explaining different pathloss models, such as Okumura-Hata or various COST model in wireless channel?

    • @iain_explains
      @iain_explains  4 года назад +2

      Good suggestions thanks. I've added them to my "to do" list.

  • @analysislearning9179
    @analysislearning9179 3 года назад

    Great video, I have a question, if the variable is its self distributed with Nakagami distribution. Then how can we compute the MLE and MAP?

    • @iain_explains
      @iain_explains  3 года назад +1

      The term f_X(x) is the density function for the random variable of interest. So, if it is Nakagami distributed, then f_X(x) equals the formula for the Nakagami p.d.f. which you can find in this video: ruclips.net/video/ztpNbE-Vpaw/видео.html

    • @analysislearning9179
      @analysislearning9179 3 года назад +1

      @@iain_explains Thankyou very much

  • @AbCd-fo6ys
    @AbCd-fo6ys 2 года назад

    What a clear explanation!
    Thank you so much.

  • @mehmetfatihayten8269
    @mehmetfatihayten8269 2 года назад

    Decent video! Thanks.

  • @adonis7419
    @adonis7419 4 года назад +1

    Very helpful

  • @rezasoleimani6636
    @rezasoleimani6636 2 года назад

    The explanation is great. The only problem is using pen and paper instead of something more comfortable. The page is too small for this amount of writing.

    • @iain_explains
      @iain_explains  2 года назад +1

      On the other hand, having everything on the one page means you don't need to scroll back and forth through the video to see the links to earlier parts, and I can simply point to the earlier parts while explaining how they link to the later parts (as I'm doing in the thumbnail image). Perhaps it doesn't work so well on small screens ...

  • @keremokyay6725
    @keremokyay6725 2 года назад

    Amazing, this was so clear to understand. Thank you very much!!!

  • @waynerooney4674
    @waynerooney4674 Год назад +1

    ty

  • @yuxinzhang4228
    @yuxinzhang4228 3 года назад

    It's very helpful thanks sooooo much

  • @NvkChaitanya777
    @NvkChaitanya777 3 года назад

    Can you please make a video on softmax regression?

    • @iain_explains
      @iain_explains  3 года назад

      Thanks for the suggestion, but I'm not familiar with it, sorry. I'll have to give it some thought.

  • @Yohan87845
    @Yohan87845 2 года назад

    Thank you very much.

  • @partheshvikassoni8601
    @partheshvikassoni8601 2 года назад

    This was really informative! Thanks.

  • @joaquinvanloon7341
    @joaquinvanloon7341 3 года назад

    Amazing video, thanks!

  • @kavineshs15
    @kavineshs15 6 месяцев назад

    I can't understand why the bell curve is shifting for every value of x.

  • @AK-yf4dp
    @AK-yf4dp 3 года назад

    Thank you so much!!! very helpful video

  • @aqeelal-shakhouri7572
    @aqeelal-shakhouri7572 3 года назад

    Thank you. you explained it clearly, just what I was looking for.

  • @amitpalkar934
    @amitpalkar934 3 года назад

    Can we have a video sometime on mmse and irc receivers ?
    Regards,
    Amit

    • @iain_explains
      @iain_explains  3 года назад

      Thanks for the suggestion. I've added them to my "to do" list. I'll see what I can do (it's starting to become a long list).

  • @loreychu7646
    @loreychu7646 3 года назад

    Good explanation of a lot of concepts in wireless communication. I'm watching your video for the preparation of QE. Hope I can pass!

  • @robsenponte3308
    @robsenponte3308 Год назад +1

    Nice

  • @СизоненкоДмитро
    @СизоненкоДмитро 4 месяца назад

    Thank you!!!

  • @ibrahimshikdaher7551
    @ibrahimshikdaher7551 4 года назад +1

    Thank you

  • @tainacoimbra1940
    @tainacoimbra1940 3 года назад

    Excelent explanation! Thank you very much :)

  • @human_agi
    @human_agi Год назад

    if x is vector ?

  • @javiergonzalezarmas8250
    @javiergonzalezarmas8250 Год назад

    Amazing

  • @donpkchannel7203
    @donpkchannel7203 2 года назад

    When your typing the screens becomes blurry because paper is moving. Please stabilize the paper.

  • @tomaguci5102
    @tomaguci5102 Год назад +1

    I think this video could be improved by providing a concrete example, also it's really mathy without much intuitive explanation

  • @egorkrapovnitskiy997
    @egorkrapovnitskiy997 2 года назад

    didn't get the idea

  • @skatingscotland64
    @skatingscotland64 3 года назад

    Iain you have offered me shelter in a howling wind, thank you - I can leave the library and go home now xo love from rory

    • @iain_explains
      @iain_explains  3 года назад

      That's great. I'm so glad you found the video helpful. Hope you mange to stay out of the wind.

  • @MohammedFarag81
    @MohammedFarag81 Год назад

    Great, thank you