Markov Decision Processes

Поделиться
HTML-код
  • Опубликовано: 9 июл 2024
  • Virginia Tech CS5804
  • НаукаНаука

Комментарии • 41

  • @hosamfikry2924
    @hosamfikry2924 5 лет назад +18

    That is the best video I watched so far to understand this topic

  • @Pexers.
    @Pexers. 3 года назад +3

    Thank you, I spent hours in this algorithm, finally understood it !

  • @hobby_coding
    @hobby_coding 3 года назад +3

    very good lecture maybe the best introduction to this topic i've ever seen on youtube

  • @syedrumman3920
    @syedrumman3920 2 года назад +2

    This is such a clear explanation!! Ty for this!! I wish I had taken your class while I was in VT!

  • @consolesblow
    @consolesblow 5 лет назад +1

    Thanks a lot! I found this very helpful.

  • @jff711
    @jff711 3 года назад +2

    Thank you very much, very well explained.

  • @quantlfc
    @quantlfc 2 года назад

    Absolutely amazing lecture!!!

  • @JustinMasayda
    @JustinMasayda 2 года назад

    This was fantastic, thank you!

  • @jub8891
    @jub8891 11 месяцев назад

    thank you so much, you explain the subject very well and have helped me to understand..

  • @coeusmaze9413
    @coeusmaze9413 4 года назад +2

    The video provides intuitive but deep understanding in MDP

  • @ismailasmcalskan2552
    @ismailasmcalskan2552 4 года назад

    Really good video about this topic. Thank you

  • @xruan6582
    @xruan6582 4 года назад +1

    can anyone explain (32:00) the switch between two modes (i.e. represented by green and red arrow). To me the green one seems like deterministic rule, the red one seems like stochastic rule. Can they exist simultaneously?

  • @sander1426-2
    @sander1426-2 4 года назад

    Thanks for the explanation!

  • @Ahmed.r.a
    @Ahmed.r.a 3 месяца назад

    thank you for this brilliant explanation. I wished there was a Question with solution to practice on.

  • @richardm5916
    @richardm5916 4 года назад

    Realy great explaintion on Machine learning

  • @seanxu6741
    @seanxu6741 Год назад

    Fantastic video! Thanks a lot!

  • @behmandtirgar
    @behmandtirgar 4 года назад

    I have a question at time 8:30
    : if we take an action to go to the left, why Pr(c | b, left) isn't 0.00? (we go to another side)

  • @srujayop
    @srujayop 2 года назад +1

    Is the reward R(s) actually R(s')?
    And should that also be multiplied with the transition probability?
    max(over a) sum P(s', r|,s, a) [r + gamma*V(s')]
    ? I am trying to relate the equation presented in the video to standard notation 4 par notation.

  • @ryanflynn386
    @ryanflynn386 5 лет назад +3

    This is a great explanation video, thanks so much. Your voice is easy to listen to too haha.

    • @berty38
      @berty38  5 лет назад +14

      Ryan Flynn Thanks! I’m glad it’s helpful. My smooth voice is a huge disadvantage when I teach morning classes and my students all fall asleep.

    • @tarik8622
      @tarik8622 3 года назад

      Very interesting topic. And i think that you will make a fortune if you use your voice in publicity field. Best regards.

  • @peterkimemiah9669
    @peterkimemiah9669 3 года назад

    Very good easy to understand.

  • @linfrancis5204
    @linfrancis5204 5 лет назад +1

    Great video. Thank you. Could you please make a similar video while we consider a two-dimensional Markov chain with more states?

  • @jaideep_yes
    @jaideep_yes 5 лет назад

    Thank you.

  • @joshuasegal4161
    @joshuasegal4161 5 лет назад

    What software are you using to make this?? It looks like you have like an infinite page which gives a really clean look

    • @berty38
      @berty38  5 лет назад +3

      Nothing too fancy. This was done with Apple Keynote, and I'm faking that scrolling effect with "Magic Move" animations. I'm always looking for better tools to build useful visuals for lectures.

  • @sanskarshrivastava5193
    @sanskarshrivastava5193 3 года назад

    Best video for MDP on youtube

  • @JebbigerJohn
    @JebbigerJohn 9 месяцев назад

    This is so good!!!

  • @_brenda4975
    @_brenda4975 3 года назад

    much better than my lecturer

  • @treegnome2371
    @treegnome2371 3 года назад

    at 17:35, why isn't it gamma = (0,1), instead of (0,1]...if gamma = 1, the influence of the actions farther down the road stays the same as all other actions, rather than shrinking the influence...right?

  • @rezadarooei248
    @rezadarooei248 4 года назад

    Thanks for your nice tutorial is it possible upload the slides?

  • @y-3084
    @y-3084 4 года назад

    excellent

  • @zenchiassassin283
    @zenchiassassin283 4 года назад

    What textbook ? thank you very much

  • @Throwingness
    @Throwingness 3 года назад +1

    Around 34:00 when there are equations on the screen you should have had a pointer or something to point at what you are talking about. It's not clear.

  • @dminn
    @dminn 4 года назад

    God bless

  • @EdupugantiAadityaaeb
    @EdupugantiAadityaaeb 10 месяцев назад

    What is the name of textbook

  • @suvinaybothra8988
    @suvinaybothra8988 4 года назад

    honesty

  • @abdullahmoiz8151
    @abdullahmoiz8151 4 года назад

    33:27

  • @ahmet9446
    @ahmet9446 5 лет назад

    The best I find is [4, 1]. I couldn't achieve [4.2, 1.2]. Does anyone achieve [4.2, 1.2]?

  • @izazkhan1640
    @izazkhan1640 5 лет назад

    jhk