Deep RL Bootcamp Lecture 4A: Policy Gradients

Поделиться
HTML-код
  • Опубликовано: 17 янв 2025

Комментарии • 43

  • @naeemajilforoushan5784
    @naeemajilforoushan5784 9 месяцев назад +5

    After 5 years still, the lecture is a great video, thank you a lot

  • @bhargav975
    @bhargav975 6 лет назад +35

    This is the best lecture I have seen on policy gradient methods. Thanks a lot.

  • @jony7779
    @jony7779 4 года назад +18

    Every time I forget how policy gradients work exactly, I just come back here and watch starting at 9:30

    • @andreasf.3930
      @andreasf.3930 4 года назад +3

      And every time you visited this video, you forgot where to start watching. Thats why you posted this comment. Smart guy!

  • @auggiewilliams3565
    @auggiewilliams3565 5 лет назад +1

    I must say that in more than 6 months, this is by far the best lecture/ material I have come across that was able to make me understand what policy gradient method actually is. I really praise this work. :) Thank you.

  • @marloncajamarca2793
    @marloncajamarca2793 6 лет назад +3

    Great Lecture!!!! Pieter's explanations are just a gem!

  • @ericsteinberger4101
    @ericsteinberger4101 7 лет назад +10

    Amazing lecture! Love how Pieter explains the math. super easy to understand.

  • @ashishj2358
    @ashishj2358 4 года назад

    Best lecture on Policy Gradients hands down. Has covered some worth noting superficial details of many papers as well.

  • @johnnylima1337
    @johnnylima1337 6 лет назад +5

    It's such a good lecture, I'm stopping to ask myself why it was so easy to cover such significant information with full understanding

  • @sharmakartikeya
    @sharmakartikeya Год назад

    I might be missing a simple concept here but how are we increasing/decreasing the grad log probability of the actions using the gradient of U(theta)? I get that positive return for a trajectory will make the gradient of U positive and so theta will be increased in favour of those trajectories but how is it increasing grad log prob?

  • @synthetic_paul
    @synthetic_paul 4 года назад +5

    Honestly I can’t keep up without seeing what he’s pointing at. Gotta pause and search around the screen each time he says “this over here”

    • @akarshrastogi3682
      @akarshrastogi3682 4 года назад +2

      Exactly. "This over here" has got to be the most uttered phrase in this lecture. So frustrating.

  • @Рамиль-ц5о
    @Рамиль-ц5о 4 года назад

    Very good lecture about policy gradient method. I have looked through a lot of articles and was understanding almost everything, but your derivation explanation is really the best. It just opened my eyes and showed the whole picture. Thank you very much!!

  • @bobsmithy3103
    @bobsmithy3103 2 года назад

    amazing work. super understandable, concise and information dense.

  • @norabelrose198
    @norabelrose198 2 года назад

    The explanation of the derivation of policy gradient is really nice and understandable here

  • @keqiaoli4617
    @keqiaoli4617 4 года назад +1

    why a good "R" would increase the probability of path??? Please help me

  • @DhruvMetha
    @DhruvMetha 3 года назад

    Wow, this is beautiful!

  • @JyoPari
    @JyoPari 5 лет назад +1

    Instead of having a baseline, why not make your reward function be negative for undesired scenarios and positive for good ones? Great lecture!

  • @ishfaqhaque1993
    @ishfaqhaque1993 5 лет назад

    23:20- Gradient of expectation is expectation of gradient "under mild assumptions". What are those assumptions?

    • @joaogui1
      @joaogui1 5 лет назад +2

      math.stackexchange.com/questions/12909/will-moving-differentiation-from-inside-to-outside-an-integral-change-the-resu

  • @dustinandrews89019
    @dustinandrews89019 7 лет назад +1

    I got a lot out of this lecture in particular. Thank you.

  • @biggeraaron
    @biggeraaron 6 лет назад +1

    Where can i buy his T-shirt?

  • @emilterman6924
    @emilterman6924 5 лет назад

    It would be nice to see what laboratories they had (what exercises)

    • @Procuste34iOSh
      @Procuste34iOSh 4 года назад

      dont know if ur still interested, but the labs are on the bootcamp website

  • @JadtheProdigy
    @JadtheProdigy 6 лет назад

    best lecturer in series

  • @mcab2222
    @mcab2222 4 года назад +3

    nice but hard to follow without knowing what "this" refers to. I hope my guesses were right :)

  • @faizanintech1909
    @faizanintech1909 6 лет назад

    Awesome instructor.

  • @isupeene
    @isupeene 4 года назад +2

    The guy in the background at 51:30

  • @ethanjyx
    @ethanjyx 5 лет назад

    wow damn this is so well explained and the last video is very entertaining.

  • @suertem1
    @suertem1 5 лет назад

    Great lecture, thanks

  • @nathanbittner8307
    @nathanbittner8307 7 лет назад

    excellent lecture. Thank you for sharing.

  • @ProfessionalTycoons
    @ProfessionalTycoons 6 лет назад

    great talk!

  • @richardteubner7364
    @richardteubner7364 7 лет назад +1

    1:11 why are DQNs and friends Dynamic Programming Methods? I mean the neural network works as functions approximator to satisfy Bellmans eqn. , but still Backprop is the workhorse. In my opinion DQNs are much more similar to PG methods than to Bellman Updates??! And another issue with RL Landscape slide is where the heck are model based RL algos?? This slide should be renamed to model free RL landscape.

  • @elzilcho222
    @elzilcho222 6 лет назад +1

    could you train a robot for 2 weeks in the real world then use those trained parameters to optimize a virtual environment? You know.. making the virtual environment very close to the real world?

    • @OfficialYunas
      @OfficialYunas 6 лет назад +1

      Of course you could. It's the opposite of what OpenAI does when they train a model in a virtual environment and deploy it in reality.

    • @soutrikband
      @soutrikband 5 лет назад

      Real world is very complicated with model uncertainties, friction, wear and tear and what have you...
      Simulators can come close , but we cannot expect them to fully mimic real world phenomena.

  • @karthik-ex4dm
    @karthik-ex4dm 6 лет назад

    PG is awesome!!!
    Doesn't depend on environment Dynamics really?? Wow
    All the pain and stress just goes away when we see our algorithms working😇😇

  • @Diablothegeek
    @Diablothegeek 7 лет назад

    Awesome!! Thanks

  • @arpitgarg5172
    @arpitgarg5172 5 лет назад +11

    If you can't explain it like Pieter Abbeel or Andrew NG then you don't understand it well enough.

  • @piyushjaininventor
    @piyushjaininventor 6 лет назад

    Can you share ppt??

    • @luxorska5143
      @luxorska5143 5 лет назад +3

      You can find all the slides and the other lectures here:
      sites.google.com/view/deep-rl-bootcamp/lectures

  • @shaz7163
    @shaz7163 7 лет назад

    very nice :)

  • @MarkoTintor
    @MarkoTintor 4 года назад

    ... you can use "a", and the math will be the same. :)