Deep RL Bootcamp Lecture 4A: Policy Gradients

Sdílet
Vložit
  • čas přidán 28. 05. 2024
  • Instructor: Pieter Abbeel
    Lecture 4A Deep RL Bootcamp Berkeley August 2017
    Policy Gradients

Komentáře • 43

  • @naeemajilforoushan5784
    @naeemajilforoushan5784 Před měsícem +2

    After 5 years still, the lecture is a great video, thank you a lot

  • @jony7779
    @jony7779 Před 4 lety +16

    Every time I forget how policy gradients work exactly, I just come back here and watch starting at 9:30

    • @andreasf.3930
      @andreasf.3930 Před 3 lety +3

      And every time you visited this video, you forgot where to start watching. Thats why you posted this comment. Smart guy!

  • @bhargav975
    @bhargav975 Před 6 lety +34

    This is the best lecture I have seen on policy gradient methods. Thanks a lot.

  • @marloncajamarca2793
    @marloncajamarca2793 Před 6 lety +3

    Great Lecture!!!! Pieter's explanations are just a gem!

  • @auggiewilliams3565
    @auggiewilliams3565 Před 4 lety +1

    I must say that in more than 6 months, this is by far the best lecture/ material I have come across that was able to make me understand what policy gradient method actually is. I really praise this work. :) Thank you.

  • @ericsteinberger4101
    @ericsteinberger4101 Před 6 lety +9

    Amazing lecture! Love how Pieter explains the math. super easy to understand.

  • @synthetic_paul
    @synthetic_paul Před 4 lety +5

    Honestly I can’t keep up without seeing what he’s pointing at. Gotta pause and search around the screen each time he says “this over here”

    • @akarshrastogi3682
      @akarshrastogi3682 Před 3 lety +2

      Exactly. "This over here" has got to be the most uttered phrase in this lecture. So frustrating.

  • @johnnylima1337
    @johnnylima1337 Před 6 lety +5

    It's such a good lecture, I'm stopping to ask myself why it was so easy to cover such significant information with full understanding

  • @ashishj2358
    @ashishj2358 Před 3 lety

    Best lecture on Policy Gradients hands down. Has covered some worth noting superficial details of many papers as well.

  • @faizanintech1909
    @faizanintech1909 Před 6 lety

    Awesome instructor.

  • @dustinandrews89019
    @dustinandrews89019 Před 6 lety +1

    I got a lot out of this lecture in particular. Thank you.

  • @user-wp6lp3ec9q
    @user-wp6lp3ec9q Před 4 lety

    Very good lecture about policy gradient method. I have looked through a lot of articles and was understanding almost everything, but your derivation explanation is really the best. It just opened my eyes and showed the whole picture. Thank you very much!!

  • @bobsmithy3103
    @bobsmithy3103 Před rokem

    amazing work. super understandable, concise and information dense.

  • @DhruvMetha
    @DhruvMetha Před 3 lety

    Wow, this is beautiful!

  • @nathanbittner8307
    @nathanbittner8307 Před 6 lety

    excellent lecture. Thank you for sharing.

  • @ethanjyx
    @ethanjyx Před 5 lety

    wow damn this is so well explained and the last video is very entertaining.

  • @norabelrose198
    @norabelrose198 Před 2 lety

    The explanation of the derivation of policy gradient is really nice and understandable here

  • @JadtheProdigy
    @JadtheProdigy Před 5 lety

    best lecturer in series

  • @suertem1
    @suertem1 Před 4 lety

    Great lecture, thanks

  • @sharmakartikeya
    @sharmakartikeya Před 5 měsíci

    I might be missing a simple concept here but how are we increasing/decreasing the grad log probability of the actions using the gradient of U(theta)? I get that positive return for a trajectory will make the gradient of U positive and so theta will be increased in favour of those trajectories but how is it increasing grad log prob?

  • @Diablothegeek
    @Diablothegeek Před 6 lety

    Awesome!! Thanks

  • @ProfessionalTycoons
    @ProfessionalTycoons Před 5 lety

    great talk!

  • @biggeraaron
    @biggeraaron Před 5 lety +1

    Where can i buy his T-shirt?

  • @JyoPari
    @JyoPari Před 4 lety +1

    Instead of having a baseline, why not make your reward function be negative for undesired scenarios and positive for good ones? Great lecture!

  • @keqiaoli4617
    @keqiaoli4617 Před 3 lety +1

    why a good "R" would increase the probability of path??? Please help me

  • @muratcan__22
    @muratcan__22 Před 3 lety +3

    nice but hard to follow without knowing what "this" refers to. I hope my guesses were right :)

  • @emilterman6924
    @emilterman6924 Před 5 lety

    It would be nice to see what laboratories they had (what exercises)

    • @Procuste34iOSh
      @Procuste34iOSh Před 3 lety

      dont know if ur still interested, but the labs are on the bootcamp website

  • @karthik-ex4dm
    @karthik-ex4dm Před 5 lety

    PG is awesome!!!
    Doesn't depend on environment Dynamics really?? Wow
    All the pain and stress just goes away when we see our algorithms working😇😇

  • @ishfaqhaque1993
    @ishfaqhaque1993 Před 4 lety

    23:20- Gradient of expectation is expectation of gradient "under mild assumptions". What are those assumptions?

    • @joaogui1
      @joaogui1 Před 4 lety +2

      math.stackexchange.com/questions/12909/will-moving-differentiation-from-inside-to-outside-an-integral-change-the-resu

  • @shaz7163
    @shaz7163 Před 6 lety

    very nice :)

  • @isupeene
    @isupeene Před 3 lety +2

    The guy in the background at 51:30

  • @elzilcho222
    @elzilcho222 Před 5 lety +1

    could you train a robot for 2 weeks in the real world then use those trained parameters to optimize a virtual environment? You know.. making the virtual environment very close to the real world?

    • @OfficialYunas
      @OfficialYunas Před 5 lety +1

      Of course you could. It's the opposite of what OpenAI does when they train a model in a virtual environment and deploy it in reality.

    • @soutrikband
      @soutrikband Před 5 lety

      Real world is very complicated with model uncertainties, friction, wear and tear and what have you...
      Simulators can come close , but we cannot expect them to fully mimic real world phenomena.

  • @richardteubner7364
    @richardteubner7364 Před 6 lety +1

    1:11 why are DQNs and friends Dynamic Programming Methods? I mean the neural network works as functions approximator to satisfy Bellmans eqn. , but still Backprop is the workhorse. In my opinion DQNs are much more similar to PG methods than to Bellman Updates??! And another issue with RL Landscape slide is where the heck are model based RL algos?? This slide should be renamed to model free RL landscape.

  • @arpitgarg5172
    @arpitgarg5172 Před 5 lety +11

    If you can't explain it like Pieter Abbeel or Andrew NG then you don't understand it well enough.

  • @piyushjaininventor
    @piyushjaininventor Před 5 lety

    Can you share ppt??

    • @luxorska5143
      @luxorska5143 Před 5 lety +3

      You can find all the slides and the other lectures here:
      sites.google.com/view/deep-rl-bootcamp/lectures

  • @MarkoTintor
    @MarkoTintor Před 3 lety

    ... you can use "a", and the math will be the same. :)