1. 程式人生 > >Deep RL Bootcamp Lecture 4A: Policy Gradients

Deep RL Bootcamp Lecture 4A: Policy Gradients

spec incr any 9.png eal match sim AD tween

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

in policy gradient, "a" is replaced by "u" usually.

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

use this new form to estimate how good the update is.

技術分享圖片

技術分享圖片

技術分享圖片

If all three path show positive reward, should the policy increase the posibility of all the sampling?

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

monte carlo estimate

技術分享圖片

TD estimate

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

技術分享圖片

2 weeks to train as respect to real world time scale.

but could be faster in emulator (MOJOCO).

we don‘t know whether a set of hyperparameter is going to work until enough interations have past. So it‘s kind of tricky, and using emulator could alleviate this problem.

question: how to transform learnt knowledge of robot to real life if we are not sure about the match between simulator and real world?

Randomly initilize many simulator and see the robustness of the algorithm

技術分享圖片

技術分享圖片

this video shows that even a robot with two years of endeavor of a group of experts still isn‘t good at locomotion

hindsight experience replay

Marcin Richard from OpenAI

the program is set to find the best way to get pizza, but when the agent find a ice cream, the agent realizes that ice cream, corresponding to a higher reward, is the exact thing it wants to get.

Deep RL Bootcamp Lecture 4A: Policy Gradients