Skip to content

Commit

Permalink
Update UNREAL.md
Browse files Browse the repository at this point in the history
  • Loading branch information
tigerneil committed Feb 23, 2017
1 parent 4b9b464 commit f4e2721
Showing 1 changed file with 1 addition and 0 deletions.
1 change: 1 addition & 0 deletions UNREAL.md
Original file line number Diff line number Diff line change
Expand Up @@ -9,3 +9,4 @@ This paper brings together the state-of-the-art Asynchronous Advantage Actor-Cri
* Auxiliary control tasks
* Auxiliary reward tasks

![(a) The base agent is a CNN-LSTM agent trained on-policy with the A3C loss (Mnih et al., 2016). Observations, rewards, and actions are stored in a small replay buffer which encapsulates a short history of agent experience. This experience is used by auxiliary learning tasks. (b) Pixel Control – auxiliary policies Qaux are trained to maximise change in pixel intensity of different regions of the input. The agent CNN and LSTM are used for this task along with an auxiliary deconvolution network. This auxiliary control task requires the agent to learn how to control the environment. (c) Reward Prediction – given three recent frames, the network must predict the reward that will be obtained in the next unobserved timestep. This task network uses instances of the agent CNN, and is trained on reward biased sequences to remove the perceptual sparsity of rewards. (d) Value Function Replay – further training of the value function using the agent network is performed to promote faster value iteration. Further visualisation of the agent can be found in https://youtu.be/Uz- zGYrYEjA](images/unreal.png)

0 comments on commit f4e2721

Please sign in to comment.