Skip to content

Implementation of NeurIPS 2018 paper "Meta-Gradient Reinforcement Learning"

License

Notifications You must be signed in to change notification settings

IandRover/meta_gradient_RL

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Meta-Gradient Reinforcement Learning

Implementation of NeurIPS 2018 paper "Meta-Gradient Reinforcement Learning". For detailed explaination and implementation techniques, please refer to an article in Medium "Meta Learning — Meta-Gradient Reinforcement Learning — An Implementation".

We implemented A2C as the base algorithm under CartPole-v1 environment and consider γ (i.e. the discounting factor) as the only hyperparameter to be meta-learned/meta-adapted.

The maximum reward of the CartPole-v1 environment is 500. We constrained the number of training steps to be 500 due to computation and time limits. For each run, the training is terminated as soon as the agent obtains rewards exceeding than 475 (475 is the reward threshold specified by the CartPole-v1 environment).

We used neural networks to implement the actor and the value network, both of which comprised three fully connected layers with a hidden size of 40 and with ReLU as activation function. Log-softmax operation was used to normalize the sum of the output of the actor network to one. The table below is the hyper-parameters used in our experiment.

The hyperparameter we use were as follows:

hyperparameter
optimizer Adam
learning rate 0.01
sample size 30
initial gamma 0.99
alpha 0.0001

About

Implementation of NeurIPS 2018 paper "Meta-Gradient Reinforcement Learning"

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published