Skip to content

icml2019-anonymous-author/Action-Robust-Reinforcement-Learning

Repository files navigation

Action-Robust-Reinforcement-Learning

Code accompanying the anonymous submission to ICML 2019 of the paper "Action Robust Reinforcement Learning and Applications in Continuous Control"

Requirements:

Howto train:

python3.6 main.py --updates_per_step 10 --env-name "Hopper-v2" --alpha 0.1 --method pr_mdp

Where method can take 3 values mdp pr_mdp or nr_mdp, where pr/nr are the probabilistic robust and noisy robust as defined in the paper.

All results are saved in the models folder.

Howto evaluate:

Once a model has been trained, run:

python3.6 test.py --eval_type model

where --eval_type model will evaluate for model (mass) uncertainty and --eval_type model_noise will create the 2d visualizations.

Howto visualize:

See Comparison_Plots.ipynb for an example of how to access and visualize your models.

Our approach is built upon the DDPG implementation by https://github.com/ikostrikov/pytorch-ddpg-naf

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages