Skip to content

Commit

Permalink
Add Adversarial Policies to list of projects using Stable Bsaelines (…
Browse files Browse the repository at this point in the history
…#375)
  • Loading branch information
AdamGleave authored and araffin committed Jun 17, 2019
1 parent 0b7726e commit 45dcfb9
Showing 1 changed file with 12 additions and 0 deletions.
12 changes: 12 additions & 0 deletions docs/misc/projects.rst
Original file line number Diff line number Diff line change
Expand Up @@ -101,3 +101,15 @@ The Proximal Policy Optimization (PPO) algorithm is applied.
| Email: 6guelden@informatik.uni-hamburg.de
| Video: https://www.youtube.com/watch?v=laGrLaMaeT4
| GitHub: https://github.com/RGring/drl_local_planner_ros_stable_baselines
Adversarial Policies: Attacking Deep Reinforcement Learning
-----------------------------------------------------------

Uses Stable Baselines to train *adversarial policies* that attack pre-trained victim policies in a zero-sum multi-agent environments.
May be useful as an example of how to integrate Stable Baselines with `Ray <https://github.com/ray-project/ray>`_ to perform distributed experiments and `Sacred <https://github.com/IDSIA/sacred>`_ for experiment configuration and monitoring.

| Authors: Adam Gleave, Michael Dennis, Neel Kant, Cody Wild
| Email: adam@gleave.me
| GitHub: https://github.com/HumanCompatibleAI/adversarial-policies
| Paper: https://arxiv.org/abs/1905.10615
| Website: https://adversarialpolicies.github.io

0 comments on commit 45dcfb9

Please sign in to comment.