Skip to content
master
Switch branches/tags
Code
This branch is 18 commits ahead, 222 commits behind openai:master.
Contribute
Fetch upstream

Latest commit

 

Git stats

Files

Permalink
Failed to load latest commit information.
Type
Name
Latest commit message
Commit time
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Important: Deprecated version. Update to latest arxiv version coming soon. Please use the resources in our main repository in the meanwhile.

RUDDER efficiently learns optimal policies in finite Markov decision processes with delayed rewards. With the following links you can find:

Baselines incl. RUDDER

Uses RUDDER on PPO to solve environments with delayed rewards. Our RUDDER paper is available at https://arxiv.org/abs/1806.07857.

See https://github.com/widmi/rudder-a-practical-tutorial for a quick tutorial on reward redistribution using a pytorch implementation and code-snippets that you can run on your CPU.

Our code is based on the OpenAI Baselines package, in which we included our implementation of RUDDER for PPO for ATARI games with delayed rewards. Additionally, the Tensorflow Layer Library (TeLL) package (v1.0.0) is required for RUDDER.

RUDDER for PPO2 and documentation for the configuration is located in the folder baselines/ppo2_rudder.

Modified/new files:

  • baselines/common/atari_wrappers.py
  • baselines/common/distributions.py
  • baselines/common/vec_env/vec_frame_stack.py
  • baselines/ppo2_rudder/
  • logger.py
  • README.md

Videos are available at:

RUDDER - Venture

RUDDER - Bowling

About

RUDDER for ATARI games with delayed rewards in OpenAI Baselines package

Resources

License

Releases

No releases published

Packages

No packages published