Reproducing the results of "Playing Atari with Deep Reinforcement Learning" by DeepMind
Switch branches/tags
Nothing to show
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Permalink
Failed to load latest commit information.
Hybrid
articles
doc
libraries
sandbox
src
test Preprocessor hopefully parallelised. May 23, 2014
tools
.gitignore Ignore data files Nov 19, 2014
LICENSE
README.md
install.sh
install_noGUI.sh

README.md

Replicating-DeepMind

Reproducing the results of "Playing Atari with Deep Reinforcement Learning" by DeepMind. All the information is in our Wiki.

Progress: System is up and running on a GPU cluster with cuda-convnet2. It can learn to play better than random but not much better yet :) It is rather fast but still about 2x slower than DeepMind's original system. It does not have RMSprop implemented at the moment which is our next goal.

Note 1: You can also check out a popular science article we wrote about the system to Robohub.

Note 2: Nathan Sprague has a implementation based on Theano. It can do fairly well. See his github for more details.