Skip to content

A replication of DeepMind's 2016 Nature publication, "Mastering the game of Go with deep neural networks and tree search," details of which can be found on their website.

License

memect/AlphaGo

 
 

Repository files navigation

AlphaGo Replication

This project is a replication/reference implementation of DeepMind's 2016 Nature publication, "Mastering the game of Go with deep neural networks and tree search," details of which can be found on their website. This implementation uses Python and Keras - a decision to prioritize code clarity, at least in the early stages.

Build Status

Current project status

This is not yet a full implementation of AlphaGo. Development is being carried out on the develop branch.

We are still early in development. There are quite a few pieces to AlphaGo that can be written in parallel. We are currently focusing our efforts on the supervised and "self-play" parts of the training pipeline because the training itself may take a very long time..

Updates were applied asynchronously on 50 GPUs... Training took around 3 weeks for 340 million training steps

-Silver et al. (page 8)

How to contribute

See the 'Contributing' document.

About

A replication of DeepMind's 2016 Nature publication, "Mastering the game of Go with deep neural networks and tree search," details of which can be found on their website.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • JavaScript 52.1%
  • Python 36.8%
  • CSS 10.6%
  • HTML 0.5%