Skip to content

Minimal implementation of clipped objective Proximal Policy Optimization (PPO) in PyTorch

License

Notifications You must be signed in to change notification settings

rhklite/Parallel-PPO-PyTorch

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Parallel PPO-PyTorch

A parallel agent training version of Proximal Policy Optimization with clipped objective.

Usage

  • To test a pre-trained network : run test.py
  • To train a new network : run parallel_PPO.py
  • All the hyperparameters are in the file, main function

Results

CartPole-v1 LunarLander-v2
cartpole lander

Dependencies

Trained and tested on:

Python 3.6
PyTorch 1.3
NumPy 1.15.3
gym 0.10.8
Pillow 5.3.0

TODO

  • implement Conv net based training

Setting up Conda Environment

  • conda env export | grep -v "^prefix: " > environment.yml to export the file environment.yml
  • conda create -f environment.yml to create the conda environment used for training

References

About

Minimal implementation of clipped objective Proximal Policy Optimization (PPO) in PyTorch

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%