Skip to content

hagerrady13/DQN-PyTorch

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

18 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

DQN-PyTorch

A PyTorch implementation of Human-level control through deep reinforcement learning

Table of Contents:

Project Structure:

├── agents
|  └── dqn.py # the main training agent for the dqn
├── graphs
|  └── models
|  |  └── dqn.py
|  └── losses
|  |  └── huber_loss.py # contains huber loss definition
├── datasets  # contains all dataloaders for the project
├── utils # utilities folder containing input extraction, replay memory, config parsing, etc
|  └── assets
|  └── replay_memory.py
|  └── env_utils.py
├── main.py
└── run.sh

Environments:

CartPole V0:

Loss during training:

alt text

Number of durations per Episode:

alt text

Usage:

  • To run the project, you need to add your configurations into the folder configs/ as fround here
  • sh run.sh
  • To run on a GPU, you need to enable cuda in the config file.

Requirements:

  • Pytorch: 0.4.0
  • torchvision: 0.2.1
  • tensorboardX: 0.8

Check requirements.txt.

Future Work:

  • Test DQN on a more complex environment such as MS-Pacman

References:

License:

This project is licensed under MIT License - see the LICENSE file for details.

Releases

No releases published

Packages

No packages published