Skip to content

HaoTse/mountain-car-v0

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

MountainCar-v0

Requirements

Enviornment

Prerequisite

pip install -r requirements.txt

Goal

  • How to use OpenAI gym to do RL research
  • How to implement RL algorithms
  • How to evaluate your algorithms

Usage

  • training
usage: train.py [-h] [-m METHOD] [-lr LEARNING_RATE] [-rd REWARD_DECAY]

optional arguments:
  -h, --help            show this help message and exit
  -m METHOD, --method METHOD
                        Choose which rl algorithm used (QL, SARSA, or DQN)
  -lr LEARNING_RATE, --learning_rate LEARNING_RATE
                        Learning rate
  -rd REWARD_DECAY, --reward_decay REWARD_DECAY
                        Reward decay

  • testing
usage: test.py [-h] [-m METHOD] [-t TEST]

optional arguments:
  -h, --help            show this help message and exit
  -m METHOD, --method METHOD
                        Choose which rl algorithm used (DQN)
  -t TEST, --test TEST  The test model path

Result

Use plot.py to plot the learning curve.

  • learning curve (steps)

leanring curve (steps)

  • learning curve (rewards)

leanring curve (rewards)

  • After 10 episode, the average step is 152.9, and the average reward is 58.8.

Analysis

See report.pdf.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages