Skip to content

Code for the paper "Learning to Ski", in which I teach a computer how to play the Atari game Skiing.

Notifications You must be signed in to change notification settings

FragLegs/learning2ski

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

How to Reproduce the Results in "Learning to Ski"

Thank you for taking the time to read my research paper. This document is intended to walk you through the process of reproducing the results I reported.

NOTE: Please keep in mind as you read the paper that this project was an attempt to learn about reinforcement learning, not break any new ground in the research.

1. Installation

The dependencies can be installed via pip install -r requirements.txt. I recommend installing into a virtual environment.

Depending on your environment, you may need to install some system level packages first. On Ubuntu you need at least:

apt-get install -y python-dev cmake zlib1g-dev libjpeg-dev xvfb libav-tools xorg-dev python-opengl libboost-all-dev libsdl2-dev swig

See gym's pypi page for more info.

2. Running the Simulation

The simulation is run via the eval.py script. Running python eval.py -h will provide information about the possible arguments.

3. Baselines

The baselines can be run with the following commands:

  • Random python eval.py -e 150 --seed 42 random

  • Straight python eval.py -e 150 --seed 42 straight

4. Oracle

The oracle was computed with

python eval.py -r -e 3 --seed 42 human

However, the output will depend on the human playing it. If you would like to play, the keys "a", "s", and "d" control the skier:

  • a: Turn to the skier's right (your left)
  • s: Do nothing
  • d: Turn to the skier's left (your right)

The game will wait for you to press a key at every step. You can hold down a key to let it play faster.

5. Learning To Ski (L2S)

The L2S result in Table 2 can be run with:

python eval.py -e 150 --seed 42 l2s

6. Changing Experience Replay Memory Length

The values in Table 3 were run with the following commands:

python eval.py -e 150 --seed 42 --agent-args '{"batch_size": 500}' l2s

python eval.py -e 150 --seed 42 --agent-args '{"batch_size": 1000}' l2s

python eval.py -e 150 --seed 42 --agent-args '{"batch_size": 2500}' l2s

python eval.py -e 150 --seed 42 --agent-args '{"batch_size": 5000}' l2s

python eval.py -e 150 --seed 42 --agent-args '{"batch_size": 10000}' l2s

7. Changing Fitted Q-Iteration Frequency

The values in Table 4 were run with the following commands:

python eval.py -e 150 --seed 42 --agent-args '{"iteration_size": 10}' l2s

python eval.py -e 150 --seed 42 --agent-args '{"iteration_size": 30}' l2s

python eval.py -e 150 --seed 42 --agent-args '{"iteration_size": 60}' l2s

python eval.py -e 150 --seed 42 --agent-args '{"iteration_size": 90}' l2s

python eval.py -e 150 --seed 42 --agent-args '{"iteration_size": 120}' l2s

8. Components of Per-Episode Reward

Every run of eval.py creates a folder ./results/{agent-name}/{timestamp} with the following files:

  • agent_args.json: A json dictionary of any args that were passed in on the command line (for repeatability)

  • rewards.csv: A csv of the per-episode rewards, with columns for total reward, reward due to elapsed time, and number of slaloms missed

You can generate the graph in Figure 1 with the following command (replacing the paths with the location of the rewards.csv for the run you wish to plot and the path where you would like the result written):

python plot_rewards.py path/to/rewards.csv path/to/output.png

About

Code for the paper "Learning to Ski", in which I teach a computer how to play the Atari game Skiing.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Languages