Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
Showing
1 changed file
with
17 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,17 @@ | ||
## Quantifying Generalization in Reinforcement Learning | ||
> Karl Cobbe, Oleg Klimov, Chris Hesse, Taehoon Kim, John Schulman | ||
### Abstract | ||
In this paper, we investigate the problem of overfitting in deep reinforcement learning. | ||
|
||
Among the most common benchmarks in RL, it is customary to use the same environments for both training and testing. | ||
|
||
This practice offers relatively little insight into an agent’s ability to generalize. | ||
|
||
We address this issue by using procedurally generated environments to construct distinct training and test sets. | ||
|
||
Most notably, we introduce a new environment called CoinRun, designed as a benchmark for generalization in RL. | ||
|
||
Using CoinRun, we find that agents overfit to surprisingly large training sets. | ||
|
||
We then show that deeper convolutional architectures improve generalization, as do methods traditionally found in supervised learning, including L2 regularization, dropout, data augmentation and batch normalization. |