Skip to content

dojeon-ai/PLASTIC

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Improving Input & Label Plasticity for Sample Efficient RL

This repository is an official PyTorch implementation of the paper, PLASTIC: Improving Input and Label Plasticity for Sample Efficient Reinforcement Learning, NeurIPS 2023.

arXiv / slide / poster

Authors: Hojoon Lee*, Hanseul Cho*, Hyunseung Kim*, Daehoon Gwak, Joonkee Kim, Jaegul Choo, Se-Young Yun, and Chulhee Yun.

Instructions

To run the synthetic experiments, please follow the instructions of Readme.md in the folder synthetic.

To run the Atari-100k experiments, please follow the instructions of Readme.md in the folder atari.

To run the DMC-M experiments, please follow the instructions of Readme.md in the folder dmc.

Results

plot

  • Left: Performance of Synthetic Experiments. Layer Normalization (LN) and SharpenessAware Minimization (SAM) considerably enhance input plasticity, while their effect on label plasticity is marginal. Conversely, Concatenated ReLU (CReLU) and periodic reinitialization (Reset) predominantly improve label plasticity with subtle benefits on input plasticity.

  • Right: Performance of RL Benchmarks. PLASTIC consistently outperforms individual methods, highlighting the synergistic benefits of its integrated approach.

Citations

@inproceedings{lee2023plastic,
  title={PLASTIC: Improving Input and Label Plasticity for Sample Efficient Reinforcement Learning},
  author={Lee, Hojoon and Cho, Hanseul and Kim, Hyunseung and Gwak, Daehoon and Kim, Joonkee and Choo, Jaegul and Yun, Se-Young and Yun, Chulhee},
  booktitle={Thirty-seventh Conference on Neural Information Processing Systems},
  year={2023}
}

Contact

For personal communication, please contact Hojoon Lee, Hanseul Cho, or Hyunseung Kim at

{joonleesky, jhs4015, mynsng}@kaist.ac.kr.

About

Code for the paper "PLASTIC: Improving Input and Label Plasticity for Sample Efficient Reinforcement Learning" (NeurIPS 2023)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published