Skip to content
/ PDS Public

Codes accompanying the paper "The Provable Benefits of Unsupervised Data Sharing for Offline Reinforcement Learning" (ICLR 2023)

License

Notifications You must be signed in to change notification settings

YiqinYang/PDS

Repository files navigation

Provable Benefits of Unsupervised Data Sharing in Offline RL

This is a jax implementation of PDS on Datasets for Deep Data-Driven Reinforcement Learning (D4RL), the corresponding paper is The provable benefits of unsupervised data sharing for offline reinforcement learning.

Quick Start

For experiments on D4RL, our code is implemented based on IQL:

$ python3 train_data_sharing.py --env_name=walker2d-expert-v2 --source_name=walker2d-random-v2 --config=configs/mujoco_config.py --data_share=learn  --target_split=0.05  --source_split=0.1

Citing

If you find this open source release useful, please reference in your paper (it is our honor):

@article{hu2023provable,
  title={The provable benefits of unsupervised data sharing for offline reinforcement learning},
  author={Hu, Hao and Yang, Yiqin and Zhao, Qianchuan and Zhang, Chongjie},
  journal={arXiv preprint arXiv:2302.13493},
  year={2023}
}

Note

About

Codes accompanying the paper "The Provable Benefits of Unsupervised Data Sharing for Offline Reinforcement Learning" (ICLR 2023)

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published