Skip to content

Data-driven offline simulation for online reinforcement learning: benchmark and baselines

License

Notifications You must be signed in to change notification settings

microsoft/rl-offline-simulation

Repository files navigation

Offline Learner Simulation

Offline Simulations for Online Reinforcement Learning

This repository contains code used in the NeurIPS 2022 Offline RL workshop paper Towards Data-Driven Offline Simulations for Online Reinforcement Learning by Shengpu Tang, Felipe Vieira Frujeri, Dipendra Misra, Alex Lamb, John Langford, Paul Mineiro, Sebastian Kochman.

As part of this research project, we have started creation of the offsim4rl library, containing algorithms for Offline Learner Simulation (OLS), like Per-State Rejection Sampling (originally proposed by Mandel et al, 2016), as well as tools for evaluating different OLS methods. We share it in the early stage of development and are looking forward to seeing how the RL research community takes it further, either via contributions or project forks.

To reproduce the experiments from the paper, please use the neurips_workshop_2022 branch, which includes experimental notebooks. The notebooks will be removed from the main branch, such that it is easier to continue developing the offsim4rl library without breaking the workshop experiments. However, if you intend to use the offsim4rl library (e.g., to run offline learner simulation using your own dataset), you're better off using the main branch.

Getting Started

To install the offsim4rl library, run the following steps.

Note: currently, we support Linux Ubuntu (can be via Windows Subsystem for Linux). Other platforms should work, but haven't been tested.

  1. Make sure the native dependencies are installed:
sudo apt update
sudo apt install libopenmpi-dev
  1. [Optional] Create a virtual environment, e.g., using conda:
conda create -n offsim4rl python=3.7
conda activate offsim4rl
  1. Clone this repository and install the library in the development mode:
git clone https://github.com/microsoft/rl-offline-simulation.git
cd rl-offline-simulation
pip install -r requirements.txt
pip install -e .
  1. [Optional] Run unit tests.
pytest

Note: many dependencies are listed in both "requirements.txt" and in "setup.py".

  • The "requirements.txt" file contains exact versions of the dependencies, some of which are required for our test pipeline to pass. If you'd like to make sure everything is running correctly, would like to reproduce our results, or would like to contribute to the project, we recommend installing the dependecies via "pip install -r requirements.txt", before installing the library.
  • The "setup.py" offers more flexibility in terms of versioning dependencies. If you intend to use offsim4rl as a library in your own project, and you'd like to use different versions of some dependencies than the ones we specified in "requirements.txt", you may skip "pip install -r requirements.txt" and run "pip install -e ." directly.

Citation

If you found this repository useful in your research, please cite our work using the following BibTeX:

@inproceedings{
    tang2022towards,
    title={Towards Data-Driven Offline Simulations for Online Reinforcement Learning},
    author={Tang, Shengpu and Frujeri, Felipe Vieira and Misra, Dipendra and Lamb, Alex and Langford, John and Mineiro, Paul and Kochman, Sebastian},
    booktitle={3rd Offline RL Workshop: Offline RL as a ''Launchpad''},
    year={2022},
    url={https://arxiv.org/abs/2211.07614}
}

Contribute

This project welcomes contributions and suggestions. Most contributions require you to agree to a Contributor License Agreement (CLA) declaring that you have the right to, and actually do, grant us the rights to use your contribution. For details, visit https://cla.opensource.microsoft.com.

When you submit a pull request, a CLA bot will automatically determine whether you need to provide a CLA and decorate the PR appropriately (e.g., status check, comment). Simply follow the instructions provided by the bot. You will only need to do this once across all repos using our CLA.

This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact opencode@microsoft.com with any additional questions or comments.

Trademarks

This project may contain trademarks or logos for projects, products, or services. Authorized use of Microsoft trademarks or logos is subject to and must follow Microsoft's Trademark & Brand Guidelines. Use of Microsoft trademarks or logos in modified versions of this project must not cause confusion or imply Microsoft sponsorship. Any use of third-party trademarks or logos are subject to those third-party's policies.

About

Data-driven offline simulation for online reinforcement learning: benchmark and baselines

Resources

License

Code of conduct

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •  

Languages