Skip to content

bmmlab/Distributional-RL-Tail-Risk

Repository files navigation


Exploiting Distributional Temporal Difference Learning
To Deal With Tail Risk

This repository contains both the environment and the agents of the leptokurtosis project.

For a detailed description of the experiment, please refer to our paper.

Environment (envs):

  • Leptokurtosis

Implemented algorithms (agents):

  • Tabular SARSA
  • Categorical Temporal Difference Learning (a tabular version of Categorical distributional reinforcement learning)
  • (Efficient) Distributional Temporal Difference Learning:
    • Integration over reward distribution (sample average) method.
    • Maximum Likelihood Estimator (EM-MLE) method.

Prerequisites

see requirements.txt


Results replication


To reference this repository

@misc{distributionalRLTailRisk,
  author = {Peter Bossaerts, Shijie Huang and Nitin Yadav},
  title = {Exploiting Distributional Temporal Difference Learning To Deal With Tail Risk},
  year = {2020}
}

Key references

Outlier and Leptokurtosis:

Statistics:

  • Casella, G., & Berger, R. L. (2002). Statistical inference (Vol. 2). Pacific Grove, CA: Duxbury.
  • Schervish, M. J. (2012). Theory of statistics. Springer Science & Business Media.

(Distributional) Reinforcement Learning:

Primary:

Others:

About

Exploiting Distributional Temporal Difference Learning To Deal With Tail Risk

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages