Skip to content

Trading with recurrent actor-critic reinforcement learning

Notifications You must be signed in to change notification settings

evgps/a3c_trading

Repository files navigation

A3C trading

Note: Sorry for misleading naming - please use A3C_trading.py for training and test_trading.py for testing.

Trading with recurrent actor-critic reinforcement learning - check paper and more detailed old report

Full_UML

Configuration: config.py

This file contains all the pathes and gloabal variables to be set up

Dataset: download from GDrive

After setting config.py please run this file to download and preprocess the data need for training and evaluation

Environment: trader_gym.py

OpenAI.gym-like environment class

Model: A3C_class.py

This file is containing AC_network, Worker and Test_Worker classes

Training: A3C_training.py

Run this file, preferrable in tmux. During training it will create files in tensorboard_dir and in model_dir

Testing: A3C_testing.ipynb

Jupyter notebook contains all for picturing

Cite as:

@article{ponomarev2019using, title={Using Reinforcement Learning in the Algorithmic Trading Problem}, author={Ponomarev, ES and Oseledets, IV and Cichocki, AS}, journal={Journal of Communications Technology and Electronics}, volume={64}, number={12}, pages={1450--1457}, year={2019}, publisher={Springer} }

About

Trading with recurrent actor-critic reinforcement learning

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published