-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
data recording and saving method #1079
Comments
I would suggest using an environment wrapper for that. At the moment tianshou is primarily an algorithm library, not focused on wrappers. In fact, you could just use the wrapper from SB3 together with tianshou. Let me know if this answers your question |
Hi, |
The best way would be to use an env wrapper. Note that in all examples you can create your own env factory with your own wrapper. I'll try to add a tutorial on how to do that soon. Apart from that, you can probably use a custom logger. You can also access the buffer directly during training through the trainer, all rewards are saved there. In very near future we will add support for callbacks during training, which then would provide the simplest way for saving custom data (see #977 #895) |
Thanks for your brilliant work! |
Hi,
When I use SubprocVectorEnv, I want to record the rewards from all environments. Do you have similar function just like VecMonitor in SB3 (https://stable-baselines3.readthedocs.io/en/master/guide/vec_envs.html#stable_baselines3.common.vec_env.VecMonitor)?
The text was updated successfully, but these errors were encountered: