Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

dose the code support MPI to speed up the training? #98

Closed
zhihaocheng opened this issue Aug 10, 2020 · 4 comments
Closed

dose the code support MPI to speed up the training? #98

zhihaocheng opened this issue Aug 10, 2020 · 4 comments

Comments

@zhihaocheng
Copy link

No description provided.

@keiohta
Copy link
Owner

keiohta commented Aug 10, 2020

APE-X type of parallelization which uses multiprocessing API has already been implemented.
https://github.com/keiohta/tf2rl/blob/master/tf2rl/algos/apex.py

Sometimes this speeds up training, but it depends on type of the environment and policy you use.

@zhihaocheng
Copy link
Author

@keiohta Thanks a lot. I want to run GAIL, but with the "--help", I still do not know how to config the number of cpu.

In addition, does the spectral normalization help to improve the performance of GAIL?

@keiohta
Copy link
Owner

keiohta commented Aug 12, 2020

@zhihaocheng Sorry for late reply.
We support distributed training only on ApeX style DDPG and DQN. We won't support inverse RL unless we find some interesting paper which uses the distributed version of IRL. However, if you do need to implement the code, you can do referring the apex code I sent you in the previous comment.

In addition, does the spectral normalization help to improve the performance of GAIL?

I'm sure it improves the performance. Please find below link and compare "accuracy" of discriminator. The lower (closer to 0.5, where discriminator cannot judge whether the inputs come from expert or current agent) the better.
#14 (comment)

@zhihaocheng
Copy link
Author

@keiohta thanks for your detailed reply, it is really helpful to me.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants