Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why training ends at epoch 50? #4

Closed
yuxudong20 opened this issue Sep 12, 2022 · 4 comments
Closed

Why training ends at epoch 50? #4

yuxudong20 opened this issue Sep 12, 2022 · 4 comments

Comments

@yuxudong20
Copy link

Hello, I have tried to reproduce ATAC's results in the paper. However, when I run the official codes, the experiment automatically ends at epoch 50. I cannot find where the problem is? Could you give me some help?
For example, I have run 'python scripts/main.py -e hopper-medium-expert-v2 --gpu_id 0 --seed 15'. Are there any other hyperparameters that need to be given?
@chinganc

@chinganc
Copy link
Collaborator

Hi @yuxudong20, there was pytorch bug created in our refactoring which causes errors when using gpu. It's fixed now. Can you please give it a try? Thanks.

@yuxudong20
Copy link
Author

Hi, @chinganc, thanks a lot. It seems that current codes can run 3000 epochs. However, the learning curves suddenly fall down, as in the following figure. Is this correct?
image

@yuxudong20
Copy link
Author

And the amplification shows that it falls down at epoch 50.
image

@chinganc
Copy link
Collaborator

This is normal: Currently, it by default uses 50 epochs to do warmstart, which is just BC for the policy, and then ATAC starts. The drop you see there is the first epoch of ATAC.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants