Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can't reproduce experimental results #69

Open
Yingdong-Hu opened this issue Jan 26, 2021 · 5 comments
Open

Can't reproduce experimental results #69

Yingdong-Hu opened this issue Jan 26, 2021 · 5 comments

Comments

@Yingdong-Hu
Copy link

I tried to run probing tasks for different Atari environments, using the following command:

python -m scripts.run_probe --method infonce-stdim --env-name {env_name}

I did not change any code, just tried different game, including PongNoFrameskip-v4, BowlingNoFrameskip-v4, BreakoutNoFrameskip-v4, HeroNoFrameskip-v4.

However, only the F1 score for pong matches the score reported in the paper. The F1 scores of the other three games are far worse than the score shown in the paper (for bowling, I got 0.22).

I check the training loss logged in wandb, it seems that training has not converged at all. See the figure below.

training loss

How to get the F1 socres reported in the paper? Am I missing something?

@ankeshanand
Copy link
Member

Hi Alex,

That's quite strange, those loss curves don't match what we saw in our runs. There definitely seems to be a bug or a numerical error in your runs since some of the loss values are blowing up to quite high values.

Can you confirm what pytorch version you are using? Also, I just looked up on our W&B dashboard, and this is how the loss curves look like for Bowling:

image

@Yingdong-Hu
Copy link
Author

Yingdong-Hu commented Jan 26, 2021

My pytorch version is 1.7.0. What's your pytorch version?

@ankeshanand
Copy link
Member

We used pytorch 1.1.0 for our runs back then

@ankeshanand
Copy link
Member

So check if an older version of pytorch gives the correct results, this does look like a numerical stability bug. Also try increasing Adam's eps value. If that doesn't fix it, it's possible (though unlikely since we were quite careful) that a bug was introduced when we were cleaning up the code. I can try to debug the source of divergence then, but might only get time after the ICML deadline.

@Yingdong-Hu
Copy link
Author

Thank you, let me try it first.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants