Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fail to Reproduce SpikFormer-8-512 on Imagenet #12

Open
CatherineCloud opened this issue Oct 8, 2023 · 4 comments
Open

Fail to Reproduce SpikFormer-8-512 on Imagenet #12

CatherineCloud opened this issue Oct 8, 2023 · 4 comments

Comments

@CatherineCloud
Copy link

Hi dear authors,

Thank you guys for your amazing work. May I ask if there is anything important that requires modification based on the current code uploaded to github? I was trying to reproduce the result for Spikformer-8-512 on Imagenet. However, the result I got was quite different from the one shown in the paper.
image
image
When it comes to epoch 25, the top1-acc shown in the paper is clearly over 50%, while my reproduced result is barely over 40%. I used the exactly same code as in this github, except I used a batch size of 24 due to my GPU limitation.

Please enlight me. Thank you so much!

@ZK-Zhou
Copy link
Owner

ZK-Zhou commented Oct 9, 2023

Please provide the curve of the full training 300 epoches, because the training process will be different with different hyperparameters, rather than just looking at the partial convergence curve

@CatherineCloud
Copy link
Author

The hyperparameters are not changed as the one shown in this github repo.
Additionally, may I ask how long did you guys take to train thie model on ImageNet?

@ZK-Zhou
Copy link
Owner

ZK-Zhou commented Oct 10, 2023

Hi, batch size is one of the most important hyperparameters 。We use 8 Nvidia V100 for 8 days.

@warming151
Copy link

Hi. May I ask which PyTorch version you installed? I couldn't install pytorch==1.10.0+cu111... Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants