Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Reproducibility of the results for the LLF (Table 1)? #4

Open
Vijayraven95 opened this issue Jun 15, 2022 · 1 comment
Open

Reproducibility of the results for the LLF (Table 1)? #4

Vijayraven95 opened this issue Jun 15, 2022 · 1 comment

Comments

@Vijayraven95
Copy link

Vijayraven95 commented Jun 15, 2022

Hello @hlml @ankitkv,

Thank you for your excellent work and in-depth analysis of the iterative training paradigm. I am currently having difficulty in reproducing the results of the experiments provided in the paper. If I run the model with the arguments (LLF mentioned in the repo) 'reset_layer_name as layer 4', I am getting around 71.37(N10) as final test accuracy in CUB dataset with smth (0.1) at the end of 10 generations. This value is around 1 percent less than the mentioned value of 72.47% (N10) (table 1). The accuracy at the end of the 3rd generation (N3) I am getting is 68.07% which is 2.7% less compared to 70.76% (Table 1). Have you resetted from block 3 or only block 4 for your LLF experiments? Any help regarding this would be better. Also in Table-1, you have mentioned LLF uses L={10,14}, corresponding to blocks 3 and 4 in ResNet18. Does that mean the results provided in Table 1 are a result of forgetting layers in block 3, block 4, and FC layer? Please provide some clarity on that.

Thank you!

The results of my reproducibility experiments are shown below with respect to each generation on CUB dataset with Label smoothening (0.1).
gen | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10
last_tst_acc1 | 59.37 | 65.21 | 68.07 | 68.55 | 70.06 | 70 | 70.54 | 70.52 | 70.3 | 70.09 | 71.37

@Vijayraven95 Vijayraven95 changed the title Reproduibility question? Reproducibility question? Jun 15, 2022
@Vijayraven95 Vijayraven95 changed the title Reproducibility question? Reproducibility of the results for the LLF (Table 1)? Jun 16, 2022
@hlml
Copy link
Owner

hlml commented Dec 22, 2022

Hi, sorry for the late reply! I believe CUB uses reset_layer_name as layer3, and uses a learning rate of 0.1. Also note that the 3rd generation equals to 4 rounds of training (the initial training + 3 generations of reset). You may be doing this already but just wanted to note that if not.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants