Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Questions about the default setting of parameter "warmup_epoch" #1

Closed
XiangQiu42 opened this issue Jun 24, 2021 · 1 comment
Closed

Comments

@XiangQiu42
Copy link

XiangQiu42 commented Jun 24, 2021

Dear dayu11,

In your code, the parameter "warmup_epoch" is -1 by default, and it seems that you did not assign a new value to this parameter in your experiment.

However, if "warmup_epoch" is -1, it seems that at the train function:

if(epoch + 1 == args.warmup_epoch):
# take a snapshot of current model for computing historical update
net.module.update_init_weight()

This part of code will never be executed.

So, Is that correct or I am missing something?

@dayu11
Copy link
Owner

dayu11 commented Jun 26, 2021

Yes, you are correct. By default, the code directly decomposes the weight matrix itself instead of the historical update.

It turns out that directly decomposing the weight matrix also works. This may because as the training progresses, the weight matrix itself contains enough information about the historical update, i.e., we do not need to subtract the weight at initialization from the current weight to get the historical update. Still, you can set warmup_epoch>1 to compute the historical update explicitly.

@dayu11 dayu11 closed this as completed Jun 26, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants