-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is it necessary to rebuild the model every train iteration? #1
Comments
Hi. Good catch. I think, it should be the same and faster. But, make sure, gradients are not accumulated. |
@arghosh Thanks for the clarification! Your implementation using higher package seems rather neat, does it support distributed training? |
My code does not support distributed training. I don't think higher supports data parallel. But, you may pass the meta batch to different GPUs, do local meta step, compute base model gradients in each node; after that, DDP can handle that, I guess. |
@arghosh Thanks for the clarification! |
Hi, thanks for the great work!
I noticed that you rebuild the meta-model every iteration (L129), and I was wondering if that is necessary?
RobustMW-Net/trainer.py
Lines 122 to 130 in cabea1f
Would it have any difference or negative impact if i were to just build the meta-model before the loop and reload the model's
state_dict
every iteration (L130) instead?The text was updated successfully, but these errors were encountered: