-
Notifications
You must be signed in to change notification settings - Fork 115
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Possible bugs #61
Comments
You may use the distilled dataset however you want. Whether to reparametrize the network shouldn't really affect results. However, other changes in training (e.g., network architectures, learning rates, epochs, etc.) can affect the results non-trivially. |
Re your original question: that shouldn't happen if you are running the code as-is, since there is this line
|
It might be a torch version issue then because I didn't see this before. I guess you can try to change code to test
Although I haven't experienced the error you are seeing so I can't guarantee correctness. |
May I ask this question? In your code, are you using real data to generate distilled dataset? and then training on the distilled dataset? However, in the function init_data_optim, I see distill_label and distill_data are generating randomly. I understand that this is the initialization, but in which part of your code updates the distilled data? would you mind if you could point it out for me? Thank you very much. |
@hailuu684 In this code, distilled data are initialized as random noise. But yes, many works show that initializing from real images can work better in certain cases. It is not implemented here. |
Hello,
I am testing your repo and I got this error. I am not sure if this is an issue in different torch versions or incompatibilities.
I got an error in d.grad.add_(g). It says 'NoneType' object has no attribute 'add_'. I printed the type of 'd', it is obviously a Tensor type. I do not understand why can cause this problem.
The text was updated successfully, but these errors were encountered: