You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I had to stop the session before it finished, again different checkpoint models were saved on the folder /local_path/logs/student/checkpoints. Including pth files for G,D, optim-0,optim-1,A-0,A1,A2 and A3
Progress seems OK on the local_path/logs/student/eval folder
I tried to resume distilling with the command line:
Load network at /local_path/logs/student/checkpoints/latest_net_G.pth
Traceback (most recent call last):
File "distill.py", line 13, in <module>
trainer = Trainer('distill')
File "/content/CAT/trainer.py", line 80, in __init__
model.setup(opt)
File "/content/CAT/distillers/base_inception_distiller.py", line 260, in setup
self.load_networks(verbose)
File "/content/CAT/distillers/inception_distiller.py", line 203, in load_networks
super(InceptionDistiller, self).load_networks()
File "/content/CAT/distillers/base_inception_distiller.py", line 368, in load_networks
self.opt.restore_student_G_path, verbose)
File "/content/CAT/utils/util.py", line 139, in load_network
net.load_state_dict(weights)
File "/usr/local/lib/python3.7/dist-packages/torch/nn/modules/module.py", line 1052, in load_state_dict
self.__class__.__name__, "\n\t".join(error_msgs)))
RuntimeError: Error(s) in loading state_dict for InceptionGenerator:
Missing key(s) in state_dict: "down_sampling.1.bias", "down_sampling.2.weight", "down_sampling.2.bias", "down_sampling.2.running_mean", "down_sampling.2.running_var", "down_sampling.2.num_batches_tracked", "down_sampling.4.bias", "down_sampling.5.weight", "down_sampling.5.bias", "down_sampling.5.running_mean", "down_sampling.5.running_var", "down_sampling.5.num_batches_tracked", "down_sampling.7.bias"
... lots of other missing layers, then
size mismatch for down_sampling.1.weight: copying a param with shape torch.Size([16, 3, 7, 7]) from checkpoint, the shape in current model is torch.Size([24, 3, 7, 7]).
size mismatch for down_sampling.4.weight: copying a param with shape torch.Size([16, 16, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 24, 3, 3]).
size mismatch for down_sampling.7.weight: copying a param with shape torch.Size([210, 16, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 48, 3, 3]).
... lots of other size mismatches.
Seems to me that there is a mismatch between the network that was created internally and the one that is being used to fill it with the previously trained model. Not sure if it is a bug or if something is wrong in the command line I am using to resume.
Any help will be appreciated.
The text was updated successfully, but these errors were encountered:
Hi @jvillegassmule, thanks for your interest in the work, and sorry for the late reply! There is a bug in the finetuning stage for the student model. May you please try this branch and let me know if you have any questions? Thanks.
After successfully train the teacher with the command line:
After training, the results on the eval/(it_number)/fake folder are acceptable.
I had to stop the session before it finished, again different checkpoint models were saved on the folder /local_path/logs/student/checkpoints. Including pth files for G,D, optim-0,optim-1,A-0,A1,A2 and A3
Progress seems OK on the local_path/logs/student/eval folder
But now I get this error:
... lots of other missing layers, then
... lots of other size mismatches.
Seems to me that there is a mismatch between the network that was created internally and the one that is being used to fill it with the previously trained model. Not sure if it is a bug or if something is wrong in the command line I am using to resume.
Any help will be appreciated.
The text was updated successfully, but these errors were encountered: