-
-
Notifications
You must be signed in to change notification settings - Fork 66
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ResNet producing non deterministic and wrong predictions #72
Comments
I have a suspicion this and related issues are due to not all weights being properly loaded. See https://julialang.zulipchat.com/#narrow/stream/237432-ml-ecosystem-coordination/topic/ResNet.20weights for details, but TL;DR no batchnorm or bias! If it hasn't done so already, this should be fixed in #70 eventually. |
Has the PR brought any change? |
Not other than discovering that the previous pre-trained model was wrong. I'm working on re-training the current models. |
So re-training is all whats needed to fix it (hopefully)? Or is there still more to it? Outputs changing all the time sounds like BN layers are continuously adapting. |
Outputs shouldn't change constantly if the normalization layers are frozen, so I assume they weren't in the example above. This definitely shouldn't be a problem with the new models, because |
Yeah this issue describes the old implementations. The new ones on master should not suffer from this issue. Training should be all that's necessary for a working model. |
Do you need help with training more models? Where’s the bottleneck? |
Primarily the lack of a multi-GPU flow for DDP. Hopefully once the JuliaCon code is released, I can use that. For now, I'll just train with a single GPU. |
If you provide the training fixture code, I can provide hardware and electrons :) Edit: Sounds like I have to watch the juliacon videos!! |
so is this still relevant, or were the models re-trained in the meantime? (and if they were, did this fix the issue?) |
No models have been re-trained so far. I can help guide someone to set it up if they have a GPU. Pre-trained weights would be a welcome contribution. |
Predictions from the ResNet pre-trained models seems to have an issue. The predicted probabilities are changing at each call and essentially look like random. VGG19 does work fine however.
I first though it might be an issue with a different preprocessing required for ResNet, thought from the randomness in the output of the model, I'd guess some layers didn't get properly defined, possibly the BatchNorm?
If using
testmode!
, resnet will then produces the same predictions after each run. However, a different set of weights seem to be initialized right at each call to ``The text was updated successfully, but these errors were encountered: