Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

pretrained models for batch normalized version of vgg #152

Closed
bodokaiser opened this issue Apr 19, 2017 · 3 comments
Closed

pretrained models for batch normalized version of vgg #152

bodokaiser opened this issue Apr 19, 2017 · 3 comments

Comments

@bodokaiser
Copy link
Contributor

According to here we should be able to reuse the retrained weight from e.g. vgg16 and apply them to vgg16bn we will just need to write a small conversion (as order in features differs due to additional nn.BatchNorm2d layers).

Do you agree? Are PR welcome on this?

@colesbury
Copy link
Member

I don't think you can take the weights from the model without batch norm and stick them in the model with batch normalization. The activation statistics will almost certainly be wrong at each batch-normalized layer if someone tries fine-tuning the model or otherwise continue training.

@bodokaiser
Copy link
Contributor Author

bodokaiser commented Apr 19, 2017

Could we then also offer pretained weights for the batch normalized versions?

Depending on how you obtain them anyway I could also look if I get a caffemodel converted to pytorch state dict.

@wangg12
Copy link
Contributor

wangg12 commented Apr 26, 2017

+1 for the request.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants