Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About fine tuning with different architecture #140

Closed
jackiechensuper opened this issue Feb 23, 2014 · 2 comments
Closed

About fine tuning with different architecture #140

jackiechensuper opened this issue Feb 23, 2014 · 2 comments

Comments

@jackiechensuper
Copy link

Hi guys,
If I want to classify 100 objects and use the convolution and max-pooling layers of pre-trained Imagenet model, I only change the imagenet.prototxt with different outputs in fully-connected network. How to initialize the network with pre-trained imagenet model in previous layers while latter layer randomized ?
Thanks

@kloudkl
Copy link
Contributor

kloudkl commented Feb 23, 2014

Please take a look at #31. You may consider searching the issues first before opening a new issue.

The secret is in Net::CopyTrainedLayersFrom. In my opinion, all you need to do is to define your desired network and name the layers that you want to copy from a pre-trained model with the names of the corresponding source layers to be copied from. Other layers will be initialized by the random fillers that you specify.

@shelhamer
Copy link
Member

@kloudkl is right. See the finetuning slide in the Caffe presentation for a miniature example.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

4 participants