Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How about peleeNet training from scratch? #30

Open
MrWhiteHomeman opened this issue Jun 1, 2018 · 14 comments
Open

How about peleeNet training from scratch? #30

MrWhiteHomeman opened this issue Jun 1, 2018 · 14 comments

Comments

@MrWhiteHomeman
Copy link

I think peleenet is similar to DSOD, how about peleeNet training from scratch, does it work???
I appreciate it if you can give me some advices!!!
Thank u!!!

@ujsyehao
Copy link

ujsyehao commented Jun 1, 2018

@MrWhiteHomeman I am trying to do it, response to you later

@ujsyehao
Copy link

ujsyehao commented Jun 1, 2018

Here is a link 链接: https://pan.baidu.com/s/1vZONIe2pBkxjo-s5wP3zAg 密码: 3ip6.

  • step 1: unzip the file in $caffeSSD/models directory

  • step 2: modify batch size in pelee_voc/train.prototxt, I modify batch size 32 to 20 because I only have 8 GB memory, you can revert batch size to 32

  • step 3:
    cd $caffeSSD
    ./build/tools/caffe train -solver=models/pelee_voc/solver.prototxt

@MrWhiteHomeman
Copy link
Author

@ujsyehao It is so nice of you for your reply !!! I have another question about the code, in the feature_extractor.py ,the 30th line, why are there two 'stage4_tb/ext/pm2/res/relu' in the Pelee.mbox_source_layers? Can you give me some advices? Thank you!!!

@ujsyehao
Copy link

ujsyehao commented Jun 8, 2018

You can use (http://ethereon.github.io/netscope/#/editor) to view peleenet-ssd network structure, you will find stage4_tb/ext/pm2/res layer is used twice to generate ext/pm1_mbox_loc layer and ext/pm2_mbox_loc layer(conf layer/priorbox layer is also the case).
The reason is that peleenet drops 38x38 feature map(you can view pelee paper) and just use the remaining 5 feature extracted layer(19x19, 10x10, 5x5, 3x3, 1x1), but SSD merges 6 layers' prior boxes, so author use 19x19 feature map(also known as stage4_tb/ext/pm2/res) twice to predict two conf/loc/priorbox layers.

@ujsyehao
Copy link

ujsyehao commented Jun 8, 2018

mobile-net ssd also follows this design pattern, I will update later.

@MrWhiteHomeman
Copy link
Author

@ujsyehao Hello,I have a question about the batchsize, in this paper , the batch-size is 32, if I change the batch-size to 64, will I get a better result about testing ? I would appreciate it if you could give me some advices.
Thank you!!!

@ujsyehao
Copy link

No, batch size can affect training time and has no direct relation with model performance.

@MrWhiteHomeman
Copy link
Author

@ujsyehao So,I always have a question about batchsize, if the batch size is too big, will it have a bad result? And I know the DetNet(旷世科技) , it set the batch size to 256, and get a greatest result...

@ujsyehao
Copy link

ujsyehao commented Jul 11, 2018

You can look it

@RainFrost1
Copy link

Could you please share the prototxt again? The link [链接: https://pan.baidu.com/s/1vZONIe2pBkxjo-s5wP3zAg 密码: 3ip6.] failed now.
Thank you very much~~~ @MrWhiteHomeman @ujsyehao

@foralliance
Copy link

@ujsyehao
所谓的batch size大小不会影响model performance,应该仅限于模型中的BN层参数固定的情况吧.
如果BN层的参数在训练过程中也进行微调,那么batch size大小还是会影响model performance的吧.

@ujsyehao
Copy link

@foralliance model performance 依赖于模型本身,batch size只是一个超参而已,如果你修改了batch size,再选定其它合适的超参比如base_lr,它最终的效果是一样的,一般而言,使用一个大batch只是训练的更快,更容易出结果,并不会从根本上决定模型的性能

@foralliance
Copy link

@ujsyehao
Copy link

@foralliance 我看了这个回答,默认accum_batch_size固定,对于无BN层/BN层参数固定情况下,batch size不影响模型性能这个观点我是认同的

@EvaneSnow
Copy link

各位大佬好,你们谁晓得Pelee训练目标检测 + 车道线,谢谢各位

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants