Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

loss doesn't decrease and keep on 40 in my dataset #24

Open
guapizyq opened this issue Jan 22, 2019 · 9 comments
Open

loss doesn't decrease and keep on 40 in my dataset #24

guapizyq opened this issue Jan 22, 2019 · 9 comments

Comments

@guapizyq
Copy link

I train with my own dataset with the default script
my dataset has three types of objection with 40000 train sample and 6000 validation sample
the batch_size is 4, my loss decrease to 49 in the first epoch, but loss keep on 40 in the 13th epoch, the loss doesn't decrease anymore.
how should i change my learning rate and others parameter

@guapizyq
Copy link
Author

it would be appreciated if you give some advise on this issue

@Adamdad
Copy link
Owner

Adamdad commented Jan 22, 2019

Can you detect anything in the testset?if not,what is your learning rate,lr decay pacience?

@guapizyq
Copy link
Author

I am sorry, it seems to my mistakes.
the first training is to get a stable loss
the second training without frozen layers is to get a lower loss?

@Adamdad
Copy link
Owner

Adamdad commented Feb 18, 2019

correct

@guapizyq
Copy link
Author

I got a lower loss than 40, but it still is 38
the result in my testdset looks ok, i do not konw how to decrease the loss

@zhangyufei1995
Copy link

@guapizyq @Adamdad I am very happy to discuss with you. What I want to ask is, 1. How much is the epoch setting of the red arrow here? 2, how much is the initial_epoch setting of the two black arrows? I look forward to your answer.
QQ图片20190531173008

@zhangyufei1995
Copy link

And why is it divided into step-by-step training?the first training? the second training ?What is their role?

@Adamdad
Copy link
Owner

Adamdad commented May 31, 2019

The first training part is for finetuning a model quickly. It freezes most layers, only to train on the last few layers. We can get an acceptable model for detection in a short period of time

The second training part is for getting a complete model. All the layers can be trained through this process.

Under most occasions, I only use the second part. Epoch under is not important here.

@gzz1529657064
Copy link

The first training part is for finetuning a model quickly. It freezes most layers, only to train on the last few layers. We can get an acceptable model for detection in a short period of time

The second training part is for getting a complete model. All the layers can be trained through this process.

Under most occasions, I only use the second part. Epoch under is not important here.

After model training, I have a model with size of 277M. It is bigger than YOLO-v3,Why?
Doesn't MobileNet reduce model parameters?
This is my training strategy in my dateset.

  1. Unfreeze all of the layers
  2. learning_rate = 0.001
  3. load_pretrained=False
  4. batch_size = 16

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants