-
Notifications
You must be signed in to change notification settings - Fork 51
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
./train.sh for TSM stop at the first log infor. : Freezing BatchNorm2D except... #1
Comments
Thanks for your feedback. You can set trainer.no_partial_bn = True if batch size >= 6 in each gpu and retry it, this will not affect the accuracy. That module exists some bug with distributed training, we will fix it quickly. |
Thanks for the reply, but it doesn't work for my problem. The changed settings in my YAML file are only dataset related: Thanks again and waiting for you suggestion. |
I test it in my env, it's ok. Could you provide your training script? and the gpu nums in default yaml is 8, you may need to check up with it. The last thing, with your img_prefix, root_dir and meta_file, the program can find the correct img path. I noticed that you only modified root_dir and meta_file, maybe img_prefix also needs to be modified. |
Thanks, it works for me when I modified 'img_prefix' : from 'image_{:05d}.jpg' to '{:05d}.jpg'; and, another problem appear when I run ./train.sh
I also have no idea about this, then I found a similar issues in the url: |
The yaml is an example, the num_classes is not for sth especially. This error comes from no_partial_bn, do you set no_partial_bn = True? Or you can git pull, I just update the code. |
Cool! It works when I set no_partial_bn=True! ( I had turn it back to False when I try to fix other problems) |
Thanks for your nice style codebase.
However, when I try to train TSM in your codebase, there is a problem which stoped me from training it.
(1)The log file stop at:
2020-04-10 xxxx094-models.py#177: Freezing BatchNorm2D except the first one
, and I wait it for 10 min but with no continue update.(2)When I use 'gpustat' check the usage of gpu, it shows only about 800M data in each gpu(I use 8 in total)
I am sorry for disturbing you, while as a green hand also would be appreaciate if you could show me some light.
The text was updated successfully, but these errors were encountered: