Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why is original resnet50 faster than octave-resnet50? #4

Closed
BlueAnthony opened this issue May 9, 2019 · 4 comments
Closed

Why is original resnet50 faster than octave-resnet50? #4

BlueAnthony opened this issue May 9, 2019 · 4 comments

Comments

@BlueAnthony
Copy link

BlueAnthony commented May 9, 2019

Why is original resnet50 faster than octave-resnet50?
In inference on gtx-1080, original resnet50 is about 70 fps and octave-resnet50 is around 42 fps.

@miguelvr
Copy link

I noticed the same thing during training.

It actually takes less GPU memory, but training is slower.

Not sure if this is an issue from pytorch (static graph vs dynamic graphs) or it is related to the fact that the authors use TVM to compile the model.

@zimenglan-sysu-512
Copy link

zimenglan-sysu-512 commented May 12, 2019

i train a model of mask-rcnn using octave-resnet-50, and find that it's slower than resnet-50 and the performance is 3 points lower as well.

@peri044
Copy link

peri044 commented May 17, 2019

@BlueAnthony I've observed the same thing from my side here facebookresearch/OctConv#4

My guess is Oct Conv graph is not well suited on GPU (Could be due to two different branch structure, upsampling layers, more number of convolutions, more kernel launches and more mem copies). On CPU, Octave conv performance is better as reported in the paper.

@BlueAnthony
Copy link
Author

@peri044 Yeah, I also try Octave Conv on CPU, it really has some improvement on CPU.
I agree with you. Maybe Oct Conv graph is not well suited on GPU.
And, thank you for ur recommendation! I will try AutoTVM.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants