Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Time profiling result is inconsistent with result from original caffe SSD #24

Closed
andyhahaha opened this issue Jun 4, 2017 · 2 comments

Comments

@andyhahaha
Copy link

I wondering the time consumption at each part. (VGG, Extra, multi_box, detection)
From the result of caffe version, the VGG part accounts for up to 80 percent of time consumption.
However, in this version, the distribution of time consumption is as follow:

Total time : 0.018(seconds) per image
VGG part 8.4%
Extra layer 2.8%
Multi_box 61%
detect 27.5%
Most of time is from Multi_box and detect.
I measure it by python time.time()

And both total time for one image is almost the same.
caffe : 19ms
pytorch : 18ms

I wondering why this inconsistence happen?

@amdegroot
Copy link
Owner

To be honest I'm not really sure. That's really cool that the time profile is actually better with pytorch though, I did not realize that. It must have something to do with the way pytorch is designed under the hood with respect to gpu acceleration (I haven't been able to keep up with pytorch development as much recently :( ).

@markusnagel
Copy link

@andyhahaha I'm curious, how do you get the 18ms? Is that on one specific image or the average over many images (VOC test set)?

I get on average over VOC test set only 70-80 ms per image (with a GTX 1080). Though the variance is very high, some images need only 20ms, others over 100ms. It also seems that the same images take always (approximately) the same time.

I'm also wondering how @amdegroot gets the 45 FPS, is that the average or for one test image? It seems I should get at least the same or faster with my GTX 1080. Could you maybe also share some numbers on training time? Mine is currently at +-130ms per iteration (again on Pascal VOC with the provided training code).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants