-
Notifications
You must be signed in to change notification settings - Fork 45.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
object_detection_tutorial test runtime #3531
Comments
Your running time is "image loading + inference + visualization", which isn't inference at all.
|
Thanks for you reply |
Typically the inference time refers to the time spent on the "session.run()" block. It's the time of a forward pass(from input layer to output layer). |
Thanks for your answer. Could I question one more? %%%%....%%%% start = time.time() |
I am guessing there is a lot of overhead for the first image e.g. loading the graph, transferring it to the GPU and so on. Your findings seem to suggest that: once the graph is loaded and data is coming in fast, the GPU is reducing the processing time a lot. Did you find out anything else related to this? Curious as I am having the exact same question/experience and run a nearly identical setup to you. |
Hi There, |
System information
(https://www.tensorflow.org/versions/r1.5/install/install_linux#InstallingNativePip)
###You can collect some of this information using our environment capture script:
2018-03-06 22:58:39.700094: I tensorflow/core/platform/cpu_feature_guard.cc:137] Your CPU supports instructions that this TensorFlow binary was not compiled to use: SSE4.1 SSE4.2 AVX AVX2 FMA
2018-03-06 22:58:39.816687: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:895] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
2018-03-06 22:58:39.816965: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1105] Found device 0 with properties:
name: GeForce GTX 1070 major: 6 minor: 1 memoryClockRate(GHz): 1.683
pciBusID: 0000:01:00.0
totalMemory: 7.93GiB freeMemory: 7.48GiB
2018-03-06 22:58:39.816981: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1195] Creating TensorFlow device (/device:GPU:0) -> (device: 0, name: GeForce GTX 1070, pci bus id: 0000:01:00.0, compute capability: 6.1)
Wrote environment to tf_env.txt. You can review the contents of that file.
and use it to populate the fields in the github issue template.
Describe the problem
It took about 3 second to detect one image. Also, It took about 130 second to detect 130 images, 253 second 400 frames. But, I think that, at least, it take 200 frames 10 second since this model has real-time performance.
GPU doesn't work properly.
--> if in this case, you can give me answer through above environment capture, system specification.
Wrong measurement test runtime method (you can see the method below picture)
--> Please, let me know how to exactly measure detection time.
Please answer me detail.. I wanna give your team help to upgrade Google Object Detection models !
Source code / logs
with detection_graph.as_default():
################################################################################
#################################################################################
P.S
If you need more information, please tell me. Thanks!
The text was updated successfully, but these errors were encountered: