Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Get Operator-wise Profiling Results #748

Closed
ZhanqiuHu opened this issue Aug 18, 2022 · 1 comment · Fixed by #749
Closed

Get Operator-wise Profiling Results #748

ZhanqiuHu opened this issue Aug 18, 2022 · 1 comment · Fixed by #749

Comments

@ZhanqiuHu
Copy link

Our group wants to profile BNNs on some ARM64 processors. We tried running ./lce_benchmark_model_aarch64 with the --enable_op_profiling set to true. The result looks like this (for QuickNet), with many floating-point operators grouped together as TfLiteXNNPackDelegate in the report. Is there a way to get the inference time for more specific operators like ADD, MUL, CONV_2D, MAX_POOL_2D, etc? Thank you!

Screen Shot 2022-08-18 at 3 11 24 PM

@CNugteren
Copy link
Contributor

At the moment I do not think this is possible, because it is the way XNNPACK behaves. However, I just saw this blog post, which indicates that XNNPACK detailed profiling is available in the next release of TensorFlow. Since the Larq Compute Engine is still based on TF 2.9.0, this is not yet supported. We plan to update LCE to TF 2.10.0 within the coming weeks, so then this should be included as well.

The alternative is to disable XNNPACK (see the command-line options for lce_benchmark_model_aarch64 ), but of course that will give different profiling results. However, it might give you a rough indication of how much time is spent in each layer.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants