Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

models works on float32 instead of uint8 #10760

Open
naarkhoo opened this issue Sep 2, 2022 · 6 comments
Open

models works on float32 instead of uint8 #10760

naarkhoo opened this issue Sep 2, 2022 · 6 comments

Comments

@naarkhoo
Copy link

naarkhoo commented Sep 2, 2022

Hi,

I am in the process of my ssd model based on ssd_mobilenet_v2_320x320_coco17_tpu and I noticed the model works on float32 and not uint8 - I am curious how I can make that change ?

Also I appreciate if you point me to other tricks that I can make my model run faster at inference level. for example larger kernel size ? or shallower model ? or some threshold ? I feel these recommendations/explanation can be helpful when it comes to optmization

here is the link to the colab notebook https://drive.google.com/file/d/1iqUgeabbTgfixehGomDoj5eHGfHd8Lvt/view?usp=sharing

@sushreebarsa sushreebarsa self-assigned this Sep 7, 2022
@sushreebarsa
Copy link
Contributor

@naarkhoo
In order to expedite the trouble-shooting process, could you please provide the entire URL of the repository which you are using. Please provide more details on the issue reported here. Thank you!

@sushreebarsa sushreebarsa added the stat:awaiting response Waiting on input from the contributor label Sep 7, 2022
@google-ml-butler
Copy link

This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you.

@naarkhoo
Copy link
Author

sorry for my late reply here is the colab https://drive.google.com/file/d/1iqUgeabbTgfixehGomDoj5eHGfHd8Lvt/view?usp=sharing and I made sure you have access to the files.

with the current code the model latency on android devices (average device) is 150ms - my goal is to make a model to work at 50ms - seems I have make sure the model works with uint8 data type.

@google-ml-butler google-ml-butler bot removed stat:awaiting response Waiting on input from the contributor stale labels Sep 14, 2022
@sushreebarsa sushreebarsa added the models:official models that come under official repository label Oct 3, 2022
@laxmareddyp laxmareddyp self-assigned this Nov 21, 2022
@saberkun
Copy link
Member

@jaeyounkim for TF-MOT problems

@jaeyounkim
Copy link
Collaborator

"ssd_mobilenet_v2_320x320_coco17_tpu" is what "TensorFlow Object Detection API" provides. It is not the model officially supported by the Model Garden team. Let me check if the TensorFlow Model Optimization Toolkit (https://github.com/tensorflow/model-optimization) team can provide some help.

@jaeyounkim jaeyounkim added models:research:odapi ODAPI and removed models:official models that come under official repository labels Nov 28, 2022
@Petros626
Copy link

The model is not quantized that's all. Read the name and compare it to the quantized model you'll find the difference. You must do post training quantization for the required result.

Additionally to run faster your model you need a tflite model and possibly a hardware accelerator like Google coral usb accelerator

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

7 participants