Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Quantized Model #26

Closed
seilassie opened this issue Jul 29, 2020 · 2 comments
Closed

Quantized Model #26

seilassie opened this issue Jul 29, 2020 · 2 comments

Comments

@seilassie
Copy link

Dear Filippo,

Thank you for the great work you do! In your paper, you achieved great performance with quantization, yet I only see a quantized model for iOS. I tried converting the frozen graph that's made with the export.py script with post-training quantization but I haven't had much luck. Would it be possible for you to upload the the quantized model as a frozen graph file? If not, would you be able to specify on how you converted your model? I followed the tutorial from https://blog.tensorflow.org/2019/06/tensorflow-integer-quantization.html using the NYU dataset, but the resulting tflite is much slower than the regular one.

Any help appreciated!

@FilippoAleotti
Copy link
Owner

Hi,
The model shouldn’t be quantized (i.e. with mlcore-quantization). The script I used to convert the models is
export.py

@seilassie
Copy link
Author

I see, my apologies for misunderstanding what you wrote in the paper. If you only used export.py, then how were you able to achieve superior performance (in terms of FPS) compared to FastDepth? Currently, I'm getting ~7 FPS for FastDepth and ~2 FPS for mobilePydnet (192x192 image). After autotuning with TVM and deploying, this jumps to ~4 FPS. Both models tested on one core for a Raspberry Pi 4 overclocked to 2GHz. I see in the paper that you deployed the FastDepth model with the same degree of optimization on the iPhone, yet I would expect slightly better performance for mobilePydnet. Then again, you mention in #1 that your model runs on the GPU; thus, may I assume mobilePydnet is much better fit for inferring on mobile GPUs? I suspect this is the case given that FastDepth is supposed to run on CPUs. I hope to make some scripts and share them with you. Closing for now, feel free to share your insights!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants