-
Notifications
You must be signed in to change notification settings - Fork 4.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Memory usage on 16bit calculation. #507
Comments
how do you measure the memory usage ? |
I want to kown what platform for your testing? |
Thanks for your reply.
And one more question. |
I haven't see the int8 implementation on arm platform, it convert to float32 when cacultion when I read the code, I don't sure if i am wrong |
ncnn now supports fp16 operations and takes up half the memory compared to fp32. |
I've tested new 8, 16bit inference.
I expected memory usage will be half on 16bit inference than 32bit, but actual result went over my expectation.
On creating net, 11MB used for 32 bit model, but 16.5MB used for 16 bit model.
I turned off winograd, sgemm options for pure memory calculation.
Caffe model size is 12.2MB
What am I wrong?
Hope to any suggestions.
The text was updated successfully, but these errors were encountered: