-
Notifications
You must be signed in to change notification settings - Fork 74k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[TF-TRT] Argument bugfix #50015
[TF-TRT] Argument bugfix #50015
Conversation
Thanks for your pull request. It looks like this may be your first contribution to a Google open source project (if not, look below for help). Before we can look at your pull request, you'll need to sign a Contributor License Agreement (CLA). 📝 Please visit https://cla.developers.google.com/ to sign. Once you've signed (or fixed any issues), please reply here with What to do if you already signed the CLAIndividual signers
Corporate signers
ℹ️ Googlers: Go here for more info. |
@googlebot I signed it! |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for adding this! A better fix would be to first normalize the string to upper cases and then construct the conversion_params. But I am approving this and we will clean up this later.
TF-TRT supports lowercase arguments.
tensorflow/tensorflow/python/compiler/tensorrt/trt_convert.py
Lines 100 to 110 in 9460c57
From official documentation API instruction,
https://docs.nvidia.com/deeplearning/frameworks/tf-trt-user-guide/index.html#int8-quantization
precision_mode
Default value is TrtPrecisionMode.FP32. This is one of TrtPrecisionMode.supported_precision_modes(), in other words, FP32, FP16 or INT8 (lowercase is also supported).
But INT8 calibration with
precision_mode='int8'
didn't work before.