-
Notifications
You must be signed in to change notification settings - Fork 621
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to run FLOAT16 OnnxRuntime models #3190
Comments
You can convert model to fp16, but you need to CUDA device. You can use the following command:
see: https://github.com/deepjavalibrary/djl/tree/master/extensions/tokenizers#use-command-line |
I converted the model to fp16 and encountered an exception while loading fp16 model with CrossEncoderBatchTranslator Caused by: java.lang.UnsupportedOperationException: type is not supported: FLOAT16
|
I added FLOAT16 in OrtUtils.toDataType and disabled CrossEncoderBatchTranslator.sigmoid(PyTorchLibrary doesn't support fp16 sigmoid op with cpu), and the problem was solved
|
It looks like FLOAT16 has not been supported by now
The text was updated successfully, but these errors were encountered: