How to load a model in bf16 format #1540
Closed
georgel2004
started this conversation in
General
Replies: 1 comment
-
turbomind doesn't hard code lmdeploy/lmdeploy/turbomind/turbomind.py Line 271 in 7c4e75b |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
In turbomind.py the data_type is hard coded to fp16.
If I change the value to bf16 I could see increase in performance speed.
How could I load a model in bf16 format using the pipeline?
Beta Was this translation helpful? Give feedback.
All reactions