generated from amazon-archives/__template_DevGuide
-
Notifications
You must be signed in to change notification settings - Fork 179
Closed
Description
Hello, I am a student who is very interested in AWS Inferentia. Through the Neuron document, I learned that if model in FP32 are compiled, they are cast as BP16 and if model in FP16 are compiled, they are cast as FP16. Therefore, I want to compile model in FP16 into neuron sdk to know performance and accuracy. But the way I can get the model in FP16 is to convert the model with TFlite. Is there a way to compile the TFlite model? As far as I know, only model in pb(protobuf) format can be compiled.
Metadata
Metadata
Assignees
Labels
No labels