Skip to content

Compile FP16 type Model #318

@workdd

Description

@workdd

Hello, I am a student who is very interested in AWS Inferentia. Through the Neuron document, I learned that if model in FP32 are compiled, they are cast as BP16 and if model in FP16 are compiled, they are cast as FP16. Therefore, I want to compile model in FP16 into neuron sdk to know performance and accuracy. But the way I can get the model in FP16 is to convert the model with TFlite. Is there a way to compile the TFlite model? As far as I know, only model in pb(protobuf) format can be compiled.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions