-
Notifications
You must be signed in to change notification settings - Fork 77
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Model that runs fine in onnxruntime throws warning and gives incorrect results in barracuda #271
Comments
Any update on this one? |
Hi @rfilkov
Indeed at the moment Barracuda does not support the "split" optional parameter from Split operator version 13. We are currently extending Barracuda range of supported operators and architectures however we don't have ETA to share yet. Something to try is to export the model as opset 9. Witch is at the moment a good target version in term of supported operator for Barracuda. Opset 9 would mean Split will be version 2 |
In regard to pose estimation for Barracuda this ressource can be helpfull imho: https://github.com/keijiro/BodyPixBarracuda |
Thank you for the reply! With opset 9 I get the same error when trying to import the ONNX model in python, I use the command:
and get the error:
Is there any feedback about the warnings of the original onnx model that I was trying to import (in my first post)? |
thanks, but this does not seem to support 3D human pose estimation. However, in the meantime I found a repository that contains a working example of 3D human pose estimation using a VRect onnx model with barracuda: |
Hi all,
I've been trying to port the following code to barracuda (Unity 2020.3.31f1 with Barracuda 3.0.0.):
https://github.com/ibaiGorordo/ONNX-Mobile-Human-Pose-3D
The model can be imported correctly in onnxruntime, but shows the following warnings in the inspector:
The onnx model can be downloaded here:
https://drive.google.com/uc?export=download&id=1vvjUsP_41Nqhj8oqRw2OpWjIdz8t5fxG
I can run the model in barracuda, but the results differ from the ones generated in onnxruntime.
Furthermore, I've tried to convert the original TFLITE model provided by the authors
https://github.com/tucan9389/PoseEstimation-TFLiteSwift/releases/download/v2.1.0/lightweight_baseline_choi.tflite
to ONNX using https://github.com/onnx/tensorflow-onnx
and the command
python -m tf2onnx.convert --opset 13 --tflite tflite--file --output model.onnx
The resulting model again runs fine in python with onnxruntime but this time I cannot import the model to unity, it throws the error:
Is there any way to make the model run in barracuda? Any help would be greatly appreciated!
Thanks,
Matthias
The text was updated successfully, but these errors were encountered: