-
Notifications
You must be signed in to change notification settings - Fork 260
predict with onnx #133
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: develop
Are you sure you want to change the base?
predict with onnx #133
Conversation
Let's wait this PR until we'll merge #129. |
@valavanisleonidas #129 got merged. We can work on this PR now. |
Yes I saw that thank you. should we also have dynamic batch size for onnx models since the predict function support batch inference ? |
@SkalskiP I think the PR is mostly ready. Two things to mention.
|
Hey @SkalskiP, any news on this PR ? |
Hi, dear developer, how is going on ? I also need to use the ONNX model to predict. |
this PR is not up to date with the latest changes. I can do it if @SkalskiP and the team is still interested in merging it. |
Description
This PR adds support for ONNX-based inference in RF-DETR, enabling faster model execution compared to PyTorch CPU inference. This change introduces an alternative inference path using the onnxruntime backend.
Ticket : #64
Type of change
How has this change been tested, please provide a testcase or example of how you tested the change?
Python test file
The bboxes and confidence scores of the image i tested are the same.
The time using GPU is
0,14 sec
for cuda with pth model and0,11 sec
for onnx modelAny specific deployment considerations
Docs