New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to deploy models where the shape of output tensor is not known #5
Comments
Did you solve this issue? And if so, could you share your solution? I am also interested in how to serve models with variably sized outputs. |
I was wrong, as the output shape is not variable, there is an upper bound for number of objects detected. So jus set the dims to this upper bound. That should work fine |
Thanks for getting back @srihari-humbarwadi. Seems defining an upper bound is fine because your model type returns a fixed size tensor, but I'm still curious if variable sizes are supported in For context: However, returning a fixed size output is not ideal for performance reasons. While it doesn't matter much for simple result types, consider the case where the served model is a MaskRCNN and the return type includes a pixel mask for each detected object. Without an output signature with variable sized tensors, the payload size would be worst-case for every return. I like to support variable outputs to reduce the payload for the common case (where less than max objects are detected). For tf-serving, this involved modifying the output before exporting a saved model, such that the return type only includes results for object's whose score exceeds some threshold. Is this behavior supported in |
TRTIS only supports variable-sized dimension for batching, but this is a common request so we are planning on fixing it. Issue #8 is tracking this request so add upvotes there to indicate that you are interested in it. |
Thanks @deadeyegoodwin ! |
hello,have you solved it? |
I have a tensorflow frozen graph of a objection detection model, i am unclear about creating a config.pbtxt file for this model since i cannot determine the output shapes before hand and i cannot start the inference server without the "dim" specified. i wanted to know how can i create a config file for this
this my config which does not work, i tried fixing the shape to the maximum proposals ie 300. Which i knew wouldn't work
The text was updated successfully, but these errors were encountered: