-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Problem when running sequence models #1123
Comments
"simple_sequence_client" assumes that the model has an input called "INPUT", and your model config shows that the model has "gpu_0/data" instead. Simply modifying |
And I don't think your model repository is organized in the right way... Are you able to start TRTIS without any errors? |
Yes, I am able to run TRTIS using this respository |
After changing "gpu_0/data" to "INPUT". I can not run TRTIS and it gives me error message like below
|
For starters you might want to use the 20.01 clientsdk container. |
I meant to change the client, not |
Cool ! thx a lot. I will try it |
I change the simple_sequence_client.cc, recompile and run, but error still happens When I am running, I get the error message like below
|
Please notice that "simple_sequence_client" is just an example to show sequence inference usage, and it is designed based on the "sequence" model, which is completely different from the stateful model you have. Thus you will have to modify the "simple_sequence_client" before you can use it on your model. |
Thank you for your clarification ! So now I just want to run sequence model. I have compiled the "sequence" model and get libsequence.so. I follow the document and create the model repository like below. Is that right? Do I need to add a config text file ? sequence_model_repository |
Yes, you will need the config.pbtxt file for the model, which can be found here |
I add the config.pbtxt file you mentioned and run the server using docker container. It still gives error message
|
Can you change the dims for both input and output to [1]? |
Thank you. It works.
So this model cannot run on GPU ? |
That's correct. |
Got it. Thank you very much |
Are there any other sequence models that support GPU ? |
Some of our qa models are sequence models (generation script). And you can certainly find more stateful model elsewhere, you just need to create appropriate model config for those models to be recognized by TRTIS. Again, all the models and client examples you can find on TRTIS repository are for demonstrating TRTIS features, if you have an existing use case, you can use the them to guide you on integrating it with TRTIS. Closing the issue as the original problem is resolved. |
@GuanLuo @zhouxuan009 while reading this issue i had a doubt , can you please let me know how the libsequence.so file is created. |
Description
I try to run sequence models but fails
I use the sample client provided in docker container
docker run -it --rm --net=host nvcr.io/nvidia/tensorrtserver:19.10-py3-clientsdk
/workspace/install/bin# ./simple_sequence_client
And it gives the error message like below
TRTIS and Model Information
I am using nvcr.io/nvidia/tensorrtserver:20.01-py3 container for server
Below is my model repository structure tree:
sequence_model_repository
---direct_stateful_resnet50_netdef
------1
----------libsequence.so
----------model.netdef
----------init_model.netdef
------config.pbtxt
------resnet50_labels.txt
Expected behavior
Could u please tell me what's going wrong ? thx very much !
The text was updated successfully, but these errors were encountered: