Bundling kserving backend and client in the same pypi#250
Bundling kserving backend and client in the same pypi#250k8s-ci-robot merged 2 commits intokserve:masterfrom jinchihe:update_setup_for_kservice_python_pkg
Conversation
|
The And user can install kfserving sdk from pip easily, and kfserving SDK works fine after installation. |
|
/assign @yuzisun |
|
/retest |
2 similar comments
|
/retest |
|
/retest |
|
mostly looks good to me, do you currently publish to PyPi manually ? |
Yes, I do that manually, since there are only one command |
|
/hold We are discussing more details in the #268 , may update once somethings decided. Thanks. |
|
Hello Reviews: From discussion and analysis from the ticket #268 , I update the PR changes, since the update is large, I would like to explain more for easy revewing. Overall, I merged the
There is a problem that the /cc @yuzisun @ellis-bigelow @animeshsingh |
|
/hold cancel |
|
CC @yuzisun |
|
Rebased to resolve the conficts, but presubmit tests faild, that's caused by no space in automation env. |
Can we make a comment on top of sdk-gen.sh to remind that folks need to replace README.md post the sdk generation? Nice job @jinchihe |
@animeshsingh Great comments! Added notes in the scripts and README file for the script. |
|
/retest |
|
Nice! |
|
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: ellis-bigelow The full list of commands accepted by this bot can be found here. The pull request process is described here DetailsNeeds approval from an approver in each of these files:
Approvers can indicate their approval by writing |
|
/retest |
1 similar comment
|
/retest |
|
/lgtm |
|
@yuzisun @ellis-bigelow final comments on this? |
@animeshsingh The PR has been approved by Ellis, and That's OK from Dan, but the presummit testing already failed since timeout or some strange reasons... /retest |
* kfservice python sdk package * merge the sdk to kfserving
Motivation The Triton runtime can be used with model-mesh to serve PyTorch torchscript models, but it does not support arbitrary PyTorch models i.e. eager mode. KServe "classic" has integration with TorchServe but it would be good to have integration with model-mesh too so that these kinds of models can be used in distributed multi-model serving contexts. Modifications The bulk of the required changes are to the adapter image, covered by PR kserve/modelmesh-runtime-adapter#34. This PR contains the minimal controller changes needed to enable the support: - TorchServe ServingRuntime spec - Add "torchserve" to the list of supported built-in runtime types - Add "ID extraction" entry for TorchServe's gRPC Predictions RPC so that model-mesh will automatically extract the model name from corresponding request messages Note the supported model format is advertised as "pytorch-mar" to distinguish from the existing "pytorch" format that refers to raw TorchScript .pt files as supported by Triton. Result TorchServe can be used seamlessly with ModelMesh Serving to serve PyTorch models, including eager mode. Resolves kserve#63 Signed-off-by: Nick Hill <nickhill@us.ibm.com>
#### Motivation Support for TorchServe was added in kserve#250 and kserve/modelmesh-runtime-adapter#34. A test should be added for it as well. #### Modifications - Adds basic FVT for load/inference with a TorchServe MAR model using the native TorchServe gRPC API - Disables OVMS runtime and tests to allow TorchServe to be tested due to resource constraints #### Result Closes kserve#280 Signed-off-by: Rafael Vasquez <raf.vasquez@ibm.com>
…ent-updates/kserve-agent-28 Update kserve-agent-28 to fc7ef75
What this PR does / why we need it:
Develop python pip package for KFServing SDK, user can directly install kfserving sdk using
pip install kfserving, and then call kfserving sdk apis to create/check/patch/delete kfservice.Which issue(s) this PR fixes :
Fixes #246
Special notes for your reviewer:
Release note:
This change is