New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
KServe does not catch Caikit runtime status correctly when subsprocess (tgis) have issues #88
Comments
We will ping IBM to see what the expected behavior is. |
I see that many
this means that Python |
This should be handled by the container splitting (this sprint) |
This is the ticket for reference :) |
When running |
@dtrifiro what about when running caikit+tgis in the single container? 🤔 |
Single container architecture was only a stopgap to the current architecture |
Verified in RHOAI 2.5 RC4, now if I cause a failure while loading a model (i.e. give the wrong path to the deployment modal) the InferenceService will set the |
When I create a ServingRuntime+InferenceService with some incorrect parameters, Caikit cannot load the model.
This part is expected. However, the
InferenceService
still shows the model asLoaded
, which is unexpected:The text was updated successfully, but these errors were encountered: