-
Notifications
You must be signed in to change notification settings - Fork 984
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Enable GPU Memory as resource requirement for InferenceService #947
Comments
Issue-Label Bot is automatically applying the labels:
Please mark this comment with 👍 or 👎 to give our bot feedback! |
Issue Label Bot is not confident enough to auto-label this issue. |
@Svendegroote91 technically inference service spec already allows |
@yuzisun Ok I can give it a try and let you know. If the resources limits is a map, I see that from a scheduling perspective it will work. |
@Svendegroote91 looks like we will need that argument, feel free to send a PR for that. |
@yuzisun I tested the setup but it doesn't work out of the box. Probably the following steps will need to be implemented:
I'll see if I can file a PR for this somewhere soon. |
@Svendegroote91 sorry for the late reply, KFServing do support gpu image for tfserving and you can specify |
I use kubeflow1.6.1 + k8s-gpushare-schd-extender:1.11-d170d8a newest version
error
Any other gpushare choice for kubeflow @yuzisun |
/kind feature
Describe the solution you'd like
[A clear and concise description of what you want to happen.]
Would it be possible to add the GPU memory as a resource requirement, similarly to the GPU count?
For example:
Is it technically already possible to try this if you have the
GPUshare scheduler extender installed on your cluster?
I noticed that they added something related in the Arena repo - kubeflow/arena#211
The text was updated successfully, but these errors were encountered: