Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update servingruntime ootb configuration #340

Merged

Conversation

lucferbux
Copy link
Contributor

@lucferbux lucferbux commented May 19, 2023

Description

Get latest changes to downstream, displaying tensorflow in OVVM and adding a new GPU field.

@lucferbux
Copy link
Contributor Author

@Xaenalt @VaishnaviHire @andrewballantyne
Can we review this before feature freeze please, in that case this can go to 1.28 without opening another pr.
thanks in advance!

@Xaenalt
Copy link
Contributor

Xaenalt commented May 25, 2023

Oh, this seems like a good place to ask, apparently despite the docs saying this isn't supported, apparently we can set target_device: HETERO:NVIDIA,CPU, so it'll fall back if it fails to compile onto the GPU, I'm not entirely sure if this will work if the Nvidia device is absent, but it's worth asking if we want to look into that

@lucferbux
Copy link
Contributor Author

Oh, this seems like a good place to ask, apparently despite the docs saying this isn't supported, apparently we can set target_device: HETERO:NVIDIA,CPU, so it'll fall back if it fails to compile onto the GPU, I'm not entirely sure if this will work if the Nvidia device is absent, but it's worth asking if we want to look into that

Hi @Xaenalt I don't fully understand this, what do you mean? In case they deploy this serving runtime and nvidia is absent it will fail? Is that it?

@Xaenalt
Copy link
Contributor

Xaenalt commented May 25, 2023

So, I need to test to make sure what happens if Nvidia is absent, but if I'm reading their docs correctly (and what their devs said correctly), it'll run some parts on the CPU and some on the Nvidia GPU, that should increase compatibility, but it's not clear to me what'll happen if one device is absent, I'd assume that it'd just run it all on the CPU, but I've been burned by that kind of assumption before xD

@lucferbux
Copy link
Contributor Author

So, I need to test to make sure what happens if Nvidia is absent, but if I'm reading their docs correctly (and what their devs said correctly), it'll run some parts on the CPU and some on the Nvidia GPU, that should increase compatibility, but it's not clear to me what'll happen if one device is absent, I'd assume that it'd just run it all on the CPU, but I've been burned by that kind of assumption before xD

But does this require something UI related? or maybe adding something in the official docs? Not sure if I need to change something based on your comments.

@lucferbux
Copy link
Contributor Author

@Xaenalt added fe33221 to support tensorflow, let me know if this is fine!

Copy link
Contributor

@andrewballantyne andrewballantyne left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

From my perspective -- this lgtm.

@lucferbux
Copy link
Contributor Author

@anishasthana @VaishnaviHire @LaVLaS could you check this out? I'm aiming to merge this before feature freeze, thanks in advance!

@lucferbux
Copy link
Contributor Author

@tarukumar I think it needs qe approval too, can you please take a look?

@openshift-ci
Copy link

openshift-ci bot commented Jun 1, 2023

[APPROVALNOTIFIER] This PR is APPROVED

Approval requirements bypassed by manually added approval.

This pull-request has been approved by: andrewballantyne, Xaenalt

The full list of commands accepted by this bot can be found here.

The pull request process is described here

Needs approval from an approver in each of these files:

Approvers can indicate their approval by writing /approve in a comment
Approvers can cancel approval by writing /approve cancel in a comment

@lucferbux lucferbux added the qe-approved The PR has been approved by the QE team label Jun 1, 2023
@openshift-merge-robot openshift-merge-robot merged commit f4938c1 into red-hat-data-services:main Jun 1, 2023
1 check passed
anishasthana pushed a commit that referenced this pull request Jun 1, 2023
* Update servingruntime ootb configuration

* Add tensorflow runtime
dimakis pushed a commit to dimakis/odh-deployer that referenced this pull request Jul 9, 2023
* Update servingruntime ootb configuration

* Add tensorflow runtime
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
approved lgtm qe-approved The PR has been approved by the QE team
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

4 participants