Skip to content

[Feature] [OIM] Make inference backend configurable #1525

@Yu-amd

Description

@Yu-amd

Priority

P2-High

OS type

Ubuntu

Hardware type

GPU-AMD

Running nodes

Single Node

Description

Based on discussion with Zhiwei at Intel, I understand there are plans to create OIM for selecting inference backend based on the models.

I'm opening this feature request to add the requirement that OIM should allow user to configure and override the backend used. The reason is that not all inference backends will be available on all hardware and software platforms due to various reasons, so we want to ensure the user can manually configure the actual inference backend for the OPEA workload.

Metadata

Metadata

Labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions