Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Using shared models/modules across inference pipelines #4272

Closed
saeid93 opened this issue Aug 14, 2022 · 1 comment
Closed

Using shared models/modules across inference pipelines #4272

saeid93 opened this issue Aug 14, 2022 · 1 comment
Labels

Comments

@saeid93
Copy link
Contributor

saeid93 commented Aug 14, 2022

Academic systems like Rim and grandslam have the ability to share a model across multiple pipelines. As there are use cases in which a single model could be used as part of two separate pipelines in a company submitted by multiple users, it would be a good idea to just reuse the already deployed model and autoscale it based on need rather than have two separate versions of it. However, as it needs some custom policy on top of K8S scheduler it might not be an urgent change.

An implementation idea for the user interface, would be to first deploy models as separate services and then just make the connection between them through the YAML file. This needs to decouple the physical placement of models from the graph representation.

@ukclivecox
Copy link
Contributor

We will be addressing this is v2 of our APIs

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

2 participants