Skip to content
This repository has been archived by the owner on Sep 13, 2023. It is now read-only.

Deploying models with seldon-core #20

Open
Tracked by #422
aguschin opened this issue Sep 16, 2021 · 3 comments
Open
Tracked by #422

Deploying models with seldon-core #20

aguschin opened this issue Sep 16, 2021 · 3 comments
Labels
deploy Related to model deployment p2-medium Medium priority plugins Plugins and extensions for MLEM!

Comments

@aguschin
Copy link
Contributor

@DavidGOrtega suggested on slack:

Hey guys. Reviewing OpenMlOps they are exposing the models via Seldon and Ambassador API Gateway. This is similar to the QuaroML stack that I did and a very convenient and easy way to expose and of course scale the models.
Maybe we can introduce MLEM inside the TPI in conjunction with Consul

We need to get back to this discussion to shape the vision for MLEM deployment part. It's better to do as soon as we are finished with the closed alpha release.

@aguschin aguschin transferred this issue from another repository Sep 30, 2021
@aguschin aguschin added the deploy Related to model deployment label Sep 30, 2021
@aguschin aguschin added this to the Beta release milestone Oct 11, 2021
@aguschin aguschin added the p2-medium Medium priority label Nov 23, 2021
@aguschin aguschin changed the title Develop understanding of advanced deployment functionality Deploying models with seldon-core Aug 1, 2022
@aguschin aguschin added plugins Plugins and extensions for MLEM! and removed p2-medium Medium priority labels Aug 1, 2022
@aguschin
Copy link
Contributor Author

aguschin commented Aug 1, 2022

Maybe we can introduce MLEM inside the TPI in conjunction with Consul

@DavidGOrtega, I'm getting back to your comment since @mike0sv is working on Deployment 2.0 in MLEM. Could you please elaborate on your thought? How do you see this?

Beside that, could MLEM integrate with TPI to deploy to AWS EC2 or GCP?

@aguschin
Copy link
Contributor Author

aguschin commented Aug 1, 2022

@casperdcl @0x2b3bfa0 @DavidGOrtega

Also, we're now implementing deployment to Sagemaker. To do that e2e we need to provision some AWS resources for the user (Roles, Policies, etc). Can we use TPI to do that? Or do you have plans to implement an option to provision this?

@casperdcl
Copy link
Contributor

Looks like 3 different feature requests:

  1. "exposing the models via Seldon and Ambassador API Gateway [...] similar to the QuaroML stack [...] introduce MLEM inside the TPI in conjunction with Consul"
    • not sure I follow. Is this a new cloud feature request on the TPI repo @DavidGOrtega?
  2. "MLEM integrate with TPI to deploy to AWS EC2 or GCP"
    • sure, TPI can: provision an instance, upload workdir, run a script on the instance, auto-recover from spot interruptions (restores wokrdir & re-runs script), auto-cleanup on script exit, download workdir.
    • requirement: provide cloud (AWS/Azure/GCP) credentials via environment variables1
    • you can use docker2
    • you can use avoid having to config/expose service ports on AWS by using a (free) port forwarding service3
    • you can use the Python wrapper (pip install tpi) which auto-detects OS, downloads/caches terraform binaries, installs TPI, and even has a Python API so you don't explicitly run CLI
  3. "deployment to Sagemaker. To do that e2e we need to provision some AWS resources for the user (Roles, Policies, etc)"

Footnotes

  1. https://registry.terraform.io/providers/iterative/iterative/latest/docs/guides/authentication 2

  2. https://dvc.org/blog/local-experiments-to-cloud-with-tpi-docker

  3. https://github.com/iterative/blog-tpi-jupyter

@aguschin aguschin added the p2-medium Medium priority label Nov 9, 2022
@aguschin aguschin removed this from the Q3 milestone Nov 9, 2022
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
deploy Related to model deployment p2-medium Medium priority plugins Plugins and extensions for MLEM!
Projects
None yet
Development

No branches or pull requests

3 participants