[FR] Adding SageMaker provider to MLflow Gateway AI #10351
Labels
area/deployments
MLflow Deployments client APIs, server, and third-party Deployments integrations
enhancement
New feature or request
Willingness to contribute
Yes. I can contribute this feature independently.
Proposal Summary
SageMaker JumpStart provides an easy way to deploy LLM endpoints on the SageMaker managed infrastructure. However, as of today, the MLflow Gateway AI does not implement a SageMaker provider.
We would like to introduce the SageMaker provider and adapter together with an example script to show how you can add a SageMaker hosted model and make it available to the MLflow Gateway AI
Motivation
SageMaker jumpstarts supports many LLMs with a one click deployment. Would be nice to have a way to interface with these models in a centralized way as offered by mlflow Gateway AI
Sagemaker is a popular tool, and some models are readily available directly there. Would open up to a great deal of possibilities
I want to have access to more providers
The provider/adapters should not be too hard, of course the evil is in the details.
Details
I would like to provide a basic SageMaker provider, which would handle the credentials and provide the basic AWS specific credentials mechanisms (similar as done already for Bedrock).
I would then provide two Adapter for two models (thinking on providing samples for the two LLama2 models, likely the 7B one).
This can be a baseline for more models deployable via sagemaker that can be extended. Users might be able to quickly integrate new adapters, requiring some sort of plugin feature, but this might be for the future.
What component(s) does this bug affect?
area/artifacts
: Artifact stores and artifact loggingarea/build
: Build and test infrastructure for MLflowarea/docs
: MLflow documentation pagesarea/examples
: Example codearea/gateway
: AI Gateway service, Gateway client APIs, third-party Gateway integrationsarea/model-registry
: Model Registry service, APIs, and the fluent client calls for Model Registryarea/models
: MLmodel format, model serialization/deserialization, flavorsarea/recipes
: Recipes, Recipe APIs, Recipe configs, Recipe Templatesarea/projects
: MLproject format, project running backendsarea/scoring
: MLflow Model server, model deployment tools, Spark UDFsarea/server-infra
: MLflow Tracking server backendarea/tracking
: Tracking Service, tracking client APIs, autologgingWhat interface(s) does this bug affect?
area/uiux
: Front-end, user experience, plotting, JavaScript, JavaScript dev serverarea/docker
: Docker use across MLflow's components, such as MLflow Projects and MLflow Modelsarea/sqlalchemy
: Use of SQLAlchemy in the Tracking Service or Model Registryarea/windows
: Windows supportWhat language(s) does this bug affect?
language/r
: R APIs and clientslanguage/java
: Java APIs and clientslanguage/new
: Proposals for new client languagesWhat integration(s) does this bug affect?
integrations/azure
: Azure and Azure ML integrationsintegrations/sagemaker
: SageMaker integrationsintegrations/databricks
: Databricks integrationsThe text was updated successfully, but these errors were encountered: