Skip to content
@bentoml

BentoML

The easiest way to run AI Inference in the cloud

Welcome to BentoML šŸ‘‹ Twitter Follow Slack

BentoML

What is BentoML? šŸ‘©ā€šŸ³

BentoML is an open-source model serving library for building model inference APIs and multi-model serving systems with any open-source or custom AI models. It comes with everything you need for serving optimization, model packaging, and simplifies production deployment via ā˜ļø BentoCloud.

Get in touch šŸ’¬

šŸ‘‰ Join our Slack community!

šŸ‘€ Follow us on X @bentomlai and LinkedIn

šŸ“– Read our blog

Pinned

  1. BentoML BentoML Public

    The easiest way to serve AI/ML models in production - Build Model Inference Service, LLM APIs, Multi-model Inference Graph/Pipelines, LLM/RAG apps, and more!

    Python 6.7k 754

  2. OpenLLM OpenLLM Public

    Run any open-source LLMs, such as Llama 2, Mistral, as OpenAI compatible API endpoint in the cloud.

    Python 9k 575

Repositories

Showing 10 of 71 repositories

Sponsoring

  • @pdm-project

Top languages

Loadingā€¦

Most used topics

Loadingā€¦