🔥 Blazingly fast ML inference server powered by Rust and Burn framework
-
Updated
Jul 18, 2025 - Rust
🔥 Blazingly fast ML inference server powered by Rust and Burn framework
A curated list of awesome open source and commercial platforms for serving models in production 🚀
Collection of OSS models that are containerized into a serving container
Serving large ml models independently and asynchronously via message queue and kv-storage for communication with other services [EXPERIMENT]
Miscellaneous codes and writings for MLOps
Big ML Project with infrastructure (MLflow, Minio, Grafana), backend (FastAPI, Catboost) and frontend (React, Maplibre)
Integrating Aporia ML model monitoring into a Bodywork serving pipeline.
Energy consumption of ML inference with Runtime Engines
🌐 Language identification for Scandinavian languages
Applied Machine Learning Projects
Heterogeneous System ML Pipeline Scheduling Framework with Triton Inference Server as Backend
Resources for serving models in production
Data extraction for Identifying architectural design decisions for achieving green ML serving
Example solution to the MLOps Case Study covering both online and batch processing.
Add a description, image, and links to the ml-serving topic page so that developers can more easily learn about it.
To associate your repository with the ml-serving topic, visit your repo's landing page and select "manage topics."