Collection of OSS models that are containerized into a serving container
-
Updated
Sep 19, 2023 - Python
Collection of OSS models that are containerized into a serving container
Serving large ml models independently and asynchronously via message queue and kv-storage for communication with other services [EXPERIMENT]
🌐 Language identification for Scandinavian languages
Resources for serving models in production
Heterogeneous System ML Pipeline Scheduling Framework with Triton Inference Server as Backend
Add a description, image, and links to the ml-serving topic page so that developers can more easily learn about it.
To associate your repository with the ml-serving topic, visit your repo's landing page and select "manage topics."