Skip to content

Reusable containerized inference service focused on throughput, latency, and observability.

Notifications You must be signed in to change notification settings

rlubenow/cv-serve

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 

Repository files navigation

CVServe

Containerized CV inference service that loads models and scores via API.

Problem

Shipping computer vision models often requires bespoke services per model, which slows iteration and makes deployment inconsistent across teams.

Solution

CVServe packages multiple CV models into a single containerized service with a unified API. It supports model hot‑swap, reproducible deployments, and low‑latency inference at scale.

Tech Stack

  • Docker
  • FastAPI
  • PyTorch
  • ONNX

Phases

Phase 1

  • Model loading and API‑based inference endpoints
  • Standardized request/response schema

Phase 2

  • Multi‑model management and versioning
  • Performance tuning and scalable deployment patterns

Status

In progress.

About

Reusable containerized inference service focused on throughput, latency, and observability.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published