Skip to content
@llm-operator

llm-operator

LLM Operatror

LLM Operator transform your GPU clusters into a powerhouse for generative AI workloads.

Key Values

  • Provide LLM as a service. LLM Operator builds a software stack that provides LLM as a service, including inference, fine-tuning, model management, and training data management.
  • Utilize GPU optimally. LLM Operator provides auto-scaling of inference-workloads, efficient scheduling of fine-tuning batch jobs, GPU sharing, etc.

Use Cases

  • Develop LLM applications with the API that is compatible with OpenAI-compatible API.
  • Fine-tune models while keeping data safely and securely in your on-premise datacenter.
  • Run fine-tuning jobs efficiently with guaranteed SLO and without interference with inference requests.

Installation

Please visit our documentation site.

A Demo Video

Please see this demo video.

Pinned

  1. llm-operator llm-operator Public

    LLM Operator

    Jupyter Notebook 1

  2. inference-manager inference-manager Public

    Inference Manager

    Go

  3. job-manager job-manager Public

    Fine-tuning Job Manager

    Go

  4. model-manager model-manager Public

    Model Manager

    Python

Repositories

Showing 10 of 16 repositories

People

This organization has no public members. You must be a member to see who’s a part of this organization.

Top languages

Loading…

Most used topics

Loading…