Skip to content

New Container: vLLM #885

@Josephrp

Description

@Josephrp

What is the new container you'd like to have?

A vLLM object container would serve testing and usage better.

Why not just use a generic container for this?

The Ollama object is not maintained against Ollama, which causes several issues:

  1. Performance: Ollama performs poorly with batch sizes of 50+ requests.
  2. Functionality: The new "search functionality" in Ollama requires ollama keys, which are not supported in the current object.
  3. Coverage & Quality: Ollama’s coverage and performance are lower compared to vLLM.

Why vLLM?

The vLLM object allows development and testing in a highly configurable, programmatic way, improving the overall testing experience.

Other references

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions