Skip to content

vllm-project/vllm projects

Search results

  • Ray

    #7 updated Mar 22, 2025
    Tracks Ray issues and pull requests in vLLM
  • DeepSeek V3/R1 Template

    #5 updated Mar 22, 2025
    2025-02-25: DeepSeek V3/R1 is supported with optimized block FP8 kernels, MLA, MTP spec decode, multi-node PP, EP, and W4A16 quantization
  • #6 updated Mar 21, 2025
    A list of onboarding tasks for first-time contributors to get started with vLLM.
  • #8 updated Mar 20, 2025
    Main tasks for the multi-modality workstream (#4194)
  • #10 updated Mar 20, 2025
    Community requests for multi-modal models
  • #2 updated Mar 19, 2025
  • #1 updated Feb 28, 2025
    [Testing] Optimize V1 PP efficiency.