Skip to content

vllm-project/vllm projects

Search results

  • DeepSeek V3/R1 Template

    #5 updated Mar 27, 2025
    2025-02-25: DeepSeek V3/R1 is supported with optimized block FP8 kernels, MLA, MTP spec decode, multi-node PP, EP, and W4A16 quantization
  • #6 updated Mar 27, 2025
    A list of onboarding tasks for first-time contributors to get started with vLLM.
  • #8 updated Mar 27, 2025
    Main tasks for the multi-modality workstream (#4194)
  • #10 updated Mar 26, 2025
    Community requests for multi-modal models
  • Ray

    #7 updated Mar 25, 2025
    Tracks Ray issues and pull requests in vLLM
  • #2 updated Mar 19, 2025
  • #1 updated Feb 28, 2025
    [Testing] Optimize V1 PP efficiency.