-
-
Notifications
You must be signed in to change notification settings - Fork 12.6k
Closed
Labels
documentationImprovements or additions to documentationImprovements or additions to documentation
Description
📚 The doc issue
Looking at https://docs.vllm.ai/en/latest/serving/env_vars.html for VLLM_ATTENTION_BACKEND:
- "FLASH_ATTN": use FlashAttention
This leaves me wondering if it's flash attention 1 or 2.
Suggest a potential alternative/fix
Appending this description with a 1 or 2, such as:
- "FLASH_ATTN": use FlashAttention-2
Before submitting a new issue...
- Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
Metadata
Metadata
Assignees
Labels
documentationImprovements or additions to documentationImprovements or additions to documentation