forked from vllm-project/vllm
-
Notifications
You must be signed in to change notification settings - Fork 0
A high-throughput and memory-efficient inference and serving engine for LLMs
License
AnswerDotAI/vllm
ErrorLooks like something went wrong!
About
A high-throughput and memory-efficient inference and serving engine for LLMs
Resources
License
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published
Languages
- Python 80.7%
- Cuda 14.0%
- C++ 2.8%
- Shell 1.0%
- C 1.0%
- CMake 0.4%
- Dockerfile 0.1%