forked from vllm-project/vllm
-
Notifications
You must be signed in to change notification settings - Fork 2
A high-throughput and memory-efficient inference and serving engine for LLMs
License
OthersideAI/vllm
ErrorLooks like something went wrong!
About
A high-throughput and memory-efficient inference and serving engine for LLMs
Resources
License
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published
Languages
- Python 82.5%
- Cuda 16.2%
- Other 1.3%