Over 10 years of experience in server-side architecture design and optimization.
-
Xiaomi
- Beijing
-
01:16
(UTC +08:00)
Pinned Loading
-
vllm-project/vllm
vllm-project/vllm PublicA high-throughput and memory-efficient inference and serving engine for LLMs
-
sgl-project/sglang
sgl-project/sglang PublicSGLang is a fast serving framework for large language models and vision language models.
-
pytorch/pytorch
pytorch/pytorch PublicTensors and Dynamic neural networks in Python with strong GPU acceleration
-
flashinfer
flashinfer PublicForked from flashinfer-ai/flashinfer
FlashInfer: Kernel Library for LLM Serving
Cuda
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.