π» AI Infra Engineer @JD.com
β‘ LLM Inference: Speculative Decoding, KV Cache Compression, Quantization & Distillation.
π΅ Prev: AI Music Training & Inference
-
@JD.com
- Beijing
Popular repositories Loading
-
smoothquant
smoothquant PublicForked from mit-han-lab/smoothquant
SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
Python
-
FasterTransformer
FasterTransformer PublicForked from void-main/FasterTransformer
Transformer related optimization, including BERT, GPT
C++
-
sglang
sglang PublicForked from jessiewei7/sglang
SGLang is a high-performance serving framework for large language models and multimodal models.
Python
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.