Skip to content
View KuntaiDu's full-sized avatar

Block or report KuntaiDu

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this userโ€™s behavior. Learn more about reporting abuse.

Report abuse
KuntaiDu/README.md

Hi, I'm Kuntai Du ๐Ÿ‘‹

I'm a PhD student @ UChicago, graduating, working in Large Language Model Inference. Check my home page for more about me!

๐Ÿ”ง Experiences

  • ๐Ÿš€ Working on vLLM project(GitHub Stars) as vLLM core team member and committer. My contributions:
    • Performance comparison with other LLM inference engines: the end of the blog.
    • Features: Disaggregated prefilling and CPU offloading.
  • ๐Ÿ’พ Contributing to the LMCache project, exploring fun ideas in KV caches.

๐ŸŽฎ Hobbies and Interests

  • ๐ŸŽฎ Gaming: League of Legends, Stardew Valley, Go
  • ๐Ÿ’ƒ Street Dance: Locking main, but I also dance waacking.
  • ๐ŸŽค Singing: Loch Lomond and ไผ ๅฅ‡ Legend

๐Ÿ“ง Contact

Pinned Loading

  1. vllm-project/vllm Public

    A high-throughput and memory-efficient inference and serving engine for LLMs

    Python 53.2k 8.9k

  2. vllm-project/production-stack Public

    vLLMโ€™s reference system for K8S-native cluster-wide deployment with community-driven performance optimization

    Python 1.5k 241

  3. LMCache/LMCache Public

    Supercharge Your LLM with the Fastest KV Cache Layer

    Python 3.3k 382