I just wondered how things were put together -- Claude Shannon
MLSys is cool, i guess. Otherwise I ramble to my IRL friends.
I just wondered how things were put together -- Claude Shannon
MLSys is cool, i guess. Otherwise I ramble to my IRL friends.
A high-throughput and memory-efficient inference and serving engine for LLMs
Run any open-source LLMs, such as DeepSeek and Llama, as OpenAI compatible API endpoint in the cloud.
Use your Neovim like using Cursor AI IDE!
🌱 a fast, batteries-included static-site generator that transforms Markdown content into fully functional websites
The easiest way to serve AI apps and models - Build Model Inference APIs, Job queues, LLM apps, Multi-model pipelines, and more!