The open-source serverless GPU container runtime.
-
Updated
May 29, 2024 - Go
The open-source serverless GPU container runtime.
Finetune LLMs on K8s by using Runbooks
A holistic way of understanding how LLaMA and its components run in practice, with code and detailed documentation.
开源的智能体项目 支持6种聊天平台 Onebotv11一对多连接 流式信息 agent 对话keyboard气泡生成 支持6种大模型接口(持续增加中) 具有将多种大模型接口转化为带有上下文的通用格式的能力.
Add a description, image, and links to the llm-inference topic page so that developers can more easily learn about it.
To associate your repository with the llm-inference topic, visit your repo's landing page and select "manage topics."