Stars
- All languages
- Assembly
- Astro
- Batchfile
- C
- C#
- C++
- CMake
- CSS
- Cuda
- Cython
- Dart
- Dockerfile
- Go
- HTML
- Java
- JavaScript
- Jinja
- Jupyter Notebook
- Kotlin
- LLVM
- Lua
- MATLAB
- Makefile
- Markdown
- Meson
- Mustache
- Objective-C
- Objective-C++
- PHP
- PowerShell
- PureBasic
- Python
- RAML
- Reason
- Ruby
- Rust
- SCSS
- Scala
- Shell
- Swift
- TeX
- TypeScript
- Vim Script
- Vue
- YAML
- Zig
Container runtimes on macOS (and Linux) with minimal setup
A toolkit to run Ray applications on Kubernetes
A Cloud Native Batch System (Project under CNCF)
Heterogeneous AI Computing Virtualization Middleware
A golang framework helps gopher to build a data visualization and admin panel in ten minutes
CLI platform to experiment with codegen. Precursor to: https://lovable.dev
go-admin外部核心代码库,为go-admin提供原生代码支持,发展go-admin自身周边生态
基于Gin + Vue + Element UI & Arco Design & Ant Design 的前后端分离权限管理系统的前端模块
基于Gin + Vue + Element UI & Arco Design & Ant Design 的前后端分离权限管理系统脚手架(包含了:多租户的支持,基础用户管理功能,jwt鉴权,代码生成器,RBAC资源控制,表单构建,定时任务等)3分钟构建自己的中后台项目;项目文档》:https://www.go-admin.pro V2 Demo: https://vue2.go-admin.d…
NVIDIA GPU Operator creates, configures, and manages GPUs in Kubernetes
A modern load testing tool, using Go and JavaScript - https://k6.io
Master programming by recreating your favorite technologies from scratch.
LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficie…
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
The official home of the Presto distributed SQL query engine for big data
An async Python micro framework for building web applications.
DeepChat - 连接强大AI与个人世界的智能助手 | DeepChat - A smart assistant that connects powerful AI to your personal world
No fortress, purely open ground. OpenManus is Coming.
Benchmarking the serving capabilities of vLLM
Autoscale LLM (vLLM, SGLang, LMDeploy) inferences on Kubernetes (and others)
Distributed ML Training and Fine-Tuning on Kubernetes
Kubernetes Operator for MPI-based applications (distributed training, HPC, etc.)
The RunPod worker template for serving our large language model endpoints. Powered by vLLM.
A web-based UI for deploying and managing applications in Kubernetes clusters
A high-performance distributed file system designed to address the challenges of AI training and inference workloads.
FlashInfer: Kernel Library for LLM Serving