Lists (2)
Sort Name ascending (A-Z)
Stars
A high-throughput and memory-efficient inference and serving engine for LLMs
Any model. Any hardware. Zero compromise. Built with @ziglang / @openxla / MLIR / @bazelbuild
Gin is a HTTP web framework written in Go (Golang). It features a Martini-like API with much better performance -- up to 40 times faster. If you need smashing performance, get yourself some Gin.
🔍 A Hex Editor for Reverse Engineers, Programmers and people who value their retinas when working at 3 AM.
You like pytorch? You like micrograd? You love tinygrad! ❤️
Python SDK, Proxy Server (LLM Gateway) to call 100+ LLM APIs in OpenAI format - [Bedrock, Azure, OpenAI, VertexAI, Cohere, Anthropic, Sagemaker, HuggingFace, Replicate, Groq]
Language model alignment-focused deep learning curriculum
Implement a ChatGPT-like LLM in PyTorch from scratch, step by step
bottle.py is a fast and simple micro-framework for python web-applications.
Code for the paper "Language Models are Unsupervised Multitask Learners"
Port of OpenAI's Whisper model in C/C++
A collection of prompts, system prompts and LLM instructions
a fast cross platform AI inference engine 🤖 using Rust 🦀 and WebGPU 🎮
21 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
Get a ChatGPT plugin up and running in under 5 minutes!
The ChatGPT Retrieval Plugin lets you easily find personal or work documents by asking questions in natural language.
tiktoken is a fast BPE tokeniser for use with OpenAI's models.
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
gunicorn 'Green Unicorn' is a WSGI HTTP Server for UNIX, fast clients and sleepy applications.
An extremely fast Python package and project manager, written in Rust.