A high-throughput and memory-efficient inference and serving engine for LLMs
-
Updated
May 26, 2024 - Python
A high-throughput and memory-efficient inference and serving engine for LLMs
☁️ Build multimodal AI applications with cloud-native stack
<⚡️> SuperAGI - A dev-first open source autonomous AI agent framework. Enabling developers to build, manage & run useful autonomous agents quickly and reliably.
RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.
The easiest way to serve AI/ML models in production - Build Model Inference Service, LLM APIs, Multi-model Inference Graph/Pipelines, LLM/RAG apps, and more!
Run any open-source LLMs, such as Llama 2, Mistral, as OpenAI compatible API endpoint in the cloud.
🔮 SuperDuperDB: Bring AI to your database! Build, deploy and manage any AI application directly with your existing data infrastructure, without moving your data. Including streaming inference, scalable model training and vector search.
ZenML 🙏: Build portable, production-ready MLOps pipelines. https://zenml.io.
AGiXT is a dynamic AI Agent Automation Platform that seamlessly orchestrates instruction management and complex task execution across diverse AI providers. Combining adaptive memory, smart features, and a versatile plugin system, AGiXT delivers efficient and comprehensive AI solutions.
Large Language Models (LLMs) tutorials & sample scripts, ft. langchain, openai, llamaindex, gpt, chromadb & pinecone
🐢 Open-Source Evaluation & Testing for LLMs and ML models
No-code multi-agent framework to build LLM Agents, workflows and applications with your data
UpTrain is an open-source unified platform to evaluate and improve Generative AI applications. We provide grades for 20+ preconfigured checks (covering language, code, embedding use-cases), perform root cause analysis on failure cases and give insights on how to resolve them.
LLMOps with Prompt Flow is a "LLMOps template and guidance" to help you build LLM-infused apps using Prompt Flow. It offers a range of features including Centralized Code Hosting, Lifecycle Management, Variant and Hyperparameter Experimentation, A/B Deployment, reporting for all runs and experiments and so on.
The all-in-one LLM developer platform: prompt management, evaluation, human feedback, and deployment all in one place.
🤖Self-Modifying Framework from the Future 🔮 World's First AMS
RAG (Retrieval Augmented Generation) Framework for building modular, open source applications for production by TrueFoundry
Add a description, image, and links to the llmops topic page so that developers can more easily learn about it.
To associate your repository with the llmops topic, visit your repo's landing page and select "manage topics."