This is the official PyTorch implementation of "LLMC: Benchmarking Large Language Model Quantization with a Versatile Compression Toolkit".
-
Updated
Sep 27, 2024 - Python
This is the official PyTorch implementation of "LLMC: Benchmarking Large Language Model Quantization with a Versatile Compression Toolkit".
Private chat with local GPT with document, images, video, etc. 100% private, Apache 2.0. Supports oLLaMa, Mixtral, llama.cpp, and more. Demo: https://gpt.h2o.ai/ https://gpt-docs.h2o.ai/
Design, conduct and analyze results of AI-powered surveys and experiments. Simulate social science and market research with large numbers of AI agents and LLMs.
On-device LLM Inference Powered by X-Bit Quantization
Firefly: 大模型训练工具,支持训练Qwen2.5、Qwen2、Yi1.5、Phi-3、Llama3、Gemma、MiniCPM、Yi、Deepseek、Orion、Xverse、Mixtral-8x7B、Zephyr、Mistral、Baichuan2、Llma2、Llama、Qwen、Baichuan、ChatGLM2、InternLM、Ziya2、Vicuna、Bloom等大模型
A snappy, keyboard-centric terminal user interface for interacting with large language models. Chat with ChatGPT, Claude, Llama 3, Phi 3, Mistral, Gemma and more.
Easy "1-line" calling of all LLMs from OpenAI, MS Azure, AWS Bedrock, GCP Vertex, and Ollama
GPT-4 level function calling models for real-world tool using use cases
🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
NoLlama is a lightweight terminal-based alternative to Ollama, enabling interaction with large language models like GPT-4o and LLaMA 70B directly from your terminal. It offers a sleek UI, multiple model choices, colorful markdown rendering, and low memory usage, all without needing a browser.
An efficient, flexible and full-featured toolkit for fine-tuning LLM (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)
🦙 Free and Open Source Large Language Model (LLM) chatbot web UI and API. Self-hosted, offline capable and easy to setup. Powered by LangChain.
LLMs prompt augmentation with RAG by integrating external custom data from a variety of sources, allowing chat with such documents
XMPP Bot designed for E2EE AI language model interactions
Use groq for evaluations
Add a description, image, and links to the mixtral topic page so that developers can more easily learn about it.
To associate your repository with the mixtral topic, visit your repo's landing page and select "manage topics."