LLM-Research
LLM for Long Text Summary (Comprehensive Bulleted Notes)
RAGFlow is a leading open-source Retrieval-Augmented Generation (RAG) engine that fuses cutting-edge RAG with Agent capabilities to create a superior context layer for LLMs
A curated compilation of AI-driven generative music resources and projects. Explore the blend of machine learning algorithms and musical creativity.
A framework for few-shot evaluation of language models.
OpenChat: Advancing Open-source Language Models with Imperfect Data
Curated tutorials and resources for Large Language Models, Text2SQL, Text2DSL、Text2API、Text2Vis and more.
Erasing concepts from neural representations with provable guarantees
[CVPR 2024] "MACE: Mass Concept Erasure in Diffusion Models" (Official Implementation)
[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
Public repo to document some SPR stuff
[EMNLP'23, ACL'24] To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.
[CCS'24] A dataset consists of 15,140 ChatGPT prompts from Reddit, Discord, websites, and open-source datasets (including 1,405 jailbreak prompts).
A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (p…
This repo contains the source code for RULER: What’s the Real Context Size of Your Long-Context Language Models?
High-speed Large Language Model Serving for Local Deployment
[NeurIPS 2024] The official implementation of paper: Chain of Preference Optimization: Improving Chain-of-Thought Reasoning in LLMs.
[ACL 2024] Progressive LLaMA with Block Expansion.
WikiChat is an improved RAG. It stops the hallucination of large language models by retrieving data from a corpus.
A pure and fast NumPy implementation of Mamba with cache support.
Repository for Meta Chameleon, a mixed-modal early-fusion foundation model from FAIR.
Code for NeurIPS'24 paper 'Grokked Transformers are Implicit Reasoners: A Mechanistic Journey to the Edge of Generalization'
Code for the paper 🌳 Tree Search for Language Model Agents
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
[Paper List] Papers integrating knowledge graphs (KGs) and large language models (LLMs)






