Skip to content

XiaoYee/Awesome_Efficient_LRM_Reasoning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

88 Commits
 
 
 
 
 
 

Repository files navigation

A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, and Beyond

Paper Github Twitter

Awesome License: MIT


🔔 News

  • [2025-05] We include LIMOPro for Efficient and Effective Reasoning in Test-time Scaling.
  • [2025-05] We update more papers on adaptive reasoning, which describe how a system/model autonomously switches between long and short reasoning chains based on problem complexity.
  • [2025-05] Welcome to our latest paper "Scaling Reasoning, Losing Control", which shows that the longer the reasoning chain, the poorer its instruction-following ability. Therefore, efficient reasoning may also be important for instruction following in LRMs.
  • [2025-04] We include AgentPrune, where efficient reasoning is important for agent systems.
  • [2025-04] We include benchmarks for Efficient Reasoning: MME-CoT, S1-Bench, DUMB500.
  • [2025-04] We add Mamba Reasoning models (e.g M1) and Hybrid models (e.g Mamba-Transformer) in Efficient Reasoning during Pre-training. It is naturally efficient to infer.
  • [2025-04] We add a new "Model Merge" category in Efficient Reasoning during Inference. It is feasible to be a promising direction.
  • [2025-04] 📢 Our work is reported by both Synced (机器之心) and Zhuanzhi (专知).
  • [2025-03] 📢 Our work is reported by both Deep Learning and NLP (深度学习自然语言处理) and Machine Learning and NLP (机器学习算法与自然语言处理).
  • [2025-03] We released our survey "A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, and Beyond". This is the first survey for efficient reasoning of Large Reasoning Models, covering language, multimodality, agent, and applications. We provide several promising future directions in our survey.
  • [2025-03] We created this repository to maintain a paper list on Awesome-Efficient-LRM-Reasoning.

Author

Taxonomy

Citation

If you find our survey useful for your research, please consider citing:

@article{qu2025survey,
  title={A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, and Beyond},
  author={Qu, Xiaoye and Li, Yafu and Su, Zhaochen and Sun, Weigao and Yan, Jianhao and Liu, Dongrui and Cui, Ganqu and Liu, Daizong and Liang, Shuxian and He, Junxian and others},
  journal={arXiv preprint arXiv:2503.21614},
  year={2025}
}


Category

🔥 Table of Contents


📜Content

👀 Introduction

In the age of LRMs, we propose that "Efficiency is the essence of intelligence." Just as a wise human knows when to stop thinking and start deciding, a wise model should know when to halt unnecessary deliberation. An intelligent model should manipulate the token economy, i.e., allocating tokens purposefully, skipping redundancy, and optimizing the path to a solution. Rather than naively traversing every possible reasoning path, it should emulate a master strategist, balancing cost and performance with elegant precision.

To summarize, this survey makes the following key contributions to the literature:

  • Instead of offering a general overview of LRMs, we focus on the emerging and critical topic of efficient reasoning in LRMs, providing an in-depth and targeted analysis.
  • We identify and characterize common patterns of reasoning inefficiency, and outline the current challenges that are unique to improving reasoning efficiency in large models.
  • We provide a comprehensive review of recent advancements aimed at enhancing reasoning efficiency, structured across the end-to-end LRM development pipeline, from pretraining and supervised fine-tuning to reinforcement learning and inference.

🚀 Papers

💭 Efficient Reasoning during Inference

Length Budgeting

System Switch

Model Switch

Model Merge

Parallel Search

💫 Efficient Reasoning with SFT

Reasoning Chain Compression

Latent-Space SFT

🧩 Efficient Reasoning with Reinforcement Learning

Efficient Reinforcement Learning with Length Reward

Efficient Reinforcement Learning without Length Reward

💬 Efficient Reasoning during Pre-training

Pretraining with Latent Space

Subquadratic Attention

Linearization

Efficient Reasoning with Subquadratic Attention

🔖 Future Directions

Efficient Multimodal Reasoning and Video Reasoning

Efficient Test-time Scaling and Infinity Thinking

Efficient and Trustworthy Reasoning

Building Efficient Reasoning Applications (RAG, Tool, Agent)

Evaluation and Benchmark


🎉 Contribution

Contributing to this paper list

⭐" Join us in improving this repository! If you know of any important works we've missed, please contribute. Your efforts are highly valued! "

Contributors


Releases

No releases published

Packages

No packages published

Contributors 5