Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibration
Zhongzhi Yu1,*, Zheng Wang1,*, Yonggan Fu1, Huihong Shi1, Khalid Shaikh1, Yingyan (Celine) Lin1
1 Georgia Institute of Technology
Accepted by ICML 2024
More info: [Paper]
The code is coming soon, stay tuned!