Skip to content
/ ACT Public

[ICML 2024] Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibration

Notifications You must be signed in to change notification settings

GATECH-EIC/ACT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 

Repository files navigation

Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibration

License: Apache 2.0

Zhongzhi Yu1,*, Zheng Wang1,*, Yonggan Fu1, Huihong Shi1, Khalid Shaikh1, Yingyan (Celine) Lin1

1 Georgia Institute of Technology

Accepted by ICML 2024

More info: [Paper]

The code is coming soon, stay tuned!

About

[ICML 2024] Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibration

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published