Skip to content

gaooooosh/Gaooooosh

Repository files navigation


👋 ABOUT

🎓 Ph.D. Student
Beijing University of Posts and Telecommunications

GitHub Twitter LinkedIn


>🔬 Research Focus

Efficient large language model architectures for long-context modeling and stability in ultra-long contexts.

Current work studies sequence-length-wise hybridization of linear and sliding-window attention to reduce computational and memory costs, while mitigating context-length-induced degradation through training-free, parameter-preserving inference-time modification.

Long-Context Linear Attention Efficient Transformers


📊 STATS

📈 CONTRIBUTIONS

今日提交 昨日提交 本月平均


🚀 RECENT ACTIVITY


About

Config files for my GitHub profile.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages