👨🌾 This is Huiqiang Jiang (姜慧强)'s homepage.
Research SDE in Microsoft Research Asia (Shanghai),
a fake MLsys/NLPer Google schoal,
Research focus on Efficient Methods (in LLMs)
A unpopular blogger Blog & Zhihu
A programming enthusiast @iofu728
Phone: +86 178 xxxx xxxx
Email: hjiang[aT]microsoft[DoT.]com
Huiqiang Jiang obtained his Master's Degree in Software Engineering from Peking University, working with A.P. Xiang Jing. And also was a research intern at the KC Group, Microsoft Research Asia (19/6-21/3) with Börje Karlsson and Guoxin Wang as well as the search group, Ant Group (20/6-20/8).
Huiqiang's research primarily concentrates on efficient methods to accelerate inference or training, including dynamic sparse attention (MInference, RetrievalAttention), prompt compression (LLMLingua), KV-cache compression, speculative decoding, model compression, sparse inference (PIT), neural architecture search (NAS), and efficient tuning, with a particular emphasis on LLMs. Additionally, he is interested in addressing typical challenges in natural language processing.
He's looking for one research intern in efficient methods. Please get in touch with him (hjiang[aT]microsoft[DoT.]com) if you are interested in the research topics.