Published multiple papers, including 'LESA: Learnable LLM Layer Scaling-Up' (ACL 2025), 'LaCo: Large language model pruning via layer collapse' (EMNLP 2025, Findings), and more, with publications in top conferences such as ACL, EMNLP, ICLR, AAAI.
Research Experience
Feb. 2024 – Jun. 2024, Research Intern in Large-scale Models at Shanghai Artificial Intelligence Laboratory, mentored by Dr. Zhi Chen and Dr. Hang Yan; Aug. 2022 – Mar. 2023, Research Intern in Webpage Entity Extraction at Microsoft STCA, mentored by Principal Applied Scientist Manager Linjun Shou.
Education
Sep. 2021 – Jun. 2026 (expected), Ph.D. in Computer Science, Shanghai Jiao Tong University (SJTU), supervised by Prof. Hai Zhao; Sep. 2017 – Jun. 2021, B.S. in Computer Science, Central South University (CSU).
Background
Research Interests: Improving the efficiency of large language models, including but not limited to model pruning, efficient training, and KV cache optimization. Actively looking for a PostDoc position.
Miscellany
Passionate about various sports, including road cycling, fitness training, and basketball.