Selected papers: 'A Survey on Model Compression for Large Language Models', 'Distilling mathematical reasoning capabilities into Small Language Models', 'Optimal Rates for Agnostic Distributed Learning', 'Optimal Convergence Rates for Distributed Nyström Approximation', 'Convolutional Spectral Kernel Learning with Generalization Guarantees', 'Optimal Convergence Rates for Agnostic Nyström Kernel Learning', 'Multi-Class Learning: From Theory to Algorithm'.
Research Experience
Associate Professor at School of Artificial Intelligence, Beijing Normal University, 2024.09 - present;
Visiting Scholar at Microsoft Research Asia (NLC Group), 2024.04 - 2024.06;
Associate Research Fellow at Institute of Information Engineering, CAS, 2023.10 - 2024.09;
Postdoc Researcher at Institute of Information Engineering, CAS, 2020.09 - 2023.10.
Education
Ph.D. in Cyber Security from University of Chinese Academy of Sciences (UCAS), 2015.09 - 2020.06, supervised by Prof. Yong Liu and Prof. Weiping Wang;
Bachelor's degree in Software Engineering (International class) from Northeastern University, 2011.09 - 2015.06.
Background
Research interests include complex reasoning capabilities, reasoning reliability, and educational applications of large language models (LLMs). The goal is to enhance the complex reasoning abilities and reasoning reliability of LLMs through fundamental theoretical exploration and key technical investigations, realizing interpretable, trustworthy, and practically deployable LLM reasoning systems.