Published several high-level papers, such as 'Don’t Take Things Out of Context: Attention Intervention for Enhancing Chain-of-Thought Reasoning in Large Language Models' (ICLR 2025) and 'CrossFormer++: A Versatile Vision Transformer Hinging on Cross-Scale Attention' (IEEE Trans. Pattern Anal. Mach. Intell. 2024).
Research Experience
Author or co-author of multiple research projects, including papers presented at ICLR 2025, ICML 2024, NeurIPS 2024, and other international conferences.
Education
Received a bachelor's degree in computer science and technology from Zhejiang University in 2017 and a Ph.D. in 2022.
Background
Assistant Professor and doctoral supervisor at the School of Software Technology, Zhejiang University. Research interests include LLM-based agent and multimodal large language model (MLLM).
Miscellany
Welcomes outstanding students (e.g., ACM/ICPC award winners, first authors of CCF-A papers, those with extensive project experience, or those with a strong passion for research) to join his lab for graduate studies or undergraduate research internships.