Published several papers, including 'Cost-Aware Contrastive Routing for LLMs' (Neurips 2025), 'Bilevel ZOFO: Bridging Parameter-Efficient and Zeroth-Order Techniques for Efficient LLM Fine-Tuning and Meta-Training' (Neurips 2025), 'Revisiting Convergence: A Study on Shuffling-Type Gradient Methods' (ICML 2025), 'Efficient Fine-Tuning and Concept Suppression for Pruned Diffusion Models' (CVPR 2025), and more. Involved in multiple research projects.
Research Experience
Served as a Postdoctoral Associate in the Department of Electrical and Computer Engineering at the University of Pittsburgh and the Department of Computer Science at the University of Maryland from 2021 to 2024, mentored by Professor Heng Huang.
Education
Received a Bachelor's degree in Mathematics and Applied Mathematics from Beijing Normal University in 2016; obtained a Doctor of Philosophy in Applied Mathematics from The Hong Kong Polytechnic University in 2021, under the supervision of Professors Ting Kei Pong and Xiaojun Chen.
Background
Research interests: Machine Learning, Optimization and Foundation models including Bilevel Optimization, Federated Learning, LLMs, diffusion models, Adversarial Training, Meta-learning, Hyper Representation Learning, Data Hyper Cleaning, Continuous Optimization. Currently an assistant professor at the Department of Computer Science and Engineering, University of Texas Arlington.