Browse publications on Google Scholar (top-right) ↗
Resume (English only)
Academic Achievements
Named ML and Systems Rising Star in 2025
Received MLSys'25 Outstanding Paper Award (Honorable Mention) for paper 'APOLLO: SGD-like Memory, AdamW-level Performance'
Oral presentation at ICML 2024 for 'GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection'
Published 'H2O: Heavy-Hitter Oracle for Efficient Generative Inference of Large Language Models' at NeurIPS 2023
Published 'R-Sparse: Rank-Aware Activation Sparsity for Efficient LLM Inference' at ICLR 2025
Co-authored multiple papers on efficient training and robustness at ICML 2021 and ICLR 2021, several with equal contribution
Background
Final-year Ph.D. student in the Department of Electrical and Computer Engineering at UT Austin
Research focuses on developing Efficient, Scalable, and Steerable machine learning algorithms and systems
Specific interests include: (i) Scalable Optimization for GenAI Model Training; (ii) Efficient Inference Algorithms; (iii) Long-Context Multimodal Modeling and Generation; (iv) Mechanistic Interpretability & Reasoning Enhancement of Foundation Models
Previously collaborated with Prof. Beidi Chen at CMU and Dr. Yuandong Tian at Meta to enhance practical AI skills
Will be on the job market starting Fall 2025, seeking full-time research positions in industry