Paper: Towards Efficient Privacy-Preserving Machine Learning: A Systematic Review from Protocol, Model, and System Perspectives, ACM Computing Survey (CSUR) Submission
Paper: H2EAL: Hybrid-Bonding Architecture with Hybrid Sparse Attention for Efficient Long-Context LLM Inference, ICCAD 2025
Paper: UniCAIM: A Unified CAM/CIM Architecture with Static-Dynamic KV Cache Pruning for Efficient Long-Context LLM Inference, DAC 2025
Paper: EQO: Exploring Ultra-Efficient Private Inference with Winograd-Based Protocol and Quantization Co-Optimization
Background
Currently a third-year master student at the Institute for Artificial Intelligence, Peking University (PKU), supervised by Prof. Meng Li and Prof. Runsheng Wang. Research interests primarily focus on efficient AI algorithms and systems, agentic AI, multimodal LLM, and LLM reasoning. Since 2022, he has also explored privacy-preserving machine learning (PPML), focusing on accelerating private inference systems via protocol-algorithm co-optimization.
Miscellany
Maintains a blog where he records his study notes and knowledge summaries about computer science since 2019 (over 300 blogs and 850,000+ visits).