- 2025.03: Paper 'A Survey on the Honesty of Large Language Models' accepted by TMLR
- 2025.02: Released a blog about KL divergence in RL algorithm (Chinese)
- 2025.01: Released a report and blog (Chinese) for the unified view of Attention and MoE
- 2024.11: Two papers accepted by COLING 2025
- 2024.11: Released a report on modeling Binary Quantization via Convex Optimization Methods
- 2024.09: One paper accepted by EMNLP 2024
- 2024.06: Released a blog (English/Chinese) about improving LoRA
- 2024.04: Released a blog (English/Chinese) about rethinking KL divergence in KD for LLMs
- 2024.03: One paper accepted by NAACL 2024 findings
- 2023.02: Two papers accepted by ICASSP 2023
- 2022.10: One research long paper accepted by WSDM 2023
Research Experience
- 2021.3~2022.5, Tencent Internship
- 2022.5~2023.5, Tencent Rahio Research Internship
Education
- Ph.D.: The University of Hong Kong, time not specified
- Master: Tsinghua University, graduated in 2023, supervised by Prof. Yujiu Yang
- Bachelor: Department of Automation, Tsinghua University, graduated in 2020
Background
- Research Interests: Efficient reasoning methods for large language models
- Professional Field: Automation
- Biography: Currently a Ph.D. student at NGai Lab, HKU. Previously, completed his master's degree at IIGroup, Tsinghua University, supervised by Prof. Yujiu Yang.
Miscellany
Information on personal interests and hobbies is insufficient