Scholar
Zhengyang Tang
Google Scholar ID: 2RRV0PQAAAAJ
CUHKSZ
Large Language Models
Mathematical Reasoning
Information Retrieval
Follow
Homepage
↗
Google Scholar
↗
Citations & Impact
All-time
Citations
3,963
H-index
8
i10-index
8
Publications
13
Co-authors
9
list available
Contact
Email
zhengyangtang@link.cuhk.edu.cn
CV
Open ↗
Twitter
Open ↗
GitHub
Open ↗
Publications
9 items
Do Phone-Use Agents Respect Your Privacy?
2026
Cited
0
Teaching Language Models to Reason with Tools
2025
Cited
0
CALM Before the STORM: Unlocking Native Reasoning for Optimization Modeling
2025
Cited
0
CoRT: Code-integrated Reasoning within Thinking
2025
Cited
0
Learning from Peers in Reasoning Models
2025
Cited
0
RealCritic: Towards Effectiveness-Driven Evaluation of Language Model Critiques
2025
Cited
0
Enabling Scalable Oversight via Self-Evolving Critic
2025
Cited
0
Second Language (Arabic) Acquisition of LLMs via Progressive Vocabulary Expansion
arXiv.org · 2024
Cited
5
Load more
Resume (English only)
Academic Achievements
NeurIPS 2025: CoRT (Code-integrated Reasoning within Thinking)
COLM 2025: SCRIT (Self-Evolving Critique Abilities in LLMs)
ICML 2024: MathScale (Scaling Instruction Tuning for Mathematical Reasoning)
TMLR 2025: GLAN (Generalized Instruction Tuning for Language Models)
ACL 2025: Second Language (Arabic) Acquisition of LLMs via Progressive Vocabulary Expansion (Oral & Panel)
COLING 2022: DPTDR (Deep Prompt Tuning for Dense Passage Retrieval)
Operations Research 2025: ORLM (Customizable Framework for Automated Optimization Modeling)
Contributed to Qwen3 Technical Report (Tool-integrated Reasoning)
Background
Ph.D. candidate at The Chinese University of Hong Kong, Shenzhen, advised by Prof. Benyou Wang.
Research focuses on developing intelligent agents capable of complex reasoning and self-improvement.
Pioneers agentic frameworks leveraging reinforcement learning (RL) for tool-integrated tasks.
Proposed SCRIT, a self-evolving critique model serving as a generative reward model for scalable, supervision-free oversight.
Designed novel instruction tuning frameworks—MathScale, GLAN, and ALAN—for scalable high-quality training data generation.
Also works on efficient information access methods (e.g., DPTDR), achieving top performance on competitive benchmarks.
Co-authors
9 total
Benyou Wang
Assistant Professor, The Chinese University of Hong Kong, Shenzhen
Xingxing Zhang
Microsoft Research
Furu Wei
Distinguished Scientist, Microsoft Research
Ziniu Li
The Chinese University of Hong Kong, Shenzhen
Dayiheng Liu (刘大一恒)
Qwen Team, Alibaba Group
Junyang Lin
Qwen Team, Alibaba Group & Peking University
Li Dong
Microsoft Research
Chengpeng Li
USTC
×
Welcome back
Sign in to Agora
Welcome back! Please sign in to continue.
Email address
Password
Forgot password?
Continue
Do not have an account?
Sign up