Scholar
Yuyang Ding
Google Scholar ID: 1E4sN7sAAAAJ
Soochow University
natural language processing
Follow
Homepage
↗
Google Scholar
↗
Citations & Impact
All-time
Citations
85
H-index
4
i10-index
4
Publications
9
Co-authors
0
Contact
Email
yyding.me@gmail.com
GitHub
Open ↗
Publications
8 items
Nex-N1: Agentic Models Trained via a Unified Ecosystem for Large-Scale Environment Construction
2025
Cited
0
FAPO: Flawed-Aware Policy Optimization for Efficient and Reliable Reasoning
2025
Cited
0
SCAN: Self-Denoising Monte Carlo Annotation for Robust Process Reward Learning
2025
Cited
0
Towards DS-NER: Unveiling and Addressing Latent Noise in Distant Annotations
IEEE Transactions on Knowledge and Data Engineering · 2025
Cited
0
A Survey of Slow Thinking-based Reasoning LLMs using Reinforced Learning and Inference-time Scaling Law
2025
Cited
0
Teaching LLMs for Step-Level Automatic Math Correction via Reinforcement Learning
2025
Cited
0
Unleashing LLM Reasoning Capability via Scalable Question Synthesis from Scratch
2024
Cited
10
Mathematical Language Models: A Survey
arXiv.org · 2023
Cited
4
Resume (English only)
Academic Achievements
NeurIPS 2025: FAPO: Flawed-Aware Policy Optimization for Efficient and Reliable Reasoning
ACL 2025: SCAN: Self-Denoising Monte Carlo Annotation for Robust Process Reward Learning
IEEE TKDE: DS-NER: Unveiling and Addressing Latent Noise in Distant Annotations
ACL 2024: ScaleQuest: Unleashing LLM Reasoning Capability via Scalable Question Synthesis from Scratch
COLING 2022: GNER: Rethinking Negative Instances for Generative Named Entity Recognition
SCIS (CCF-A): SelfMix: Robust Learning Against Textual Label Noise with Self-Mixup Training (*equal contribution)
EMNLP 2023: OpenBA: An Open-sourced 15B Bilingual Asymmetric seq2seq Model Pre-trained from Scratch (*equal contribution)
EMNLP 2022: CMD: a framework for Context-aware Model self-Detoxification
CCF Elite Collegiate Award
Gold Medal, ICPC National Invitational Programming Contest
Silver Medal, ICPC Asia-East Continent Final Contest (EC-Final)
Background
Third-year Ph.D. student at the Institute of Artificial Intelligence, Soochow University
Research focuses on LLM Reasoning
Particular interests in reinforcement learning, test-time scaling, and robust learning
Currently a research intern at Seed-Infrastructures, ByteDance, contributing to verl, an RL framework for LLMs
Research emphasizes joint optimization of algorithms and infrastructure for scalable and efficient reinforcement learning
Co-authors
0 total
Co-authors: 0 (list not available)
×
Welcome back
Sign in to Agora
Welcome back! Please sign in to continue.
Email address
Password
Forgot password?
Continue
Do not have an account?
Sign up