[{'PaperTitle': 'FlashRL: 8Bit Rollouts, Full Power RL', 'Authors': 'Liyuan Liu*, Feng Yao* et al.', 'PublicationStatus': 'Preprint', 'PublicationYear': '2025'}, {'PaperTitle': 'Your Efficient RL Framework Secretly Brings You Off-Policy RL Training', 'Authors': 'Feng Yao*, Liyuan Liu* et al.', 'PublicationStatus': 'Preprint', 'PublicationYear': '2025'}, {'PaperTitle': 'DenseMixer: Improving MoE Post-Training with Precise Router Gradients', 'Authors': 'Feng Yao et al.', 'PublicationStatus': 'Preprint', 'PublicationYear': '2025'}, {'PaperTitle': 'Training Language Models to Generate Quality Code with Program Analysis Feedback', 'Authors': 'Feng Yao*, Zilong Wang* et al.', 'PublicationStatus': 'NeurIPS 2025', 'PublicationYear': '2025'}, {'PaperTitle': 'Revisiting Reinforcement Learning for LLM Reasoning from A Cross-Domain Perspective', 'Authors': 'Zhoujun Cheng*, Shibo Hao* et al.', 'PublicationStatus': 'Preprint', 'PublicationYear': '2025'}, {'PaperTitle': 'Data Contamination Can Cross Language Barriers', 'Authors': 'Feng Yao*, Yufan Zhuang* et al.', 'PublicationStatus': 'EMNLP 2024', 'PublicationYear': '2024'}, {'PaperTitle': 'Configurable Foundation Models: Building LLMs from a Modular Perspective', 'Authors': 'Chaojun Xiao, Zhengyan Zhang et al.', 'PublicationStatus': 'Preprint', 'PublicationYear': '2024'}]
Background
Research Interests: Intersection of Natural Language Processing and Deep Learning. Recently focused on training Mixture-of-Experts (MoE) and improving the efficiency of large-scale Reinforcement Learning.
Miscellany
Invited Talks:
- Applied Compute (ex-OpenAI startup), November 10, 2025