Rubric-Based Benchmarking and Reinforcement Learning for Advancing LLM Instruction Following

📅 2025-11-13
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Current large language models (LLMs) face significant challenges in evaluating and improving instruction following (IF) capabilities—particularly for complex, multi-turn, and system-level instructions—due to the lack of high-quality benchmarks and unreliable, uninterpretable reward signals. To address these limitations, we propose AdvancedIF, the first fine-grained, human-annotated benchmark specifically designed for advanced IF evaluation. Complementing this, we introduce RIFL, a novel training framework that, for the first time, decouples expert-crafted scoring rubrics into learnable, structured reward signals. RIFL integrates a rubric-verification model, reward shaping, and reinforcement learning-based post-training to enable precise, interpretable IF modeling. Our approach balances annotation reliability with scalable automated feedback. On AdvancedIF, RIFL achieves a 6.7% absolute improvement; it also demonstrates strong generalization across multiple public benchmarks, validating both its effectiveness and interpretability.

Technology Category

Application Category

📝 Abstract
Recent progress in large language models (LLMs) has led to impressive performance on a range of tasks, yet advanced instruction following (IF)-especially for complex, multi-turn, and system-prompted instructions-remains a significant challenge. Rigorous evaluation and effective training for such capabilities are hindered by the lack of high-quality, human-annotated benchmarks and reliable, interpretable reward signals. In this work, we introduce AdvancedIF (we will release this benchmark soon), a comprehensive benchmark featuring over 1,600 prompts and expert-curated rubrics that assess LLMs ability to follow complex, multi-turn, and system-level instructions. We further propose RIFL (Rubric-based Instruction-Following Learning), a novel post-training pipeline that leverages rubric generation, a finetuned rubric verifier, and reward shaping to enable effective reinforcement learning for instruction following. Extensive experiments demonstrate that RIFL substantially improves the instruction-following abilities of LLMs, achieving a 6.7% absolute gain on AdvancedIF and strong results on public benchmarks. Our ablation studies confirm the effectiveness of each component in RIFL. This work establishes rubrics as a powerful tool for both training and evaluating advanced IF in LLMs, paving the way for more capable and reliable AI systems.
Problem

Research questions and friction points this paper is trying to address.

Developing rigorous evaluation benchmarks for complex LLM instruction following
Creating reliable reward signals for training instruction-following capabilities
Improving LLMs' ability to handle multi-turn and system-level instructions
Innovation

Methods, ideas, or system contributions that make the work stand out.

Rubric-based benchmarking for complex instruction evaluation
Reinforcement learning pipeline with rubric-guided reward shaping
Finetuned verifier model for automated rubric assessment
🔎 Similar Papers
No similar papers found.
Yun He
Yun He
Research Scientist at Meta
data mininginformation retrieval and recommender systems
Wenzhe Li
Wenzhe Li
Princeton University
Hejia Zhang
Hejia Zhang
Skild AI
AIMachine LearningReinforcement LearningRoboticsAI/Robot Hardware Design
S
Songlin Li
Meta Superintelligence Labs
K
Karishma Mandyam
Meta Superintelligence Labs
Sopan Khosla
Sopan Khosla
Meta AI
Machine LearningNatural Language ProcessingArtificial Intelligence
Yuanhao Xiong
Yuanhao Xiong
Meta Superintelligence Labs
Nanshu Wang
Nanshu Wang
Meta
AI AssistantsRAGAI-powered searchLLM post-training
S
Selina Peng
Meta Superintelligence Labs
Beibin Li
Beibin Li
Meta
Language modelmachine learningcomputer vision
S
Shengjie Bi
Meta Superintelligence Labs
Shishir G. Patil
Shishir G. Patil
UC Berkeley
LLMsMachine LearningSystems
Q
Qi Qi
Meta Superintelligence Labs
Shengyu Feng
Shengyu Feng
Carnegie Mellon University
Combinatorial OptimizationLanguage Models
Julian Katz-Samuels
Julian Katz-Samuels
University of Wisconsin
Machine Learning
Richard Yuanzhe Pang
Richard Yuanzhe Pang
Meta, New York University
Natural Language ProcessingMachine Learning
Sujan Kumar Gonugondla
Sujan Kumar Gonugondla
Meta Superintelligence Labs
Hunter Lang
Hunter Lang
Meta GenAI, MIT
machine learningsemi-supervised learningweak supervisionnlp
Y
Yue Yu
Meta Superintelligence Labs
Y
Yundi Qian
Meta Superintelligence Labs
Maryam Fazel-Zarandi
Maryam Fazel-Zarandi
Meta Superintelligence Labs
L
Licheng Yu
Meta Superintelligence Labs
A
Amine Benhalloum
Meta Superintelligence Labs
H
H. Awadalla
Meta Superintelligence Labs
Manaal Faruqui
Manaal Faruqui
Research Scientist, Meta
natural language processing