Improving Instruction Following in Language Models through Proxy-Based Uncertainty Estimation

📅 2024-05-10
🏛️ International Conference on Machine Learning
📈 Citations: 1
Influential: 0
📄 PDF
🤖 AI Summary
To address ambiguous response quality assessment and inconsistent scoring in instruction-following evaluation, this paper proposes the Uncertainty-Aware Reward Model (URM), the first approach to incorporate Bayesian approximate uncertainty modeling into instruction-following reward learning. URM jointly optimizes response scoring and uncertainty quantification on preference data. It enables dual-dimensional evaluation—assessing both response quality and confidence—thereby providing a unified framework for high-quality data filtering and robust policy optimization. Empirical results on benchmarks including Vicuna and MT-Bench demonstrate that reinforcement learning guided by URM significantly improves instruction-following capability over state-of-the-art methods. Moreover, URM enhances the model’s ability to discriminate training data quality and improves policy convergence stability.

Technology Category

Application Category

📝 Abstract
Assessing response quality to instructions in language models is vital but challenging due to the complexity of human language across different contexts. This complexity often results in ambiguous or inconsistent interpretations, making accurate assessment difficult. To address this issue, we propose a novel Uncertainty-aware Reward Model (URM) that introduces a robust uncertainty estimation for the quality of paired responses based on Bayesian approximation. Trained with preference datasets, our uncertainty-enabled proxy not only scores rewards for responses but also evaluates their inherent uncertainty. Empirical results demonstrate significant benefits of incorporating the proposed proxy into language model training. Our method boosts the instruction following capability of language models by refining data curation for training and improving policy optimization objectives, thereby surpassing existing methods by a large margin on benchmarks such as Vicuna and MT-bench. These findings highlight that our proposed approach substantially advances language model training and paves a new way of harnessing uncertainty within language models.
Problem

Research questions and friction points this paper is trying to address.

Language Model Accuracy
Complexity and Ambiguity in Language
Response Quality Evaluation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Agent-based Uncertainty Estimation
Uncertainty-aware Reward Model (URM)
Enhanced Instruction Understanding and Execution
🔎 Similar Papers
No similar papers found.
J
Joonho Lee
Samsung SDS Technology Research, Seoul, Korea
Jae Oh Woo
Jae Oh Woo
GenAI Innovation Center - Amazon Web Services (AWS)
Data ScienceMathematical Foundations of Deep LearningInformation TheoryStochastic Geometry
J
Juree Seok
Samsung SDS Technology Research, Seoul, Korea
Parisa Hassanzadeh
Parisa Hassanzadeh
Samsung SDS America, San Jose, California, USA
Wooseok Jang
Wooseok Jang
KAIST
J
JuYoun Son
Samsung SDS Technology Research, Seoul, Korea
S
Sima Didari
Samsung SDS America, San Jose, California, USA
B
Baruch Gutow
Samsung SDS America, San Jose, California, USA
H
Heng Hao
Samsung SDS America, San Jose, California, USA
H
Hankyu Moon
Samsung SDS America, San Jose, California, USA
Wenjun Hu
Wenjun Hu
Assistant Professor of Electrical Engineering and Computer Science, Yale University
WirelessMobileNetworkingSystems
Y
Yeong-Dae Kwon
Samsung SDS Technology Research, Seoul, Korea
Taehee Lee
Taehee Lee
Seoul National University Hospital and Georgia Institute of Technology
Thoracic RadiologyLung CancerData ScienceMachine Learning
S
Seungjai Min
Samsung SDS America, San Jose, California, USA