Auxiliary-Hyperparameter-Free Sampling: Entropy Equilibrium for Text Generation

📅 2025-11-30
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing large language models (LLMs) commonly rely on manually tuned hyperparameters for token sampling, resulting in complex deployment and poor generalization. To address this, we propose Entropy-Equilibrium Sampling (EES), the first dynamic sampling mechanism that operates **without auxiliary hyperparameters**. Grounded in information theory, EES jointly models normalized entropy and probability mass to adaptively construct candidate token sets, thereby simultaneously improving accuracy, coherence, and diversity across varying temperature settings. EES is architecture-agnostic and integrates seamlessly with mainstream LLMs. Empirical evaluation across diverse reasoning and text generation benchmarks demonstrates consistent and significant improvements over standard sampling methods—including Top-k and Nucleus sampling—across multiple models and temperature configurations. Crucially, EES maintains robust performance without task-specific tuning, substantially simplifying deployment and enhancing generation reliability.

Technology Category

Application Category

📝 Abstract
Token sampling strategies critically influence text generation quality in large language models (LLMs). However, existing methods introduce additional hyperparameters, requiring extensive tuning and complicating deployment. We present Entropy Equilibrium Sampling (EES), an auxiliary hyperparameter-free approach inspired by information theory that can dynamically adjust candidate sets by balancing normalized entropy with probability mass. We evaluate EES on both reasoning and generation tasks across a range of model architectures. Our results show that EES consistently performs well across temperature settings, delivering competitive accuracy and coherence while maintaining diversity. By eliminating the need for hyperparameter tuning, EES greatly simplifies deployment while improving performance. Code is available at https://github.com/shuanncai/EES
Problem

Research questions and friction points this paper is trying to address.

Eliminates hyperparameter tuning in text generation
Dynamically adjusts candidate sets using entropy balance
Improves performance across reasoning and generation tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

EES dynamically adjusts candidate sets using entropy equilibrium
EES eliminates hyperparameter tuning for simplified deployment
EES balances accuracy, coherence, and diversity across tasks
🔎 Similar Papers
No similar papers found.
X
Xiaodong Cai
Shenzhen International Graduate School, Tsinghua University, China
Hai Lin
Hai Lin
Electrical Engineering, University of Notre Dame
Cyber-Physical SystemsHybrid Dynamical SystemsDistributed Cooperative Systems
Shaoxiong Zhan
Shaoxiong Zhan
Tsinghua University
Natural Language ProcessingLarge Language Model
Weiqi Luo
Weiqi Luo
School of Computer, Sun Yat-Sen Univ. Guangzhou, P.R. China
Steganography and SteganalysisMultimedia ForensicsAI Security
H
Hong-Gee Kim
Seoul National University, South Korea
H
Hongyan Hao
Meituan, China
Y
Yu Yang
Meituan, China
H
Hai-Tao Zheng
Shenzhen International Graduate School, Tsinghua University, China