🤖 AI Summary
Language models frequently fail due to misinterpretation of the initial input—particularly idiomatic, metaphorical, or context-sensitive expressions—rather than errors in output generation. This paper proposes a lightweight, input-only method for pre-emptively detecting such failures. It leverages token-level likelihood features derived from surprisal and information density hypotheses, jointly modeling span-localized uncertainty and global statistical patterns—without requiring access to model internals or generated outputs. The approach is model-scale adaptive: larger models rely more on local features, while smaller models emphasize global patterns. Evaluated on five challenging language understanding benchmarks, it significantly outperforms strong baselines, demonstrating effectiveness, cross-model generalizability, and computational efficiency.
📝 Abstract
Language models often struggle with idiomatic, figurative, or context-sensitive inputs, not because they produce flawed outputs, but because they misinterpret the input from the outset. We propose an input-only method for anticipating such failures using token-level likelihood features inspired by surprisal and the Uniform Information Density hypothesis. These features capture localized uncertainty in input comprehension and outperform standard baselines across five linguistically challenging datasets. We show that span-localized features improve error detection for larger models, while smaller models benefit from global patterns. Our method requires no access to outputs or hidden activations, offering a lightweight and generalizable approach to pre-generation error prediction.