🤖 AI Summary
This study investigates the cognitive mechanisms underlying human deviations from information-theoretically optimal guessing strategies in Wordle. Methodologically, it integrates semantic (word embedding), phonological (phoneme-based encoding), and orthographic (edit distance) representations to computationally disentangle and quantify the distinct influences of these three linguistic biases on real-time lexical decision-making. Leveraging large-scale behavioral data from actual players and a near-optimal baseline strategy derived from entropy maximization, results show that semantic and phonological similarity significantly increase repetition probability by 37% and 29%, respectively, whereas orthographic similarity exerts a comparatively weaker effect. This work constitutes the first computational decomposition of linguistic cognitive dimensions within a gamified constrained inference task, thereby bridging NLP modeling with cognitive linguistic theory and revealing how deep linguistic representations systematically shape everyday inductive reasoning.
📝 Abstract
We show that human players'gameplay in the game of Wordle is influenced by the semantics, orthography, and phonology of the player's previous guesses. We compare actual human players'guesses with near-optimal guesses using NLP techniques. We study human language use in the constrained environment of Wordle, which is situated between natural language use and the artificial word association task