đ¤ AI Summary
This work identifies a âdehumanized reasoningâ problem in large language models (LLMs), wherein overreliance on formal logic leads to outputs violating human commonsense under logicâcommonsense conflicts. To address this, we introduce the first benchmark dataset targeting the ârulebreakerâ phenomenonâdefined here as a systematic cognitive biasâand propose a novel evaluation paradigm that jointly enforces logical rigor and human-aligned plausibility. Methodologically, we design adversarial logicâcommonsense conflict instances grounded in cognitive science principles, and perform multi-dimensional diagnostics via human annotation, attention distribution analysis, and world-knowledge utilization assessment. Experiments across seven state-of-the-art LLMsâincluding GPT-4o (achieving only moderate accuracy)âdemonstrate the ubiquity of this flaw; we pinpoint imbalanced attention allocation and insufficient world-knowledge integration as primary causal factors. This work establishes a new benchmark and cognitive calibration framework for trustworthy AI reasoning.
đ Abstract
Formal logic enables computers to reason in natural language by representing sentences in symbolic forms and applying rules to derive conclusions. However, in what our study characterizes as"rulebreaker"scenarios, this method can lead to conclusions that are typically not inferred or accepted by humans given their common sense and factual knowledge. Inspired by works in cognitive science, we create RULEBREAKERS, the first dataset for rigorously evaluating the ability of large language models (LLMs) to recognize and respond to rulebreakers (versus non-rulebreakers) in a human-like manner. Evaluating seven LLMs, we find that most models, including GPT-4o, achieve mediocre accuracy on RULEBREAKERS and exhibit some tendency to over-rigidly apply logical rules unlike what is expected from typical human reasoners. Further analysis suggests that this apparent failure is potentially associated with the models' poor utilization of their world knowledge and their attention distribution patterns. Whilst revealing a limitation of current LLMs, our study also provides a timely counterbalance to a growing body of recent works that propose methods relying on formal logic to improve LLMs' general reasoning capabilities, highlighting their risk of further increasing divergence between LLMs and human-like reasoning.