🤖 AI Summary
This study addresses a critical limitation in existing neuro-symbolic fact-checking systems, which rely solely on formal logic and consequently fail to detect conclusions that, while logically valid, are cognitively misleading to humans. To bridge this gap, the authors integrate perspectives from cognitive science and pragmatics to develop a taxonomy of such “logically sound yet cognitively deceptive” cases. They further propose leveraging large language models to simulate human-like reasoning tendencies, thereby evaluating and refining the outputs of neuro-symbolic systems. This approach transcends the constraints of purely formal logic by incorporating human cognitive consistency as a validation criterion, offering a novel paradigm that harmonizes logical rigor with cognitive plausibility in automated fact-checking.
📝 Abstract
As large language models (LLMs) are increasing integrated into fact-checking pipelines, formal logic is often proposed as a rigorous means by which to mitigate bias, errors and hallucinations in these models' outputs. For example, some neurosymbolic systems verify claims by using LLMs to translate natural language into logical formulae and then checking whether the proposed claims are logically sound, i.e. whether they can be validly derived from premises that are verified to be true. We argue that such approaches structurally fail to detect misleading claims due to systematic divergences between conclusions that are logically sound and inferences that humans typically make and accept. Drawing on studies in cognitive science and pragmatics, we present a typology of cases in which logically sound conclusions systematically elicit human inferences that are unsupported by the underlying premises. Consequently, we advocate for a complementary approach: leveraging the human-like reasoning tendencies of LLMs as a feature rather than a bug, and using these models to validate the outputs of formal components in neurosymbolic systems against potentially misleading conclusions.