🤖 AI Summary
Existing causal relation extraction research largely overlooks “anti-causal claims”—statements explicitly denying causal relationships—resulting in datasets that only contain pro-causal claims and causing models to misclassify anti-causal statements as valid causal relations. Method: This work formally defines the anti-causal concept for the first time and constructs the first causal extraction dataset with explicit anti-causal annotations. Drawing on a systematic literature review, we establish rigorous annotation guidelines, extend the Causal News Corpus accordingly, and achieve high inter-annotator agreement (Cohen’s κ = 0.74). Contribution/Results: Experiments demonstrate that incorporating anti-causal samples substantially improves Transformer-based models’ ability to discriminate causal direction, significantly reducing misclassification rates. This addresses a critical gap in conventional causal inference frameworks, which systematically neglect refutative or negating statements, thereby enhancing robustness and fidelity in causal claim classification.
📝 Abstract
Research on causality extraction from text has so far almost entirely neglected counterclaims. Existing causality extraction datasets focus solely on "procausal" claims, i.e., statements that support a relationship. "Concausal" claims, i.e., statements that refute a relationship, are entirely ignored or even accidentally annotated as procausal. We address this shortcoming by developing a new dataset that integrates concausality. Based on an extensive literature review, we first show that concausality is an integral part of causal reasoning on incomplete knowledge. We operationalize this theory in the form of a rigorous guideline for annotation and then augment the Causal News Corpus with concausal statements, obtaining a substantial inter-annotator agreement of Cohen's $κ=0.74$. To demonstrate the importance of integrating concausal statements, we show that models trained without concausal relationships tend to misclassify these as procausal instead. Based on our new dataset, this mistake can be mitigated, enabling transformers to effectively distinguish pro- and concausality.