🤖 AI Summary
This paper addresses the lack of consensus and semantic fragmentation surrounding counterfactual definitions across causal inference (CI) and eXplainable Artificial Intelligence (XAI). We propose the first unified formal framework for counterfactuals that spans both domains. By systematically comparing their definitional logic, generation mechanisms, evaluation criteria, and application paradigms, we identify fundamental distinctions—namely, divergent semantic objectives (causal effect identification vs. model behavior explanation), contrasting constraints (structural assumptions in CI vs. feasibility or minimal perturbation in XAI), and differing practical requirements—while also uncovering complementary synergies. Our analysis clarifies core conceptual alignments and mismatches, and establishes a theoretical bridge enabling cross-domain methodological transfer: for instance, integrating structural causal models and robustness constraints from CI into XAI counterfactual generation enhances both causal fidelity and generalizability of explanations.
📝 Abstract
Counterfactuals play a pivotal role in the two distinct data science fields of causal inference (CI) and explainable artificial intelligence (XAI). While the core idea behind counterfactuals remains the same in both fields--the examination of what would have happened under different circumstances--there are key differences in how they are used and interpreted. We introduce a formal definition that encompasses the multi-faceted concept of the counterfactual in CI and XAI. We then discuss how counterfactuals are used, evaluated, generated, and operationalized in CI vs. XAI, highlighting conceptual and practical differences. By comparing and contrasting the two, we hope to identify opportunities for cross-fertilization across CI and XAI.