🤖 AI Summary
Current large language models (LLMs) lack genuine causal reasoning capabilities, relying instead on pattern matching and memorization; they struggle to reliably evaluate counterfactuals and are constrained by the unavailability of observational data, often resorting to synthetic benchmarks for evaluation.
Method: We propose a graph-augmented causal knowledge extraction framework and a causally constrained counterfactual reasoning agent. Our approach integrates causal graph modeling, constraint-based logical reasoning, and retrieval-augmented generation (RAG) to automatically construct an interpretable, structured causal knowledge base directly from real-world data, enabling out-of-distribution counterfactual generation.
Contribution/Results: The method significantly reduces hallucination and spurious correlations. Experiments demonstrate strong generalization in real-world counterfactual assessment, with simultaneous improvements in reasoning robustness and efficiency. It endows LLMs with verifiable, low-cost, and interpretable causal reasoning—marking the first such capability grounded in empirical data.
📝 Abstract
Causal world models are systems that can answer counterfactual questions about an environment of interest, i.e. predict how it would have evolved if an arbitrary subset of events had been realized differently. It requires understanding the underlying causes behind chains of events and conducting causal inference for arbitrary unseen distributions. So far, this task eludes foundation models, notably large language models (LLMs), which do not have demonstrated causal reasoning capabilities beyond the memorization of existing causal relationships. Furthermore, evaluating counterfactuals in real-world applications is challenging since only the factual world is observed, limiting evaluation to synthetic datasets. We address these problems by explicitly extracting and modeling causal relationships and propose the Causal Cartographer framework. First, we introduce a graph retrieval-augmented generation agent tasked to retrieve causal relationships from data. This approach allows us to construct a large network of real-world causal relationships that can serve as a repository of causal knowledge and build real-world counterfactuals. In addition, we create a counterfactual reasoning agent constrained by causal relationships to perform reliable step-by-step causal inference. We show that our approach can extract causal knowledge and improve the robustness of LLMs for causal reasoning tasks while reducing inference costs and spurious correlations.