🤖 AI Summary
Social science causal modeling faces challenges in intervening on macro-level variables (e.g., age, gender, race) and satisfying modularity assumptions. This paper proposes an extended Causal Feature Learning (CFL) framework that, for the first time, systematically establishes causal mappings between non-intervenable macro-level abstractions and intervenable micro-level features, thereby reconstructing social constructs as SCM-compatible, low-level operational representations. CFL integrates structural causal models with deep representation learning to jointly discover macro-states and identify causal structures from multi-source observational data. Experiments demonstrate that CFL-derived macro-states significantly improve downstream causal estimation: average treatment effect (ATE) accuracy increases by 12.7% over baselines. Moreover, the learned representations enhance model interpretability and cross-dataset generalizability, offering a principled bridge between high-level social theory and low-level causal inference.
📝 Abstract
Variable selection poses a significant challenge in causal modeling, particularly within the social sciences, where constructs often rely on inter-related factors such as age, socioeconomic status, gender, and race. Indeed, it has been argued that such attributes must be modeled as macro-level abstractions of lower-level manipulable features, in order to preserve the modularity assumption essential to causal inference. This paper accordingly extends the theoretical framework of Causal Feature Learning (CFL). Empirically, we apply the CFL algorithm to diverse social science datasets, evaluating how CFL-derived macrostates compare with traditional microstates in downstream modeling tasks.