🤖 AI Summary
To address weak interpretability and poor clinical consistency of counterfactual explanations in medical imaging (e.g., chest X-rays), this paper proposes an unsupervised, label-agnostic, clinically concept-driven counterfactual generation framework. Methodologically, it introduces the first human-annotation-free Concept Activation Vector (CAV) construction for chest radiographs, aligning clinical concepts with latent features via a reconstruction-based autoencoder; counterfactual images are then generated by directional traversal along concept directions in the latent space. Key contributions include: (1) the first unsupervised CAV paradigm enabling stable cross-dataset transfer; and (2) an editable, interpretable clinical concept-guided generation mechanism. Experiments demonstrate clinically consistent counterfactuals for large pathologies (e.g., cardiomegaly), though small lesions remain limited by reconstruction fidelity—highlighting a need for further refinement.
📝 Abstract
An essential step in deploying medical imaging models is ensuring alignment with clinical knowledge and interpretability. We focus on mapping clinical concepts into the latent space of generative models to identify Concept Activation Vectors (CAVs). Using a simple reconstruction autoencoder, we link user-defined concepts to image-level features without explicit label training. The extracted concepts are stable across datasets, enabling visual explanations that highlight clinically relevant features. By traversing latent space along concept directions, we produce counterfactuals that exaggerate or reduce specific clinical features. Preliminary results on chest X-rays show promise for large pathologies like cardiomegaly, while smaller pathologies remain challenging due to reconstruction limits. Although not outperforming baselines, this approach offers a path toward interpretable, concept-based explanations aligned with clinical knowledge.