🤖 AI Summary
This work investigates the intrinsic relationship between deep model merging and neural network interpretability, addressing the lack of a unified geometric foundation for understanding their interplay.
Method: We systematically analyze empirical phenomena in model merging and identify four geometric properties of the loss landscape—mode convexity, determinacy, directionality, and connectivity—as shared structural determinants of merging efficacy and generalization robustness. We develop a unified framework integrating loss landscape analysis, neural representation modeling, and interpretability evaluation.
Contribution/Results: Theoretically, we establish how landscape geometry governs the interpretability of internal representations and their stability across tasks. Our framework provides the first geometric interpretation paradigm for model merging and pioneers a novel interdisciplinary research direction at the intersection of “landscape geometry–interpretability–robustness,” offering principled insights into both model composition and trustworthy AI.
📝 Abstract
We survey the model merging literature through the lens of loss landscape geometry to connect observations from empirical studies on model merging and loss landscape analysis to phenomena that govern neural network training and the emergence of their inner representations. We distill repeated empirical observations from the literature in these fields into descriptions of four major characteristics of loss landscape geometry: mode convexity, determinism, directedness, and connectivity. We argue that insights into the structure of learned representations from model merging have applications to model interpretability and robustness, subsequently we propose promising new research directions at the intersection of these fields.