🤖 AI Summary
Current AI architectures—such as Transformers and diffusion models—lack a unified theoretical foundation linking them to classical associative memory models (e.g., Hopfield networks), hindering interpretability and principled design.
Method: We propose a Lagrangian dynamical framework that formalizes memory storage and retrieval as energy minimization, unifying associative computation across paradigms. Our approach implements differentiable recursive联想 dynamics via end-to-end optimization and integrates continuous-time dynamics with discrete neural computations.
Contribution/Results: We establish theoretical equivalence between attention mechanisms and associative retrieval through shared energy function structure, and reveal an intrinsic correspondence between diffusion processes and memory reactivation dynamics. The project delivers open-source pedagogical code, rigorous mathematical derivations, and interactive visualization tools, substantially enhancing the explanatory power and practical utility of associative memory theory in AI fundamentals and education.
📝 Abstract
Associative Memories like the famous Hopfield Networks are elegant models for describing fully recurrent neural networks whose fundamental job is to store and retrieve information. In the past few years they experienced a surge of interest due to novel theoretical results pertaining to their information storage capabilities, and their relationship with SOTA AI architectures, such as Transformers and Diffusion Models. These connections open up possibilities for interpreting the computation of traditional AI networks through the theoretical lens of Associative Memories. Additionally, novel Lagrangian formulations of these networks make it possible to design powerful distributed models that learn useful representations and inform the design of novel architectures. This tutorial provides an approachable introduction to Associative Memories, emphasizing the modern language and methods used in this area of research, with practical hands-on mathematical derivations and coding notebooks.