Toward a Physics of Deep Learning and Brains

📅 2025-09-26
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Can a unified theoretical framework explain learning mechanisms in both deep neural networks and biological brains? Method: Leveraging nonequilibrium statistical physics, we investigate the shared dynamical regime—quasi-criticality—situated between absorbing and active phases. We employ neural avalanche analysis, finite-size scaling, and Barkhausen noise modeling to characterize critical dynamics across artificial and biological neural systems. Contribution/Results: We demonstrate that crackling noise scaling laws approximately hold in the quasi-critical regime; maximum susceptibility—not the critical point itself—serves as a superior predictor of generalization performance and distinguishes distinct universality classes (e.g., directed percolation). Experiments span diverse weight initialization schemes in deep network training. This work establishes the first physically grounded, unified framework bridging artificial and biological neural networks, offering a novel paradigm for designing high-performance learning systems based on fundamental physical principles.

Technology Category

Application Category

📝 Abstract
Deep neural networks and brains both learn and share superficial similarities: processing nodes are likened to neurons and adjustable weights are likened to modifiable synapses. But can a unified theoretical framework be found to underlie them both? Here we show that the equations used to describe neuronal avalanches in living brains can also be applied to cascades of activity in deep neural networks. These equations are derived from non-equilibrium statistical physics and show that deep neural networks learn best when poised between absorbing and active phases. Because these networks are strongly driven by inputs, however, they do not operate at a true critical point but within a quasi-critical regime -- one that still approximately satisfies crackling noise scaling relations. By training networks with different initializations, we show that maximal susceptibility is a more reliable predictor of learning than proximity to the critical point itself. This provides a blueprint for engineering improved network performance. Finally, using finite-size scaling we identify distinct universality classes, including Barkhausen noise and directed percolation. This theoretical framework demonstrates that universal features are shared by both biological and artificial neural networks.
Problem

Research questions and friction points this paper is trying to address.

Developing unified physics theory for deep learning and brain function
Identifying quasi-critical regimes for optimal neural network learning
Establishing universal scaling relations across biological and artificial networks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Applied neuronal avalanche equations to neural networks
Identified quasi-critical regime for optimal learning
Used finite-size scaling to classify universality classes
🔎 Similar Papers
No similar papers found.
Arsham Ghavasieh
Arsham Ghavasieh
Posdoctoral Fellow
Statistical PhysicsArtificial IntelligenceTheoretical NeuroscienceNetwork Science
M
Meritxell Vila-Minana
Center for Complex Networks and Systems Research, Luddy School of Informatics, Computing, and Engineering, Indiana University, Bloomington, Indiana 47408, USA
A
Akanksha Khurd
Center for Complex Networks and Systems Research, Luddy School of Informatics, Computing, and Engineering, Indiana University, Bloomington, Indiana 47408, USA
J
John Beggs
Department of Physics, Indiana University, Bloomington, Indiana 47405, USA
Gerardo Ortiz
Gerardo Ortiz
Indiana University
Physics
Santo Fortunato
Santo Fortunato
James H. Rudy Professor of Informatics and Computing, Indiana University Bloomington, USA
complex systemsnetwork sciencecomputational social sciencescience of scienceclimate change