Shannon invariants: A scalable approach to information decomposition

📅 2025-04-22
📈 Citations: 0
✹ Influential: 0
📄 PDF
đŸ€– AI Summary
Quantifying higher-order information interactions in distributed systems—such as biological and artificial neural networks—remains challenging due to semantic ambiguities in existing frameworks like Partial Information Decomposition (PID). Method: This paper introduces the “Shannon Invariant” framework, grounded exclusively in Shannon’s entropy axioms, to establish the first axiomatically rigorous multivariate information decomposition. It eliminates long-standing semantic inconsistencies by deriving decomposition solely from fundamental entropy principles. The framework is inherently scalable and interpretable, enabling cross-scale and cross-architecture analysis of information processing. Coupled with higher-order dependency modeling and optimized algorithms, it supports efficient computation. Results: We successfully identify layer-specific information processing signatures across diverse deep neural networks, characterize the dynamic evolution of information flow during training, and achieve substantial improvements in both decomposition efficiency and theoretical consistency on systems with up to one thousand nodes.

Technology Category

Application Category

📝 Abstract
Distributed systems, such as biological and artificial neural networks, process information via complex interactions engaging multiple subsystems, resulting in high-order patterns with distinct properties across scales. Investigating how these systems process information remains challenging due to difficulties in defining appropriate multivariate metrics and ensuring their scalability to large systems. To address these challenges, we introduce a novel framework based on what we call"Shannon invariants"-- quantities that capture essential properties of high-order information processing in a way that depends only on the definition of entropy and can be efficiently calculated for large systems. Our theoretical results demonstrate how Shannon invariants can be used to resolve long-standing ambiguities regarding the interpretation of widely used multivariate information-theoretic measures. Moreover, our practical results reveal distinctive information-processing signatures of various deep learning architectures across layers, which lead to new insights into how these systems process information and how this evolves during training. Overall, our framework resolves fundamental limitations in analyzing high-order phenomena and offers broad opportunities for theoretical developments and empirical analyses.
Problem

Research questions and friction points this paper is trying to address.

Defining scalable metrics for high-order information processing
Resolving ambiguities in multivariate information-theoretic measures
Analyzing information-processing signatures in deep learning architectures
Innovation

Methods, ideas, or system contributions that make the work stand out.

Introduces Shannon invariants for scalable analysis
Resolves ambiguities in multivariate information measures
Reveals information-processing signatures in deep learning
🔎 Similar Papers
No similar papers found.
A
A. Gutknecht
Göttingen Campus Institute for Dynamics of Biological Networks, Georg-August University Göttingen
F
Fernando Rosas
Sussex Centre for Consciousness Science and Sussex AI, Department of Informatics, University of Sussex; Center for Psychedelic Research and Centre for Complexity Science, Department of Brain Science, Imperial College London; Center for Eudaimonia and Human Flourishing, University of Oxford; Principles of Intelligent Behaviour in Biological and Social Systems (PIBBSS)
D
David A. Ehrlich
Göttingen Campus Institute for Dynamics of Biological Networks, Georg-August University Göttingen; Max Planck Institute for Dynamics and Self-Organization, Göttingen
Abdullah Makkeh
Abdullah Makkeh
Postdoc, University of Göttingen
Information TheoryNeuroscienceOptimization#unitartucs#unigoe
P
Pedro Mediano
Department of Computing, Imperial College London; Division of Psychology and Language Sciences, University College London
Michael Wibral
Michael Wibral
Professor for Data Driven Analysis of Biological Networks, Georg-August University, Göttingen
Information theoryInformation dynamicsPartial Information DecompositionPredictive CodingNeuroscience