Minimal Sufficient Representations for Self-interpretable Deep Neural Networks

📅 2026-03-25
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work proposes DeepIn, a novel framework that addresses the poor interpretability of over-parameterized deep neural networks and the challenge of identifying minimal sufficient architectures. By integrating minimal sufficient representation learning with statistical inference, DeepIn simultaneously performs variable selection and network pruning through adaptive representation learning, non-asymptotic error analysis, and hypothesis testing. The resulting model retains the expressive power of standard DNNs while offering statistical rigor and self-explainability. Evaluated on biomedical and visual tasks, DeepIn reduces prediction error by up to 30% and automatically uncovers human-interpretable discriminative patterns, substantially enhancing both model interpretability and generalization performance.

Technology Category

Application Category

📝 Abstract
Deep neural networks (DNNs) achieve remarkable predictive performance but remain difficult to interpret, largely due to overparameterization that obscures the minimal structure required for interpretation. Here we introduce DeepIn, a self-interpretable neural network framework that adaptively identifies and learns the minimal representation necessary for preserving the full expressive capacity of standard DNNs. We show that DeepIn can correctly identify the minimal representation dimension, select relevant variables, and recover the minimal sufficient network architecture for prediction. The resulting estimator achieves optimal non-asymptotic error rates that adapt to the learned minimal dimension, demonstrating that recovering minimal sufficient structure fundamentally improves generalization error. Building on these guarantees, we further develop hypothesis testing procedures for both selected variables and learned representations, bridging deep representation learning with formal statistical inference. Across biomedical and vision benchmarks, DeepIn improves both predictive accuracy and interpretability, reducing error by up to 30% on real-world datasets while automatically uncovering human-interpretable discriminative patterns. Our results suggest that interpretability and statistical rigor can be embedded directly into deep architectures without sacrificing performance.
Problem

Research questions and friction points this paper is trying to address.

interpretability
minimal sufficient representation
overparameterization
deep neural networks
statistical inference
Innovation

Methods, ideas, or system contributions that make the work stand out.

minimal sufficient representation
self-interpretable neural networks
adaptive dimensionality reduction
statistical inference in deep learning
non-asymptotic error rates
🔎 Similar Papers
No similar papers found.
Z
Zhiyao Tan
Center of Statistical Research, School of Statistics, Southwestern University of Finance and Economics, Chengdu, China
Liu Li
Liu Li
Imperial College London
Medical image analysisComputer vision
Huazhen Lin
Huazhen Lin
Southwestern University of Finance and Economics
Nonparametric methodFunctional Data Analysis