Integral Signatures of Activation Functions: A 9-Dimensional Taxonomy and Stability Theory for Deep Learning

📅 2025-10-09
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Activation function selection has long relied on heuristic experience without rigorous theoretical foundations, particularly regarding stability and expressivity trade-offs. Method: We propose the first nine-dimensional integral feature system unifying stability and kernel properties to jointly characterize expressivity and dynamical stability. Our framework integrates Gaussian propagation statistics, Lyapunov dynamical analysis, dimension-free Hessian bounds, and total variation-based smoothness measures, yielding an affine reparameterization-invariant classification scheme with provable dynamic stability guarantees. Results: The theory identifies a precise variance-stabilizing region, enabling sharp categorization of saturation-type, linearly growing, and smooth activation functions. Classification of eight mainstream activations—including ReLU, Swish, and GELU—aligns closely with Gauss–Hermite quadrature and Monte Carlo numerical validation. This work establishes the first theoretically grounded, provably stable criterion for activation function selection.

Technology Category

Application Category

📝 Abstract
Activation functions govern the expressivity and stability of neural networks, yet existing comparisons remain largely heuristic. We propose a rigorous framework for their classification via a nine-dimensional integral signature S_sigma(phi), combining Gaussian propagation statistics (m1, g1, g2, m2, eta), asymptotic slopes (alpha_plus, alpha_minus), and regularity measures (TV(phi'), C(phi)). This taxonomy establishes well-posedness, affine reparameterization laws with bias, and closure under bounded slope variation. Dynamical analysis yields Lyapunov theorems with explicit descent constants and identifies variance stability regions through (m2', g2). From a kernel perspective, we derive dimension-free Hessian bounds and connect smoothness to bounded variation of phi'. Applying the framework, we classify eight standard activations (ReLU, leaky-ReLU, tanh, sigmoid, Swish, GELU, Mish, TeLU), proving sharp distinctions between saturating, linear-growth, and smooth families. Numerical Gauss-Hermite and Monte Carlo validation confirms theoretical predictions. Our framework provides principled design guidance, moving activation choice from trial-and-error to provable stability and kernel conditioning.
Problem

Research questions and friction points this paper is trying to address.

Classifying activation functions via nine-dimensional integral signatures
Establishing stability theorems and explicit descent constants
Providing principled design guidance for activation selection
Innovation

Methods, ideas, or system contributions that make the work stand out.

Integral signature framework classifies activation functions
Dynamical analysis yields Lyapunov stability theorems
Dimension-free Hessian bounds derived from kernel perspective
🔎 Similar Papers
No similar papers found.
Ankur Mali
Ankur Mali
Assistant Professor, University of South Florida
Formal languageMemory NetworksPredictive CodingNatural Language Processinglifelong machine
L
Lawrence Hall
Bellini College of AI, Cybersecurity and Computing, University of South Florida, Tampa, FL 33620
J
Jake Williams
College of Computing and Informatics, Drexel University, Philadelphia, PA 19104
Gordon Richards
Gordon Richards
Departments of Physics, Drexel University, Philadelphia, PA 19104