Spectral Architecture Search for Neural Networks

📅 2025-04-01
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the discrete nature and high computational cost of neural architecture search (NAS), this paper proposes SPARCS—the first NAS method incorporating spectral analysis. SPARCS models the eigen-spectrum of inter-layer propagation matrices and parameterizes the architecture space on a continuous, differentiable manifold, enabling gradient-based end-to-end optimization. By imposing constraints on the spectral distribution, it jointly optimizes architectural expressivity and parameter efficiency, thereby inducing task-adapted, compact architectures automatically. On standard benchmarks, architectures discovered by SPARCS achieve significantly reduced parameter counts and improved inference latency, while matching or surpassing the accuracy of state-of-the-art differentiable NAS methods. This demonstrates the effectiveness and novelty of adopting a spectral-geometric perspective for encoding architectural priors in NAS.

Technology Category

Application Category

📝 Abstract
Architecture design and optimization are challenging problems in the field of artificial neural networks. Working in this context, we here present SPARCS (SPectral ARchiteCture Search), a novel architecture search protocol which exploits the spectral attributes of the inter-layer transfer matrices. SPARCS allows one to explore the space of possible architectures by spanning continuous and differentiable manifolds, thus enabling for gradient-based optimization algorithms to be eventually employed. With reference to simple benchmark models, we show that the newly proposed method yields a self-emerging architecture with a minimal degree of expressivity to handle the task under investigation and with a reduced parameter count as compared to other viable alternatives.
Problem

Research questions and friction points this paper is trying to address.

Optimizing neural network architecture design challenges
Exploring architectures via spectral layer transfer matrices
Reducing parameter count while maintaining task expressivity
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses spectral attributes of transfer matrices
Enables gradient-based optimization on manifolds
Yields self-emerging architectures with fewer parameters
🔎 Similar Papers
No similar papers found.
G
Gianluca Peri
Department of Physics and Astronomy, University of Florence, Sesto Fiorentino, Italy
L
Lorenzo Giambagli
Department of Physics, Freie Universität Berlin, Arnimallee 12, 14195, Berlin, Germany
L
Lorenzo Chicchi
Department of Physics and Astronomy, University of Florence, Sesto Fiorentino, Italy
Duccio Fanelli
Duccio Fanelli
University of Florence
Physics