🤖 AI Summary
Conventional inverse problem algorithms in signal processing rely heavily on hand-crafted designs, suffering from poor generalizability and limited adaptability across diverse data distributions and algorithmic families.
Method: This paper introduces neural architecture search (NAS) to automated sparse recovery algorithm discovery—the first such application. We propose a differentiable meta-learning framework that jointly optimizes the iterative update structure, soft-thresholding operator, and acceleration mechanism over a search space of >50,000 candidate operations, effectively reconstructing the core update logic of ISTA/FISTA.
Contribution/Results: Our method precisely reproduces the convergence behavior and acceleration properties of classical algorithms while demonstrating strong cross-distribution and cross-algorithm-family generalization. It breaks away from heuristic design paradigms, establishing the first scalable and interpretable automated framework for inverse problem algorithm discovery—enabling principled, data-aware, and architecture-agnostic solver design.
📝 Abstract
The design of novel algorithms for solving inverse problems in signal processing is an incredibly difficult, heuristic-driven, and time-consuming task. In this short paper, we the idea of automated algorithm discovery in the signal processing context through meta-learning tools such as Neural Architecture Search (NAS). Specifically, we examine the Iterative Shrinkage Thresholding Algorithm (ISTA) and its accelerated Fast ISTA (FISTA) variant as candidates for algorithm rediscovery. We develop a meta-learning framework which is capable of rediscovering (several key elements of) the two aforementioned algorithms when given a search space of over 50,000 variables. We then show how our framework can apply to various data distributions and algorithms besides ISTA/FISTA.