Accelerated First-Order Optimization under Nonlinear Constraints

📅 2023-02-01
🏛️ Mathematical programming
📈 Citations: 3
Influential: 0
📄 PDF
🤖 AI Summary
This paper addresses first-order optimization under nonlinear constraints—including nonconvex feasible sets—by proposing a novel accelerated algorithm grounded in nonsmooth dynamical systems. Methodologically, it models constraints in the **velocity space**, rather than the conventional position space, yielding sparse, local, and convex approximations of the feasible set and eliminating the need for expensive global projections at each iteration. Theoretically, the algorithm converges to stable points under nonconvex objectives and nonconvex constraints; under convexity, it achieves optimal acceleration rates in both continuous- and discrete-time settings. Its computational complexity scales nearly linearly with problem dimension and constraint count. Empirically, the method efficiently solves ℓ^p (p < 1) nonconvex regularized problems in compressed sensing and sparse regression: at p = 1, it matches state-of-the-art performance and substantially outperforms existing approaches.
📝 Abstract
We exploit analogies between first-order algorithms for constrained optimization and non-smooth dynamical systems to design a new class of accelerated first-order algorithms for constrained optimization. Unlike Frank–Wolfe or projected gradients, these algorithms avoid optimization over the entire feasible set at each iteration. We prove convergence to stationary points even in a nonconvex setting and we derive accelerated rates for the convex setting both in continuous time, as well as in discrete time. An important property of these algorithms is that constraints are expressed in terms of velocities instead of positions, which naturally leads to sparse, local and convex approximations of the feasible set (even if the feasible set is nonconvex). Thus, the complexity tends to grow mildly in the number of decision variables and in the number of constraints, which makes the algorithms suitable for machine learning applications. We apply our algorithms to a compressed sensing and a sparse regression problem, showing that we can treat nonconvex $$ell ^p$$ ℓ p constraints ( $$p<1$$ p < 1 ) efficiently, while recovering state-of-the-art performance for $$p=1$$ p = 1 .
Problem

Research questions and friction points this paper is trying to address.

Design accelerated first-order algorithms for constrained optimization
Avoid full feasible set optimization per iteration
Handle nonconvex constraints efficiently with sparse approximations
Innovation

Methods, ideas, or system contributions that make the work stand out.

Analogies between first-order algorithms and dynamical systems
Constraints expressed in terms of velocities
Sparse local convex approximations of feasible sets
🔎 Similar Papers
No similar papers found.