đ¤ AI Summary
FrankâWolfe (FW) algorithms suffer from conservative step sizes and lack rigorous primalâdual coordination guarantees.
Method: We propose the first optimistic FW algorithm with a provably convergent primalâdual framework. It introduces an optimistic predictionâcorrection mechanism and a generalized short-step strategy, extending gap-driven adaptive updatesâpreviously limited to primal-only methodsâto gradient-based algorithms. By integrating smoothness exploitation and convergence-rate tightening techniques, our method ensures computable primalâdual gaps and verifiable step sizes.
Contribution/Results: We establish the tightest known $O(1/k)$ primalâdual convergence rate for FW-type methods. Experiments on structured convex optimization tasksâincluding Lasso, matrix completion, and Wasserstein distance computationâdemonstrate that our algorithm significantly outperforms classical FW and existing variants in accuracy, speed, and practicality.
đ Abstract
We introduce novel techniques to enhance Frank-Wolfe algorithms by leveraging function smoothness beyond traditional short steps. Our study focuses on Frank-Wolfe algorithms with step sizes that incorporate primal-dual guarantees, offering practical stopping criteria. We present a new Frank-Wolfe algorithm utilizing an optimistic framework and provide a primal-dual convergence proof. Additionally, we propose a generalized short-step strategy aimed at optimizing a computable primal-dual gap. Interestingly, this new generalized short-step strategy is also applicable to gradient descent algorithms beyond Frank-Wolfe methods. As a byproduct, our work revisits and refines primal-dual techniques for analyzing Frank-Wolfe algorithms, achieving tighter primal-dual convergence rates. Empirical results demonstrate that our optimistic algorithm outperforms existing methods, highlighting its practical advantages.