Revisit CP Tensor Decomposition: Statistical Optimality and Fast Convergence

πŸ“… 2025-05-29
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This paper addresses two fundamental limitations in CANDECOMP/PARAFAC (CP) tensor decomposition: the lack of statistical optimality guarantees under noise, non-orthogonality, and high rank, and the weak convergence theory for the alternating least squares (ALS) algorithm. To resolve these, we propose a joint framework integrating the Tensor Alternating Spectral Decomposition (TASD) initialization with ALS. We establish, for the first time, non-asymptotic minimax-optimal error bounds for CP decomposition under general order, dimension, and rank. We further characterize ALS’s two-stage convergence dynamics: quadratic convergence initially, followed by linear refinement; in the rank-one case, only 1–2 iterations suffice to achieve statistical optimality. Experiments demonstrate that TASD+ALS significantly improves estimation stability and accuracy in noisy settings. Our core contribution is a unified theoretical characterization of both the statistical limits and algorithmic convergence rates of CP decomposition, thereby filling a critical gap in high-dimensional tensor decomposition theory.

Technology Category

Application Category

πŸ“ Abstract
Canonical Polyadic (CP) tensor decomposition is a fundamental technique for analyzing high-dimensional tensor data. While the Alternating Least Squares (ALS) algorithm is widely used for computing CP decomposition due to its simplicity and empirical success, its theoretical foundation, particularly regarding statistical optimality and convergence behavior, remain underdeveloped, especially in noisy, non-orthogonal, and higher-rank settings. In this work, we revisit CP tensor decomposition from a statistical perspective and provide a comprehensive theoretical analysis of ALS under a signal-plus-noise model. We establish non-asymptotic, minimax-optimal error bounds for tensors of general order, dimensions, and rank, assuming suitable initialization. To enable such initialization, we propose Tucker-based Approximation with Simultaneous Diagonalization (TASD), a robust method that improves stability and accuracy in noisy regimes. Combined with ALS, TASD yields a statistically consistent estimator. We further analyze the convergence dynamics of ALS, identifying a two-phase pattern-initial quadratic convergence followed by linear refinement. We further show that in the rank-one setting, ALS with an appropriately chosen initialization attains optimal error within just one or two iterations.
Problem

Research questions and friction points this paper is trying to address.

Analyzing statistical optimality of CP tensor decomposition
Improving ALS convergence in noisy, non-orthogonal settings
Developing robust initialization for accurate tensor decomposition
Innovation

Methods, ideas, or system contributions that make the work stand out.

Tucker-based Approximation with Simultaneous Diagonalization (TASD)
Non-asymptotic minimax-optimal error bounds
Two-phase convergence dynamics analysis
πŸ”Ž Similar Papers
No similar papers found.
R
Runshi Tang
Department of Statistics, University of Wisconsin-Madison
Julien Chhor
Julien Chhor
Postdoc - Harvard University
High-dimensional and non-parametric statistics
O
Olga Klopp
ESSEC Business School and Statistics Department, CREST-ENSAE Paris
A
Anru R. Zhang
Department of Biostatistics & Bioinformatics and Department of Computer Science, Duke University