🤖 AI Summary
This work addresses the long-standing open problem of tightness of approximation guarantees for higher-order singular value decomposition (HOSVD) and its variants—sequential truncated HOSVD (ST-HOSVD) and the higher-order orthogonal iteration (HOOI). We construct, for the first time, a family of pathological tensors that rigorously demonstrates that, for any ε > 0, all three algorithms achieve an approximation ratio of *N* / (1 + ε) in the worst case, matching the previously known upper bounds. This establishes the exact tightness of their approximation ratios, confirming that classical worst-case analyses are optimal and cannot be improved. Methodologically, our approach integrates tensor-based counterexample construction, multilinear algebraic analysis, and low-rank approximation theory. The result provides the first tight characterization of the theoretical limits of these widely used high-dimensional data dimensionality reduction algorithms.
📝 Abstract
We prove that the classic approximation guarantee for the higher-order singular value decomposition (HOSVD) is tight by constructing a tensor for which HOSVD achieves an approximation ratio of $N/(1+varepsilon)$, for any $varepsilon > 0$. This matches the upper bound of De Lathauwer et al. (2000a) and shows that the approximation ratio of HOSVD cannot be improved. Using a more advanced construction, we also prove that the approximation guarantees for the ST-HOSVD algorithm of Vannieuwenhoven et al. (2012) and higher-order orthogonal iteration (HOOI) of De Lathauwer et al. (2000b) are tight by showing that they can achieve their worst-case approximation ratio of $N / (1 + varepsilon)$, for any $varepsilon > 0$.