🤖 AI Summary
This work addresses the problem of deriving expectation upper bounds for the ℓₚ-injective norm of sums of sub-Gaussian random tensors—a fundamental estimation challenge at the intersection of high-dimensional probability and tensor analysis. Methodologically, we introduce the PAC-Bayesian lemma to control random tensor norms, marking the first application of this technique in sub-Gaussian tensor analysis and overcoming key limitations of classical moment-based approaches. Our main contribution is a universal expectation upper bound that strictly improves upon the seminal results of Bandeira et al. and Latała. In particular, when *p* = 2, our bound sharpens Latała’s theorem and yields a concise, elementary derivation of all moments of Gaussian chaos. Moreover, the result unifies treatment across all *p* ≥ 1, substantially broadening the scope and applicability of sub-Gaussian tensor analysis.
📝 Abstract
We prove an upper bound on the expected $ell_p$ injective norm of sums of subgaussian random tensors. Our proof is simple and does not rely on any explicit geometric or chaining arguments. Instead, it follows from a simple application of the PAC-Bayesian lemma, a tool that has proven effective at controlling the suprema of certain ``smooth'' empirical processes in recent years. Our bound strictly improves a very recent result of Bandeira, Gopi, Jiang, Lucca, and Rothvoss. In the Euclidean case ($p=2$), our bound sharpens a result of Lata{l}a that was central to proving his estimates on the moments of Gaussian chaoses. As a consequence, we obtain an elementary proof of this fundamental result.