Linear Discriminant Analysis with Gradient Optimization on Covariance Inverse

📅 2025-06-07
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In high-dimensional settings, classical Linear Discriminant Analysis (LDA) suffers severe performance degradation due to unstable sample covariance matrix estimation. To address this, we propose a novel paradigm that directly optimizes the inverse covariance matrix: modeling it as a learnable parameter constrained via Cholesky decomposition to ensure positive definiteness, augmented with a low-rank expansion to improve estimability, and initialized via a multi-start strategy—including both the identity matrix and a warm start from classical LDA—to enhance convergence robustness. Crucially, our approach bypasses explicit covariance estimation, achieving a favorable trade-off between statistical stability and gradient-based optimization feasibility. Extensive experiments on multivariate simulations and real-world high-dimensional datasets demonstrate that our method significantly outperforms classical LDA and leading regularized variants (e.g., rLDA, sLDA) under small-sample conditions, delivering consistent improvements in classification accuracy and generalization robustness.

Technology Category

Application Category

📝 Abstract
Linear discriminant analysis (LDA) is a fundamental method in statistical pattern recognition and classification, achieving Bayes optimality under Gaussian assumptions. However, it is well-known that classical LDA may struggle in high-dimensional settings due to instability in covariance estimation. In this work, we propose LDA with gradient optimization (LDA-GO), a new approach that directly optimizes the inverse covariance matrix via gradient descent. The algorithm parametrizes the inverse covariance matrix through Cholesky factorization, incorporates a low-rank extension to reduce computational complexity, and considers a multiple-initialization strategy, including identity initialization and warm-starting from the classical LDA estimates. The effectiveness of LDA-GO is demonstrated through extensive multivariate simulations and real-data experiments.
Problem

Research questions and friction points this paper is trying to address.

Improves LDA stability in high-dimensional data
Optimizes inverse covariance via gradient descent
Reduces complexity with low-rank Cholesky factorization
Innovation

Methods, ideas, or system contributions that make the work stand out.

Optimizes inverse covariance via gradient descent
Uses Cholesky factorization for parametrization
Incorporates low-rank extension for efficiency
🔎 Similar Papers
No similar papers found.
Cencheng Shen
Cencheng Shen
University of Delaware
Machine LearningCorrelation and Dependence
Y
Yuexiao Dong
Department of Statistics, Operations, and Data Science, Temple University, Pennsylvania, PA 19122, USA