🤖 AI Summary
Traditional Linear Discriminant Analysis (LDA) is sensitive to noise and fails when the within-class scatter matrix is singular; its stepwise feature selection relies on Wilks’ Λ, which tends to terminate prematurely and degrades discriminative performance. This paper proposes a novel forward discriminant analysis framework. Methodologically, it integrates Pillai’s trace criterion with Uncorrelated LDA (ULDA) for the first time, establishing a unified and interpretable forward feature selection mechanism that avoids premature termination inherent to Wilks’ Λ and naturally accommodates perfectly separable classes. Furthermore, Type I error calibration is incorporated to ensure statistical significance control. Empirical evaluation on both synthetic and real-world datasets demonstrates substantial improvements in classification accuracy and robust false positive rate control, particularly excelling in scenarios of complete class separability.
📝 Abstract
Linear discriminant analysis (LDA), a traditional classification tool, suffers from limitations such as sensitivity to noise and computational challenges when dealing with non-invertible within-class scatter matrices. Traditional stepwise LDA frameworks, which iteratively select the most informative features, often exacerbate these issues by relying heavily on Wilks' $Lambda$, potentially causing premature stopping of the selection process. This paper introduces a novel forward discriminant analysis framework that integrates Pillai's trace with Uncorrelated Linear Discriminant Analysis (ULDA) to address these challenges, and offers a unified and stand-alone classifier. Through simulations and real-world datasets, the new framework demonstrates effective control of Type I error rates and improved classification accuracy, particularly in cases involving perfect group separations. The results highlight the potential of this approach as a robust alternative to the traditional stepwise LDA framework.