Distribution Shift Aware Neural Tabular Learning

📅 2025-08-26
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address performance degradation in tabular learning caused by distribution shift between training and test data, this paper formally defines the novel task of Distribution-Shifted Tabular Learning (DSTL). We propose SAFT (Shift-Aware Feature Transformation), a differentiable, end-to-end framework that replaces discrete feature search with continuous, learnable feature generation. SAFT integrates four core mechanisms: embedding decorrelation, sample reweighting, suboptimal embedding averaging, and normalized distribution alignment. Extensive experiments across diverse real-world distribution shift scenarios—including covariate shift, concept drift, and domain shift—demonstrate that SAFT significantly enhances model robustness, generalization, and predictive accuracy. Compared to state-of-the-art tabular learning methods, SAFT achieves consistent and substantial improvements across multiple benchmarks and evaluation metrics, establishing a new foundation for distribution-robust tabular modeling.

Technology Category

Application Category

📝 Abstract
Tabular learning transforms raw features into optimized spaces for downstream tasks, but its effectiveness deteriorates under distribution shifts between training and testing data. We formalize this challenge as the Distribution Shift Tabular Learning (DSTL) problem and propose a novel Shift-Aware Feature Transformation (SAFT) framework to address it. SAFT reframes tabular learning from a discrete search task into a continuous representation-generation paradigm, enabling differentiable optimization over transformed feature sets. SAFT integrates three mechanisms to ensure robustness: (i) shift-resistant representation via embedding decorrelation and sample reweighting, (ii) flatness-aware generation through suboptimal embedding averaging, and (iii) normalization-based alignment between training and test distributions. Extensive experiments show that SAFT consistently outperforms prior tabular learning methods in terms of robustness, effectiveness, and generalization ability under diverse real-world distribution shifts.
Problem

Research questions and friction points this paper is trying to address.

Addressing performance deterioration in tabular learning under distribution shifts
Transforming discrete feature search into continuous differentiable optimization
Ensuring robustness through decorrelation, reweighting and distribution alignment
Innovation

Methods, ideas, or system contributions that make the work stand out.

Shift-Aware Feature Transformation framework
Differentiable optimization over feature sets
Embedding decorrelation and sample reweighting
🔎 Similar Papers
No similar papers found.