Continual Contrastive Learning on Tabular Data with Out of Distribution

📅 2025-03-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the poor out-of-distribution (OOD) generalization of tabular models under distributional shift, this paper proposes Tabular Continual Contrastive Learning (TCCL)—the first framework to synergistically integrate contrastive learning with continual learning. TCCL introduces an incrementally updatable encoder–decoder–learning-head architecture that enables robust representation evolution in response to dynamic distribution drift. The method jointly optimizes self-supervised autoencoding reconstruction, representation alignment, and multi-task learning, supporting online OOD detection and regression. Evaluated on eight real-world tabular datasets, TCCL consistently outperforms 14 state-of-the-art baselines—including gradient-boosted decision trees and leading deep models—achieving superior performance in both OOD classification and regression. Moreover, it demonstrates significantly enhanced generalization capability and training stability under continual distribution shifts.

Technology Category

Application Category

📝 Abstract
Out-of-distribution (OOD) prediction remains a significant challenge in machine learning, particularly for tabular data where traditional methods often fail to generalize beyond their training distribution. This paper introduces Tabular Continual Contrastive Learning (TCCL), a novel framework designed to address OOD challenges in tabular data processing. TCCL integrates contrastive learning principles with continual learning mechanisms, featuring a three-component architecture: an Encoder for data transformation, a Decoder for representation learning, and a Learner Head. We evaluate TCCL against 14 baseline models, including state-of-the-art deep learning approaches and gradient-boosted decision trees (GBDT), across eight diverse tabular datasets. Our experimental results demonstrate that TCCL consistently outperforms existing methods in both classification and regression tasks on OOD data, with particular strength in handling distribution shifts. These findings suggest that TCCL represents a significant advancement in handling OOD scenarios for tabular data.
Problem

Research questions and friction points this paper is trying to address.

Addresses out-of-distribution prediction challenges in tabular data.
Introduces a novel framework combining contrastive and continual learning.
Outperforms existing methods in classification and regression on OOD data.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Integrates contrastive and continual learning
Three-component architecture: Encoder, Decoder, Learner
Outperforms 14 baselines on OOD tabular data
🔎 Similar Papers
No similar papers found.
A
Achmad Ginanjar
School of Electrical Engineering and Computer Science, The University of Queensland, Queensland, Australia
X
Xue Li
School of Electrical Engineering and Computer Science, The University of Queensland, Queensland, Australia
Priyanka Singh
Priyanka Singh
Lecturer of Cyber Security, University of Queensland
Cyber SecurityMultimedia ForensicsEncrypted Domain Processing
Wen Hua
Wen Hua
The Hong Kong Polytechnic University
DatabaseInformation SystemData MiningDeep Learning