Open-World Test-Time Adaptation with Hierarchical Feature Aggregation and Attention Affine

📅 2025-11-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the insufficient robustness of test-time adaptation (TTA) to out-of-distribution (OOD) samples in open-world settings—particularly its tendency to misclassify unknown classes as known—the paper proposes a hierarchical feature-driven TTA framework. Methodologically: (1) a Hierarchical Ladder Network integrates multi-level Transformer class tokens for fine-grained OOD detection; (2) an Attention Affine Network dynamically recalibrates self-attention weights to adapt to domain shifts; and (3) a weighted entropy mechanism suppresses interference from low-confidence predictions. Evaluated on multiple benchmarks, the approach significantly improves in-distribution (ID) classification accuracy and OOD detection performance while ensuring robust prediction for both ID and OOD samples. This work establishes a scalable and highly stable new paradigm for open-world TTA.

Technology Category

Application Category

📝 Abstract
Test-time adaptation (TTA) refers to adjusting the model during the testing phase to cope with changes in sample distribution and enhance the model's adaptability to new environments. In real-world scenarios, models often encounter samples from unseen (out-of-distribution, OOD) categories. Misclassifying these as known (in-distribution, ID) classes not only degrades predictive accuracy but can also impair the adaptation process, leading to further errors on subsequent ID samples. Many existing TTA methods suffer substantial performance drops under such conditions. To address this challenge, we propose a Hierarchical Ladder Network that extracts OOD features from class tokens aggregated across all Transformer layers. OOD detection performance is enhanced by combining the original model prediction with the output of the Hierarchical Ladder Network (HLN) via weighted probability fusion. To improve robustness under domain shift, we further introduce an Attention Affine Network (AAN) that adaptively refines the self-attention mechanism conditioned on the token information to better adapt to domain drift, thereby improving the classification performance of the model on datasets with domain shift. Additionally, a weighted entropy mechanism is employed to dynamically suppress the influence of low-confidence samples during adaptation. Experimental results on benchmark datasets show that our method significantly improves the performance on the most widely used classification datasets.
Problem

Research questions and friction points this paper is trying to address.

Adapting models to distribution shifts during testing with hierarchical feature aggregation
Detecting out-of-distribution samples while maintaining in-distribution classification accuracy
Enhancing robustness to domain drift through attention-based affine transformations
Innovation

Methods, ideas, or system contributions that make the work stand out.

Hierarchical Ladder Network extracts OOD features
Attention Affine Network adaptively refines self-attention mechanism
Weighted entropy suppresses low-confidence samples influence
🔎 Similar Papers
No similar papers found.
Ziqiong Liu
Ziqiong Liu
MIND
Multimedia SearchComputer VisionMachine Learning
Y
Yushun Tang
Southern University of Science and Technology, Shenzhen, China
J
Junyang Ji
Southern University of Science and Technology, Shenzhen, China; Shenzhen International Graduate School, Tsinghua University, Shenzhen, China
Zhihai He
Zhihai He
Southern University of Science and Technology
Deep learningcomputer visionmachine learningsmart cyber-physical systems