FOCUS: Frequency-Optimized Conditioning of DiffUSion Models for mitigating catastrophic forgetting during Test-Time Adaptation

📅 2025-08-20
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address catastrophic forgetting of task-relevant knowledge under domain shift in test-time adaptation (TTA), this paper proposes a frequency-optimized input adaptation framework based on diffusion models. The method operates without updating model parameters, preserving original task knowledge while adapting to corrupted inputs. Key contributions include: (i) the first introduction of a frequency-domain conditioning mechanism; (ii) a lightweight Y-shaped Frequency Prediction Network (Y-FPN) that decouples and retains semantically critical low-frequency components; (iii) FrequencyMix for enhanced frequency-domain robustness; and (iv) pseudo-label co-optimization for improved consistency. Extensive experiments across 15 noise types and three benchmark datasets demonstrate state-of-the-art performance on semantic segmentation and monocular depth estimation, with reduced computational overhead compared to existing TTA approaches.

Technology Category

Application Category

📝 Abstract
Test-time adaptation enables models to adapt to evolving domains. However, balancing the tradeoff between preserving knowledge and adapting to domain shifts remains challenging for model adaptation methods, since adapting to domain shifts can induce forgetting of task-relevant knowledge. To address this problem, we propose FOCUS, a novel frequency-based conditioning approach within a diffusion-driven input-adaptation framework. Utilising learned, spatially adaptive frequency priors, our approach conditions the reverse steps during diffusion-driven denoising to preserve task-relevant semantic information for dense prediction. FOCUS leverages a trained, lightweight, Y-shaped Frequency Prediction Network (Y-FPN) that disentangles high and low frequency information from noisy images. This minimizes the computational costs involved in implementing our approach in a diffusion-driven framework. We train Y-FPN with FrequencyMix, a novel data augmentation method that perturbs the images across diverse frequency bands, which improves the robustness of our approach to diverse corruptions. We demonstrate the effectiveness of FOCUS for semantic segmentation and monocular depth estimation across 15 corruption types and three datasets, achieving state-of-the-art averaged performance. In addition to improving standalone performance, FOCUS complements existing model adaptation methods since we can derive pseudo labels from FOCUS-denoised images for additional supervision. Even under limited, intermittent supervision with the pseudo labels derived from the FOCUS denoised images, we show that FOCUS mitigates catastrophic forgetting for recent model adaptation methods.
Problem

Research questions and friction points this paper is trying to address.

Mitigating catastrophic forgetting during test-time domain adaptation
Balancing knowledge preservation and domain shift adaptation
Preserving task-relevant semantic information in diffusion models
Innovation

Methods, ideas, or system contributions that make the work stand out.

Frequency-based conditioning in diffusion models
Y-shaped Frequency Prediction Network for efficiency
FrequencyMix data augmentation for robustness
🔎 Similar Papers
No similar papers found.
G
Gabriel Tjio
Centre for Frontier AI Research (CFAR), Agency for Science, Technology and Research (A*STAR), 1 Fusionopolis Way, #16-16 Connexis, Singapore, 138632, Republic of Singapore.
J
Jie Zhang
Centre for Frontier AI Research (CFAR), Agency for Science, Technology and Research (A*STAR), 1 Fusionopolis Way, #16-16 Connexis, Singapore, 138632, Republic of Singapore.
Xulei Yang
Xulei Yang
Principal Scientist & Group Leader, A*STAR, Singapore
3D VisionArtificial IntelligenceMedical Imaging
Yun Xing
Yun Xing
School of Computer Science and Engineering, Nanyang Technological University
Computer Vision
N
Nhat Chung
Centre for Frontier AI Research (CFAR), Agency for Science, Technology and Research (A*STAR), 1 Fusionopolis Way, #16-16 Connexis, Singapore, 138632, Republic of Singapore.
X
Xiaofeng Cao
School of Computer Science and Technology, Tongji University, 1238 Gonghexin Road, Shanghai, 200070, China.
I
Ivor W. Tsang
Centre for Frontier AI Research (CFAR), Agency for Science, Technology and Research (A*STAR), 1 Fusionopolis Way, #16-16 Connexis, Singapore, 138632, Republic of Singapore.
C
Chee Keong Kwoh
College of Computing and Data Science, Nanyang Technological University, 50 Nanyang Avenue, Singapore, 639798, Republic of Singapore.
Q
Qing Guo
Centre for Frontier AI Research (CFAR), Agency for Science, Technology and Research (A*STAR), 1 Fusionopolis Way, #16-16 Connexis, Singapore, 138632, Republic of Singapore.