π€ AI Summary
This work addresses the failure of synergistic integration between masked image modeling (MIM) and diffusion-style denoising in visual self-supervised pretraining. We systematically identify three root causes: insufficient noise injection in the encoderβs feature space, lack of explicit decoupling between masked and noisy tokens, and objective misalignment between the two paradigms. To resolve these issues, we propose a novel unified framework that injects additive noise directly into the intermediate feature representations of the Transformer encoder and introduces explicit token-type identifiers alongside a decoupled attention mechanism to distinguish and process masked versus noisy tokens separately. Experiments demonstrate substantial improvements in fine-grained recognition and high-frequency detail reconstruction. Our method consistently outperforms state-of-the-art baselines across multi-scale downstream image classification tasks, validating the critical role of feature-space noise modeling and semantic token decoupling in representation learning.
π Abstract
In this work, we dive deep into the impact of additive noise in pre-training deep networks. While various methods have attempted to use additive noise inspired by the success of latent denoising diffusion models, when used in combination with masked image modeling, their gains have been marginal when it comes to recognition tasks. We thus investigate why this would be the case, in an attempt to find effective ways to combine the two ideas. Specifically, we find three critical conditions: corruption and restoration must be applied within the encoder, noise must be introduced in the feature space, and an explicit disentanglement between noised and masked tokens is necessary. By implementing these findings, we demonstrate improved pre-training performance for a wide range of recognition tasks, including those that require fine-grained, high-frequency information to solve.