🤖 AI Summary
Image Aesthetic Assessment (IAA) faces challenges including severe label scarcity, high subjectivity, and the incompatibility of large CLIP-based models with lightweight deployment scenarios. To address these, we propose CSKD, a two-stage CLIP-driven semi-supervised knowledge distillation framework. In Stage I, cross-encoder feature alignment mitigates feature collapse during CLIP fine-tuning. In Stage II, multi-source unlabeled data are leveraged via consistency regularization and attention-distance constraints for collaborative optimization. Our method introduces the first unified, scalable CLIP feature transfer paradigm, enabling plug-and-play integration of arbitrary vision encoders. Evaluated on multiple mainstream IAA benchmarks, CSKD achieves state-of-the-art performance: student models attain significant accuracy gains while reducing inference latency, demonstrating strong efficacy in resource-constrained and low-label regimes.
📝 Abstract
The success of pre-training approaches on a variety of downstream tasks has revitalized the field of computer vision. Image aesthetics assessment (IAA) is one of the ideal application scenarios for such methods due to subjective and expensive labeling procedure. In this work, an unified and flexible two-phase extbf{C}LIP-based extbf{S}emi-supervised extbf{K}nowledge extbf{D}istillation paradigm is proposed, namely extbf{ extit{CSKD}}. Specifically, we first integrate and leverage a multi-source unlabeled dataset to align rich features between a given visual encoder and an off-the-shelf CLIP image encoder via feature alignment loss. Notably, the given visual encoder is not limited by size or structure and, once well-trained, it can seamlessly serve as a better visual aesthetic learner for both student and teacher. In the second phase, the unlabeled data is also utilized in semi-supervised IAA learning to further boost student model performance when applied in latency-sensitive production scenarios. By analyzing the attention distance and entropy before and after feature alignment, we notice an alleviation of feature collapse issue, which in turn showcase the necessity of feature alignment instead of training directly based on CLIP image encoder. Extensive experiments indicate the superiority of CSKD, which achieves state-of-the-art performance on multiple widely used IAA benchmarks.