🤖 AI Summary
This study investigates whether multimodal large language models (MLLMs) can unsupervisedly recapitulate human affective perception of images—specifically valence, arousal, and basic discrete emotions. Using standardized affective image datasets (e.g., IAPS) and state-of-the-art vision–language pre-trained models, we employ contrastive image–text learning without any explicit affective supervision, thereby implicitly modeling affective concepts through cross-modal statistical regularities. Results demonstrate strong alignment (r > 0.8) between model-predicted affective dimensions and human population-level ground-truth ratings—providing the first empirical evidence that affective perception can spontaneously emerge from unsupervised multimodal representation learning. Crucially, ablation analyses reveal that the language modality plays a pivotal role in affective concept formation, uncovering a structural guidance mechanism whereby textual priors shape visual affect understanding. These findings offer novel, cognitively grounded evidence for the interpretability and psychological plausibility of AI-based affective intelligence.
📝 Abstract
Affective reactions have deep biological foundations, however in humans the development of emotion concepts is also shaped by language and higher-order cognition. A recent breakthrough in AI has been the creation of multimodal language models that exhibit impressive intellectual capabilities, but their responses to affective stimuli have not been investigated. Here we study whether state-of-the-art multimodal systems can emulate human emotional ratings on a standardized set of images, in terms of affective dimensions and basic discrete emotions. The AI judgements correlate surprisingly well with the average human ratings: given that these systems were not explicitly trained to match human affective reactions, this suggests that the ability to visually judge emotional content can emerge from statistical learning over large-scale databases of images paired with linguistic descriptions. Besides showing that language can support the development of rich emotion concepts in AI, these findings have broad implications for sensitive use of multimodal AI technology.