MER 2025: When Affective Computing Meets Large Language Models

📅 2025-04-28
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses the limitations of conventional affective computing—namely, its reliance on predefined emotion categories and poor generalizability—by proposing a large language model (LLM)-driven generative paradigm for emotion understanding. Methodologically, we introduce the first LLM-adapted generative emotion modeling framework, integrating semi-supervised learning, multimodal (text, speech, vision) fusion, fine-grained label expansion, and interpretable attention mechanisms; we further pioneer a cross-task collaborative learning paradigm linking emotion and personality. Key contributions include: (1) releasing MERTools—an open-source toolkit; (2) establishing a unified evaluation protocol; and (3) introducing an open benchmark dataset covering four tracks: semi-supervised emotion recognition, fine-grained emotion analysis, multimodal interpretable prediction, and emotion-enhanced personality recognition. Experimental results demonstrate significant improvements in accuracy and cross-domain generalization across multiple metrics, advancing the practical deployment of generative affective AI.

Technology Category

Application Category

📝 Abstract
MER2025 is the third year of our MER series of challenges, aiming to bring together researchers in the affective computing community to explore emerging trends and future directions in the field. Previously, MER2023 focused on multi-label learning, noise robustness, and semi-supervised learning, while MER2024 introduced a new track dedicated to open-vocabulary emotion recognition. This year, MER2025 centers on the theme"When Affective Computing Meets Large Language Models (LLMs)".We aim to shift the paradigm from traditional categorical frameworks reliant on predefined emotion taxonomies to LLM-driven generative methods, offering innovative solutions for more accurate and reliable emotion understanding. The challenge features four tracks: MER-SEMI focuses on fixed categorical emotion recognition enhanced by semi-supervised learning; MER-FG explores fine-grained emotions, expanding recognition from basic to nuanced emotional states; MER-DES incorporates multimodal cues (beyond emotion words) into predictions to enhance model interpretability; MER-PR investigates whether emotion prediction results can improve personality recognition performance. For the first three tracks, baseline code is available at MERTools, and datasets can be accessed via Hugging Face. For the last track, the dataset and baseline code are available on GitHub.
Problem

Research questions and friction points this paper is trying to address.

Shifting from categorical frameworks to LLM-driven emotion understanding
Enhancing emotion recognition with semi-supervised and fine-grained methods
Investigating multimodal cues for interpretable affective computing
Innovation

Methods, ideas, or system contributions that make the work stand out.

LLM-driven generative emotion understanding methods
Multimodal cues enhance model interpretability
Semi-supervised learning for emotion recognition
🔎 Similar Papers
No similar papers found.
Zheng Lian
Zheng Lian
Associate Professor, IEEE/CCF Senior Member, Institute of Automation, Chinese Academy of Sciences
Affective ComputingSentiment AnalysisMachine Learning
R
Rui Liu
Inner Mongolia University, Inner Mongolia, China
K
Kele Xu
National University of Defense Technology, Changsha, China
B
Bin Liu
Institute of Automation, CAS, Beijing, China
X
Xuefei Liu
Tianjin Normal University, Tianjin, China
Yazhou Zhang
Yazhou Zhang
Associate Professor, Tianjin University
Sentiment AnalysisQuantum CognitionSarcasm DetectionHumor Analysis
X
Xin Liu
Lappeenranta-Lahti University of Technology, Lappeenranta, Finland
Y
Yong Li
Southeast University, Nanjing, China
Zebang Cheng
Zebang Cheng
Shenzhen University
AICVMLLMAffective Computing
H
Haolin Zuo
Inner Mongolia University, Inner Mongolia, China
Z
Ziyang Ma
Shanghai Jiaotong University, Shanghai, China
Xiaojiang Peng
Xiaojiang Peng
Shenzhen Technology University
Computer VisionFacial Expression RecognitionMultimodal Emotion Recognition
X
Xie Chen
Shanghai Jiaotong University, Shanghai, China
Y
Ya Li
Beijing University of Posts and Telecommunications, Beijing, China
Erik Cambria
Erik Cambria
Professor @ NTU CCDS & Visiting @ MIT Media Lab
Neurosymbolic AIMultimodal InteractionNLPAffective ComputingSentiment Analysis
Guoying Zhao
Guoying Zhao
Academy Professor, IEEE Fellow, Professor of Computer Science and Engineering, University of Oulu
Affective ComputingArtificial IntelligenceComputer VisionPattern Recognition
B
Bjorn W. Schuller
Technical University of Munich, Munich, Germany
J
Jianhua Tao
Tsinghua University, Beijing, China