Realtime Multimodal Emotion Estimation using Behavioral and Neurophysiological Data

📅 2025-08-12
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address emotion recognition and expression challenges faced by neurodiverse individuals—particularly those with autism spectrum disorder (ASD)—this study proposes an interpretable, personalized, real-time multimodal affect estimation framework. The method integrates physiological signals (EEG, ECG, BVP, GSR/EDA) with behavioral modalities (facial expressions, speech) and models affect dynamically in the two-dimensional arousal–valence space. Key contributions include: (1) a cross-modal unified representation mechanism supporting both naturalistic (passive video viewing) and interactive (semi-structured dialogue) scenarios; and (2) an individual adaptation module coupled with neuroadaptive feedback, enabling affective education and inclusive human–machine interaction. Experimental results demonstrate significant improvements in real-time affect estimation accuracy and inter-subject specificity. This work establishes a novel paradigm for affective computing tailored to neurodiverse users, advancing both theoretical understanding and practical deployment in assistive and educational technologies.

Technology Category

Application Category

📝 Abstract
Many individuals especially those with autism spectrum disorder (ASD), alexithymia, or other neurodivergent profiles face challenges in recognizing, expressing, or interpreting emotions. To support more inclusive and personalized emotion technologies, we present a real-time multimodal emotion estimation system that combines neurophysiological EEG, ECG, blood volume pulse (BVP), and galvanic skin response (GSR/EDA) and behavioral modalities (facial expressions, and speech) in a unified arousal-valence 2D interface to track moment-to-moment emotional states. This architecture enables interpretable, user-specific analysis and supports applications in emotion education, neuroadaptive feedback, and interaction support for neurodiverse users. Two demonstration scenarios illustrate its application: (1) passive media viewing (2D or VR videos) reveals cortical and autonomic responses to affective content, and (2) semi-scripted conversations with a facilitator or virtual agent capture real-time facial and vocal expressions. These tasks enable controlled and naturalistic emotion monitoring, making the system well-suited for personalized feedback and neurodiversity-informed interaction design.
Problem

Research questions and friction points this paper is trying to address.

Real-time emotion estimation for neurodiverse individuals
Multimodal data fusion for accurate emotion tracking
Personalized feedback and interaction design support
Innovation

Methods, ideas, or system contributions that make the work stand out.

Combines EEG, ECG, BVP, GSR for realtime emotion tracking
Integrates facial and speech data in 2D arousal-valence interface
Enables neurodiverse-inclusive feedback via multimodal user-specific analysis
🔎 Similar Papers
No similar papers found.
V
Von Ralph Dane Marquez Herbuela
International Research Center for Neurointelligence (WPI-IRCN), The University of Tokyo, Tokyo, Japan
Yukie Nagai
Yukie Nagai
The University of Tokyo
cognitive developmental roboticscomputational neurosciencehuman-robot interaction