🤖 AI Summary
This study investigates how embodied emotional expressions—facial, bodily, and their multimodal fusion—modulate socio-emotional behaviors of 7–11-year-old children during turn-taking games with the QTrobot, focusing specifically on empathy responses and behavioral alignment. Using a controlled experimental design, we combined systematic behavioral coding with qualitative analysis to conduct the first systematic comparison of differential effects across emotional expression modalities in child-robot interaction. Results demonstrate that children significantly align their behavior with the robot’s expressed emotional state, confirming that robotic emotional cues can elicit rudimentary empathic responses. However, multimodal (face + body) expression did not enhance behavioral alignment relative to unimodal conditions—challenging the intuitive “more modalities, greater efficacy” assumption. These findings provide empirical grounding and a novel design paradigm for child-robot social relationships: effective emotional expression must prioritize appropriateness and parsimony over intensity or modality proliferation.
📝 Abstract
The growing development of robots with artificial emotional expressiveness raises important questions about their persuasive potential in children's behavior. While research highlights the pragmatic value of emotional expressiveness in human social communication, the extent to which robotic expressiveness can or should influence empathic responses in children is grounds for debate. In a pilot study with 22 children (aged 7-11) we begin to explore the ways in which different levels of embodied expressiveness (body only, face only, body and face) of two basic emotions (happiness and sadness) displayed by an anthropomorphic robot (QTRobot) might modify children's behavior in a child-robot cooperative turn-taking game. We observed that children aligned their behavior to the robot's inferred emotional state. However, higher levels of expressiveness did not result in increased alignment. The preliminary results reported here provide a starting point for reflecting on robotic expressiveness and its role in shaping children's social-emotional behavior toward robots as social peers in the near future.