Human-Like Gaze Behavior in Social Robots: A Deep Learning Approach Integrating Human and Non-Human Stimuli

πŸ“… 2026-02-12
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This study addresses the challenge that current social robots struggle to naturally emulate human gaze behavior in complex social scenarios, particularly in responding to non-human stimuli such as falling objects or opening doors. Moving beyond prior work focused solely on interpersonal interactions, this research presents the first systematic model of integrated human gaze responses to both social and non-social events. Using Unity-generated 3D animations and 360Β° real-world videos within a virtual reality environment, eye-tracking data were collected to train LSTM and Transformer models for predicting gaze direction. The models achieved prediction accuracies of 70.4% and 72% in animated and real-world scenes, respectively, significantly outperforming existing approaches. When deployed on a NAO robot, the system received high user ratings from 275 participants, demonstrating a marked improvement in the naturalness of human–robot interaction.

Technology Category

Application Category

πŸ“ Abstract
Nonverbal behaviors, particularly gaze direction, play a crucial role in enhancing effective communication in social interactions. As social robots increasingly participate in these interactions, they must adapt their gaze based on human activities and remain receptive to all cues, whether human-generated or not, to ensure seamless and effective communication. This study aims to increase the similarity between robot and human gaze behavior across various social situations, including both human and non-human stimuli (e.g., conversations, pointing, door openings, and object drops). A key innovation in this study, is the investigation of gaze responses to non-human stimuli, a critical yet underexplored area in prior research. These scenarios, were simulated in the Unity software as a 3D animation and a 360-degree real-world video. Data on gaze directions from 41 participants were collected via virtual reality (VR) glasses. Preprocessed data, trained two neural networks-LSTM and Transformer-to build predictive models based on individuals'gaze patterns. In the animated scenario, the LSTM and Transformer models achieved prediction accuracies of 67.6% and 70.4%, respectively; In the real-world scenario, the LSTM and Transformer models achieved accuracies of 72% and 71.6%, respectively. Despite the gaze pattern differences among individuals, our models outperform existing approaches in accuracy while uniquely considering non-human stimuli, offering a significant advantage over previous literature. Furthermore, deployed on the NAO robot, the system was evaluated by 275 participants via a comprehensive questionnaire, with results demonstrating high satisfaction during interactions. This work advances social robotics by enabling robots to dynamically mimic human gaze behavior in complex social contexts.
Problem

Research questions and friction points this paper is trying to address.

human-like gaze
social robots
non-human stimuli
gaze behavior
social interaction
Innovation

Methods, ideas, or system contributions that make the work stand out.

gaze behavior
non-human stimuli
deep learning
social robotics
Transformer
πŸ”Ž Similar Papers
No similar papers found.
F
Faezeh Vahedi
Social and Cognitive Robotics Laboratory, Center of Excellence in Design, Robotics, and Automation (CEDRA), Sharif University of Technology, Tehran, Iran
M
Morteza Memari
Social and Cognitive Robotics Laboratory, Center of Excellence in Design, Robotics, and Automation (CEDRA), Sharif University of Technology, Tehran, Iran
Ramtin Tabatabaei
Ramtin Tabatabaei
PhD Student, University of Melbourne
Robot FailureHuman Robot CollaborationHuman Gaze
Alireza Taheri
Alireza Taheri
PhD in Mechanical Engineering, Associate Professor, Sharif University of Technology
Social RoboticsCognitive RoboticsHuman-Robot InteractionChildren with Special Needs