Baby Sophia: A Developmental Approach to Self-Exploration through Self-Touch and Hand Regard

📅 2025-11-12
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses the challenge of robotic autonomous development by proposing a curiosity-driven reinforcement learning framework inspired by infant self-exploration. Methodologically, it integrates an intrinsic reward mechanism, high-dimensional tactile encoding, skin-color- and shape-guided visual feature learning, and a staged curriculum learning strategy to model visuomotor coordination development—from unimanual to bimanual control—in simulation. Its key contribution is the first demonstration of fully unsupervised, end-to-end acquisition of whole-body tactile coverage and hand–eye coordination in robots, achieved solely through intrinsically motivated behaviors such as self-touch and hand-focused visual attention—closely mirroring early infant developmental trajectories. Experimental results demonstrate that the framework effectively supports progressive, self-organized emergence of multimodal sensorimotor capabilities, enabling hierarchical integration of vision, touch, and motor control without external supervision or task-specific rewards.

Technology Category

Application Category

📝 Abstract
Inspired by infant development, we propose a Reinforcement Learning (RL) framework for autonomous self-exploration in a robotic agent, Baby Sophia, using the BabyBench simulation environment. The agent learns self-touch and hand regard behaviors through intrinsic rewards that mimic an infant's curiosity-driven exploration of its own body. For self-touch, high-dimensional tactile inputs are transformed into compact, meaningful representations, enabling efficient learning. The agent then discovers new tactile contacts through intrinsic rewards and curriculum learning that encourage broad body coverage, balance, and generalization. For hand regard, visual features of the hands, such as skin-color and shape, are learned through motor babbling. Then, intrinsic rewards encourage the agent to perform novel hand motions, and follow its hands with its gaze. A curriculum learning setup from single-hand to dual-hand training allows the agent to reach complex visual-motor coordination. The results of this work demonstrate that purely curiosity-based signals, with no external supervision, can drive coordinated multimodal learning, imitating an infant's progression from random motor babbling to purposeful behaviors.
Problem

Research questions and friction points this paper is trying to address.

Develops robotic self-exploration through infant-inspired touch and vision learning
Enables autonomous discovery of body interactions using intrinsic reward systems
Achieves coordinated multimodal behaviors without external supervision or guidance
Innovation

Methods, ideas, or system contributions that make the work stand out.

Reinforcement Learning framework for autonomous self-exploration
Intrinsic rewards drive curiosity-based multimodal learning
Curriculum learning enables complex visual-motor coordination
🔎 Similar Papers
No similar papers found.
S
Stelios Zarifis
Robotics Institute, Athena Research Center, Athens, Greece
I
Ioannis Chalkiadakis
Institute for Language and Speech Processing, Athena Research Center, Athens, Greece
A
Artemis Chardouveli
Institute for Language and Speech Processing, Athena Research Center, Athens, Greece
V
Vasiliki Moutzouri
Institute for Language and Speech Processing, Athena Research Center, Athens, Greece
A
Aggelos Sotirchos
Institute for Language and Speech Processing, Athena Research Center, Athens, Greece
K
Katerina Papadimitriou
Robotics Institute, Athena Research Center, Athens, Greece
P
P. Filntisis
Robotics Institute, Athena Research Center, Athens, Greece
Niki Efthymiou
Niki Efthymiou
Athena Research Center, National Technical Uneversity of Athens
human-robot interactioncomputer visionmachine learning
Petros Maragos
Petros Maragos
Professor of Electrical and Computer Engineering, National Technical University of Athens
computer visionsignal processingspeech&languagemachine learningrobotics
K
Katerina Pastra
Institute for Language and Speech Processing, Athena Research Center, Athens, Greece