🤖 AI Summary
This study addresses the challenge of robotic autonomous development by proposing a curiosity-driven reinforcement learning framework inspired by infant self-exploration. Methodologically, it integrates an intrinsic reward mechanism, high-dimensional tactile encoding, skin-color- and shape-guided visual feature learning, and a staged curriculum learning strategy to model visuomotor coordination development—from unimanual to bimanual control—in simulation. Its key contribution is the first demonstration of fully unsupervised, end-to-end acquisition of whole-body tactile coverage and hand–eye coordination in robots, achieved solely through intrinsically motivated behaviors such as self-touch and hand-focused visual attention—closely mirroring early infant developmental trajectories. Experimental results demonstrate that the framework effectively supports progressive, self-organized emergence of multimodal sensorimotor capabilities, enabling hierarchical integration of vision, touch, and motor control without external supervision or task-specific rewards.
📝 Abstract
Inspired by infant development, we propose a Reinforcement Learning (RL) framework for autonomous self-exploration in a robotic agent, Baby Sophia, using the BabyBench simulation environment. The agent learns self-touch and hand regard behaviors through intrinsic rewards that mimic an infant's curiosity-driven exploration of its own body. For self-touch, high-dimensional tactile inputs are transformed into compact, meaningful representations, enabling efficient learning. The agent then discovers new tactile contacts through intrinsic rewards and curriculum learning that encourage broad body coverage, balance, and generalization. For hand regard, visual features of the hands, such as skin-color and shape, are learned through motor babbling. Then, intrinsic rewards encourage the agent to perform novel hand motions, and follow its hands with its gaze. A curriculum learning setup from single-hand to dual-hand training allows the agent to reach complex visual-motor coordination. The results of this work demonstrate that purely curiosity-based signals, with no external supervision, can drive coordinated multimodal learning, imitating an infant's progression from random motor babbling to purposeful behaviors.