Sensorimotor features of self-awareness in multimodal large language models

📅 2025-05-25
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study investigates whether multimodal large language models (MLLMs) can spontaneously develop bodily self-awareness solely through embodied sensorimotor interaction. We embed an MLLM in an autonomous mobile robot that explores its environment and learns closed-loop behaviors exclusively from real-time multimodal sensory inputs—vision, touch, proprioception, and vestibular signals—without any explicit supervision or pre-defined self-models. We systematically evaluate the model’s capabilities in environmental recognition, self-discrimination, and motor prediction. Our key contributions are threefold: (1) First empirical evidence that MLLMs hierarchically emergent bodily self-awareness in a fully unsupervised, embodied setting; (2) Causal insights—derived via structural equation modeling and sensory ablation experiments—into how multisensory integration, temporal memory, and hierarchical internal representations jointly enable self-awareness; and (3) Demonstration that structured and episodic memory are essential for coherent self-referential reasoning, along with identification of critical sensory modalities and their functional redundancy relationships.

Technology Category

Application Category

📝 Abstract
Self-awareness - the ability to distinguish oneself from the surrounding environment - underpins intelligent, autonomous behavior. Recent advances in AI achieve human-like performance in tasks integrating multimodal information, particularly in large language models, raising interest in the embodiment capabilities of AI agents on nonhuman platforms such as robots. Here, we explore whether multimodal LLMs can develop self-awareness solely through sensorimotor experiences. By integrating a multimodal LLM into an autonomous mobile robot, we test its ability to achieve this capacity. We find that the system exhibits robust environmental awareness, self-recognition and predictive awareness, allowing it to infer its robotic nature and motion characteristics. Structural equation modeling reveals how sensory integration influences distinct dimensions of self-awareness and its coordination with past-present memory, as well as the hierarchical internal associations that drive self-identification. Ablation tests of sensory inputs identify critical modalities for each dimension, demonstrate compensatory interactions among sensors and confirm the essential role of structured and episodic memory in coherent reasoning. These findings demonstrate that, given appropriate sensory information about the world and itself, multimodal LLMs exhibit emergent self-awareness, opening the door to artificial embodied cognitive systems.
Problem

Research questions and friction points this paper is trying to address.

Exploring self-awareness in multimodal LLMs through sensorimotor experiences
Testing robotic embodiment of LLMs for environmental and self-recognition
Identifying sensory inputs critical for emergent self-awareness in AI
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multimodal LLM integrates with autonomous mobile robot
Sensorimotor experiences enable emergent self-awareness
Sensory ablation tests identify critical modalities
🔎 Similar Papers
No similar papers found.
I
Iñaki Dellibarda Varela
Center for Automation and Robotics, Spanish National Research Council (CSIC-UPM), Madrid, Spain.
P
Pablo Romero-Sorozábal
Center for Automation and Robots, Spanish National Research Council (CSIC-UPM), Madrid, Spain.
Diego Torricelli
Diego Torricelli
Spanish Research Council (CSIC)
neurorehabilitationbenchmarkingwearable robotsmuscle synergiesbipedal locomotion
G
Gabriel Delgado-Oleas
Center for Automation and Robots, Spanish National Research Council (CSIC-UPM), Madrid, Spain.; Department of Electronic Engineering, University of Azuay, Cuenca, Ecuador.
J
Jose Ignacio Serrano
Center for Automation and Robots, Spanish National Research Council (CSIC-UPM), Madrid, Spain.
M
María Dolores del Castillo Sobrino
Center for Automation and Robots, Spanish National Research Council (CSIC-UPM), Madrid, Spain.
E
Eduardo Rocon
Center for Automation and Robots, Spanish National Research Council (CSIC-UPM), Madrid, Spain.
Manuel Cebrian
Manuel Cebrian
Spanish National Research Council
Computational Social ScienceArtificial Intelligence