PRAM-R: A Perception-Reasoning-Action-Memory Framework with LLM-Guided Modality Routing for Adaptive Autonomous Driving

📅 2026-03-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the high computational cost of multimodal perception in autonomous driving and its limited adaptability to dynamic environmental changes. The authors propose PRAM-R, a framework featuring an asynchronous dual-loop architecture that jointly coordinates perception, reasoning, action, and memory. Central to this approach is a large language model (LLM)-driven, context-aware modality routing mechanism coupled with a hierarchical memory system, enabling adaptive modality selection and long-term consistency preservation. Evaluated on the nuScenes dataset, PRAM-R reduces modality usage by 6.22% while improving memory recall by 20%, all without compromising trajectory prediction accuracy. In synthetic stress tests, the framework demonstrates an 87.2% reduction in routing oscillations, substantially enhancing both computational efficiency and system robustness.

Technology Category

Application Category

📝 Abstract
Multimodal perception enables robust autonomous driving but incurs unnecessary computational cost when all sensors remain active. This paper presents PRAM-R, a unified Perception-Reasoning-Action-Memory framework with LLM-Guided Modality Routing for adaptive autonomous driving. PRAM-R adopts an asynchronous dual-loop design: a fast reactive loop for perception and control, and a slow deliberative loop for reasoning-driven modality selection and memory updates. An LLM router selects and weights modalities using environmental context and sensor diagnostics, while a hierarchical memory module preserves temporal consistency and supports long-term adaptation. We conduct a two-stage evaluation: (1) synthetic stress tests for stability analysis and (2) real-world validation on the nuScenes dataset. Synthetic stress tests confirm 87.2% reduction in routing oscillations via hysteresis-based stabilization. Real-world validation on nuScenes shows 6.22% modality reduction with 20% memory recall while maintaining comparable trajectory accuracy to full-modality baselines in complex urban scenarios. Our work demonstrates that LLM-augmented architectures with hierarchical memory achieve efficient, adaptive multimodal perception in autonomous driving.
Problem

Research questions and friction points this paper is trying to address.

multimodal perception
autonomous driving
modality selection
computational efficiency
adaptive perception
Innovation

Methods, ideas, or system contributions that make the work stand out.

LLM-Guided Modality Routing
Adaptive Autonomous Driving
Hierarchical Memory
Asynchronous Dual-Loop Architecture
Multimodal Perception
🔎 Similar Papers
No similar papers found.
Y
Yi Zhang
Chair of Robotics, Artificial Intelligence and Embedded Systems, Technical University of Munich (TUM), Munich, Germany
X
Xian Zhang
Chair of Robotics, Artificial Intelligence and Embedded Systems, Technical University of Munich (TUM), Munich, Germany
S
Saisi Zhao
Chair of Robotics, Artificial Intelligence and Embedded Systems, Technical University of Munich (TUM), Munich, Germany
Y
Yinglei Song
Chair of Robotics, Artificial Intelligence and Embedded Systems, Technical University of Munich (TUM), Munich, Germany
C
Chengdong Wu
Chair of Robotics, Artificial Intelligence and Embedded Systems, Technical University of Munich (TUM), Munich, Germany
Nenad Petrovic
Nenad Petrovic
Faculty of Electronic Engineering, University of Nis
Semantic TechnologyModel-Driven Software EngineeringDomain-Specific LanguagesLLM
Alois Knoll
Alois Knoll
Technische Universität München
RoboticsAISensor Data FusionAutonomous DrivingCyber Physical Systems