Designing Memory-Augmented AR Agents for Spatiotemporal Reasoning in Personalized Task Assistance

📅 2025-08-12
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing AR agents struggle with complex, multi-step tasks primarily due to their inability to model users’ long-term experiences and preferences—specifically, they lack persistent spatiotemporal awareness, memory retention, and historical interaction reasoning. This work introduces the first AR agent framework integrating a spatiotemporal memory mechanism, combining multimodal large language models, multi-sensor fusion perception, spatiotemporal graph neural networks, and persistent memory storage to enable cumulative cross-spatiotemporal reasoning. Its modular architecture shifts AR systems from transient response paradigms toward sustained, personalized adaptation. Experiments in smart home and cognitive assistance scenarios demonstrate significant improvements in task completion rate and user adaptation fidelity. The framework establishes a foundational technical basis for long-term, personalized AR interaction. (138 words)

Technology Category

Application Category

📝 Abstract
Augmented Reality (AR) systems are increasingly integrating foundation models, such as Multimodal Large Language Models (MLLMs), to provide more context-aware and adaptive user experiences. This integration has led to the development of AR agents to support intelligent, goal-directed interactions in real-world environments. While current AR agents effectively support immediate tasks, they struggle with complex multi-step scenarios that require understanding and leveraging user's long-term experiences and preferences. This limitation stems from their inability to capture, retain, and reason over historical user interactions in spatiotemporal contexts. To address these challenges, we propose a conceptual framework for memory-augmented AR agents that can provide personalized task assistance by learning from and adapting to user-specific experiences over time. Our framework consists of four interconnected modules: (1) Perception Module for multimodal sensor processing, (2) Memory Module for persistent spatiotemporal experience storage, (3) Spatiotemporal Reasoning Module for synthesizing past and present contexts, and (4) Actuator Module for effective AR communication. We further present an implementation roadmap, a future evaluation strategy, a potential target application and use cases to demonstrate the practical applicability of our framework across diverse domains. We aim for this work to motivate future research toward developing more intelligent AR systems that can effectively bridge user's interaction history with adaptive, context-aware task assistance.
Problem

Research questions and friction points this paper is trying to address.

Enhancing AR agents for complex multi-step task scenarios
Improving long-term user experience and preference understanding
Enabling spatiotemporal reasoning for personalized task assistance
Innovation

Methods, ideas, or system contributions that make the work stand out.

Memory-augmented AR agents for personalized assistance
Multimodal sensor processing in Perception Module
Spatiotemporal Reasoning Module synthesizes past and present
🔎 Similar Papers
No similar papers found.
Dongwook Choi
Dongwook Choi
Yonsei University M.S.
T
Taeyoon Kwon
Language & AGI Lab, Yonsei University
D
Dongil Yang
Language & AGI Lab, Yonsei University
H
Hyojun Kim
Language & AGI Lab, Yonsei University
Jinyoung Yeo
Jinyoung Yeo
Yonsei University
Natural Language ProceesingLarge Language ModelsAI Agents