Visionary Co-Driver: Enhancing Driver Perception of Potential Risks with LLM and HUD

📅 2025-11-18
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses drivers’ insufficient perception of non-collision road hazards—such as illegal parking, construction zones, and anomalous pedestrian behavior—by proposing a novel driver assistance system integrating large language models (LLMs), eye-tracking, and adaptive head-up display (HUD). Methodologically, it pioneers the use of LLMs for driving risk reasoning, combining video understanding algorithms to detect latent hazards and dynamically modulating HUD cue location, timing, and semantic intensity based on real-time gaze data—enabling context-aware, personalized hazard alerts. A user study (N=41) demonstrates significant improvements in both hazard identification accuracy and situation awareness for non-collision risks. Key contributions include: (1) establishing an LLM-driven paradigm for semantic risk inference; (2) closing the “vision–cognition–interaction” loop; and (3) providing a foundational framework for next-generation explainable, adaptive intelligent driver assistance systems.

Technology Category

Application Category

📝 Abstract
Drivers' perception of risky situations has always been a challenge in driving. Existing risk-detection methods excel at identifying collisions but face challenges in assessing the behavior of road users in non-collision situations. This paper introduces Visionary Co-Driver, a system that leverages large language models to identify non-collision roadside risks and alert drivers based on their eye movements. Specifically, the system combines video processing algorithms and LLMs to identify potentially risky road users. These risks are dynamically indicated on an adaptive heads-up display interface to enhance drivers' attention. A user study with 41 drivers confirms that Visionary Co-Driver improves drivers' risk perception and supports their recognition of roadside risks.
Problem

Research questions and friction points this paper is trying to address.

Enhancing driver perception of non-collision roadside risks
Leveraging LLMs to identify potentially risky road user behavior
Alerting drivers dynamically via adaptive heads-up display interface
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses large language models for non-collision risk identification
Combines video processing algorithms with LLMs
Displays dynamic alerts on adaptive heads-up interface
🔎 Similar Papers
No similar papers found.
W
Wei Xiang
International Design Institute, Zhejiang University, Hangzhou, China
Z
Ziyue Lei
International Design Institute, Zhejiang University, Hangzhou, China
J
Jie Wang
GRASP Lab, University of Pennsylvania, PA, USA
Y
Yingying Huang
International Design Institute, Zhejiang University, Hangzhou, China
Q
Qi Zheng
The Hong Kong University of Science and Technology (Guangzhou), Guangzhou, China
T
Tianyi Zhang
Chu Kochen Honors College, Zhejiang University, Hangzhou, China
An Zhao
An Zhao
University College London
Medical Image AnalysisDeep Learning
Lingyun Sun
Lingyun Sun
Zhejiang University
Design IntelligenceHCIArtificial IntelligenceIndustrial DesignAIGC