🤖 AI Summary
This study addresses the usability bottlenecks of gaze-based interaction on mobile devices under seated versus walking conditions. We propose and empirically validate a multimodal gaze-coordinated paradigm that decouples navigation (performed via gaze gestures) from selection (executed via dwell time or pursuit motions). To our knowledge, this is the first systematic investigation of such a functional separation in gaze-based interaction. Using eye-tracking, gaze gesture recognition, and dual-task user studies, we quantitatively characterize how locomotion degrades gaze accuracy—revealing that pursuit-based selection induces significant visual interference, whereas gaze gestures substantially reduce false-selection rates. Results show that our approach improves task accuracy by 18% and reduces completion time by 23% over baseline gaze-only methods. It thus significantly enhances the robustness and efficiency of hands-free interaction in dynamic environments, establishing a novel paradigm for ubiquitous gaze interaction under motion.
📝 Abstract
The potential of gaze for hands-free mobile interaction is increasingly evident. While each gaze input technique presents distinct advantages and limitations, a combination can amplify strengths and mitigate challenges. We report on the results of a user study (N=24), in which we compared the usability and performance of pairing three popular gaze input techniques: Dwell Time, Pursuits, and Gaze Gestures, for navigation and selection tasks while sitting and walking. Results show that pairing gestures for navigation with either Dwell time or Pursuits for selection improves task completion time and rate compared to using either individually. We discuss the implications of pairing gaze input techniques, such as how Pursuits may negatively impact other techniques, likely due to the visual clutter it adds, how integrating gestures for navigation reduces the chances of unintentional selections, and the impact of motor activity on performance. Our findings provide insights for effective gaze-enabled interfaces.