π€ AI Summary
This work addresses the significant accessibility barriers faced by screen reader users due to the poor adaptation of mainstream graphical user interfaces, which impose steep learning curves and operational challenges for people with visual impairments. To overcome these limitations, the authors propose AskEaseβthe first on-demand AI assistant that integrates large language models with multimodal context awareness. By dynamically fusing on-screen content, interaction history, and user task goals, AskEase infers user intent in real time and delivers low-interruption, step-by-step contextual voice guidance. In a controlled study with 12 blind and visually impaired participants, AskEase significantly improved task success rates while markedly reducing physical demand, mental effort, and frustration, thereby demonstrating its effectiveness and innovation in empowering visually impaired users to independently accomplish computer-based tasks.
π Abstract
Equal access to digital technologies is critical for education, employment, and social participation. However, mainstream interfaces are visually oriented, creating steep learning curves and frequent obstacles for screen reader users, and limiting their independence and opportunities. Existing support is inadequate -- tutorials mainly target sighted users, while human assistance lacks real-time availability. We introduce AskEase, an on-demand AI assistant that provides step-by-step, screen reader user-friendly guidance for computer use. AskEase manages multiple sources of context to infer user intent and deliver precise, situation-specific guidance. Its seamless interaction design minimizes disruption and reduces the effort of seeking help. We demonstrated its effectiveness through representative usage scenarios and robustness tests. In a within-subjects study with 12 screen reader users, AskEase significantly improved task success while reducing perceived workload, including physical demand, effort, and frustration. These results demonstrate the potential of LLM-powered assistants to promote accessible computing and expand opportunities for users with visual impairments.