🤖 AI Summary
Existing UI agents predominantly employ end-to-end automation, neglecting user involvement and contextual awareness at critical decision points—thereby undermining user autonomy (e.g., blind and low-vision (BLV) users remain unaware of equally priced but superior alternatives, such as higher-rated or better-tasting sparkling water options). Method: We propose a hybrid-proactivity interaction paradigm: for the first time, enabling UI agents to autonomously identify decision points—without explicit prompting—by jointly analyzing user queries, UI source code, and screenshots via large multimodal models, then pausing execution to solicit user preferences. Contribution/Results: Evaluated on real-world web tasks, our approach significantly improves task completion rate and option alignment fidelity. In BLV-user scenarios, it outperforms baseline systems including OpenAI Operator, effectively balancing automation efficiency with user agency and control.
📝 Abstract
User interface (UI) agents promise to make inaccessible or complex UIs easier to access for blind and low-vision (BLV) users. However, current UI agents typically perform tasks end-to-end without involving users in critical choices or making them aware of important contextual information, thus reducing user agency. For example, in our field study, a BLV participant asked to buy the cheapest available sparkling water, and the agent automatically chose one from several equally priced options, without mentioning alternative products with different flavors or better ratings. To address this problem, we introduce Morae, a UI agent that automatically identifies decision points during task execution and pauses so that users can make choices. Morae uses large multimodal models to interpret user queries alongside UI code and screenshots, and prompt users for clarification when there is a choice to be made. In a study over real-world web tasks with BLV participants, Morae helped users complete more tasks and select options that better matched their preferences, as compared to baseline agents, including OpenAI Operator. More broadly, this work exemplifies a mixed-initiative approach in which users benefit from the automation of UI agents while being able to express their preferences.