đ¤ AI Summary
This work addresses the challenge of modeling operating-system-level GUI dynamics end-to-end for pixel-level screen frame generation from raw user inputs (mouse/keyboard events). We propose the first end-to-end neural simulation framework that jointly models GUI state evolutionâusing a recurrent neural networkâand high-fidelity image synthesisâvia a diffusion-based neural rendererâtrained on large-scale real-world Ubuntu XFCE screen recordings. Our contributions are threefold: (1) unified modeling of GUI state transitions and corresponding visual outputs; (2) integration of AI agents to synthesize high-quality interactive trajectories, mitigating scarcity of human-annotated interaction data; and (3) superior performance over baselines on complex state-transition tasks such as application launching and window switching. Experiments demonstrate photorealistic reproduction of fine-grained interaction detailsâincluding cursor motion and animated transitionsâadvancing the frontier of generative humanâcomputer interface modeling.
đ Abstract
We introduce NeuralOS, a neural framework that simulates graphical user interfaces (GUIs) of operating systems by directly predicting screen frames in response to user inputs such as mouse movements, clicks, and keyboard events. NeuralOS combines a recurrent neural network (RNN), which tracks computer state, with a diffusion-based neural renderer that generates screen images. The model is trained on a large-scale dataset of Ubuntu XFCE recordings, which include both randomly generated interactions and realistic interactions produced by AI agents. Experiments show that NeuralOS successfully renders realistic GUI sequences, accurately captures mouse interactions, and reliably predicts state transitions like application launches. Although modeling fine-grained keyboard interactions precisely remains challenging, NeuralOS offers a step toward creating fully adaptive, generative neural interfaces for future human-computer interaction systems.