π€ AI Summary
This paper proposes a general multimodal agent framework tailored for cross-platform gaming scenarios, addressing key limitations of existing game AIβincluding poor generalization, heterogeneous action spaces, and low training efficiency. Methodologically, it introduces a unified, extensible action space grounded in native keyboard-mouse inputs, enabling seamless operation across diverse environments such as operating systems, web browsers, and emulators; further, it incorporates a causal decay loss function and a sparse reasoning strategy to support large-scale continual pretraining and efficient inference. Contributions include: (1) the first cross-domain multimodal trajectory modeling with human-aligned action representations; (2) a ~2Γ improvement in success rate over SOTA on open-ended Minecraft tasks; (3) achieving novice-human-level performance on unseen web-based 3D games; and (4) outperforming GPT-5, Gemini-2.5-Pro, and Claude-4-Sonnet on FPS benchmarks.
π Abstract
We present Game-TARS, a generalist game agent trained with a unified, scalable action space anchored to human-aligned native keyboard-mouse inputs. Unlike API- or GUI-based approaches, this paradigm enables large-scale continual pre-training across heterogeneous domains, including OS, web, and simulation games. Game-TARS is pre-trained on over 500B tokens with diverse trajectories and multimodal data. Key techniques include a decaying continual loss to reduce causal confusion and an efficient Sparse-Thinking strategy that balances reasoning depth and inference cost. Experiments show that Game-TARS achieves about 2 times the success rate over the previous sota model on open-world Minecraft tasks, is close to the generality of fresh humans in unseen web 3d games, and outperforms GPT-5, Gemini-2.5-Pro, and Claude-4-Sonnet in FPS benchmarks. Scaling results on training-time and test-time confirm that the unified action space sustains improvements when scaled to cross-game and multimodal data. Our results demonstrate that simple, scalable action representations combined with large-scale pre-training provide a promising path toward generalist agents with broad computer-use abilities.