VisionCreator: A Native Visual-Generation Agentic Model with Understanding, Thinking, Planning and Creation

📅 2026-03-03
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the limitations of existing general-purpose models, which struggle to balance design constraints with creative workflows in visual generation, and current agents, which lack autonomous creative planning capabilities. To overcome these challenges, we propose VisionCreator—an end-to-end learnable visual generation agent that integrates Understanding, Thinking, Planning, and Creation (UTPC) to autonomously produce complex visual content. Our key contributions include the construction of the VisGenData-4k dataset, a metacognition-based creative trajectory generation mechanism, and the introduction of Progressive Specialization Training (PST) and Virtual Reinforcement Learning (VRL), all optimized within a high-fidelity simulation environment. The resulting VisionCreator-8B/32B models outperform larger closed-source counterparts on the VisGenBench benchmark, demonstrating superior performance in multi-step visual creation tasks.

Technology Category

Application Category

📝 Abstract
Visual content creation tasks demand a nuanced understanding of design conventions and creative workflows-capabilities challenging for general models, while workflow-based agents lack specialized knowledge for autonomous creative planning. To overcome these challenges, we propose VisionCreator, a native visual-generation agentic model that unifies Understanding, Thinking, Planning, and Creation (UTPC) capabilities within an end-to-end learnable framework. Our work introduces four key contributions: (i) VisGenData-4k and its construction methodology using metacognition-based VisionAgent to generate high-quality creation trajectories with explicit UTPC structures; (ii) The VisionCreator agentic model, optimized through Progressive Specialization Training (PST) and Virtual Reinforcement Learning (VRL) within a high-fidelity simulated environment, enabling stable and efficient acquisition of UTPC capabilities for complex creation tasks; (iii) VisGenBench, a comprehensive benchmark featuring 1.2k test samples across diverse scenarios for standardized evaluation of multi-step visual creation capabilities; (iv) Remarkably, our VisionCreator-8B/32B models demonstrate superior performance over larger closed-source models across multiple evaluation dimensions. Overall, this work provides a foundation for future research in visual-generation agentic systems.
Problem

Research questions and friction points this paper is trying to address.

visual content creation
creative workflows
autonomous creative planning
design conventions
agentic model
Innovation

Methods, ideas, or system contributions that make the work stand out.

Visual-generation agentic model
UTPC framework
Progressive Specialization Training
Virtual Reinforcement Learning
VisGenBench
🔎 Similar Papers
No similar papers found.
Jinxiang Lai
Jinxiang Lai
Hong Kong University of Science and Technology (HKUST)
Multimodal LLMFew-Shot LearningComputer Vision
Zexin Lu
Zexin Lu
Sichuan University
Jiajun He
Jiajun He
PhD Student, University of Cambridge
Probabilistic MethodsMachine Learning
R
Rongwei Quan
Tencent Hunyuan
W
Wenzhe Zhao
Tencent Hunyuan
Q
Qinyu Yang
Tencent Hunyuan
Q
Qi Chen
Tencent Hunyuan
Q
Qin Lin
Tencent Hunyuan
C
Chuyue Li
Tencent Hunyuan
T
Tao Gao
Tencent Hunyuan
Y
Yuhao Shan
Tencent Hunyuan
Shuai Shao
Shuai Shao
Tencent
Computer VisionMultimediaAIGC
Song Guo
Song Guo
Chair Professor of CSE, HKUST
Large Language ModelEdge AIMachine Learning Systems
Q
Qinglin Lu
Tencent Hunyuan