🤖 AI Summary
This work addresses the challenges of view consistency and conditional controllability in multi-view image generation. To this end, we propose MV-AR, a multi-view autoregressive generative framework. Methodologically, MV-AR introduces a unified X-to-multi-view modeling paradigm and incorporates a conditional injection module to fuse heterogeneous modalities—including text, pose, reference images, and shape priors. A progressive training strategy is adopted to enhance geometric and textural consistency across views, while a novel “Shuffle View” data augmentation technique mitigates the scarcity of high-quality multi-view training data. Experiments demonstrate that MV-AR achieves strong cross-view geometric and appearance consistency, supports diverse conditional inputs, and attains generation quality competitive with state-of-the-art diffusion-based methods. Moreover, it exhibits superior generalization, adaptability, and training stability on multi-view synthesis tasks.
📝 Abstract
Generating multi-view images from human instructions is crucial for 3D content creation. The primary challenges involve maintaining consistency across multiple views and effectively synthesizing shapes and textures under diverse conditions. In this paper, we propose the Multi-View Auto-Regressive (MV-AR) method, which leverages an auto-regressive model to progressively generate consistent multi-view images from arbitrary prompts. Firstly, the next-token-prediction capability of the AR model significantly enhances its effectiveness in facilitating progressive multi-view synthesis. When generating widely-separated views, MV-AR can utilize all its preceding views to extract effective reference information. Subsequently, we propose a unified model that accommodates various prompts via architecture designing and training strategies. To address multiple conditions, we introduce condition injection modules for text, camera pose, image, and shape. To manage multi-modal conditions simultaneously, a progressive training strategy is employed. This strategy initially adopts the text-to-multi-view (t2mv) model as a baseline to enhance the development of a comprehensive X-to-multi-view (X2mv) model through the randomly dropping and combining conditions. Finally, to alleviate the overfitting problem caused by limited high-quality data, we propose the "Shuffle View" data augmentation technique, thus significantly expanding the training data by several magnitudes. Experiments demonstrate the performance and versatility of our MV-AR, which consistently generates consistent multi-view images across a range of conditions and performs on par with leading diffusion-based multi-view image generation models. Code and models will be released at https://github.com/MILab-PKU/MVAR.