DiffusionX: Efficient Edge-Cloud Collaborative Image Generation with Multi-Round Prompt Evolution

📅 2025-10-17
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Diffusion-based image generation faces bottlenecks including high computational overhead, time-consuming prompt tuning, and excessive cloud resource demand. To address these, we propose the first edge–cloud collaborative framework supporting multi-round prompt evolution: a lightweight diffusion model on the edge rapidly generates coarse previews, while a large-scale model in the cloud performs fine-grained refinement. We introduce a novel noise-level predictor that dynamically allocates computational tasks to optimize the trade-off between end-to-end latency and cloud load. Experiments show that our framework reduces average generation time by 15.8% over Stable Diffusion v1.5 with comparable image quality (FID and CLIP-Score), and incurs only 0.9% higher latency than Tiny-SD while significantly improving FID. This work is the first to deeply integrate iterative prompt refinement into the edge–cloud generation pipeline, achieving a balanced design across efficiency, fidelity, and scalability.

Technology Category

Application Category

📝 Abstract
Recent advances in diffusion models have driven remarkable progress in image generation. However, the generation process remains computationally intensive, and users often need to iteratively refine prompts to achieve the desired results, further increasing latency and placing a heavy burden on cloud resources. To address this challenge, we propose DiffusionX, a cloud-edge collaborative framework for efficient multi-round, prompt-based generation. In this system, a lightweight on-device diffusion model interacts with users by rapidly producing preview images, while a high-capacity cloud model performs final refinements after the prompt is finalized. We further introduce a noise level predictor that dynamically balances the computation load, optimizing the trade-off between latency and cloud workload. Experiments show that DiffusionX reduces average generation time by 15.8% compared with Stable Diffusion v1.5, while maintaining comparable image quality. Moreover, it is only 0.9% slower than Tiny-SD with significantly improved image quality, thereby demonstrating efficiency and scalability with minimal overhead.
Problem

Research questions and friction points this paper is trying to address.

Reduces computational load of diffusion models
Minimizes latency in iterative prompt refinement
Optimizes edge-cloud workload distribution efficiently
Innovation

Methods, ideas, or system contributions that make the work stand out.

Cloud-edge collaborative framework for prompt-based image generation
Lightweight on-device model interacts with users rapidly
Noise level predictor dynamically balances computation load
🔎 Similar Papers
No similar papers found.
Y
Yi Wei
College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
Shunpu Tang
Shunpu Tang
PhD student at Zhejiang University
Semantic CommunicationsEdge IntelligenceGenerative AI
L
Liang Zhao
College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
Qianqian Yang
Qianqian Yang
Zhejiang University
Information TheoryWireless AISemantic CommunicationMachine Learning