BokehFlow: Depth-Free Controllable Bokeh Rendering via Flow Matching

📅 2025-11-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses controllable bokeh rendering without requiring depth maps. Method: We propose the first end-to-end, text-guided generative framework that eliminates reliance on depth estimation. Instead, foreground–background motion relationships are modeled via optical flow matching, while cross-attention mechanisms precisely map textual prompts (e.g., “focus on the flower with strong background blur”) to spatial focus regions and per-pixel blur intensity distributions. Contribution/Results: To support training and evaluation, we introduce a dedicated bokeh dataset comprising four scene categories. Extensive experiments demonstrate that our method surpasses existing depth-based approaches and general-purpose generative models in visual fidelity, semantic controllability, and inference efficiency. It significantly lowers deployment barriers and enhances user interactivity, enabling intuitive, text-driven bokeh control.

Technology Category

Application Category

📝 Abstract
Bokeh rendering simulates the shallow depth-of-field effect in photography, enhancing visual aesthetics and guiding viewer attention to regions of interest. Although recent approaches perform well, rendering controllable bokeh without additional depth inputs remains a significant challenge. Existing classical and neural controllable methods rely on accurate depth maps, while generative approaches often struggle with limited controllability and efficiency. In this paper, we propose BokehFlow, a depth-free framework for controllable bokeh rendering based on flow matching. BokehFlow directly synthesizes photorealistic bokeh effects from all-in-focus images, eliminating the need for depth inputs. It employs a cross-attention mechanism to enable semantic control over both focus regions and blur intensity via text prompts. To support training and evaluation, we collect and synthesize four datasets. Extensive experiments demonstrate that BokehFlow achieves visually compelling bokeh effects and offers precise control, outperforming existing depth-dependent and generative methods in both rendering quality and efficiency.
Problem

Research questions and friction points this paper is trying to address.

Rendering controllable bokeh effects without depth input requirements
Overcoming limited controllability and efficiency in generative bokeh methods
Enabling semantic control of focus regions and blur intensity
Innovation

Methods, ideas, or system contributions that make the work stand out.

Depth-free bokeh rendering via flow matching
Semantic control using cross-attention and text prompts
Direct synthesis from all-in-focus images without depth
🔎 Similar Papers
No similar papers found.
Y
Yachuan Huang
School of AIA, Huazhong University of Science and Technology
Xianrui Luo
Xianrui Luo
Tsinghua University
3D visionmulti-modalconstruction automationcomputational photography
Q
Qiwen Wang
School of AIA, Huazhong University of Science and Technology
Liao Shen
Liao Shen
Huazhong University of Science and Technology
computer vision
J
Jiaqi Li
School of AIA, Huazhong University of Science and Technology
H
Huiqiang Sun
School of AIA, Huazhong University of Science and Technology
Z
Zihao Huang
School of AIA, Huazhong University of Science and Technology
W
Wei Jiang
School of AIA, Huazhong University of Science and Technology
Zhiguo Cao
Zhiguo Cao
Huazhong University of Science and Technology
Pattern RecognitionComputer Vision