| license: mit | |
| base_model: | |
| - Manojb/stable-diffusion-2-1-base | |
| pipeline_tag: image-to-image | |
| library_name: diffusers | |
| # Towards Photorealistic and Efficient Bokeh Rendering via Diffusion Framework | |
| This repository contains the model and code for **Towards Photorealistic and Efficient Bokeh Rendering via Diffusion Framework**, as presented in the paper: | |
| [**Towards Photorealistic and Efficient Bokeh Rendering via Diffusion Framework**](https://arxiv.org/abs/2605.07429) | |
| ## Abstract | |
| Existing mobile devices are constrained by compact optical designs, such as small apertures, which make it difficult to produce natural, optically realistic bokeh effects. Although recent learning-based methods have shown promising results, they still struggle with photos captured under high digital zoom levels, which often suffer from reduced resolution and loss of fine details. A naive solution is to enhance image quality before applying bokeh rendering, yet this two-stage pipeline reduces efficiency and introduces unnecessary error accumulation. To overcome these limitations, we propose MagicBokeh, a unified diffusion-based framework designed for high-quality and efficient bokeh rendering. Through an alternative training strategy and a focus-aware masked attention mechanism, our method jointly optimizes bokeh rendering and super-resolution, substantially improving both controllability and visual fidelity. Furthermore, we introduce degradation-aware depth module to enable more accurate depth estimation from low-quality inputs. Experimental results demonstrate that MagicBokeh efficiently produces photorealistic bokeh effects, particularly on real-world low-resolution images, paving the way for future advancements in bokeh rendering. | |
| ## Code and Usage | |
| The official code and model are available at the following GitHub repository: | |
| [https://github.com/vivoCameraResearch/MagicBokeh](https://github.com/vivoCameraResearch/MagicBokeh) |