🤖 AI Summary
This work addresses the inverse ISP problem: reconstructing high-fidelity RAW sensor data from metadata-free smartphone sRGB images—bypassing the irreversibility inherent in conventional ISP modeling. We introduce the first large-scale, end-to-end sRGB→RAW inverse ISP benchmark and establish a camera-agnostic reconstruction paradigm. Our method employs a conditional generative deep neural network, incorporating physics-inspired constraints and multi-scale perceptual loss to significantly enhance cross-device generalization. Evaluated on a benchmark with over 150 participating teams, our approach sets new state-of-the-art scores across PSNR, SSIM, and LPIPS. Reconstructed RAW images are rigorously validated via forward ISP rendering and real-sensor response simulation, demonstrating strong fidelity and consistency. This work opens new avenues for low-level vision modeling and synthetic RAW data generation.
📝 Abstract
Numerous low-level vision tasks operate in the RAW domain due to its linear properties, bit depth, and sensor designs. Despite this, RAW image datasets are scarce and more expensive to collect than the already large and public sRGB datasets. For this reason, many approaches try to generate realistic RAW images using sensor information and sRGB images. This paper covers the second challenge on RAW Reconstruction from sRGB (Reverse ISP). We aim to recover RAW sensor images from smartphones given the corresponding sRGB images without metadata and, by doing this, ``reverse"the ISP transformation. Over 150 participants joined this NTIRE 2025 challenge and submitted efficient models. The proposed methods and benchmark establish the state-of-the-art for generating realistic RAW data.