A Cycle Ride to HDR: Semantics Aware Self-Supervised Framework for Unpaired LDR-to-HDR Image Translation

📅 2024-10-19
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the limitations of paired-data dependency and degraded performance in unpaired LDR-to-HDR image translation, this paper proposes a semantic-aware self-supervised cycle-consistent framework. Methodologically, it introduces (1) a novel semantic consistency encoder with an associated loss to explicitly enforce cross-domain semantic preservation; (2) a gradient-aware generator jointly optimized via adversarial loss, cycle-consistency loss, perceptual loss, and gradient-domain regularization—effectively suppressing blur and color-shift artifacts; and (3) the first systematic realization of high-fidelity HDR reconstruction under fully unpaired settings. Extensive experiments demonstrate state-of-the-art performance across multiple benchmarks, with significant improvements in detail fidelity, global contrast, and visual naturalness.

Technology Category

Application Category

📝 Abstract
Low Dynamic Range (LDR) to High Dynamic Range (HDR) image translation is an important computer vision problem. There is a significant amount of research utilizing both conventional non-learning methods and modern data-driven approaches, focusing on using both single-exposed and multi-exposed LDR for HDR image reconstruction. However, most current state-of-the-art methods require high-quality paired {LDR,HDR} datasets for model training. In addition, there is limited literature on using unpaired datasets for this task where the model learns a mapping between domains, i.e., LDR to HDR. To address limitations of current methods, such as the paired data constraint , as well as unwanted blurring and visual artifacts in the reconstructed HDR, we propose a method that uses a modified cycle-consistent adversarial architecture and utilizes unpaired {LDR,HDR} datasets for training. The method introduces novel generators to address visual artifact removal and an encoder and loss to address semantic consistency, another under-explored topic. The method achieves state-of-the-art results across several benchmark datasets and reconstructs high-quality HDR images.
Problem

Research questions and friction points this paper is trying to address.

Unpaired LDR-to-HDR image reconstruction using LLM-based perception.
Addressing visual artifacts and semantic consistency in HDR reconstruction.
Achieving state-of-the-art performance without paired LDR-HDR datasets.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Integrates LLM perception into adversarial architecture
Uses unpaired LDR-HDR datasets for training
Introduces artifact- and exposure-aware generators
🔎 Similar Papers
No similar papers found.
H
H. Barua
Monash University & TCS Research
Kalin Stefanov
Kalin Stefanov
Monash University
Affective ComputingAccessible ComputingHuman-Centered Computing
L
Lemuel Lai En Che
Monash University
A
A. Dhall
Flinders University
K
Koksheik Wong
Monash University
Ganesh Krishnasamy
Ganesh Krishnasamy
Monash University Malaysia
Machine learningcomputer visiondeep learning