🤖 AI Summary
Grayscale image colorization is a highly ill-posed problem due to the absence of two-dimensional chrominance information, leading to multimodal color mapping ambiguities. This paper proposes a semantic-aware classification-based colorization framework that integrates convolutional neural networks (CNNs) with generative adversarial networks (GANs), formulating colorization as a classification task over a discretized color space. We jointly optimize a semantic-guided classification loss and an adversarial loss. Crucially, semantic segmentation priors are incorporated to constrain color distributions and mitigate color ambiguity, while adversarial training enhances texture fidelity and perceptual realism. Evaluated on standard benchmarks—including ImageNet and Places2—our method significantly outperforms regression-based colorization approaches, achieving state-of-the-art performance in color accuracy, semantic plausibility, and output diversity.
📝 Abstract
Image colorization, the task of adding colors to grayscale images, has been the focus of significant research efforts in computer vision in recent years for its various application areas such as color restoration and automatic animation colorization [15, 1]. The colorization problem is challenging as it is highly ill-posed with two out of three image dimensions lost, resulting in large degrees of freedom. However, semantics of the scene as well as the surface texture could provide important cues for colors: the sky is typically blue, the clouds are typically white and the grass is typically green, and there are huge amounts of training data available for learning such priors since any colored image could serve as a training data point [20].
Colorization is initially formulated as a regression task[5], which ignores the multi-modal nature of color prediction. In this project, we explore automatic image colorization via classification and adversarial learning. We will build our models on prior works, apply modifications for our specific scenario and make comparisons.