🤖 AI Summary
Existing autoencoder-based image compression methods suffer from suboptimal rate-distortion performance at high bitrates and lack single-model support for broad-rate adaptive coding. To address this, we propose a reversible-transform-based variable-rate image compression model. Our method introduces a lightweight multi-scale invertible neural network that establishes a bijective mapping between the image and a hierarchical latent space, ensuring lossless invertibility and full-rate adaptability. We further design a novel entropy estimation mechanism that jointly leverages multi-scale spatial–semantic context modeling and an expansion gain unit. Evaluated across an exceptionally wide bitrate range (0.01–1.0 bpp), our single model consistently outperforms the VVC standard—particularly at high bitrates—and achieves state-of-the-art variable-rate performance, matching or exceeding multi-model approaches. The source code is publicly available.
📝 Abstract
Autoencoder-based structures have dominated recent learned image compression methods. However, the inherent information loss associated with autoencoders limits their rate-distortion performance at high bit rates and restricts their flexibility of rate adaptation. In this paper, we present a variable-rate image compression model based on invertible transform to overcome these limitations. Specifically, we design a lightweight multi-scale invertible neural network, which bijectively maps the input image into multi-scale latent representations. To improve the compression efficiency, a multi-scale spatial-channel context model with extended gain units is devised to estimate the entropy of the latent representation from high to low levels. Experimental results demonstrate that the proposed method achieves state-of-the-art performance compared to existing variable-rate methods, and remains competitive with recent multi-model approaches. Notably, our method is the first learned image compression solution that outperforms VVC across a very wide range of bit rates using a single model, especially at high bit rates.The source code is available at href{https://github.com/hytu99/MSINN-VRLIC}{https://github.com/hytu99/MSINN-VRLIC}.