DynaQuant: Dynamic Mixed-Precision Quantization for Learned Image Compression

๐Ÿ“… 2025-11-11
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Existing learned image compression (LIC) methods typically employ static, uniform quantization bit-widths, which fail to accommodate the heterogeneous and sensitivity-diverse feature distributions within modelsโ€”leading to suboptimal trade-offs between rate-distortion performance and computational efficiency. To address this, we propose DynaQuant, a dynamic mixed-precision quantization framework featuring a novel dual-level bit-width selection mechanism: content-aware quantization at the feature level and data-driven adaptation at the channel level. We further introduce a distance-aware gradient modulator (DGM) to enable end-to-end differentiable optimization. Leveraging learnable scaling/offset parameters and a lightweight bit-width selection network, DynaQuant adaptively allocates precision at both feature- and channel-granularities. Experiments demonstrate that DynaQuant preserves the rate-distortion performance of full-precision models while significantly reducing computational cost and memory footprint, thereby enhancing deployment flexibility across diverse hardware platforms.

Technology Category

Application Category

๐Ÿ“ Abstract
Prevailing quantization techniques in Learned Image Compression (LIC) typically employ a static, uniform bit-width across all layers, failing to adapt to the highly diverse data distributions and sensitivity characteristics inherent in LIC models. This leads to a suboptimal trade-off between performance and efficiency. In this paper, we introduce DynaQuant, a novel framework for dynamic mixed-precision quantization that operates on two complementary levels. First, we propose content-aware quantization, where learnable scaling and offset parameters dynamically adapt to the statistical variations of latent features. This fine-grained adaptation is trained end-to-end using a novel Distance-aware Gradient Modulator (DGM), which provides a more informative learning signal than the standard Straight-Through Estimator. Second, we introduce a data-driven, dynamic bit-width selector that learns to assign an optimal bit precision to each layer, dynamically reconfiguring the network's precision profile based on the input data. Our fully dynamic approach offers substantial flexibility in balancing rate-distortion (R-D) performance and computational cost. Experiments demonstrate that DynaQuant achieves rd performance comparable to full-precision models while significantly reducing computational and storage requirements, thereby enabling the practical deployment of advanced LIC on diverse hardware platforms.
Problem

Research questions and friction points this paper is trying to address.

Dynamic mixed-precision quantization adapts to diverse data distributions in image compression
Content-aware quantization with learnable parameters adjusts to latent feature variations
Dynamic bit-width selector optimizes layer precision based on input data characteristics
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dynamic mixed-precision quantization adapts to data distributions
Content-aware quantization with learnable scaling and offset parameters
Data-driven dynamic bit-width selector optimizes layer precision
๐Ÿ”Ž Similar Papers
No similar papers found.
Y
Youneng Bao
College of Electronics and Information Engineering, Shenzhen University
Y
Yulong Cheng
Harbin Institute of Technology, Shenzhen
Yiping Liu
Yiping Liu
Hunan University
Evolutionary ComputationMulti-Objective OptimizationMachine LearningDrug Discorvery
Yichen Yang
Yichen Yang
EX-PhD Student at University of Michigan
Computer Architecture
P
Peng Qin
China Telecom Group Qinhuangdao Branch
M
Mu Li
Harbin Institute of Technology, Shenzhen
Yongsheng Liang
Yongsheng Liang
Harbin Institute of Technology
Image ProcessingSource CodingChannel Coding