Ultra-lightweight Neural Video Representation Compression

📅 2025-12-03
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing lightweight video compression methods based on implicit neural representations (INRs) suffer from slow encoding speed and limited performance due to autoregressive entropy coding. To address this, we propose an ultra-lightweight neural video representation compression framework. Our key contributions are: (1) a high-resolution multi-scale feature grid that significantly improves reconstruction quality under low computational complexity; and (2) an octree-based context modeling mechanism that replaces conventional autoregressive models, enabling substantially faster entropy coding of high-dimensional feature grids. The framework is trained end-to-end to jointly optimize rate-distortion performance and computational efficiency. Experiments demonstrate that, compared to the state-of-the-art lightweight method C3, our approach achieves BD-rate reductions of 21.03% (PSNR) and 23.06% (MS-SSIM), while accelerating encoding and decoding by 8.4× and 2.5×, respectively.

Technology Category

Application Category

📝 Abstract
Recent works have demonstrated the viability of utilizing over-fitted implicit neural representations (INRs) as alternatives to autoencoder-based models for neural video compression. Among these INR-based video codecs, Neural Video Representation Compression (NVRC) was the first to adopt a fully end-to-end compression framework that compresses INRs, achieving state-of-the-art performance. Moreover, some recently proposed lightweight INRs have shown comparable performance to their baseline codecs with computational complexity lower than 10kMACs/pixel. In this work, we extend NVRC toward lightweight representations, and propose NVRC-Lite, which incorporates two key changes. Firstly, we integrated multi-scale feature grids into our lightweight neural representation, and the use of higher resolution grids significantly improves the performance of INRs at low complexity. Secondly, we address the issue that existing INRs typically leverage autoregressive models for entropy coding: these are effective but impractical due to their slow coding speed. In this work, we propose an octree-based context model for entropy coding high-dimensional feature grids, which accelerates the entropy coding module of the model. Our experimental results demonstrate that NVRC-Lite outperforms C3, one of the best lightweight INR-based video codecs, with up to 21.03% and 23.06% BD-rate savings when measured in PSNR and MS-SSIM, respectively, while achieving 8.4x encoding and 2.5x decoding speedup. The implementation of NVRC-Lite will be made available.
Problem

Research questions and friction points this paper is trying to address.

Compressing video with lightweight neural representations efficiently
Improving entropy coding speed for neural video compression
Enhancing performance of low-complexity implicit neural video codecs
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses multi-scale feature grids for lightweight neural video compression
Proposes octree-based context model to accelerate entropy coding
Achieves significant BD-rate savings and faster encoding/decoding speeds
🔎 Similar Papers
No similar papers found.
Ho Man Kwan
Ho Man Kwan
University of Bristol
Deep LearningVideo Compression
T
Tianhao Peng
Visual Information Laboratory, University of Bristol, UK
G
Ge Gao
Visual Information Laboratory, University of Bristol, UK
F
Fan Zhang
Visual Information Laboratory, University of Bristol, UK
M
Mike Nilsson
Network Connectivity Services Research, BT, UK
A
Andrew Gower
Network Connectivity Services Research, BT, UK
David Bull
David Bull
Professor of Signal Processing, Director Bristol Vision Insitute, University of Bristol
imagevideosignal processingcommunicationsoptimisation