NTIRE 2025 Challenge on UGC Video Enhancement: Methods and Results

📅 2025-05-05
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Real-world user-generated content (UGC) videos suffer from multiple degradations—including noise, blur, color fading, and compression artifacts—posing significant challenges for no-reference video enhancement. To address this, we introduce UGC-Video, the first large-scale, multi-degradation, no-reference benchmark comprising 150短视频 clips tailored for short-video platforms, accompanied by a large-scale no-reference enhancement challenge. We innovatively employ over 8,000 crowd-sourced subjective evaluations to establish a perception-driven assessment framework independent of PSNR/SSIM. Methodologically, we propose an end-to-end spatiotemporal joint modeling framework integrating degradation-aware feature extraction and perceptual consistency loss. The challenge attracted participation from 25+ teams, with seven successfully validating their methods via source-code reproduction. We fully open-source all enhanced results, subjective scores, and evaluation protocols—constituting, to date, the largest publicly available no-reference UGC video enhancement benchmark.

Technology Category

Application Category

📝 Abstract
This paper presents an overview of the NTIRE 2025 Challenge on UGC Video Enhancement. The challenge constructed a set of 150 user-generated content videos without reference ground truth, which suffer from real-world degradations such as noise, blur, faded colors, compression artifacts, etc. The goal of the participants was to develop an algorithm capable of improving the visual quality of such videos. Given the widespread use of UGC on short-form video platforms, this task holds substantial practical importance. The evaluation was based on subjective quality assessment in crowdsourcing, obtaining votes from over 8000 assessors. The challenge attracted more than 25 teams submitting solutions, 7 of which passed the final phase with source code verification. The outcomes may provide insights into the state-of-the-art in UGC video enhancement and highlight emerging trends and effective strategies in this evolving research area. All data, including the processed videos and subjective comparison votes and scores, is made publicly available at https://github.com/msu-video-group/NTIRE25_UGC_Video_Enhancement.
Problem

Research questions and friction points this paper is trying to address.

Enhancing visual quality of user-generated videos
Addressing real-world degradations like noise and blur
Developing algorithms for UGC video improvement
Innovation

Methods, ideas, or system contributions that make the work stand out.

UGC video enhancement without ground truth
Subjective quality assessment via crowdsourcing
Public dataset with processed videos and scores
🔎 Similar Papers
No similar papers found.
N
Nikolay Safonov
Department of Graphics and Media Lab, Moscow State University
A
Alexey Bryncev
Andrey Moskalenko
Andrey Moskalenko
PhD Student @ MSU
Computer VisionRoboticsVideo ProcessingCrowdsourcing
D
Dmitry Kulikov
Department of Graphics and Media Lab, Moscow State University
D
Dmitry Vatolin
Department of Graphics and Media Lab, Moscow State University
R
R. Timofte
University of Würzburg
H
Haibo Lei
Q
Qifan Gao
Qing Luo
Qing Luo
Institute of Microelectronics, Chinese Academy of Sciences
memory
Y
Yaqing Li
J
Jie Song
S
Shaozhe Hao
M
Meisong Zheng
J
Jingyi Xu
C
Chengbin Wu
Jiahui Liu
Jiahui Liu
Fujitsu Research of America
Quantum ComputingCryptographyQuantum Cryptography
Y
Ying Chen
X
Xin Deng
Mai Xu
Mai Xu
Beihang Univeristy, Tsinghua Univeristy, Imperial College London
P
Peipei Liang
J
Jie Ma
J
Junjie Jin
Y
Yingxue Pang
Fang Luo
Fang Luo
Stony Brook University (SUNY Stony Brook), NY
Power ElectronicsEMIPower Module Packaging
K
Kai Chen
S
Shijie Zhao
Mingyang Wu
Mingyang Wu
Texas A&M University
Generative AI
R
Renjie Li
Yushen Zuo
Yushen Zuo
The Hong Kong Polytechnic University
Computer visionDeep learningImage Generation
S
Shengyun Zhong
Zhengzhong Tu
Zhengzhong Tu
Texas A&M University, Google Research, University of Texas at Austin
Agentic AITrustworthy AIEmbodied AI