Transformer-Progressive Mamba Network for Lightweight Image Super-Resolution

๐Ÿ“… 2025-11-05
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Existing Mamba-based super-resolution methods lack fine-grained multi-scale modeling capability, limiting feature representation efficiency. To address this, we propose a lightweight multi-scale receptive field interaction framework that synergistically integrates windowed self-attention with a progressive Mamba mechanism, enabling joint global contextual awareness and local detail preservation under linear computational complexity. Furthermore, we design an adaptive high-frequency refinement module to ensure smooth multi-scale feature transitions and precise recovery of high-frequency details. Our approach seamlessly unifies the long-range modeling strength of Transformers with the efficient state-space properties of Mamba. Extensive experiments demonstrate that the method consistently outperforms state-of-the-art Transformer- and Mamba-based baselines across multiple benchmarks, achieving superior PSNR and SSIM scores with significantly lower computational overheadโ€”thus striking an optimal balance between accuracy and efficiency.

Technology Category

Application Category

๐Ÿ“ Abstract
Recently, Mamba-based super-resolution (SR) methods have demonstrated the ability to capture global receptive fields with linear complexity, addressing the quadratic computational cost of Transformer-based SR approaches. However, existing Mamba-based methods lack fine-grained transitions across different modeling scales, which limits the efficiency of feature representation. In this paper, we propose T-PMambaSR, a lightweight SR framework that integrates window-based self-attention with Progressive Mamba. By enabling interactions among receptive fields of different scales, our method establishes a fine-grained modeling paradigm that progressively enhances feature representation with linear complexity. Furthermore, we introduce an Adaptive High-Frequency Refinement Module (AHFRM) to recover high-frequency details lost during Transformer and Mamba processing. Extensive experiments demonstrate that T-PMambaSR progressively enhances the model's receptive field and expressiveness, yielding better performance than recent Transformer- or Mamba-based methods while incurring lower computational cost. Our codes will be released after acceptance.
Problem

Research questions and friction points this paper is trying to address.

Addresses quadratic computational cost in Transformer-based super-resolution
Enables fine-grained transitions across different modeling scales
Recovers high-frequency details lost during Transformer and Mamba processing
Innovation

Methods, ideas, or system contributions that make the work stand out.

Combines window self-attention with Progressive Mamba
Enables multi-scale receptive field interactions progressively
Introduces Adaptive High-Frequency Refinement Module for details
๐Ÿ”Ž Similar Papers
No similar papers found.
S
Sichen Guo
Bell Honors School, Nanjing University of Posts and Telecommunications, Nanjing, China
W
Wenjie Li
Pattern Recognition and Intelligent System Laboratory, School of Artificial Intelligence, Beijing University of Posts and Telecommunications, Beijing 100080, China
Y
Yuanyang Liu
Bell Honors School, Nanjing University of Posts and Telecommunications, Nanjing, China
Guangwei Gao
Guangwei Gao
Professor of PCALab@NJUST, IEEE/CCF/CSIG/CAAI/CAA Senior Member
Pattern RecognitionImage UnderstandingMachine Learning
J
Jian Yang
PCA Lab, Key Lab of Intelligent Perception and Systems for High-Dimensional Information of Ministry of Education, School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
C
Chia-Wen Lin
Department of Electrical Engineering and the Institute of Communications Engineering, National Tsing Hua University, Hsinchu 300044, Taiwan