Convolutional Rectangular Attention Module

📅 2025-03-13
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This paper addresses the poor generalizability, training instability, and weak interpretability of conventional spatial attention mechanisms in convolutional neural networks (CNNs), which stem from irregular, pixel-level attention regions. To this end, we propose a parametric Rectangular Spatial Attention Module (RSAM) that explicitly defines a rectangular attention region using only five learnable parameters. RSAM is fully differentiable and enables end-to-end joint optimization, serving as a plug-and-play component compatible with arbitrary CNN architectures. Our key contribution is the first explicit geometric constraint of spatial attention to a rectangle—enhancing boundary regularity, training stability, and cross-sample generalization, while improving semantic interpretability of attended locations. Extensive experiments on multiple benchmarks demonstrate that RSAM consistently outperforms pixel-wise attention methods, achieving significant gains in classification accuracy, robustness to input perturbations, and visual localization consistency.

Technology Category

Application Category

📝 Abstract
In this paper, we introduce a novel spatial attention module, that can be integrated to any convolutional network. This module guides the model to pay attention to the most discriminative part of an image. This enables the model to attain a better performance by an end-to-end training. In standard approaches, a spatial attention map is generated in a position-wise fashion. We observe that this results in very irregular boundaries. This could make it difficult to generalize to new samples. In our method, the attention region is constrained to be rectangular. This rectangle is parametrized by only 5 parameters, allowing for a better stability and generalization to new samples. In our experiments, our method systematically outperforms the position-wise counterpart. Thus, this provides us a novel useful spatial attention mechanism for convolutional models. Besides, our module also provides the interpretability concerning the ``where to look"question, as it helps to know the part of the input on which the model focuses to produce the prediction.
Problem

Research questions and friction points this paper is trying to address.

Introduces a spatial attention module for convolutional networks.
Improves model performance by focusing on discriminative image parts.
Enhances generalization with rectangular attention regions using 5 parameters.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Rectangular attention module for convolutional networks
Five-parameter rectangle for stable attention regions
Enhanced interpretability by identifying focus areas
🔎 Similar Papers
No similar papers found.
H
Hai-Vy Nguyen
Ampere Software Technology
Fabrice Gamboa
Fabrice Gamboa
IMT
Probabilités et Statistiques
S
Sixin Zhang
Institut de Recherche en Informatique de Toulouse
R
Reda Chhaibi
Laboratoire Jean Alexandre Dieudonné, Université Côte d'Azur
Serge Gratton
Serge Gratton
Toulouse INP - IRIT - Scientific Director ANITI
OptimizationData AssimilationMachine Learning
T
Thierry Giaccone
Ampere Software Technology