MirrorLA: Reflecting Feature Map for Vision Linear Attention

📅 2026-02-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the performance gap between linear attention and standard Softmax attention, which arises because kernel feature mappings in linear attention discard critical semantic information in the negative domain due to non-negativity constraints. To overcome this limitation, the authors propose MirrorLA, a geometric framework that leverages learnable Householder reflections to actively redirect features into the non-negative quadrant, replacing the conventional passive truncation. This approach maximizes information retention while preserving strict linear complexity. By integrating multi-scale design, block-wise isometric transformations, and variance-aware modulation, MirrorLA jointly enhances local discriminability and long-range dependency modeling. Extensive experiments demonstrate that MirrorLA achieves state-of-the-art performance among linear attention methods on major vision benchmarks, confirming its ability to simultaneously attain high efficiency and representation fidelity.

Technology Category

Application Category

📝 Abstract
Linear attention significantly reduces the computational complexity of Transformers from quadratic to linear, yet it consistently lags behind softmax-based attention in performance. We identify the root cause of this degradation as the non-negativity constraint imposed on kernel feature maps: standard projections like ReLU act as"passive truncation"operators, indiscriminately discarding semantic information residing in the negative domain. We propose MirrorLA, a geometric framework that substitutes passive truncation with active reorientation. By leveraging learnable Householder reflections, MirrorLA rotates the feature geometry into the non-negative orthant to maximize information retention. Our approach restores representational density through a cohesive, multi-scale design: it first optimizes local discriminability via block-wise isometries, stabilizes long-context dynamics using variance-aware modulation to diversify activations, and finally, integrates dispersed subspaces via cross-head reflections to induce global covariance mixing. MirrorLA achieves state-of-the-art performance across standard benchmarks, demonstrating that strictly linear efficiency can be achieved without compromising representational fidelity.
Problem

Research questions and friction points this paper is trying to address.

linear attention
non-negativity constraint
feature map
semantic information loss
computational complexity
Innovation

Methods, ideas, or system contributions that make the work stand out.

Linear Attention
Householder Reflection
Feature Map Geometry
Non-negativity Constraint
Multi-scale Representation
🔎 Similar Papers
No similar papers found.
W
Weikang Meng
Harbin Institute of Technology, Shenzhen; Pengcheng Laboratory
L
Liangyu Huo
Harbin Institute of Technology, Shenzhen
Yadan Luo
Yadan Luo
ARC DECRA and Senior Lecturer, University of Queensland
Generalization3D VisionAutonomous Driving
Yaowei Wang
Yaowei Wang
The Hong Kong Polytechnic University
Y
Yingjian Li
Pengcheng Laboratory
Zheng Zhang
Zheng Zhang
HIT, SLAI
Multimodal LearningEfficient Deep LearningAI Security