Transformer-based Scalable Beamforming Optimization via Deep Residual Learning

📅 2025-10-14
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the real-time optimization challenge of large-scale MU-MISO downlink beamforming under dynamic channels, this paper proposes a Transformer-based unsupervised learning-to-optimize (L2O) framework. Methodologically, it integrates residual iterative optimization, curriculum learning, semi-amortized training, and a sliding-window strategy to jointly model channel and beamforming features; it employs a multi-layer Transformer coupled with feed-forward networks, enabling offline training and millisecond-level online inference. Compared to state-of-the-art methods, the framework significantly outperforms baselines at low-to-moderate SNRs, approaches WMMSE performance at high SNRs, and accelerates inference by one to two orders of magnitude—achieving both high accuracy and real-time efficiency. This work establishes a scalable, deep-learning-driven paradigm for physical-layer communication design.

Technology Category

Application Category

📝 Abstract
We develop an unsupervised deep learning framework for downlink beamforming in large-scale MU-MISO channels. The model is trained offline, allowing real-time inference through lightweight feedforward computations in dynamic communication environments. Following the learning-to-optimize (L2O) paradigm, a multi-layer Transformer iteratively refines both channel and beamformer features via residual connections. To enhance training, three strategies are introduced: (i) curriculum learning (CL) to improve early-stage convergence and avoid local optima, (ii) semi-amortized learning to refine each Transformer block with a few gradient ascent steps, and (iii) sliding-window training to stabilize optimization by training only a subset of Transformer blocks at a time. Extensive simulations show that the proposed scheme outperforms existing baselines at low-to-medium SNRs and closely approaches WMMSE performance at high SNRs, while achieving substantially faster inference than iterative and online learning approaches.
Problem

Research questions and friction points this paper is trying to address.

Optimizing downlink beamforming in large-scale MU-MISO channels
Enabling real-time inference through offline-trained deep learning
Improving beamformer performance via Transformer-based residual learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Unsupervised deep learning framework for beamforming
Transformer refines features via residual connections
Curriculum and semi-amortized learning enhance training
Y
Yubo Zhang
Department of Electrical Engineering, Columbia University, New York, NY 10027
Xiao-Yang Liu
Xiao-Yang Liu
Columbia University
TensorDeep LearningReinforcement LearningBig Data
X
Xiaodong Wang
Department of Electrical Engineering, Columbia University, New York, NY 10027