🤖 AI Summary
To address the high computational overhead and poor real-time performance of conventional Transformers when processing high-dimensional, non-stationary time-series data—including KPIs and IQ samples—in 6G Open Radio Access Networks (O-RAN), this paper proposes the Reservoir-Enhanced Masked Autoencoder Transformer (RES-MAE). RES-MAE integrates an Echo State Network (ESN) to enable linear-complexity dynamic temporal chunking and projection, thereby avoiding the quadratic complexity of self-attention. It employs a masked autoencoding objective to reconstruct 30% of masked segments, enabling efficient, backpropagation-free pretraining. Downstream tasks are accomplished by freezing the backbone and fine-tuning only a lightweight head. Evaluated on multiple O-RAN KPI forecasting benchmarks, RES-MAE achieves mean squared error below 0.06, significantly improving inference latency and energy efficiency. This work establishes a low-latency, highly adaptable foundation model paradigm for 6G networks.
📝 Abstract
Next-generation open radio access networks (O-RAN) continuously stream tens of key performance indicators (KPIs) together with raw in-phase/quadrature (IQ) samples, yielding ultra-high-dimensional, non-stationary time series that overwhelm conventional transformer architectures. We introduce a reservoir-augmented masked autoencoding transformer (RA-MAT). This time series foundation model employs echo state network (ESN) computing with masked autoencoding to satisfy the stringent latency, energy efficiency, and scalability requirements of 6G O-RAN testing. A fixed, randomly initialized ESN rapidly projects each temporal patch into a rich dynamical embedding without backpropagation through time, converting the quadratic self-attention bottleneck into a lightweight linear operation. These embeddings drive a patch-wise masked autoencoder that reconstructs 30% randomly masked patches, compelling the encoder to capture both local dynamics and long-range structure from unlabeled data. After self-supervised pre-training, RA-MAT is fine-tuned with a shallow task head while keeping the reservoir and most transformer layers frozen, enabling low-footprint adaptation to diverse downstream tasks such as O-RAN KPI forecasting. In a comprehensive O-RAN KPI case study, RA-MAT achieved sub-0.06 mean squared error (MSE) on several continuous and discrete KPIs. This work positions RA-MAT as a practical pathway toward real-time, foundation-level analytics in future 6G networks.