BrainRVQ: A High-Fidelity EEG Foundation Model via Dual-Domain Residual Quantization and Hierarchical Autoregression

📅 2026-02-18
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the limitations of existing EEG foundation models, which struggle to effectively model the low signal-to-noise ratio and complex time-frequency non-stationarity of neural signals while neglecting their intrinsic hierarchical structure, thereby constraining fine-grained reconstruction performance. To overcome these challenges, the authors propose BrainRVQ, a general-purpose foundation model pretrained on large-scale clinical EEG data. Its key innovations include a dual-domain residual vector quantization (DD-RVQ) tokenizer that disentangles time-domain waveforms and spectral patterns into hierarchical discrete codes, coupled with a self-supervised pretraining strategy featuring importance-guided curriculum masking and coarse-to-fine hierarchical autoregressive objectives. Evaluated across eight diverse downstream tasks, BrainRVQ substantially outperforms current state-of-the-art methods, demonstrating its superior capability in learning robust and generalizable neural representations.

Technology Category

Application Category

📝 Abstract
Developing foundation models for electroencephalography (EEG) remains challenging due to the signal's low signal-to-noise ratio and complex spectro-temporal non-stationarity. Existing approaches often overlook the hierarchical latent structure inherent in neural dynamics, leading to suboptimal reconstruction of fine-grained information. In this work, we propose BrainRVQ, a general-purpose EEG foundation model pre-trained on a large-scale corpus of clinical EEG data. Unlike standard masked modeling, BrainRVQ features a Dual-Domain Residual Vector Quantization (DD-RVQ) tokenizer that disentangles temporal waveforms and spectral patterns into hierarchical discrete codes. We further introduce a hierarchical autoregressive pre-training objective that learns to reconstruct these codes in a coarse-to-fine manner, utilizing an importance-guided curriculum masking strategy to prioritize information-rich neural events over background noise. Extensive experiments across 8 diverse downstream datasets demonstrate that BrainRVQ consistently outperforms state-of-the-art baselines, validating its effectiveness in learning robust and generalizable neural representations. Our code and model weights are available:https://github.com/keqicmz/BrainRVQ
Problem

Research questions and friction points this paper is trying to address.

EEG foundation model
low signal-to-noise ratio
spectro-temporal non-stationarity
hierarchical latent structure
fine-grained reconstruction
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dual-Domain Residual Vector Quantization
Hierarchical Autoregression
EEG Foundation Model
Coarse-to-Fine Reconstruction
Curriculum Masking
🔎 Similar Papers
No similar papers found.
M
Mingzhe Cui
State Key Laboratory of Industrial Control Technology, Zhejiang University, Hangzhou, China
Tao Chen
Tao Chen
Zhejiang University
Natural Language Processing
Y
Yang Jiao
Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
Y
Yiqin Wang
State Key Laboratory of Industrial Control Technology, Zhejiang University, Hangzhou, China
Lei Xie
Lei Xie
Zhejiang University
Statistical Process MonitoringControl Performance Assessment
Yi Pan
Yi Pan
Shenzhen University of Advanced Technology, China
BioinformaticsBig DataMachine LearningCloud ComputingWireless Networks
Luca Mainardi
Luca Mainardi
Full Professor, Politecnico di Milano
Biomedical signal processingHeart rate VariabilityCardiac repolarizationImage processingRadiomics