Decoding Covert Speech from EEG Using a Functional Areas Spatio-Temporal Transformer

📅 2025-04-02
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses the challenge of EEG-based decoding of covert speech—hindered by poorly understood neuroarticulatory mapping mechanisms and low EEG signal-to-noise ratio. To this end, we introduce the first large-scale, synchronized EEG dataset comprising both covert and overt speech from 57 participants, covering multi-word utterances. We propose the Functional-Area Spatio-Temporal Transformer (FAST), the first model to integrate brain functional parcellation priors for guided time-frequency feature extraction, channel-adaptive weighting, and spatio-temporal Transformer encoding—enabling high-accuracy, character-level sequential decoding. Experiments reveal functionally specific activation patterns in frontal and temporal regions during covert speech, with interpretable visualizations offering novel neuroscientific evidence. FAST achieves state-of-the-art performance while maintaining model interpretability. The code is publicly released, advancing both brain–computer interface development and neural language decoding research.

Technology Category

Application Category

📝 Abstract
Covert speech involves imagining speaking without audible sound or any movements. Decoding covert speech from electroencephalogram (EEG) is challenging due to a limited understanding of neural pronunciation mapping and the low signal-to-noise ratio of the signal. In this study, we developed a large-scale multi-utterance speech EEG dataset from 57 right-handed native English-speaking subjects, each performing covert and overt speech tasks by repeating the same word in five utterances within a ten-second duration. Given the spatio-temporal nature of the neural activation process during speech pronunciation, we developed a Functional Areas Spatio-temporal Transformer (FAST), an effective framework for converting EEG signals into tokens and utilizing transformer architecture for sequence encoding. Our results reveal distinct and interpretable speech neural features by the visualization of FAST-generated activation maps across frontal and temporal brain regions with each word being covertly spoken, providing new insights into the discriminative features of the neural representation of covert speech. This is the first report of such a study, which provides interpretable evidence for speech decoding from EEG. The code for this work has been made public at https://github.com/Jiang-Muyun/FAST
Problem

Research questions and friction points this paper is trying to address.

Decoding covert speech from EEG signals
Overcoming low signal-to-noise ratio in EEG
Mapping neural activation during speech pronunciation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Functional Areas Spatio-temporal Transformer (FAST) framework
EEG signals converted into transformer-compatible tokens
Visualization of brain activation maps for interpretability
🔎 Similar Papers
No similar papers found.
Muyun Jiang
Muyun Jiang
Nanyang Technological University
Y
Yi Ding
College of Computing and Data Science, Nanyang Technological University, Singapore
W
Wei Zhang
Cognitive Neuroimaging Centre and Lee Kong Chian School of Medicine, Nanyang Technological University, Singapore
K
Kok Ann Colin Teo
Cognitive Neuroimaging Centre and Lee Kong Chian School of Medicine, Nanyang Technological University, Singapore
L
LaiGuan Fong
Cognitive Neuroimaging Centre and Lee Kong Chian School of Medicine, Nanyang Technological University, Singapore
Shuailei Zhang
Shuailei Zhang
Research fellow, College of Computing and Data Science, Nanyang Technological University
Brain computer interfaceNeurorehabilitation
Z
Zhiwei Guo
College of Computing and Data Science, Nanyang Technological University, Singapore
C
Chenyu Liu
College of Computing and Data Science, Nanyang Technological University, Singapore
R
Raghavan Bhuvanakantham
Cognitive Neuroimaging Centre and Lee Kong Chian School of Medicine, Nanyang Technological University, Singapore
W
Wei Khang Jeremy Sim
Cognitive Neuroimaging Centre and Lee Kong Chian School of Medicine, Nanyang Technological University, Singapore
C
Chuan Huat Vince Foo
DSO National Laboratories, Singapore
R
Rong Hui Jonathan Chua
DSO National Laboratories, Singapore
P
Parasuraman Padmanabhan
Cognitive Neuroimaging Centre and Lee Kong Chian School of Medicine, Nanyang Technological University, Singapore
V
Victoria Leong
Division of Psychology, Nanyang Technological University, Singapore, and the Department of Pediatrics, University of Cambridge, United Kingdom
Jia Lu
Jia Lu
Professor of Journalism and Communication, Tsinghua University
New ICTs and Social Change
Balazs Gulyas
Balazs Gulyas
Lee Kong Chian School of Medicine, Nanyang Technological University, Singapore; Karolinska Institute
Cuntai Guan
Cuntai Guan
President's Chair Professor, CCDS, Nanyang Technological University
Brain-Computer InterfaceBrain-Computer InterfacesMachine LearningArtificial Intelligence