🤖 AI Summary
This work addresses the challenge of high-fidelity semantic decoding of continuous speech from intracranial electroencephalography (iEEG). We propose a two-stage transfer learning framework: first, a lightweight LSTM adapter aligns iEEG features to the semantic embedding space of a pre-trained language model (e.g., BERT); second, an end-to-end correction module generates fluent, unconstrained continuous text. Trained on only 30 minutes of neural data, our method significantly outperforms existing state-of-the-art approaches in low-resource settings, achieving high-accuracy semantic reconstruction. Its core innovations lie in (1) a novel semantic-space alignment mechanism that bridges neural dynamics and linguistic representations, and (2) a task-adapted, parameter-efficient architecture. The framework enables scalable, clinically viable real-time neural decoding—advancing practical brain–computer interfaces for speech restoration.
📝 Abstract
Decoding continuous language from neural signals remains a significant challenge in the intersection of neuroscience and artificial intelligence. We introduce Neuro2Semantic, a novel framework that reconstructs the semantic content of perceived speech from intracranial EEG (iEEG) recordings. Our approach consists of two phases: first, an LSTM-based adapter aligns neural signals with pre-trained text embeddings; second, a corrector module generates continuous, natural text directly from these aligned embeddings. This flexible method overcomes the limitations of previous decoding approaches and enables unconstrained text generation. Neuro2Semantic achieves strong performance with as little as 30 minutes of neural data, outperforming a recent state-of-the-art method in low-data settings. These results highlight the potential for practical applications in brain-computer interfaces and neural decoding technologies.