🤖 AI Summary
To address the challenge of decoding EEG signals across heterogeneous EEG headsets—differing in electrode count and spatial layout—this paper proposes a robust, target-label-free cross-device transfer method. The method introduces spatial knowledge distillation into cross-device transfer for the first time and establishes a three-level joint distribution alignment framework—spanning input, feature, and output spaces—integrated with multi-layer Maximum Mean Discrepancy (MMD) and Wasserstein distance metrics, a deep EEG encoder, and a cross-paradigm domain adaptation training strategy. Extensive evaluation is conducted across six public EEG datasets and two major BCI paradigms: motor imagery (MI) and event-related potentials (ERP). The approach achieves state-of-the-art performance in both offline unsupervised and online supervised settings, significantly outperforming ten baseline methods.
📝 Abstract
A non-invasive brain-computer interface (BCI) enables direct interaction between the user and external devices, typically via electroencephalogram (EEG) signals. However, decoding EEG signals across different headsets remains a significant challenge due to differences in the number and locations of the electrodes. To address this challenge, we propose a spatial distillation based distribution alignment (SDDA) approach for heterogeneous cross-headset transfer in non-invasive BCIs. SDDA uses first spatial distillation to make use of the full set of electrodes, and then input/feature/output space distribution alignments to cope with the significant differences between the source and target domains. To our knowledge, this is the first work to use knowledge distillation in cross-headset transfers. Extensive experiments on six EEG datasets from two BCI paradigms demonstrated that SDDA achieved superior performance in both offline unsupervised domain adaptation and online supervised domain adaptation scenarios, consistently outperforming 10 classical and state-of-the-art transfer learning algorithms.