🤖 AI Summary
To address the challenges of population-level neural representation learning and generalization—stemming from sparse, irregular, cross-subject and cross-device neural data with inconsistent electrode distributions—this paper introduces Population Transformer (PopT), the first self-supervised representation learning framework specifically designed for neural populations. PopT integrates spatiotemporal self-supervised pretraining, dynamic multi-channel sparse feature aggregation, and cross-modal time-series embedding alignment to construct transferable, population-level neural representations. Its key contributions include: (1) zero-shot cross-subject decoding capability; (2) substantial reduction in downstream annotation requirements (average 62% fewer labeled samples) and improved decoding accuracy (+8.3%); and (3) lightweight architecture, high interpretability, and strong generalization across subjects and devices. The code and pretrained models are publicly released to advance the practical deployment of invasive brain–computer interfaces.
📝 Abstract
We present a self-supervised framework that learns population-level codes for arbitrary ensembles of neural recordings at scal. We address two key challenges in scaling models with neural time-series data: sparse and variable electrode distribution across subjects and datasets. The Population Transformer (PopT) stacks on top of pretrained representations and enhances downstream decoding by enabling learned aggregation of multiple spatially-sparse data channels. The pretrained PopT lowers the amount of data required for downstream decoding experiments, while increasing accuracy, even on held-out subjects and tasks. Compared to end-to-end methods, this approach is computationally lightweight and more interpretable, while still retaining competitive performance. We further show how our framework is generalizable to multiple time-series embeddings and neural data modalities. Beyond decoding, we interpret the pretrained PopT and fine-tuned models to show how they can be used to extract neuroscience insights from massive amounts of data. We release our code as well as a pretrained PopT to enable off-the-shelf improvements in multi-channel intracranial data decoding and interpretability.