π€ AI Summary
To address the challenges of scarce labeled data, high computational cost, and poor model generalization in ECG analysis, this paper proposes HeartBERTβthe first self-supervised pretraining framework specifically designed for physiological time-series signals. Methodologically, it adapts the RoBERTa architecture for ECG modeling by introducing ECG-specific segmented normalization and a masked signal reconstruction pretraining objective, while incorporating a bidirectional LSTM head for downstream tasks. Experimentally, HeartBERT achieves superior performance on sleep staging and heartbeat classification: using only 30% of labeled data, it surpasses fully supervised state-of-the-art methods. It reduces parameter count by 42% and accelerates inference by 3.1Γ, significantly enhancing few-shot generalization capability.
π Abstract
The HeartBert model is introduced with three primary objectives: reducing the need for labeled data, minimizing computational resources, and simultaneously improving performance in machine learning systems that analyze Electrocardiogram (ECG) signals. Inspired by Bidirectional Encoder Representations from Transformers (BERT) in natural language processing and enhanced with a self-supervised learning approach, the HeartBert model-built on the RoBERTa architecture-generates sophisticated embeddings tailored for ECG-based projects in the medical domain. To demonstrate the versatility, generalizability, and efficiency of the proposed model, two key downstream tasks have been selected: sleep stage detection and heartbeat classification. HeartBERT-based systems, utilizing bidirectional LSTM heads, are designed to address complex challenges. A series of practical experiments have been conducted to demonstrate the superiority and advancements of HeartBERT, particularly in terms of its ability to perform well with smaller training datasets, reduced learning parameters, and effective performance compared to rival models. The code and data are publicly available at https://github.com/ecgResearch/HeartBert.