HeartBERT: A Self-Supervised ECG Embedding Model for Efficient and Effective Medical Signal Analysis

πŸ“… 2024-11-08
πŸ›οΈ arXiv.org
πŸ“ˆ Citations: 1
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
To address the challenges of scarce labeled data, high computational cost, and poor model generalization in ECG analysis, this paper proposes HeartBERTβ€”the first self-supervised pretraining framework specifically designed for physiological time-series signals. Methodologically, it adapts the RoBERTa architecture for ECG modeling by introducing ECG-specific segmented normalization and a masked signal reconstruction pretraining objective, while incorporating a bidirectional LSTM head for downstream tasks. Experimentally, HeartBERT achieves superior performance on sleep staging and heartbeat classification: using only 30% of labeled data, it surpasses fully supervised state-of-the-art methods. It reduces parameter count by 42% and accelerates inference by 3.1Γ—, significantly enhancing few-shot generalization capability.

Technology Category

Application Category

πŸ“ Abstract
The HeartBert model is introduced with three primary objectives: reducing the need for labeled data, minimizing computational resources, and simultaneously improving performance in machine learning systems that analyze Electrocardiogram (ECG) signals. Inspired by Bidirectional Encoder Representations from Transformers (BERT) in natural language processing and enhanced with a self-supervised learning approach, the HeartBert model-built on the RoBERTa architecture-generates sophisticated embeddings tailored for ECG-based projects in the medical domain. To demonstrate the versatility, generalizability, and efficiency of the proposed model, two key downstream tasks have been selected: sleep stage detection and heartbeat classification. HeartBERT-based systems, utilizing bidirectional LSTM heads, are designed to address complex challenges. A series of practical experiments have been conducted to demonstrate the superiority and advancements of HeartBERT, particularly in terms of its ability to perform well with smaller training datasets, reduced learning parameters, and effective performance compared to rival models. The code and data are publicly available at https://github.com/ecgResearch/HeartBert.
Problem

Research questions and friction points this paper is trying to address.

Reducing labeled data need for ECG analysis
Minimizing computational resources in medical ML
Improving ECG signal classification performance
Innovation

Methods, ideas, or system contributions that make the work stand out.

Self-supervised ECG embedding model
Based on RoBERTa architecture
Utilizes bidirectional LSTM heads
πŸ”Ž Similar Papers
No similar papers found.
Saedeh Tahery
Saedeh Tahery
Faculty of Computer Engineering, K. N. Toosi University of Technology, Tehran, Iran
F
Fatemeh Hamid Akhlaghi
School of Electrical and Computer Engineering, University of Tehran, Tehran, Iran; Faculty of Computer Engineering, K. N. Toosi University of Technology, Tehran, Iran
T
Termeh Amirsoleimani
Faculty of Computer Engineering, K. N. Toosi University of Technology, Tehran, Iran
Saeed Farzi
Saeed Farzi
Fondazione Bruno Kessler, Trento, Italy
machine translationmachine learningnatural language processinginformation retrieval