Toward Interpretable Sleep Stage Classification Using Cross-Modal Transformers

📅 2022-08-15
🏛️ IEEE transactions on neural systems and rehabilitation engineering
📈 Citations: 26
Influential: 2
📄 PDF
🤖 AI Summary
To address the limited interpretability and clinical applicability of deep learning models in sleep staging, this paper proposes an interpretable cross-modal Transformer architecture. It is the first to introduce a cross-modal Transformer to sleep staging, integrating multi-scale 1D convolutions with self-attention mechanisms for end-to-end automatic feature learning and stage classification from multi-channel physiological signals (e.g., EEG, EOG, EMG). The model provides traceable decision rationales via attention weights and supports visual attribution analysis. Experiments demonstrate performance on par with state-of-the-art methods while reducing parameter count by 40% and training time by 35%. The proposed approach thus achieves a favorable trade-off among high accuracy, computational efficiency, and clinical interpretability. Code and an interactive demo are publicly available.
📝 Abstract
Accurate sleep stage classification is significant for sleep health assessment. In recent years, several machine-learning based sleep staging algorithms have been developed, and in particular, deep-learning based algorithms have achieved performance on par with human annotation. Despite improved performance, a limitation of most deep-learning based algorithms is their black-box behavior, which have limited their use in clinical settings. Here, we propose a cross-modal transformer, which is a transformer-based method for sleep stage classification. The proposed cross-modal transformer consists of a cross-modal transformer encoder architecture along with a multi-scale one-dimensional convolutional neural network for automatic representation learning. The performance of our method is on-par with the state-of-the-art methods and eliminates the black-box behavior of deep-learning models by utilizing the interpretability aspect of the attention modules. Furthermore, our method provides considerable reductions in the number of parameters and training time compared to the state-of-the-art methods. Our code is available at https://github.com/Jathurshan0330/Cross-Modal-Transformer. A demo of our work can be found at https://bit.ly/Cross_modal_transformer_demo.
Problem

Research questions and friction points this paper is trying to address.

Improving interpretability in sleep stage classification models
Overcoming black-box limitations of deep-learning algorithms
Reducing parameters and training time for efficiency
Innovation

Methods, ideas, or system contributions that make the work stand out.

Cross-modal transformer for sleep staging
Multi-scale 1D CNN for representation learning
Interpretable attention modules reduce parameters
🔎 Similar Papers
No similar papers found.
Jathurshan Pradeepkumar
Jathurshan Pradeepkumar
PhD Student at University of Illinois Urbana-Champaign
AI for HealthcareDeep LearningEEG AnalysisBiosignal Processing
M
M. Anandakumar
Department of Electronic and Telecommunication Engineering, University of Moratuwa, Sri Lanka and currently affiliated to Faculty of Arts and Sciences at Harvard University
Vinith Kugathasan
Vinith Kugathasan
Department of Electronic and Telecommunication Engineering, University of Moratuwa, Sri Lanka
D
Dhinesh Suntharalingham
Department of Electronic and Telecommunication Engineering, University of Moratuwa, Sri Lanka
S
S. L. Kappel
Department of Electrical and Computer Engineering, Aarhus University, DK-8200 Aarhus, Denmark
A
A. D. Silva
Department of Electronic and Telecommunication Engineering, University of Moratuwa, Sri Lanka
Chamira U. S. Edussooriya
Chamira U. S. Edussooriya
University of Moratuwa
Multi-dimensional Signal ProcessingDigital FiltersLight FieldsGraph Signal ProcessingApplications of Machine Learning