TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Do Not Sleep on Traditional Machine Learning: Simple and I...

Do Not Sleep on Traditional Machine Learning: Simple and Interpretable Techniques Are Competitive to Deep Learning for Sleep Scoring

Jeroen Van Der Donckt, Jonas Van Der Donckt, Emiel Deprost, Nicolas Vandenbussche, Michael Rademaker, Gilles Vandewiele, Sofie Van Hoecke

2022-07-15Sleep StagingSleep Stage DetectionAutomatic Sleep Stage ClassificationMultimodal Sleep Stage DetectionDeep LearningBIG-bench Machine Learning
PaperPDFCode(official)

Abstract

Over the last few years, research in automatic sleep scoring has mainly focused on developing increasingly complex deep learning architectures. However, recently these approaches achieved only marginal improvements, often at the expense of requiring more data and more expensive training procedures. Despite all these efforts and their satisfactory performance, automatic sleep staging solutions are not widely adopted in a clinical context yet. We argue that most deep learning solutions for sleep scoring are limited in their real-world applicability as they are hard to train, deploy, and reproduce. Moreover, these solutions lack interpretability and transparency, which are often key to increase adoption rates. In this work, we revisit the problem of sleep stage classification using classical machine learning. Results show that competitive performance can be achieved with a conventional machine learning pipeline consisting of preprocessing, feature extraction, and a simple machine learning model. In particular, we analyze the performance of a linear model and a non-linear (gradient boosting) model. Our approach surpasses state-of-the-art (that uses the same data) on two public datasets: Sleep-EDF SC-20 (MF1 0.810) and Sleep-EDF ST (MF1 0.795), while achieving competitive results on Sleep-EDF SC-78 (MF1 0.775) and MASS SS3 (MF1 0.817). We show that, for the sleep stage scoring task, the expressiveness of an engineered feature vector is on par with the internally learned representations of deep learning models. This observation opens the door to clinical adoption, as a representative feature vector allows to leverage both the interpretability and successful track record of traditional machine learning models.

Results

TaskDatasetMetricValueModel
Sleep QualitySleep-EDFCohen's kappa0.816CatBoost
Sleep QualitySleep-EDFMacro-F10.81CatBoost
Sleep QualitySleep-EDFCohen's kappa0.813Linear model
Sleep QualitySleep-EDFMacro-F10.805Linear model
Sleep QualityMASS SS3Cohen's kappa0.803CatBoost
Sleep QualityMASS SS3Macro-F10.817CatBoost
Sleep QualitySleep-EDF-SCCohen's kappa0.812CatBoost
Sleep QualitySleep-EDF-SCMacro-F10.802CatBoost
Sleep QualitySleep-EDF-SCCohen's kappa0.806Linear model
Sleep QualitySleep-EDF-SCMacro-F10.809Linear model
Sleep QualitySleep-EDF-ST Macro-F10.795CatBoost
Sleep QualitySleep-EDF-STCohen's kappa0.765CatBoost
Sleep QualitySleep-EDF-ST Macro-F10.792Linear model
Sleep QualitySleep-EDF-STCohen's kappa0.759Linear model
Sleep Stage DetectionSleep-EDFCohen's kappa0.816CatBoost
Sleep Stage DetectionSleep-EDFMacro-F10.81CatBoost
Sleep Stage DetectionSleep-EDFCohen's kappa0.813Linear model
Sleep Stage DetectionSleep-EDFMacro-F10.805Linear model
Sleep Stage DetectionMASS SS3Cohen's kappa0.803CatBoost
Sleep Stage DetectionMASS SS3Macro-F10.817CatBoost

Related Papers

Automatic Classification and Segmentation of Tunnel Cracks Based on Deep Learning and Visual Explanations2025-07-18A Survey of Deep Learning for Geometry Problem Solving2025-07-16Uncertainty Quantification for Motor Imagery BCI -- Machine Learning vs. Deep Learning2025-07-10eegFloss: A Python package for refining sleep EEG recordings using machine learning models2025-07-08Chat-Ghosting: A Comparative Study of Methods for Auto-Completion in Dialog Systems2025-07-08Deep Learning Optimization of Two-State Pinching Antennas Systems2025-07-08AXLearn: Modular Large Model Training on Heterogeneous Infrastructure2025-07-07Determination Of Structural Cracks Using Deep Learning Frameworks2025-07-03