AI-Driven sleep staging from actigraphy and heart rate.

Sleep is an important indicator of a person's health, and its accurate and cost-effective quantification is of great value in healthcare. The gold standard for sleep assessment and the clinical diagnosis of sleep disorders is polysomnography (PSG). However, PSG requires an overnight clinic visi...

Full description

Bibliographic Details
Main Authors: Tzu-An Song, Samadrita Roy Chowdhury, Masoud Malekzadeh, Stephanie Harrison, Terri Blackwell Hoge, Susan Redline, Katie L Stone, Richa Saxena, Shaun M Purcell, Joyita Dutta
Format: Article
Language:English
Published: Public Library of Science (PLoS) 2023-01-01
Series:PLoS ONE
Online Access:https://doi.org/10.1371/journal.pone.0285703
_version_ 1827345907954221056
author Tzu-An Song
Samadrita Roy Chowdhury
Masoud Malekzadeh
Stephanie Harrison
Terri Blackwell Hoge
Susan Redline
Katie L Stone
Richa Saxena
Shaun M Purcell
Joyita Dutta
author_facet Tzu-An Song
Samadrita Roy Chowdhury
Masoud Malekzadeh
Stephanie Harrison
Terri Blackwell Hoge
Susan Redline
Katie L Stone
Richa Saxena
Shaun M Purcell
Joyita Dutta
author_sort Tzu-An Song
collection DOAJ
description Sleep is an important indicator of a person's health, and its accurate and cost-effective quantification is of great value in healthcare. The gold standard for sleep assessment and the clinical diagnosis of sleep disorders is polysomnography (PSG). However, PSG requires an overnight clinic visit and trained technicians to score the obtained multimodality data. Wrist-worn consumer devices, such as smartwatches, are a promising alternative to PSG because of their small form factor, continuous monitoring capability, and popularity. Unlike PSG, however, wearables-derived data are noisier and far less information-rich because of the fewer number of modalities and less accurate measurements due to their small form factor. Given these challenges, most consumer devices perform two-stage (i.e., sleep-wake) classification, which is inadequate for deep insights into a person's sleep health. The challenging multi-class (three, four, or five-class) staging of sleep using data from wrist-worn wearables remains unresolved. The difference in the data quality between consumer-grade wearables and lab-grade clinical equipment is the motivation behind this study. In this paper, we present an artificial intelligence (AI) technique termed sequence-to-sequence LSTM for automated mobile sleep staging (SLAMSS), which can perform three-class (wake, NREM, REM) and four-class (wake, light, deep, REM) sleep classification from activity (i.e., wrist-accelerometry-derived locomotion) and two coarse heart rate measures-both of which can be reliably obtained from a consumer-grade wrist-wearable device. Our method relies on raw time-series datasets and obviates the need for manual feature selection. We validated our model using actigraphy and coarse heart rate data from two independent study populations: the Multi-Ethnic Study of Atherosclerosis (MESA; N = 808) cohort and the Osteoporotic Fractures in Men (MrOS; N = 817) cohort. SLAMSS achieves an overall accuracy of 79%, weighted F1 score of 0.80, 77% sensitivity, and 89% specificity for three-class sleep staging and an overall accuracy of 70-72%, weighted F1 score of 0.72-0.73, 64-66% sensitivity, and 89-90% specificity for four-class sleep staging in the MESA cohort. It yielded an overall accuracy of 77%, weighted F1 score of 0.77, 74% sensitivity, and 88% specificity for three-class sleep staging and an overall accuracy of 68-69%, weighted F1 score of 0.68-0.69, 60-63% sensitivity, and 88-89% specificity for four-class sleep staging in the MrOS cohort. These results were achieved with feature-poor inputs with a low temporal resolution. In addition, we extended our three-class staging model to an unrelated Apple Watch dataset. Importantly, SLAMSS predicts the duration of each sleep stage with high accuracy. This is especially significant for four-class sleep staging, where deep sleep is severely underrepresented. We show that, by appropriately choosing the loss function to address the inherent class imbalance, our method can accurately estimate deep sleep time (SLAMSS/MESA: 0.61±0.69 hours, PSG/MESA ground truth: 0.60±0.60 hours; SLAMSS/MrOS: 0.53±0.66 hours, PSG/MrOS ground truth: 0.55±0.57 hours;). Deep sleep quality and quantity are vital metrics and early indicators for a number of diseases. Our method, which enables accurate deep sleep estimation from wearables-derived data, is therefore promising for a variety of clinical applications requiring long-term deep sleep monitoring.
first_indexed 2024-03-07T23:20:32Z
format Article
id doaj.art-0a2d9d15799f4749a35f8ab22daaa9c8
institution Directory Open Access Journal
issn 1932-6203
language English
last_indexed 2024-03-07T23:20:32Z
publishDate 2023-01-01
publisher Public Library of Science (PLoS)
record_format Article
series PLoS ONE
spelling doaj.art-0a2d9d15799f4749a35f8ab22daaa9c82024-02-21T05:31:55ZengPublic Library of Science (PLoS)PLoS ONE1932-62032023-01-01185e028570310.1371/journal.pone.0285703AI-Driven sleep staging from actigraphy and heart rate.Tzu-An SongSamadrita Roy ChowdhuryMasoud MalekzadehStephanie HarrisonTerri Blackwell HogeSusan RedlineKatie L StoneRicha SaxenaShaun M PurcellJoyita DuttaSleep is an important indicator of a person's health, and its accurate and cost-effective quantification is of great value in healthcare. The gold standard for sleep assessment and the clinical diagnosis of sleep disorders is polysomnography (PSG). However, PSG requires an overnight clinic visit and trained technicians to score the obtained multimodality data. Wrist-worn consumer devices, such as smartwatches, are a promising alternative to PSG because of their small form factor, continuous monitoring capability, and popularity. Unlike PSG, however, wearables-derived data are noisier and far less information-rich because of the fewer number of modalities and less accurate measurements due to their small form factor. Given these challenges, most consumer devices perform two-stage (i.e., sleep-wake) classification, which is inadequate for deep insights into a person's sleep health. The challenging multi-class (three, four, or five-class) staging of sleep using data from wrist-worn wearables remains unresolved. The difference in the data quality between consumer-grade wearables and lab-grade clinical equipment is the motivation behind this study. In this paper, we present an artificial intelligence (AI) technique termed sequence-to-sequence LSTM for automated mobile sleep staging (SLAMSS), which can perform three-class (wake, NREM, REM) and four-class (wake, light, deep, REM) sleep classification from activity (i.e., wrist-accelerometry-derived locomotion) and two coarse heart rate measures-both of which can be reliably obtained from a consumer-grade wrist-wearable device. Our method relies on raw time-series datasets and obviates the need for manual feature selection. We validated our model using actigraphy and coarse heart rate data from two independent study populations: the Multi-Ethnic Study of Atherosclerosis (MESA; N = 808) cohort and the Osteoporotic Fractures in Men (MrOS; N = 817) cohort. SLAMSS achieves an overall accuracy of 79%, weighted F1 score of 0.80, 77% sensitivity, and 89% specificity for three-class sleep staging and an overall accuracy of 70-72%, weighted F1 score of 0.72-0.73, 64-66% sensitivity, and 89-90% specificity for four-class sleep staging in the MESA cohort. It yielded an overall accuracy of 77%, weighted F1 score of 0.77, 74% sensitivity, and 88% specificity for three-class sleep staging and an overall accuracy of 68-69%, weighted F1 score of 0.68-0.69, 60-63% sensitivity, and 88-89% specificity for four-class sleep staging in the MrOS cohort. These results were achieved with feature-poor inputs with a low temporal resolution. In addition, we extended our three-class staging model to an unrelated Apple Watch dataset. Importantly, SLAMSS predicts the duration of each sleep stage with high accuracy. This is especially significant for four-class sleep staging, where deep sleep is severely underrepresented. We show that, by appropriately choosing the loss function to address the inherent class imbalance, our method can accurately estimate deep sleep time (SLAMSS/MESA: 0.61±0.69 hours, PSG/MESA ground truth: 0.60±0.60 hours; SLAMSS/MrOS: 0.53±0.66 hours, PSG/MrOS ground truth: 0.55±0.57 hours;). Deep sleep quality and quantity are vital metrics and early indicators for a number of diseases. Our method, which enables accurate deep sleep estimation from wearables-derived data, is therefore promising for a variety of clinical applications requiring long-term deep sleep monitoring.https://doi.org/10.1371/journal.pone.0285703
spellingShingle Tzu-An Song
Samadrita Roy Chowdhury
Masoud Malekzadeh
Stephanie Harrison
Terri Blackwell Hoge
Susan Redline
Katie L Stone
Richa Saxena
Shaun M Purcell
Joyita Dutta
AI-Driven sleep staging from actigraphy and heart rate.
PLoS ONE
title AI-Driven sleep staging from actigraphy and heart rate.
title_full AI-Driven sleep staging from actigraphy and heart rate.
title_fullStr AI-Driven sleep staging from actigraphy and heart rate.
title_full_unstemmed AI-Driven sleep staging from actigraphy and heart rate.
title_short AI-Driven sleep staging from actigraphy and heart rate.
title_sort ai driven sleep staging from actigraphy and heart rate
url https://doi.org/10.1371/journal.pone.0285703
work_keys_str_mv AT tzuansong aidrivensleepstagingfromactigraphyandheartrate
AT samadritaroychowdhury aidrivensleepstagingfromactigraphyandheartrate
AT masoudmalekzadeh aidrivensleepstagingfromactigraphyandheartrate
AT stephanieharrison aidrivensleepstagingfromactigraphyandheartrate
AT terriblackwellhoge aidrivensleepstagingfromactigraphyandheartrate
AT susanredline aidrivensleepstagingfromactigraphyandheartrate
AT katielstone aidrivensleepstagingfromactigraphyandheartrate
AT richasaxena aidrivensleepstagingfromactigraphyandheartrate
AT shaunmpurcell aidrivensleepstagingfromactigraphyandheartrate
AT joyitadutta aidrivensleepstagingfromactigraphyandheartrate