Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines

BackgroundIntensive care unit (ICU) readmissions are associated with mortality and poor outcomes. To improve discharge decisions, machine learning (ML) could help to identify patients at risk of ICU readmission. However, as many models are black boxes, dangerous properties may remain unnoticed. Wide...

Full description

Bibliographic Details
Main Authors: Stefan Hegselmann, Christian Ertmer, Thomas Volkert, Antje Gottschalk, Martin Dugas, Julian Varghese
Format: Article
Language:English
Published: Frontiers Media S.A. 2022-08-01
Series:Frontiers in Medicine
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/fmed.2022.960296/full
_version_ 1811339482320863232
author Stefan Hegselmann
Christian Ertmer
Thomas Volkert
Antje Gottschalk
Martin Dugas
Julian Varghese
author_facet Stefan Hegselmann
Christian Ertmer
Thomas Volkert
Antje Gottschalk
Martin Dugas
Julian Varghese
author_sort Stefan Hegselmann
collection DOAJ
description BackgroundIntensive care unit (ICU) readmissions are associated with mortality and poor outcomes. To improve discharge decisions, machine learning (ML) could help to identify patients at risk of ICU readmission. However, as many models are black boxes, dangerous properties may remain unnoticed. Widely used post hoc explanation methods also have inherent limitations. Few studies are evaluating inherently interpretable ML models for health care and involve clinicians in inspecting the trained model.MethodsAn inherently interpretable model for the prediction of 3 day ICU readmission was developed. We used explainable boosting machines that learn modular risk functions and which have already been shown to be suitable for the health care domain. We created a retrospective cohort of 15,589 ICU stays and 169 variables collected between 2006 and 2019 from the University Hospital Münster. A team of physicians inspected the model, checked the plausibility of each risk function, and removed problematic ones. We collected qualitative feedback during this process and analyzed the reasons for removing risk functions. The performance of the final explainable boosting machine was compared with a validated clinical score and three commonly used ML models. External validation was performed on the widely used Medical Information Mart for Intensive Care version IV database.ResultsThe developed explainable boosting machine used 67 features and showed an area under the precision-recall curve of 0.119 ± 0.020 and an area under the receiver operating characteristic curve of 0.680 ± 0.025. It performed on par with state-of-the-art gradient boosting machines (0.123 ± 0.016, 0.665 ± 0.036) and outperformed the Simplified Acute Physiology Score II (0.084 ± 0.025, 0.607 ± 0.019), logistic regression (0.092 ± 0.026, 0.587 ± 0.016), and recurrent neural networks (0.095 ± 0.008, 0.594 ± 0.027). External validation confirmed that explainable boosting machines (0.221 ± 0.023, 0.760 ± 0.010) performed similarly to gradient boosting machines (0.232 ± 0.029, 0.772 ± 0.018). Evaluation of the model inspection showed that explainable boosting machines can be useful to detect and remove problematic risk functions.ConclusionsWe developed an inherently interpretable ML model for 3 day ICU readmission prediction that reached the state-of-the-art performance of black box models. Our results suggest that for low- to medium-dimensional datasets that are common in health care, it is feasible to develop ML models that allow a high level of human control without sacrificing performance.
first_indexed 2024-04-13T18:27:21Z
format Article
id doaj.art-42d673fc214843d0a8d972bd450510d3
institution Directory Open Access Journal
issn 2296-858X
language English
last_indexed 2024-04-13T18:27:21Z
publishDate 2022-08-01
publisher Frontiers Media S.A.
record_format Article
series Frontiers in Medicine
spelling doaj.art-42d673fc214843d0a8d972bd450510d32022-12-22T02:35:12ZengFrontiers Media S.A.Frontiers in Medicine2296-858X2022-08-01910.3389/fmed.2022.960296960296Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machinesStefan Hegselmann0Christian Ertmer1Thomas Volkert2Antje Gottschalk3Martin Dugas4Julian Varghese5Institute of Medical Informatics, University of Münster, Münster, GermanyDepartment of Anesthesiology, Intensive Care and Pain Medicine, University Hospital Münster, Münster, GermanyDepartment of Anesthesiology, Intensive Care and Pain Medicine, University Hospital Münster, Münster, GermanyDepartment of Anesthesiology, Intensive Care and Pain Medicine, University Hospital Münster, Münster, GermanyInstitute of Medical Informatics, Heidelberg University Hospital, Heidelberg, GermanyInstitute of Medical Informatics, University of Münster, Münster, GermanyBackgroundIntensive care unit (ICU) readmissions are associated with mortality and poor outcomes. To improve discharge decisions, machine learning (ML) could help to identify patients at risk of ICU readmission. However, as many models are black boxes, dangerous properties may remain unnoticed. Widely used post hoc explanation methods also have inherent limitations. Few studies are evaluating inherently interpretable ML models for health care and involve clinicians in inspecting the trained model.MethodsAn inherently interpretable model for the prediction of 3 day ICU readmission was developed. We used explainable boosting machines that learn modular risk functions and which have already been shown to be suitable for the health care domain. We created a retrospective cohort of 15,589 ICU stays and 169 variables collected between 2006 and 2019 from the University Hospital Münster. A team of physicians inspected the model, checked the plausibility of each risk function, and removed problematic ones. We collected qualitative feedback during this process and analyzed the reasons for removing risk functions. The performance of the final explainable boosting machine was compared with a validated clinical score and three commonly used ML models. External validation was performed on the widely used Medical Information Mart for Intensive Care version IV database.ResultsThe developed explainable boosting machine used 67 features and showed an area under the precision-recall curve of 0.119 ± 0.020 and an area under the receiver operating characteristic curve of 0.680 ± 0.025. It performed on par with state-of-the-art gradient boosting machines (0.123 ± 0.016, 0.665 ± 0.036) and outperformed the Simplified Acute Physiology Score II (0.084 ± 0.025, 0.607 ± 0.019), logistic regression (0.092 ± 0.026, 0.587 ± 0.016), and recurrent neural networks (0.095 ± 0.008, 0.594 ± 0.027). External validation confirmed that explainable boosting machines (0.221 ± 0.023, 0.760 ± 0.010) performed similarly to gradient boosting machines (0.232 ± 0.029, 0.772 ± 0.018). Evaluation of the model inspection showed that explainable boosting machines can be useful to detect and remove problematic risk functions.ConclusionsWe developed an inherently interpretable ML model for 3 day ICU readmission prediction that reached the state-of-the-art performance of black box models. Our results suggest that for low- to medium-dimensional datasets that are common in health care, it is feasible to develop ML models that allow a high level of human control without sacrificing performance.https://www.frontiersin.org/articles/10.3389/fmed.2022.960296/fullintensive care unitreadmissionartificial intelligencemachine learningexplainable AIinterpretable machine learning
spellingShingle Stefan Hegselmann
Christian Ertmer
Thomas Volkert
Antje Gottschalk
Martin Dugas
Julian Varghese
Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines
Frontiers in Medicine
intensive care unit
readmission
artificial intelligence
machine learning
explainable AI
interpretable machine learning
title Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines
title_full Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines
title_fullStr Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines
title_full_unstemmed Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines
title_short Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines
title_sort development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines
topic intensive care unit
readmission
artificial intelligence
machine learning
explainable AI
interpretable machine learning
url https://www.frontiersin.org/articles/10.3389/fmed.2022.960296/full
work_keys_str_mv AT stefanhegselmann developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines
AT christianertmer developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines
AT thomasvolkert developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines
AT antjegottschalk developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines
AT martindugas developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines
AT julianvarghese developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines