Generalizability of deep neural networks for vertical cup-to-disc ratio estimation in ultra-widefield and amartphone-based fundus images
Purpose: To develop and validate a deep learning system (DLS) for estimation of vertical cup-to-disc ratio (vCDR) in ultra-widefield (UWF) and smartphone-based fundus images. Methods: A DLS consisting of two sequential convolutional neural networks (CNNs) to delineate optic disc (OD) and optic cup (...
Main Authors: | , , , , , , , , |
---|---|
Other Authors: | |
Format: | Journal Article |
Language: | English |
Published: |
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/179838 |
_version_ | 1824455802585350144 |
---|---|
author | Yap, Boon Peng Li, Kelvin Zhenghao Toh, En Qi Low, Kok Yao Rani, Sumaya Khan Goh, Eunice Jin Hui Hui, Vivien Yip Cherng Ng, Beng Koon Lim, Tock Han |
author2 | School of Electrical and Electronic Engineering |
author_facet | School of Electrical and Electronic Engineering Yap, Boon Peng Li, Kelvin Zhenghao Toh, En Qi Low, Kok Yao Rani, Sumaya Khan Goh, Eunice Jin Hui Hui, Vivien Yip Cherng Ng, Beng Koon Lim, Tock Han |
author_sort | Yap, Boon Peng |
collection | NTU |
description | Purpose: To develop and validate a deep learning system (DLS) for estimation of vertical cup-to-disc ratio (vCDR) in ultra-widefield (UWF) and smartphone-based fundus images. Methods: A DLS consisting of two sequential convolutional neural networks (CNNs) to delineate optic disc (OD) and optic cup (OC) boundaries was developed using 800 standard fundus images from the public REFUGE data set. The CNNs were tested on 400 test images from the REFUGE data set and 296 UWF and 300 smartphone-based images from a teleophthalmology clinic. vCDRs derived from the delineated OD/OC boundaries were compared with optometrists’ annotations using mean absolute error (MAE). Subgroup analysis was conducted to study the impact of peripapillary atrophy (PPA), and correlation study was performed to investigate potential correlations between sectoral CDR (sCDR) and retinal nerve fiber layer (RNFL) thickness. Results: The system achieved MAEs of 0.040 (95% CI, 0.037–0.043) in the REFUGE test images, 0.068 (95% CI, 0.061–0.075) in the UWF images, and 0.084 (95% CI, 0.075–0.092) in the smartphone-based images. There was no statistical significance in differences between PPA and non-PPA images. Weak correlation (r = −0.4046, P < 0.05) between sCDR and RNFL thickness was found only in the superior sector. Conclusions: We developed a deep learning system that estimates vCDR from standard, UWF, and smartphone-based images. We also described anatomic peripapillary adversarial lesion and its potential impact on OD/OC delineation. Translational Relevance: Artificial intelligence can estimate vCDR from different types of fundus images and may be used as a general and interpretable screening tool to improve community reach for diagnosis and management of glaucoma. |
first_indexed | 2025-02-19T03:44:00Z |
format | Journal Article |
id | ntu-10356/179838 |
institution | Nanyang Technological University |
language | English |
last_indexed | 2025-02-19T03:44:00Z |
publishDate | 2024 |
record_format | dspace |
spelling | ntu-10356/1798382024-08-30T15:40:22Z Generalizability of deep neural networks for vertical cup-to-disc ratio estimation in ultra-widefield and amartphone-based fundus images Yap, Boon Peng Li, Kelvin Zhenghao Toh, En Qi Low, Kok Yao Rani, Sumaya Khan Goh, Eunice Jin Hui Hui, Vivien Yip Cherng Ng, Beng Koon Lim, Tock Han School of Electrical and Electronic Engineering Lee Kong Chian School of Medicine (LKCMedicine) Tan Tock Seng Hospital National Healthcare Group Eye Institute Engineering Deep learning Optic disc Purpose: To develop and validate a deep learning system (DLS) for estimation of vertical cup-to-disc ratio (vCDR) in ultra-widefield (UWF) and smartphone-based fundus images. Methods: A DLS consisting of two sequential convolutional neural networks (CNNs) to delineate optic disc (OD) and optic cup (OC) boundaries was developed using 800 standard fundus images from the public REFUGE data set. The CNNs were tested on 400 test images from the REFUGE data set and 296 UWF and 300 smartphone-based images from a teleophthalmology clinic. vCDRs derived from the delineated OD/OC boundaries were compared with optometrists’ annotations using mean absolute error (MAE). Subgroup analysis was conducted to study the impact of peripapillary atrophy (PPA), and correlation study was performed to investigate potential correlations between sectoral CDR (sCDR) and retinal nerve fiber layer (RNFL) thickness. Results: The system achieved MAEs of 0.040 (95% CI, 0.037–0.043) in the REFUGE test images, 0.068 (95% CI, 0.061–0.075) in the UWF images, and 0.084 (95% CI, 0.075–0.092) in the smartphone-based images. There was no statistical significance in differences between PPA and non-PPA images. Weak correlation (r = −0.4046, P < 0.05) between sCDR and RNFL thickness was found only in the superior sector. Conclusions: We developed a deep learning system that estimates vCDR from standard, UWF, and smartphone-based images. We also described anatomic peripapillary adversarial lesion and its potential impact on OD/OC delineation. Translational Relevance: Artificial intelligence can estimate vCDR from different types of fundus images and may be used as a general and interpretable screening tool to improve community reach for diagnosis and management of glaucoma. Published version Supported by the Ng Teng Fong Healthcare Innovation Programme (Innovation Track), project code NTF_DEC2021_1_C1_D_03. 2024-08-27T03:11:31Z 2024-08-27T03:11:31Z 2024 Journal Article Yap, B. P., Li, K. Z., Toh, E. Q., Low, K. Y., Rani, S. K., Goh, E. J. H., Hui, V. Y. C., Ng, B. K. & Lim, T. H. (2024). Generalizability of deep neural networks for vertical cup-to-disc ratio estimation in ultra-widefield and amartphone-based fundus images. Translational Vision Science & Technology, 13(4), 6-. https://dx.doi.org/10.1167/tvst.13.4.6 2164-2591 https://hdl.handle.net/10356/179838 10.1167/tvst.13.4.6 38568608 2-s2.0-85190086013 4 13 6 en NTF_DEC2021_1_C1_D_03 Translational Vision Science & Technology © 2024 The Authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. application/pdf |
spellingShingle | Engineering Deep learning Optic disc Yap, Boon Peng Li, Kelvin Zhenghao Toh, En Qi Low, Kok Yao Rani, Sumaya Khan Goh, Eunice Jin Hui Hui, Vivien Yip Cherng Ng, Beng Koon Lim, Tock Han Generalizability of deep neural networks for vertical cup-to-disc ratio estimation in ultra-widefield and amartphone-based fundus images |
title | Generalizability of deep neural networks for vertical cup-to-disc ratio estimation in ultra-widefield and amartphone-based fundus images |
title_full | Generalizability of deep neural networks for vertical cup-to-disc ratio estimation in ultra-widefield and amartphone-based fundus images |
title_fullStr | Generalizability of deep neural networks for vertical cup-to-disc ratio estimation in ultra-widefield and amartphone-based fundus images |
title_full_unstemmed | Generalizability of deep neural networks for vertical cup-to-disc ratio estimation in ultra-widefield and amartphone-based fundus images |
title_short | Generalizability of deep neural networks for vertical cup-to-disc ratio estimation in ultra-widefield and amartphone-based fundus images |
title_sort | generalizability of deep neural networks for vertical cup to disc ratio estimation in ultra widefield and amartphone based fundus images |
topic | Engineering Deep learning Optic disc |
url | https://hdl.handle.net/10356/179838 |
work_keys_str_mv | AT yapboonpeng generalizabilityofdeepneuralnetworksforverticalcuptodiscratioestimationinultrawidefieldandamartphonebasedfundusimages AT likelvinzhenghao generalizabilityofdeepneuralnetworksforverticalcuptodiscratioestimationinultrawidefieldandamartphonebasedfundusimages AT tohenqi generalizabilityofdeepneuralnetworksforverticalcuptodiscratioestimationinultrawidefieldandamartphonebasedfundusimages AT lowkokyao generalizabilityofdeepneuralnetworksforverticalcuptodiscratioestimationinultrawidefieldandamartphonebasedfundusimages AT ranisumayakhan generalizabilityofdeepneuralnetworksforverticalcuptodiscratioestimationinultrawidefieldandamartphonebasedfundusimages AT goheunicejinhui generalizabilityofdeepneuralnetworksforverticalcuptodiscratioestimationinultrawidefieldandamartphonebasedfundusimages AT huivivienyipcherng generalizabilityofdeepneuralnetworksforverticalcuptodiscratioestimationinultrawidefieldandamartphonebasedfundusimages AT ngbengkoon generalizabilityofdeepneuralnetworksforverticalcuptodiscratioestimationinultrawidefieldandamartphonebasedfundusimages AT limtockhan generalizabilityofdeepneuralnetworksforverticalcuptodiscratioestimationinultrawidefieldandamartphonebasedfundusimages |