Scoring reading parameters: An inter-rater reliability study using the MNREAD chart.

<h4>Purpose</h4>First, to evaluate inter-rater reliability when human raters estimate the reading performance of visually impaired individuals using the MNREAD acuity chart. Second, to evaluate the agreement between computer-based scoring algorithms and compare them with human rating.<...

Full description

Bibliographic Details
Main Authors: Karthikeyan Baskaran, Antonio Filipe Macedo, Yingchen He, Laura Hernandez-Moreno, Tatiana Queirós, J Stephen Mansfield, Aurélie Calabrèse
Format: Article
Language:English
Published: Public Library of Science (PLoS) 2019-01-01
Series:PLoS ONE
Online Access:https://doi.org/10.1371/journal.pone.0216775
_version_ 1819037251470884864
author Karthikeyan Baskaran
Antonio Filipe Macedo
Yingchen He
Laura Hernandez-Moreno
Tatiana Queirós
J Stephen Mansfield
Aurélie Calabrèse
author_facet Karthikeyan Baskaran
Antonio Filipe Macedo
Yingchen He
Laura Hernandez-Moreno
Tatiana Queirós
J Stephen Mansfield
Aurélie Calabrèse
author_sort Karthikeyan Baskaran
collection DOAJ
description <h4>Purpose</h4>First, to evaluate inter-rater reliability when human raters estimate the reading performance of visually impaired individuals using the MNREAD acuity chart. Second, to evaluate the agreement between computer-based scoring algorithms and compare them with human rating.<h4>Methods</h4>Reading performance was measured for 101 individuals with low vision, using the Portuguese version of the MNREAD test. Seven raters estimated the maximum reading speed (MRS) and critical print size (CPS) of each individual MNREAD curve. MRS and CPS were also calculated automatically for each curve using two different algorithms: the original standard deviation method (SDev) and a non-linear mixed effects (NLME) modeling. Intra-class correlation coefficients (ICC) were used to estimate absolute agreement between raters and/or algorithms.<h4>Results</h4>Absolute agreement between raters was 'excellent' for MRS (ICC = 0.97; 95%CI [0.96, 0.98]) and 'moderate' to 'good' for CPS (ICC = 0.77; 95%CI [0.69, 0.83]). For CPS, inter-rater reliability was poorer among less experienced raters (ICC = 0.70; 95%CI [0.57, 0.80]) when compared to experienced ones (ICC = 0.82; 95%CI [0.76, 0.88]). Absolute agreement between the two algorithms was 'excellent' for MRS (ICC = 0.96; 95%CI [0.91, 0.98]). For CPS, the best possible agreement was found for CPS defined as the print size sustaining 80% of MRS (ICC = 0.77; 95%CI [0.68, 0.84]). Absolute agreement between raters and automated methods was 'excellent' for MRS (ICC = 0.96; 95% CI [0.88, 0.98] for SDev; ICC = 0.97; 95% CI [0.95, 0.98] for NLME). For CPS, absolute agreement between raters and SDev ranged from 'poor' to 'good' (ICC = 0.66; 95% CI [0.3, 0.80]), while agreement between raters and NLME was 'good' (ICC = 0.83; 95% CI [0.76, 0.88]).<h4>Conclusion</h4>For MRS, inter-rater reliability is excellent, even considering the possibility of noisy and/or incomplete data collected in low-vision individuals. For CPS, inter-rater reliability is lower. This may be problematic, for instance in the context of multisite investigations or follow-up examinations. The NLME method showed better agreement with the raters than the SDev method for both reading parameters. Setting up consensual guidelines to deal with ambiguous curves may help improve reliability. While the exact definition of CPS should be chosen on a case-by-case basis depending on the clinician or researcher's motivations, evidence suggests that estimating CPS as the smallest print size sustaining about 80% of MRS would increase inter-rater reliability.
first_indexed 2024-12-21T08:18:27Z
format Article
id doaj.art-b1abdc4bd955489e99a897431a415d89
institution Directory Open Access Journal
issn 1932-6203
language English
last_indexed 2024-12-21T08:18:27Z
publishDate 2019-01-01
publisher Public Library of Science (PLoS)
record_format Article
series PLoS ONE
spelling doaj.art-b1abdc4bd955489e99a897431a415d892022-12-21T19:10:31ZengPublic Library of Science (PLoS)PLoS ONE1932-62032019-01-01146e021677510.1371/journal.pone.0216775Scoring reading parameters: An inter-rater reliability study using the MNREAD chart.Karthikeyan BaskaranAntonio Filipe MacedoYingchen HeLaura Hernandez-MorenoTatiana QueirósJ Stephen MansfieldAurélie Calabrèse<h4>Purpose</h4>First, to evaluate inter-rater reliability when human raters estimate the reading performance of visually impaired individuals using the MNREAD acuity chart. Second, to evaluate the agreement between computer-based scoring algorithms and compare them with human rating.<h4>Methods</h4>Reading performance was measured for 101 individuals with low vision, using the Portuguese version of the MNREAD test. Seven raters estimated the maximum reading speed (MRS) and critical print size (CPS) of each individual MNREAD curve. MRS and CPS were also calculated automatically for each curve using two different algorithms: the original standard deviation method (SDev) and a non-linear mixed effects (NLME) modeling. Intra-class correlation coefficients (ICC) were used to estimate absolute agreement between raters and/or algorithms.<h4>Results</h4>Absolute agreement between raters was 'excellent' for MRS (ICC = 0.97; 95%CI [0.96, 0.98]) and 'moderate' to 'good' for CPS (ICC = 0.77; 95%CI [0.69, 0.83]). For CPS, inter-rater reliability was poorer among less experienced raters (ICC = 0.70; 95%CI [0.57, 0.80]) when compared to experienced ones (ICC = 0.82; 95%CI [0.76, 0.88]). Absolute agreement between the two algorithms was 'excellent' for MRS (ICC = 0.96; 95%CI [0.91, 0.98]). For CPS, the best possible agreement was found for CPS defined as the print size sustaining 80% of MRS (ICC = 0.77; 95%CI [0.68, 0.84]). Absolute agreement between raters and automated methods was 'excellent' for MRS (ICC = 0.96; 95% CI [0.88, 0.98] for SDev; ICC = 0.97; 95% CI [0.95, 0.98] for NLME). For CPS, absolute agreement between raters and SDev ranged from 'poor' to 'good' (ICC = 0.66; 95% CI [0.3, 0.80]), while agreement between raters and NLME was 'good' (ICC = 0.83; 95% CI [0.76, 0.88]).<h4>Conclusion</h4>For MRS, inter-rater reliability is excellent, even considering the possibility of noisy and/or incomplete data collected in low-vision individuals. For CPS, inter-rater reliability is lower. This may be problematic, for instance in the context of multisite investigations or follow-up examinations. The NLME method showed better agreement with the raters than the SDev method for both reading parameters. Setting up consensual guidelines to deal with ambiguous curves may help improve reliability. While the exact definition of CPS should be chosen on a case-by-case basis depending on the clinician or researcher's motivations, evidence suggests that estimating CPS as the smallest print size sustaining about 80% of MRS would increase inter-rater reliability.https://doi.org/10.1371/journal.pone.0216775
spellingShingle Karthikeyan Baskaran
Antonio Filipe Macedo
Yingchen He
Laura Hernandez-Moreno
Tatiana Queirós
J Stephen Mansfield
Aurélie Calabrèse
Scoring reading parameters: An inter-rater reliability study using the MNREAD chart.
PLoS ONE
title Scoring reading parameters: An inter-rater reliability study using the MNREAD chart.
title_full Scoring reading parameters: An inter-rater reliability study using the MNREAD chart.
title_fullStr Scoring reading parameters: An inter-rater reliability study using the MNREAD chart.
title_full_unstemmed Scoring reading parameters: An inter-rater reliability study using the MNREAD chart.
title_short Scoring reading parameters: An inter-rater reliability study using the MNREAD chart.
title_sort scoring reading parameters an inter rater reliability study using the mnread chart
url https://doi.org/10.1371/journal.pone.0216775
work_keys_str_mv AT karthikeyanbaskaran scoringreadingparametersaninterraterreliabilitystudyusingthemnreadchart
AT antoniofilipemacedo scoringreadingparametersaninterraterreliabilitystudyusingthemnreadchart
AT yingchenhe scoringreadingparametersaninterraterreliabilitystudyusingthemnreadchart
AT laurahernandezmoreno scoringreadingparametersaninterraterreliabilitystudyusingthemnreadchart
AT tatianaqueiros scoringreadingparametersaninterraterreliabilitystudyusingthemnreadchart
AT jstephenmansfield scoringreadingparametersaninterraterreliabilitystudyusingthemnreadchart
AT aureliecalabrese scoringreadingparametersaninterraterreliabilitystudyusingthemnreadchart