Biostatistics series module 7: The statistics of diagnostic tests
Crucial therapeutic decisions are based on diagnostic tests. Therefore, it is important to evaluate such tests before adopting them for routine use. Although things such as blood tests, cultures, biopsies, and radiological imaging are obvious diagnostic tests, it is not to be forgotten that specific...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
Wolters Kluwer Medknow Publications
2017-01-01
|
Series: | Indian Journal of Dermatology |
Subjects: | |
Online Access: | http://www.e-ijd.org/article.asp?issn=0019-5154;year=2017;volume=62;issue=1;spage=18;epage=24;aulast=Hazra |
_version_ | 1818318199388635136 |
---|---|
author | Avijit Hazra Nithya Gogtay |
author_facet | Avijit Hazra Nithya Gogtay |
author_sort | Avijit Hazra |
collection | DOAJ |
description | Crucial therapeutic decisions are based on diagnostic tests. Therefore, it is important to evaluate such tests before adopting them for routine use. Although things such as blood tests, cultures, biopsies, and radiological imaging are obvious diagnostic tests, it is not to be forgotten that specific clinical examination procedures, scoring systems based on physiological or psychological evaluation, and ratings based on questionnaires are also diagnostic tests and therefore merit similar evaluation. In the simplest scenario, a diagnostic test will give either a positive (disease likely) or negative (disease unlikely) result. Ideally, all those with the disease should be classified by a test as positive and all those without the disease as negative. Unfortunately, practically no test gives 100% accurate results. Therefore, leaving aside the economic question, the performance of diagnostic tests is evaluated on the basis of certain indices such as sensitivity, specificity, positive predictive value, and negative predictive value. Likelihood ratios combine information on specificity and sensitivity to expresses the likelihood that a given test result would occur in a subject with a disorder compared to the probability that the same result would occur in a subject without the disorder. Not all test can be categorized simply as “positive” or “negative.” Physicians are frequently exposed to test results on a numerical scale, and in such cases, judgment is required in choosing a cutoff point to distinguish normal from abnormal. Naturally, a cutoff value should provide the greatest predictive accuracy, but there is a trade-off between sensitivity and specificity here - if the cutoff is too low, it will identify most patients who have the disease (high sensitivity) but will also incorrectly identify many who do not (low specificity). A receiver operating characteristic curve plots pairs of sensitivity versus (1 − specificity) values and helps in selecting an optimum cutoff – the one lying on the “elbow” of the curve. Cohen's kappa (κ) statistic is a measure of inter-rater agreement for categorical variables. It can also be applied to assess how far two tests agree with respect to diagnostic categorization. It is generally thought to be a more robust measure than simple percent agreement calculation since kappa takes into account the agreement occurring by chance. |
first_indexed | 2024-12-13T09:49:25Z |
format | Article |
id | doaj.art-2229967797db472b97ee31e4b1a58cfa |
institution | Directory Open Access Journal |
issn | 0019-5154 1998-3611 |
language | English |
last_indexed | 2024-12-13T09:49:25Z |
publishDate | 2017-01-01 |
publisher | Wolters Kluwer Medknow Publications |
record_format | Article |
series | Indian Journal of Dermatology |
spelling | doaj.art-2229967797db472b97ee31e4b1a58cfa2022-12-21T23:51:57ZengWolters Kluwer Medknow PublicationsIndian Journal of Dermatology0019-51541998-36112017-01-01621182410.4103/0019-5154.198047Biostatistics series module 7: The statistics of diagnostic testsAvijit HazraNithya GogtayCrucial therapeutic decisions are based on diagnostic tests. Therefore, it is important to evaluate such tests before adopting them for routine use. Although things such as blood tests, cultures, biopsies, and radiological imaging are obvious diagnostic tests, it is not to be forgotten that specific clinical examination procedures, scoring systems based on physiological or psychological evaluation, and ratings based on questionnaires are also diagnostic tests and therefore merit similar evaluation. In the simplest scenario, a diagnostic test will give either a positive (disease likely) or negative (disease unlikely) result. Ideally, all those with the disease should be classified by a test as positive and all those without the disease as negative. Unfortunately, practically no test gives 100% accurate results. Therefore, leaving aside the economic question, the performance of diagnostic tests is evaluated on the basis of certain indices such as sensitivity, specificity, positive predictive value, and negative predictive value. Likelihood ratios combine information on specificity and sensitivity to expresses the likelihood that a given test result would occur in a subject with a disorder compared to the probability that the same result would occur in a subject without the disorder. Not all test can be categorized simply as “positive” or “negative.” Physicians are frequently exposed to test results on a numerical scale, and in such cases, judgment is required in choosing a cutoff point to distinguish normal from abnormal. Naturally, a cutoff value should provide the greatest predictive accuracy, but there is a trade-off between sensitivity and specificity here - if the cutoff is too low, it will identify most patients who have the disease (high sensitivity) but will also incorrectly identify many who do not (low specificity). A receiver operating characteristic curve plots pairs of sensitivity versus (1 − specificity) values and helps in selecting an optimum cutoff – the one lying on the “elbow” of the curve. Cohen's kappa (κ) statistic is a measure of inter-rater agreement for categorical variables. It can also be applied to assess how far two tests agree with respect to diagnostic categorization. It is generally thought to be a more robust measure than simple percent agreement calculation since kappa takes into account the agreement occurring by chance.http://www.e-ijd.org/article.asp?issn=0019-5154;year=2017;volume=62;issue=1;spage=18;epage=24;aulast=HazraKappa statisticlikelihood rationegative predictive valuepositive predictive valuepositivity criterionreceiver operating characteristic curvereference rangesensitivityspecificity |
spellingShingle | Avijit Hazra Nithya Gogtay Biostatistics series module 7: The statistics of diagnostic tests Indian Journal of Dermatology Kappa statistic likelihood ratio negative predictive value positive predictive value positivity criterion receiver operating characteristic curve reference range sensitivity specificity |
title | Biostatistics series module 7: The statistics of diagnostic tests |
title_full | Biostatistics series module 7: The statistics of diagnostic tests |
title_fullStr | Biostatistics series module 7: The statistics of diagnostic tests |
title_full_unstemmed | Biostatistics series module 7: The statistics of diagnostic tests |
title_short | Biostatistics series module 7: The statistics of diagnostic tests |
title_sort | biostatistics series module 7 the statistics of diagnostic tests |
topic | Kappa statistic likelihood ratio negative predictive value positive predictive value positivity criterion receiver operating characteristic curve reference range sensitivity specificity |
url | http://www.e-ijd.org/article.asp?issn=0019-5154;year=2017;volume=62;issue=1;spage=18;epage=24;aulast=Hazra |
work_keys_str_mv | AT avijithazra biostatisticsseriesmodule7thestatisticsofdiagnostictests AT nithyagogtay biostatisticsseriesmodule7thestatisticsofdiagnostictests |