Developing non-response weights to account for attrition-related bias in a longitudinal pregnancy cohort.

Objective The prospective cohort study design is ideal for examining diseases of public health importance. A main source of potential bias for longitudinal studies is attrition. In this study, we compare the performance of two models developed to predict sources of attrition and develop weights to a...

Full description

Bibliographic Details
Main Authors: Tona Pitt, Kamala Adhikari, Shainur Premji, Sheila McDonald
Format: Article
Language:English
Published: Swansea University 2022-08-01
Series:International Journal of Population Data Science
Subjects:
Online Access:https://ijpds.org/article/view/1925
_version_ 1797430693633982464
author Tona Pitt
Kamala Adhikari
Shainur Premji
Sheila McDonald
author_facet Tona Pitt
Kamala Adhikari
Shainur Premji
Sheila McDonald
author_sort Tona Pitt
collection DOAJ
description Objective The prospective cohort study design is ideal for examining diseases of public health importance. A main source of potential bias for longitudinal studies is attrition. In this study, we compare the performance of two models developed to predict sources of attrition and develop weights to adjust for potential bias. Approach This study used the All Our Families longitudinal pregnancy cohort of 3351 maternal-infant pairs. Logistic regression models were developed to predict study continuation versus drop-out from baseline to the three-year data collection wave. Two methods of variable selection took place. One method used previous knowledge and content expertise while the second used Least Absolute Shrinkage and Selection Operator (LASSO). Model performance for both methods were compared using area under the receiver operator curve values (AUROC) and calibration plots. Stabilized inverse probability weights were generated using predicted probabilities. Weight performance was assessed using standardized differences with and without weights (unadjusted estimates). Results LASSO and investigator prediction models had good and fair discrimination with AUROC of 0.73 (95% Confidence Interval [CI]: 0.71 – 0.75) and 0.69 ( 95% CI: 0.67 – 0.71), respectively. Calibration plots and non significant Hosmer-Lemeshow Goodness of Fit Tests indicated that both the LASSO model (p = 0.10) and investigator model (p = 0.50) were well-calibrated. Unweighted results indicated large (>10%) standardized differences in 15 demographic data variables (range: 11% - 29%), when comparing those who continued in study with those that did not. Weights derived from the LASSO and investigator models reduced standardized differences relative to unadjusted estimates, with ranges of 0.1% - 5.3% and 0.3% - 12.7%, respectively. Conclusion The data-driven approach produced robust weights that addressed non-response bias more than the knowledge-driven approach. The data driven approach, did, however still require content knowledge in how data were grouped, combined, or split. The weights can be applied to analyses across multiple waves of data collection to reduce bias.
first_indexed 2024-03-09T09:31:15Z
format Article
id doaj.art-bdf81ff38c7247ae912de9f005233f0a
institution Directory Open Access Journal
issn 2399-4908
language English
last_indexed 2024-03-09T09:31:15Z
publishDate 2022-08-01
publisher Swansea University
record_format Article
series International Journal of Population Data Science
spelling doaj.art-bdf81ff38c7247ae912de9f005233f0a2023-12-02T03:51:49ZengSwansea UniversityInternational Journal of Population Data Science2399-49082022-08-017310.23889/ijpds.v7i3.1925Developing non-response weights to account for attrition-related bias in a longitudinal pregnancy cohort.Tona Pitt0Kamala Adhikari1Shainur Premji2Sheila McDonald3University of CalgaryAlberta Health ServicesUniversity of YorkAlberta Health ServicesObjective The prospective cohort study design is ideal for examining diseases of public health importance. A main source of potential bias for longitudinal studies is attrition. In this study, we compare the performance of two models developed to predict sources of attrition and develop weights to adjust for potential bias. Approach This study used the All Our Families longitudinal pregnancy cohort of 3351 maternal-infant pairs. Logistic regression models were developed to predict study continuation versus drop-out from baseline to the three-year data collection wave. Two methods of variable selection took place. One method used previous knowledge and content expertise while the second used Least Absolute Shrinkage and Selection Operator (LASSO). Model performance for both methods were compared using area under the receiver operator curve values (AUROC) and calibration plots. Stabilized inverse probability weights were generated using predicted probabilities. Weight performance was assessed using standardized differences with and without weights (unadjusted estimates). Results LASSO and investigator prediction models had good and fair discrimination with AUROC of 0.73 (95% Confidence Interval [CI]: 0.71 – 0.75) and 0.69 ( 95% CI: 0.67 – 0.71), respectively. Calibration plots and non significant Hosmer-Lemeshow Goodness of Fit Tests indicated that both the LASSO model (p = 0.10) and investigator model (p = 0.50) were well-calibrated. Unweighted results indicated large (>10%) standardized differences in 15 demographic data variables (range: 11% - 29%), when comparing those who continued in study with those that did not. Weights derived from the LASSO and investigator models reduced standardized differences relative to unadjusted estimates, with ranges of 0.1% - 5.3% and 0.3% - 12.7%, respectively. Conclusion The data-driven approach produced robust weights that addressed non-response bias more than the knowledge-driven approach. The data driven approach, did, however still require content knowledge in how data were grouped, combined, or split. The weights can be applied to analyses across multiple waves of data collection to reduce bias. https://ijpds.org/article/view/1925weightingdrop-outlongitudinal cohortLeast Absolute ShrinkageSelection Operator
spellingShingle Tona Pitt
Kamala Adhikari
Shainur Premji
Sheila McDonald
Developing non-response weights to account for attrition-related bias in a longitudinal pregnancy cohort.
International Journal of Population Data Science
weighting
drop-out
longitudinal cohort
Least Absolute Shrinkage
Selection Operator
title Developing non-response weights to account for attrition-related bias in a longitudinal pregnancy cohort.
title_full Developing non-response weights to account for attrition-related bias in a longitudinal pregnancy cohort.
title_fullStr Developing non-response weights to account for attrition-related bias in a longitudinal pregnancy cohort.
title_full_unstemmed Developing non-response weights to account for attrition-related bias in a longitudinal pregnancy cohort.
title_short Developing non-response weights to account for attrition-related bias in a longitudinal pregnancy cohort.
title_sort developing non response weights to account for attrition related bias in a longitudinal pregnancy cohort
topic weighting
drop-out
longitudinal cohort
Least Absolute Shrinkage
Selection Operator
url https://ijpds.org/article/view/1925
work_keys_str_mv AT tonapitt developingnonresponseweightstoaccountforattritionrelatedbiasinalongitudinalpregnancycohort
AT kamalaadhikari developingnonresponseweightstoaccountforattritionrelatedbiasinalongitudinalpregnancycohort
AT shainurpremji developingnonresponseweightstoaccountforattritionrelatedbiasinalongitudinalpregnancycohort
AT sheilamcdonald developingnonresponseweightstoaccountforattritionrelatedbiasinalongitudinalpregnancycohort