Developing non-response weights to account for attrition-related bias in a longitudinal pregnancy cohort.
Objective The prospective cohort study design is ideal for examining diseases of public health importance. A main source of potential bias for longitudinal studies is attrition. In this study, we compare the performance of two models developed to predict sources of attrition and develop weights to a...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Swansea University
2022-08-01
|
Series: | International Journal of Population Data Science |
Subjects: | |
Online Access: | https://ijpds.org/article/view/1925 |
_version_ | 1797430693633982464 |
---|---|
author | Tona Pitt Kamala Adhikari Shainur Premji Sheila McDonald |
author_facet | Tona Pitt Kamala Adhikari Shainur Premji Sheila McDonald |
author_sort | Tona Pitt |
collection | DOAJ |
description | Objective
The prospective cohort study design is ideal for examining diseases of public health importance. A main source of potential bias for longitudinal studies is attrition. In this study, we compare the performance of two models developed to predict sources of attrition and develop weights to adjust for potential bias.
Approach
This study used the All Our Families longitudinal pregnancy cohort of 3351 maternal-infant pairs. Logistic regression models were developed to predict study continuation versus drop-out from baseline to the three-year data collection wave.
Two methods of variable selection took place. One method used previous knowledge and content expertise while the second used Least Absolute Shrinkage and Selection Operator (LASSO). Model performance for both methods were compared using area under the receiver operator curve values (AUROC) and calibration plots. Stabilized inverse probability weights were generated using predicted probabilities. Weight performance was assessed using standardized differences with and without weights (unadjusted estimates).
Results
LASSO and investigator prediction models had good and fair discrimination with AUROC of 0.73 (95% Confidence Interval [CI]: 0.71 – 0.75) and 0.69 ( 95% CI: 0.67 – 0.71), respectively. Calibration plots and non significant Hosmer-Lemeshow Goodness of Fit Tests indicated that both the LASSO model (p = 0.10) and investigator model (p = 0.50) were well-calibrated.
Unweighted results indicated large (>10%) standardized differences in 15 demographic data variables (range: 11% - 29%), when comparing those who continued in study with those that did not. Weights derived from the LASSO and investigator models reduced standardized differences relative to unadjusted estimates, with ranges of 0.1% - 5.3% and 0.3% - 12.7%, respectively.
Conclusion
The data-driven approach produced robust weights that addressed non-response bias more than the knowledge-driven approach. The data driven approach, did, however still require content knowledge in how data were grouped, combined, or split. The weights can be applied to analyses across multiple waves of data collection to reduce bias.
|
first_indexed | 2024-03-09T09:31:15Z |
format | Article |
id | doaj.art-bdf81ff38c7247ae912de9f005233f0a |
institution | Directory Open Access Journal |
issn | 2399-4908 |
language | English |
last_indexed | 2024-03-09T09:31:15Z |
publishDate | 2022-08-01 |
publisher | Swansea University |
record_format | Article |
series | International Journal of Population Data Science |
spelling | doaj.art-bdf81ff38c7247ae912de9f005233f0a2023-12-02T03:51:49ZengSwansea UniversityInternational Journal of Population Data Science2399-49082022-08-017310.23889/ijpds.v7i3.1925Developing non-response weights to account for attrition-related bias in a longitudinal pregnancy cohort.Tona Pitt0Kamala Adhikari1Shainur Premji2Sheila McDonald3University of CalgaryAlberta Health ServicesUniversity of YorkAlberta Health ServicesObjective The prospective cohort study design is ideal for examining diseases of public health importance. A main source of potential bias for longitudinal studies is attrition. In this study, we compare the performance of two models developed to predict sources of attrition and develop weights to adjust for potential bias. Approach This study used the All Our Families longitudinal pregnancy cohort of 3351 maternal-infant pairs. Logistic regression models were developed to predict study continuation versus drop-out from baseline to the three-year data collection wave. Two methods of variable selection took place. One method used previous knowledge and content expertise while the second used Least Absolute Shrinkage and Selection Operator (LASSO). Model performance for both methods were compared using area under the receiver operator curve values (AUROC) and calibration plots. Stabilized inverse probability weights were generated using predicted probabilities. Weight performance was assessed using standardized differences with and without weights (unadjusted estimates). Results LASSO and investigator prediction models had good and fair discrimination with AUROC of 0.73 (95% Confidence Interval [CI]: 0.71 – 0.75) and 0.69 ( 95% CI: 0.67 – 0.71), respectively. Calibration plots and non significant Hosmer-Lemeshow Goodness of Fit Tests indicated that both the LASSO model (p = 0.10) and investigator model (p = 0.50) were well-calibrated. Unweighted results indicated large (>10%) standardized differences in 15 demographic data variables (range: 11% - 29%), when comparing those who continued in study with those that did not. Weights derived from the LASSO and investigator models reduced standardized differences relative to unadjusted estimates, with ranges of 0.1% - 5.3% and 0.3% - 12.7%, respectively. Conclusion The data-driven approach produced robust weights that addressed non-response bias more than the knowledge-driven approach. The data driven approach, did, however still require content knowledge in how data were grouped, combined, or split. The weights can be applied to analyses across multiple waves of data collection to reduce bias. https://ijpds.org/article/view/1925weightingdrop-outlongitudinal cohortLeast Absolute ShrinkageSelection Operator |
spellingShingle | Tona Pitt Kamala Adhikari Shainur Premji Sheila McDonald Developing non-response weights to account for attrition-related bias in a longitudinal pregnancy cohort. International Journal of Population Data Science weighting drop-out longitudinal cohort Least Absolute Shrinkage Selection Operator |
title | Developing non-response weights to account for attrition-related bias in a longitudinal pregnancy cohort. |
title_full | Developing non-response weights to account for attrition-related bias in a longitudinal pregnancy cohort. |
title_fullStr | Developing non-response weights to account for attrition-related bias in a longitudinal pregnancy cohort. |
title_full_unstemmed | Developing non-response weights to account for attrition-related bias in a longitudinal pregnancy cohort. |
title_short | Developing non-response weights to account for attrition-related bias in a longitudinal pregnancy cohort. |
title_sort | developing non response weights to account for attrition related bias in a longitudinal pregnancy cohort |
topic | weighting drop-out longitudinal cohort Least Absolute Shrinkage Selection Operator |
url | https://ijpds.org/article/view/1925 |
work_keys_str_mv | AT tonapitt developingnonresponseweightstoaccountforattritionrelatedbiasinalongitudinalpregnancycohort AT kamalaadhikari developingnonresponseweightstoaccountforattritionrelatedbiasinalongitudinalpregnancycohort AT shainurpremji developingnonresponseweightstoaccountforattritionrelatedbiasinalongitudinalpregnancycohort AT sheilamcdonald developingnonresponseweightstoaccountforattritionrelatedbiasinalongitudinalpregnancycohort |