TY - JOUR
T1 - Changing predictor measurement procedures affected the performance of prediction models in clinical examples
AU - Luijken, Kim
AU - Wynants, Laure
AU - van Smeden, Maarten
AU - Van Calster, Ben
AU - Steyerberg, Ewout W.
AU - Groenwold, Rolf H.H.
AU - Timmerman, Dirk
AU - Bourne, Tom
AU - Ukaegbu, Chinedu
N1 - Funding Information:
Funding: E.W.S. was funded partially through two Patient-Centered Outcomes Research Institute (PCORI) grants (the Predictive Analytics Resource Center [PARC] [SA.Tufts.PARC.OSCO.2018.01.25] and Methods Award [ME-1606-35,555]). R.H.H.G. was supported by the Netherlands Organisation for Scientific Research [ZonMW, project 917.16.430]. L.W. is a postdoctoral fellow of the Research Foundation–Flanders. B.V.C. was supported by the Research Foundation–Flanders (grant G0B4716N) and Internal Funds KU Leuven (grant C24/15/037).
Publisher Copyright:
© 2019 The Authors
PY - 2020/3
Y1 - 2020/3
N2 - Objectives: The aim of this study was to quantify the impact of predictor measurement heterogeneity on prediction model performance. Predictor measurement heterogeneity refers to variation in the measurement of predictor(s) between the derivation of a prediction model and its validation or application. It arises, for instance, when predictors are measured using different measurement instruments or protocols. Study Design and Setting: We examined the effects of various scenarios of predictor measurement heterogeneity in real-world clinical examples using previously developed prediction models for diagnosis of ovarian cancer, mutation carriers for Lynch syndrome, and intrauterine pregnancy. Results: Changing the measurement procedure of a predictor influenced the performance at validation of the prediction models in nine clinical examples. Notably, it induced model miscalibration. The calibration intercept at validation ranged from −0.70 to 1.43 (0 for good calibration), whereas the calibration slope ranged from 0.50 to 1.67 (1 for good calibration). The difference in C-statistic and scaled Brier score between derivation and validation ranged from −0.08 to +0.08 and from −0.40 to +0.16, respectively. Conclusion: This study illustrates that predictor measurement heterogeneity can influence the performance of a prediction model substantially, underlining that predictor measurements used in research settings should resemble clinical practice. Specification of measurement heterogeneity can help researchers explaining discrepancies in predictive performance between derivation and validation setting.
AB - Objectives: The aim of this study was to quantify the impact of predictor measurement heterogeneity on prediction model performance. Predictor measurement heterogeneity refers to variation in the measurement of predictor(s) between the derivation of a prediction model and its validation or application. It arises, for instance, when predictors are measured using different measurement instruments or protocols. Study Design and Setting: We examined the effects of various scenarios of predictor measurement heterogeneity in real-world clinical examples using previously developed prediction models for diagnosis of ovarian cancer, mutation carriers for Lynch syndrome, and intrauterine pregnancy. Results: Changing the measurement procedure of a predictor influenced the performance at validation of the prediction models in nine clinical examples. Notably, it induced model miscalibration. The calibration intercept at validation ranged from −0.70 to 1.43 (0 for good calibration), whereas the calibration slope ranged from 0.50 to 1.67 (1 for good calibration). The difference in C-statistic and scaled Brier score between derivation and validation ranged from −0.08 to +0.08 and from −0.40 to +0.16, respectively. Conclusion: This study illustrates that predictor measurement heterogeneity can influence the performance of a prediction model substantially, underlining that predictor measurements used in research settings should resemble clinical practice. Specification of measurement heterogeneity can help researchers explaining discrepancies in predictive performance between derivation and validation setting.
KW - Calibration
KW - External validation
KW - Measurement error
KW - Measurement heterogeneity
KW - Prediction model
KW - Predictive performance
UR - http://www.scopus.com/inward/record.url?scp=85076402004&partnerID=8YFLogxK
U2 - 10.1016/j.jclinepi.2019.11.001
DO - 10.1016/j.jclinepi.2019.11.001
M3 - Article
C2 - 31706963
AN - SCOPUS:85076402004
SN - 0895-4356
VL - 119
SP - 7
EP - 18
JO - Journal of Clinical Epidemiology
JF - Journal of Clinical Epidemiology
ER -