Calibration The process of fitting the model to evidence on modeled outcomes by adjusting input parameters
Validation Assessing model quality by comparing model predictions to evidence on modeled outcomes
Calibration vs. Validation:
“Calibrate until the model validates…”
SSE = \sum_{i=1}^{n} (y_i - \hat{y}_i)^2 Error = distance between model result and target
Where:
- n: number of calibration targets
- y_i: calibration target i
- \hat{y}_i: model outcome for target i
Potential issues
- How to weight different targets?
- Are some targets more important?
- Are some targets less uncertain?
Three hikers search for the highest point using only elevation: reflect, expand, contract, shrink, stop when converged.
Imagine three hikers searching for the highest point on a hill (optimal solution) using only elevation (function values). They follow these steps:
Start with a Triangle (Simplex): Each hiker stands at a different location and checks their elevation. The lowest hiker (worst point) needs to move.
Reflect the Worst Hiker: Move them across the midpoint of the other two.
Expand if Promising: If the new spot is higher, take a bigger step in that direction.
Contract if Reflection Fails: If the move wasn’t helpful, take a smaller step closer to the midpoint.
Shrink if Stuck: If there is no progress, all hikers move closer together.
Converge & Stop: Once all hikers meet at the same spot, they’ve found the best location.
Specify function describing model fit criteria (target windows, SSE, likelihood)
Generate multiple parameter sets consistent with model fit criteria
Use sample of fitted parameter sets to produce results
Menzies NA, Soeteman DI, Pandya A, Kim JJ. Bayesian Methods for Calibrating Health Policy Models: A Tutorial. Pharmacoeconomics. 2017 Jun;35(6):613–24.
This is exactly why we need calibration
These parameters were estimated using a Bayesian calibration approach – a method that allows us to match the model to real-world outcomes and quantify uncertainty in those estimates.
Transparency:
Clearly describing the model structure, equations, parameter values, and assumptions to enable interested parties to understand the model
Validation:
Comparing model results with events observed in reality
Eddy DM, Hollingworth W, Caro JJ, Tsevat J, McDonald KM, Wong JB. Model Transparency and Validation: A Report of the ISPOR-SMDM Modeling Good Research Practices Task Force–7. Med Decis Making. 2012 Sep;32(5):733–43.
Transparency:
Clearly describing the model structure, equations, parameter values, and assumptions to enable interested parties to understand the model
Validation:
Comparing model results with events observed in reality
Eddy DM, Hollingworth W, Caro JJ, Tsevat J, McDonald KM, Wong JB. Model Transparency and Validation: A Report of the ISPOR-SMDM Modeling Good Research Practices Task Force–7. Med Decis Making. 2012 Sep;32(5):733–43.
Face Validity
The model reflects current scientific understanding, as judged by experts.
Internal Validation (verification)
The model is implemented correctly and behaves as expected (e.g., code verification).
External Validation
Model outputs are compared with empirical observations not used in model development.
Predictive Validation
Can the model reproduce data that wasn’t available during development?
Cross-model Validation
Do different models give similar results for the same question?
Eddy DM, Hollingworth W, Caro JJ, Tsevat J, McDonald KM, Wong JB. Model Transparency and Validation: A Report of the ISPOR-SMDM Modeling Good Research Practices Task Force–7. Med Decis Making. 2012 Sep;32(5):733–43.
Modeled projections of HIV prevalence, South Africa for 2012
→ 10 models fit to earlier data, then compared to 2012 prevalence survey
Eaton, Bacaër et al, Lancet Global Health, 2015
Eaton, Bacaër et al, Lancet Global Health, 2015
Vital Strategies & CDC Foundation Health Economics Fellowship