Calibrating models in economic evaluation: A comparison of alternative measures of goodness of fit, parameter search strategies and convergence criteria.

Jonathan Karnon, Tazio Vanni

Research output: Contribution to journalArticlepeer-review

45 Citations (Scopus)


Background: The importance of assessing the accuracy of health economic decision models is widely recognized. Many applied decision models (implicitly) assume that the process of identifying relevant values for a model's input parameters is sufficient to prove the model's accuracy. The selection of infeasible combinations of input parameter values is most likely in the context of probabilistic sensitivity analysis (PSA), where parameter values are drawn from independently specified probability distributions for each model parameter. Model calibration involves the identification of input parameter values that produce model output parameters that best predict observed data.

Methods: An empirical comparison of three key calibration issues is presented: the applied measure of goodness of fit (GOF); the search strategy for selecting sets of input parameter values; and the convergence criteria for determining acceptable GOF. The comparisons are presented in the context of probabilistic calibration, a widely applicable approach to calibration that can be easily integrated with PSA. The appendix provides a user's guide to probabilistic calibration, with the reader invited to download the Microsoft-Excel-based model reported in this article.

Results: The calibrated models consistently provided higher mean estimates of the models' output parameter, illustrating the potential gain in accuracy derived from calibrating decision models. Model uncertainty was also reduced. The chi-squared GOF measure differentiated between the accuracy of different parameter sets to a far greater degree than the likelihood GOF measure. The guided search strategy produced higher mean estimates of the models' output parameter, as well as a narrower range of predicted output values, which may reflect greater precision in the identification of candidate parameter sets or more limited coverage of the parameter space. The broader convergence threshold resulted in lower mean estimates of the models' output, and slightly wider ranges, which were closer to the outputs associated with the non-calibrated approach.

Conclusions: Probabilistic calibration provides a broadly applicable method that will improve the relevance of health economic decision models, and simultaneously reduce model uncertainty. The analyses reported in this paper inform the more efficient and accurate application of calibration methods for health economic decision models.

Original languageEnglish
Pages (from-to)51-62
Number of pages12
Issue number1
Publication statusPublished - Jan 2011
Externally publishedYes


  • probability weight
  • calibration target
  • health economics
  • Modelling
  • Cost-effectiveness


Dive into the research topics of 'Calibrating models in economic evaluation: A comparison of alternative measures of goodness of fit, parameter search strategies and convergence criteria.'. Together they form a unique fingerprint.

Cite this