 Research article
 Open Access
 Published:
Sequentially calibrating a Bayesian microsimulation model to incorporate new information and assumptions
BMC Medical Informatics and Decision Making volume 22, Article number: 12 (2022)
Abstract
Background
Microsimulation models are mathematical models that simulate event histories for individual members of a population. They are useful for policy decisions because they simulate a large number of individuals from an idealized population, with features that change over time, and the resulting event histories can be summarized to describe key populationlevel outcomes. Model calibration is the process of incorporating evidence into the model. Calibrated models can be used to make predictions about population trends in disease outcomes and effectiveness of interventions, but calibration can be challenging and computationally expensive.
Methods
This paper develops a technique for sequentially updating models to take full advantage of earlier calibration results, to ultimately speed up the calibration process. A Bayesian approach to calibration is used because it combines different sources of evidence and enables uncertainty quantification which is appealing for decisionmaking. We develop this method in order to recalibrate a microsimulation model for the natural history of colorectal cancer to include new targets that better inform the time from initiation of preclinical cancer to presentation with clinical cancer (sojourn time), because model exploration and validation revealed that more information was needed on sojourn time, and that the predicted percentage of patients with cancers detected via colonoscopy screening was too low.
Results
The sequential approach to calibration was more efficient than recalibrating the model from scratch. Incorporating new information on the percentage of patients with cancers detected upon screening changed the estimated sojourn time parameters significantly, increasing the estimated mean sojourn time for cancers in the colon and rectum, providing results with more validity.
Conclusions
A sequential approach to recalibration can be used to efficiently recalibrate a microsimulation model when new information becomes available that requires the original targets to be supplemented with additional targets.
Background
Microsimulation models (MSMs) are mathematical models that simulate event histories for individual members of a population. They can be used to simulate a large number of individuals from an idealized population, with features that change over time. MSMs are useful for policy decisions because life histories can be simulated under alternative scenarios or assumptions, and aggregated to predict populationlevel outcomes. MSMs can be used to estimate effects of interventions that may influence multiple components of the model in complex ways [9].
Health policy MSMs represent social or biological mechanisms that determine health or economic outcomes and rely on assumptions about how policy choices will affect those outcomes [25]. Measures or outcomes that are relevant to the health policy question at hand can then be summarized by aggregating individuallevel predictions to understand the effects of different policies. MSMs have been applied to health policy questions since 1985 [44]. Cancer is one area of application where MSMs have been heavily used to inform policy, with the U.S. Preventive Services Task Force considering the effectiveness of different screening regimens predicted by MSMs when developing guidelines for breast [22], cervical [14], colorectal [17], and lung cancer [4].
Information may be directly available on some parameters that govern the model predictions, such as the sensitivity and specificity of a screening exam or survival time following diagnosis. Such parameters are referred to as inputs, which are fixed rather than estimated. However, most parameters will be related to observable outcomes or available evidence in a complex way. For example, populationlevel outcomes such as colorectal cancer (CRC) incidence rates can be observed from national cancer registries, but these are functions of the unobservable process of developing CRC, which depends on the rate of adenoma (precursor lesion) initiation, and the growth and transition to preclinical (asymptomatic) and clinical (detected) cancer. The functional relationships between the observed data and the MSM parameters is not known. Evidence provided by screening trials and epidemiological studies informs MSM parameters, but the function of this relationship is not known and parameters may interact in complex ways to generate the observed outcomes.
At its simplest, model calibration is the process of incorporating evidence into the model [25], and is accomplished by selecting parameter values that result in model predictions that are consistent with observed or expected results based on observed data or expert knowledge. Calibrated MSMs can be used to make predictions about population trends in disease outcomes and effectiveness of interventions. However, calibration can be a challenging and time consuming process because it involves searching a high dimensional parameter space to predict many targets simultaneously, and it is sometimes difficult to find suitable parameter values. There are many methods for calibration, ranging from simple oneatatime parameter perturbation [37], to grid search algorithms from engineering [15, 19], which may break down when calibrating a large number of parameters. Directed searches use the derivative of the likelihood function, or an approximation to the derivative, to move toward areas of improved fit (e.g., [30, 35]), however for MSMs, the likelihood function is generally not available in closed form.
Bayesian approaches to calibration are particularly useful in the context of MSMs for policymaking because they combine different sources of evidence and enable estimation and uncertainty quantification of parameters, functions of parameters, and model predictions. Such uncertainty arises from variability in calibration data (sampling variability), simulation (Monte Carlo) variability, and parameter estimation. Markov Chain Monte Carlo (MCMC) is a commonly used class of algorithms for estimating the parameters of a Bayesian model [41, 50]. Rutter et al. [41] developed an approximate MetropolisHastings algorithm that includes an embedded simulation to estimate the likelihood function used to calculate the acceptance ratio. The MCMC algorithm was used to simulate draws from the posterior distribution of the parameters of a MSM for CRC given calibration targets. However, MCMC can be costly to apply to MSM calibration because the likelihood is often intractable and computationally expensive to estimate [25], and because it is based on a process of sequentially updating draws. The calibration process can also require a large number of samples from a model’s parameter space to be used for propagating uncertainties of the resulting parameter estimates, which can make certain nonparametric, Gaussian processbased Bayesian approaches less feasible [8].
In order to take advantage of modern computing resources, Rutter et al. [43] developed a likelihoodfree approximate Bayesian computation (ABC) algorithm [23, 46], referred to as Incremental Mixture ABC (IMABC), that was successfully applied to calibrate a MSM for CRC, referred to as CRCSPIN 2.0 (ColoRectal Cancer Simulated Population model for Incidence and Natural history). Once CRCSPIN 2.0 had been calibrated, posterior distributions were examined, and predictions from the model were compared to external validation targets that were not used in model calibration [40]. This examination revealed that very little information on sojourn time, the time from initiation of preclinical cancer to presentation with clinical CRC, was provided by the calibration targets. The calibrated model also resulted in screen detected cancer rates (the percentage of patients with cancer detected via colonoscopy screening) that were too low and indicated that additional calibration targets should be incorporated to better inform this aspect of the disease process. One approach is to perform another full calibration, including one or more new targets in addition to the previous set. However, this is time consuming and does not build on information about parameters obtained from the original calibration. This paper develops a technique for sequentially updating MSMs to take full advantage of earlier calibration results that are based on a similar set of targets, with the objective being to ultimately speed up the calibration process.
The remainder of this paper is organized as follows. In “Calibrating the microsimulation model” section, we present the CRCSPIN 2.0 model, the calibration targets included, and the resulting estimates of parameters and predictions arising from this calibrated model. In “Recalibrating the microsimulation model to incorporate new study targets” section, we explore methods for incorporating additional (new) calibration targets and updating the parameter draws so that the model calibrates to the new set of calibration targets. “Discussion” section concludes with a discussion of other applications of this method, implications for modelers and remaining questions to be addressed in future work.
Calibrating the microsimulation model
The microsimulation model for colorectal cancer
The ColoRectal Cancer Simulated Population Incidence and Natural history model [41, 42] describes the natural history of CRC based on the adenomacarcinoma sequence [20, 26]. Four model components describe the natural history of CRC: (1) adenoma risk; (2) adenoma growth; (3) transition from adenoma to preclinical cancer; and (4) transition from preclinical to clinical cancer (sojourn time). CRCSPIN has been used to provide guidance to the Centers for Medicare and Medicaid Services (CMS) [51] and to inform U.S. Preventive Services Task Force CRC screening guidelines [17]. We provide an overview of CRCSPIN 2.0 [43], which contains 22 calibrated parameters, \(\theta\), and prior distributions which are truncated normal or uniform, as informed by prior knowledge from published literature and previous calibration exercises. We refer the reader to [43] and online at cisnet.cancer.gov [28] for more detail.
Adenoma risk
Adenomas are assumed to arise according to a nonhomogeneous Poisson process with a piecewise linear ageeffect. The ith agent’s baseline instantaneous risk of an adenoma at age \(a=20\) years is given by \(\psi _i(20)= \exp (\alpha _{0i} + \alpha _1 \text{ female}_i\)) where \(\alpha _{0i} \sim N(A,\sigma _\alpha )\) and \(\alpha _1\) captures the difference in risk for women (female\(_i=1\) indicates agent i is female). Adenoma risk changes over time, generally increasing with age, a process we model using a piecewise linear function for logrisk with knots at ages 50, 60, and 70 and assuming zero risk before age 20 [18]:
Adenoma growth
For each adenoma, we simulate a hypothetical time to reach 10mm, assuming that \(t_{10mm}\) has a Frèchet distribution with shape parameter \(\beta _1\), scale parameter \(\beta _2\), and cumulative distribution function given by
for \(t \ge 0\). We allow different scale and shape parameters for adenomas in the colon and rectum, using the notation \(\beta _{1c}\) and \(\beta _{2c}\) for the colon, and \(\beta _{1r}\) and \(\beta _{2r}\) for the rectum.
Adenoma size at any point in time is simulated using the Richard’s growth model, with a calibrated parameter that allows for a wide range of sigmoidal growth patterns [48]. The diameter of the jth adenoma in the ith agent at time t after initiation is given by
where \(d_0=1\) mm is the minimum adenoma diameter in millimeters (mm) and \(d_\infty =50\) is the maximum adenoma diameter. The calibrated parameter p determines the shape of the growth curve. The growth rate for the jth adenoma within the ith agent, \(\lambda _{ij}\), is calculated by setting \(t=t_{\mathrm{10\,mm}}\) and \(d_{ij}=10\) in Eq. (3).
Transition from adenoma to preclinical invasive cancer
For the jth adenoma in the ith agent, the size at transition to preclinical cancer (in mm) is simulated using a lognormal distribution; the underlying (exponentiated) normal distribution is assumed to have standard deviation \(\sigma _{\gamma }\) and mean
where rectum\(_{ij}\) is an indicator of rectal versus colon location and age\(_{ij}\) is the age at adenoma initiation in decades, centered at 50 years. Based on this model, the probability that an adenoma transitions to preclinical cancer increases with size. Most adenomas do not reach transition size and small adenomas are unlikely to transition to cancer.
Sojourn time
Sojourn time is the time between the transition from preclinical (asymptomatic) CRC to clinical (symptomatic and detected) cancer. We simulate sojourn time using a Weibull distribution with survival function
for preclinical cancer in the colon, and assume a proportional hazards model, with hazard ratio \(\exp (\lambda _3\text{ rectum}_{ij})\), to allow sojourn time to systematically differ for preclinical cancers in the colon and rectum.
Simulation of lifespan and colorectal cancer survival
Once a cancer becomes clinically detectable, we simulate stage and tumor size at clinical detection based on Surveillance, epidemiology, and end results (SEER) data from 1975 to 1979, the most recent period prior to widespread dissemination of CRC screening tests [27]. Survival time after CRC diagnosis is based on the first diagnosed CRC and depends on age, sex, cancer location, and stage, and is simulated using relative survival estimates from analysis of SEER data from individuals diagnosed with CRC from 1975 through 2003 [39]. We assume proportional hazards of CRC and othercause mortality within sex and birthyear cohorts. Othercause mortality is modeled using survival probabilities based on productlimit estimates for age and birthyear cohorts from the National Center for Health Statistics Databases [29].
Calibration data
Calibration data consist of individuallevel data that are reported in aggregate in published studies. Calibration targets therefore take the form of summary statistics. Because targets come from small and larger studies, as well as registry data that results in very precisely estimated targets, the level of uncertainty varies across targets. Generating calibration targets requires simulating a set of agents with risk that is similar to the study population based on age, gender, prior screening patterns, and the time period of the study, which may affect both overall and cancerspecific mortality. This simulation can be computationally demanding, depending on the number of agents and the process used to simulate the particular target. We calibrated to 40 targets from six sources: SEER registry data ([27], 20 targets) and five published studies (20 targets). Let \(y=(y_1,\dots ,y_{J})\) denote these \(J=40\) calibration targets.
SEER colon and rectal cancer incidence rates in 1975–1979 are reported per 100,000 individuals and are a key calibration target. We assumed that the number of incident CRC cases in any year follows a binomial distribution with number of trials equal to the SEER population size. To simulate SEER incidence rates, we generated a population of individuals from aged 20 to 100 who are free from clinically detected CRC, with an age and sexdistribution that matches the SEER 1978 population. Modelpredicted CRC incidence is based on the number of people who develop CRC in the next year.
To simulate additional targets from published studies, we generated separate populations for each study that match the age and gender distribution of the sample during the timeperiod of the study. We assume that study participants are free from symptomatic (clinically detectable) CRC and have not been screened for CRC prior to the study. This is a reasonable assumption because studies used for model calibration were conducted prior to widespread screening, or were based on minimally screened samples.
Simulation of targets also requires simulating the detection and removal of adenomas and preclinical cancers. The probability of detection, or test sensitivity, is a function of lesion size, and is informed by backtoback colonoscopy studies [10, 38]. We specify the probability of not detecting (or missing) an adenoma of size s that produces miss rates that are consistent with observed findings [10, 38] and were successfully used in [43].
The miss rate functions result in sensitivities of 0.81 for adenomas of 5mm, 0.92 for adenomas of 10mm, and 0.98 for adenomas of 15mm. For preclinical cancers, we assume sensitivity that is the maximum of 0.95 and sensitivity based on adenoma size, so that colonoscopy sensitivity is 0.95 for preclinical cancers 12mm or smaller, and sensitivity is greater than 0.95 for larger preclinical cancers.
Calibration results and motivation for recalibration
A Bayesian approach to model calibration is preferred in this context to enable uncertainty quantification in model parameters and because of the ability to incorporate different sources of information through calibration targets and prior distributions. We use the Incremental Mixture Approximate Bayesian Computation (IMABC) algorithm developed by [43] and used successfully to calibrate the CRCSPIN model with 22 parameters and 40 targets. We refer the reader to the Additional file 1 for more description of the algorithm. This algorithm generates a sample of MSM parameter draws that are from an approximate posterior distribution. This approach is an approximate Bayesian version of adaptive importance sampling [36, 47] and similar to the Population Monte Carlo ABC algorithm of Beaumont et al. [1]. While While McKinley et al. [24] showed that popular ABC methods can be inefficient or fail to converge when applied to complex, highdimensional models (in their example there are 22 parameters and 18 outputs, which we refer to here as targets), the Incremental Mixture Approximate Bayesian Computation (IMABC) algorithm successfully calibrated the CRCSPIN model with 22 parameters and 40 targets.
The application of IMABC in the experiments described here was implemented using the EMEWS framework [31] and run on the Midway2 cluster at the University of Chicago Research Computing Center and the Bebop cluster at Argonne National Laboratory. To take advantage of parallel processing, we used 80 node job allocations to execute up to 318 concurrent CRCSPIN instances.
Internal model validation indicated that, as expected, the model simulated (predicted) targets were within the tolerance intervals of the observed calibration targets. However, the posterior distributions for two parameters of the sojourn time (time spent in the preclinical cancer phase) distribution largely reflected the prior distributions [43], and suggested too little time spent in the preclinical cancer phase. This finding was also noted by Rutter et al. [40] for an earlier version of the CRCSPIN model, based on the results of a model validation, and comparison to other CRC models [16, 49].
We hypothesize that sojourn time is not well estimated because sojourn time is informed by screening studies, and our targets include only a single calibration target from a screening study that is imprecise and therefore has extremely wide tolerance intervals [11]. Sojourn time and preclinical cancer detection rates are closely related, as a longer sojourn time implies more time to detect preclinical cancers. In light of this finding, we sought to include an additional calibration target that would provide more information about sojourn time.
Recalibrating the microsimulation model to incorporate new study targets
The United Kingdom Flexible Sigmoidoscopy Screening (UKFSS) Trial [12] was a large, randomized controlled trial that examined the effectiveness of a onetime flexible sigmoidoscopy. All adults aged between 55 and 64 years who were registered with participating practices were eligible to take part unless they met standard exclusion criteria. Eligible, interested respondents were randomly allocated to the screening or control groups of the trial. No screening program was in place in the UK during the period under study, so it is reasonable for the model to make the assumption of no prior screening for trial participants. One key baseline outcome reported from this study is percentage of patients with cancers detected upon screening. The rate of CRC detection was 0.0046 for men (derived from a sample size of 20,519 men) and 0.0017 for women (derived from a sample size of 20,155 women). This corresponds to an overall screen detection rate of 0.0032 based on 40,674 people. The previously included screen detection rate (the Imperiale target) was based on a small study that included only 1994 people, hence the UKFSS study is based on a sample size of more than 20 times the size, and therefore provides a much more precisely estimated target that will yield more information for estimating the model parameters.
Methods
To revise the set of calibration targets and include a more precisely estimated set of targets (this is equivalent to replacing the imprecise target with the new targets since in this case the tolerance or confidence intervals around the precisely estimated target are fully contained in the other set), one option is to “start from scratch”, applying the IMABC algorithm or a different algorithm to calibrate the model to this revised set of targets. However, this may be inefficient and does not take advantage of what has already been learned about model parameters from the first calibration. An alternative is to build from the original calibration results, and update them in light of new information.
Let \(z=(z_1,\dots ,z_{K})\) denote the new K calibration targets, to be added to the original targets y. We want to simulate from the posterior distribution of \(\theta\), given original targets y and new targets z. Note that \(p(\theta \mid y,z)\) is proportional to \(p(\theta )p(y\mid \theta )p(z\mid \theta ,y)\), or \(p(\theta \mid y)p(z\mid \theta ,y)\), where \(p(\theta \mid y)\) is the posterior distribution from the last calibration, by standard Bayesian updating rules.
We explore two different approaches to calibration, the first being a “start from scratch” method that incorporates all of the original targets as well as the new genderspecific screen detection rate targets, and the second being a sequential calibration that begins with the current set of posterior samples and updates them to fit the new targets (as well as the original ones). The structure of the IMABC algorithm makes it amenable to a “warm start”. Because we have already successfully applied the IMABC algorithm to obtain samples from an approximate posterior fit to y, our sequential approach continues to iterate through the IMABC algorithm until the target ESS of 5000 is reached. For comparative purposes, the start from scratch method begins the IMABC algorithm in the standard way by drawing from the prior. The idea behind this sequential algorithm is that, although none of the posterior samples from the original calibrated model fit the genderspecific screendetection targets well (e.g., none of the current samples are from the full posterior \(p(\theta \mid y,z)\)), they are still informative about where the posterior lies, since they are samples from \(p(\theta \mid y)\).
Let the posterior distribution approximated by the original calibration be denoted by \(p_{orig}(\theta \mid y)\). The start from scratch method and sequential calibration method are both used to calibrate the model to targets y and z, where z represents the more precise screendetection rates. Let these posteriors be denoted by \(p_{scratch}(\theta \mid y,z)\) and \(p_{seq}(\theta \mid y,z)\). We focus on two main comparisons. The first is to compare \(p_{orig}(\theta \mid y)\) to \(p_{scratch}(\theta \mid y,z)\) and then determine whether, and to what degree, adding in new study targets z changes the model parameter estimates. The second is to compare \(p_{scratch}(\theta \mid y,z)\) and \(p_{seq}(\theta \mid y,z)\) to assess whether the method applied for recalibration, e.g., the start from scratch and sequential approaches, yield the same estimated posterior distributions.
To facilitate these comparisons, we compute a 95% credible interval overlap measure (e.g., [5, 13]) for each parameter and each pair of posterior distributions (Table 2). The overlap measure based on two (overlapping) credible intervals \((L_1,U_1)\) and \((L_2,U_2)\) is computed as \(0.5(U_IL_I)/(U_1L_1)+0.5(U_IL_I)/(U_2L_2)\), where \((L_I,U_I)\) is the intersection of the two overlapping intervals. When the intervals are identical, this is equal to 1. When the intervals do not overlap at all, the measure is defined as equal to 0. This overlap measure averages the proportion of the first interval that is contained in the second, and the proportion of the second interval that is contained in the first. We also estimate the overlapping area between the two empirical posterior distributions (e.g., [6, 32, 41]). Values closer to 1 indicate more similar empirical distributions and values closer to 0 indicate the distributions overlap very little. We compute the standardized difference in means (SMD), which we define as the difference in posterior means, standardized by the estimated posterior standard deviation from the original calibration. Finally, we calculate an estimate of the Hellinger distance between the two distributions, using the “statip” package in R [34].
Results
The start from scratch approach required 73 iterations and 737,715 simulated parameters to reach the target ESS of 5000. Starting from the previous iterations, the sequential approach required 50 iterations and 500,500 simulated parameters to reach the target ESS, which translates to a significant time savings given that microsimulation model evaluations can be very expensive and each parameter draw requires multiple model evaluations as a large number of agents need to be simulated via the natural history model in order to produce a single simulated target without too much stochastic variability. The exact time of each iteration is problemspecific, as it depends on the number of targets and how expensive they are to simulate. However, in this application, the start from scratch approach required 70.6 h, whereas the sequential approach required 48.1 h. Thus the sequential approach is significantly more efficient requiring only 68% of the time required by the start from scratch approach.
In Table 1, posterior means and 95% credible intervals are displayed for the model parameters for sojourn time from each of the three calibration results. (Results for all parameters are included in the “Appendix”.) Comparing the posterior estimates shown in Table 1 and the overlap or similarity measures based on \(p_{orig}(\theta \mid y)\) and \(p_{scratch}(\theta \mid y,z)\) in Table 2 indicates that the estimates of the parameters of the sojourn time function are significantly changed by the addition of new targets z. The scale parameter and rectal location effect for the distribution of sojourn time essentially do not overlap at all, and the posterior means are very different. In Fig. 1, we compare the marginal posterior distributions for the three sojourn time parameters from \(p_{orig}(\theta \mid y)\) and \(p_{scratch}(\theta \mid y,z)\). The posterior distribution for the shape parameter and rectal location effect for sojourn time become more peaked as a consequence of the more precise information provided by the new screendetection rate targets. The sojourn time parameters are all meaningfully shifted in terms of location.
We also compare the two methodological approaches to recalibrating the model to match the new, more precise, study targets. The start from scratch approach and the sequential approach yield posterior distributions \(p_{seq}(\theta \mid y,z)\) and \(p_{scratch}(\theta \mid y,z)\) that are very similar. In Fig. 2, we compare the posterior distributions for the three sojourn time parameters previously shown, noting the similarity in the estimated distributions, which is expected given that both methods should be approximating the same posterior distribution, albeit using different approaches to get to that place. Given these findings, the sequential approach to recalibrating the model is preferred, as it takes far fewer iterations to reach convergence.
While it is clear that the new screendetection rate targets change the location and shape of the posterior distributions for the sojourn time distribution parameters, in order to more fully understand the implications of these changes on our estimates of sojourn time and cancer screening detection rates, we compute mean sojourn time (MST) for cancers in the colon and the rectum. MST is defined as \(\lambda _1\Gamma (1+1/\lambda _2)\) for cancers in the colon and \(\lambda _1\exp (\lambda _3)\Gamma (1+1/\lambda _2)\) for cancers in the rectum, where \(\Gamma\) represents the gamma function. The posterior mean and 95% credible intervals for MST for cancers in the colon based on the original calibrated model that includes the imprecise screendetection rate target of Imperiale et al. [11] are 2.32 (2.05, 2.74), whereas when calibrated to the more precise and genderspecific UKFSS targets (using the start from scratch method) they are 3.35 (2.97, 3.70). The posterior mean and 95% credible intervals for MST for cancers in the rectum based on the original calibration are 1.83 (0.85, 4.65), and calibrating to the UKFSS targets yields 8.01 (6.52, 9.33). Therefore, the more precisely estimated UKFSS targets significantly increase the estimated sojourn time for both types of cancers. This is a desirable outcome given that prior model validation work [40, 43] had suggested that CRCSPIN underestimated the length of the preclinical cancer phase, providing sojourn time estimates that were too short per validation results and comparison to other competing models [16, 49]. Additionally, as expected, the sequential recalibration method yielded estimates for MST that were virtually identical to the start from scratch method: 3.37 (2.98, 3.71) for the colon and 8.00 (6.38, 9.37) for the rectum. Figure 3 compares posterior distributions for MST for cancers in the colon and rectum across all three sets of results. As expected, the posterior distributions from the recalibrated models that include the additional genderspecific screendetection rate targets look virtually identical regardless of whether the start from scratch or sequential method is applied.
While each of the three calibrations provides a set of posterior samples of model parameters, targets are also simulated from the models during the process of model fitting. Since the models were calibrated to these targets, the simulated targets all “fit” the observed targets, however they may not be identical since the observed targets are estimated with error and therefore simulated targets are accepted as long as they fall within a specified tolerance interval of the observed targets (refer to the IMABC algorithm of Rutter et al. [43] for details). We therefore compare simulated targets (rather than model parameters) across the three sets of results. Tables 3 and 4 provide the overlap and standardized difference in means measures used to compare the original calibrated model to the updated model that incorporates the new screendetected cancer rates, and to compare the two methods for recalibration. There are some simulated targets that change when the more precisely estimated screen detection rates are added as calibration targets. The most notable differences are in the number of preclinical cancers per large lesion [2, 21]. The number of preclinical CRCs per 1,000 adenomas greater than 10mm was estimated as 16 (95% CI of (13, 22)), but with the addition of new targets, these estimates increased to 33 (95% CI of (28, 37)) (identical under the start from scratch and sequential approach to calibration).
Discussion
We have proposed a method for recalibrating a microsimulation model to incorporate new calibration targets. This method can be used when new information becomes available that requires the original targets to be supplemented with additional targets. We applied the algorithm to recalibrate a natural history model for CRC to include additional targets based on larger studies, and found that it was more efficient than starting over from scratch.
There are other settings in which this method is useful for recalibration, not just when adding new calibration targets. Sensitivity of colonoscopy is a fixed model input (noncalibrated parameter) that is specified based on direct evidence. A recent metaanalysis [52] examining the sensitivity of colonoscopy indicated that the sensitivity we assumed for a colonoscopy exam, which was based on earlier studies [10, 38], may have been too high. Test sensitivity is a critical input for any simulations of screening effects, since it determines whether an adenoma of a given size is detected and removed upon screening. Information from this recent metaanalysis suggests we consider respecifying the probability of not detecting an adenoma to incorporate lower test sensitivities for small adenomas. We used the sequential approach to recalibration to generate posterior samples under a lower test sensitivity value for small adenomas. That is, we began with the posterior samples obtained from calibration to the original, higher, assumed sensitivity, since evidence from “Recalibrating the microsimulation model to incorporate new study targets” section indicates this can speed up the calibration process relative to starting over from scratch. While the original calibration based on initially specified test sensitivity inputs required 73 iterations, this process ultimately requires only another 24 algorithm iterations to achieve an ESS that exceeds 5000.
This work suggests that modelers may be able to substantially speed up the time needed to recalibrate their models. In many MSM or agent based modeling applications, calibration is a challenging and time consuming process due to the stochastic nature of the models and the need for embedded simulations to evaluate the computationally expensive model a large number of times (e.g., [25]). This prevents modelers from fully exploring the sensitivity of the model to inputs, or assumptions, and the choice of calibration data. Thus, a model is typically only calibrated once, and that calibrated model is ultimately used for making predictions about population trends in disease outcomes and effectiveness of interventions. The methods we propose here reduce the computing and person time required to recalibrate a model, and therefore aid in the ability of modelers to more fully explore sensitivity to assumptions and make changes to data or inputs when the evidence changes. Furthermore, in modeling applications where new data are continuously made available through, for example daily or weekly case counts for infectious diseases, the sequential approach becomes even more relevant for tracking the changing underlying contexts in a timely manner by efficiently generating updated parameter estimates.
While we focused on speeding up calibration within an approximate Bayesian framework, another promising approach to efficient calibration includes the use of an emulator (e.g., [7, 45]). An emulator is statistical representation of the MSM model that is less expensive to compute. The MSM can be evaluated at a series of design points, and the emulator can then be “trained” to fit these simulated outputs. Developing an emulator for computationally expensive MSMs is appealing, however the emulator would have to be complex and accurate enough to fit the model well across the entire parameter space, which is difficult in highdimensional settings. This represents a promising area for future research.
While the methods developed here apply to situations when additional calibration data is included as a supplement to the original calibration data, future work will focus on settings where the original calibration targets are not a subset of the updated set. That is, when one wants to remove some of the original targets completely, possibly replacing them with other targets that are inconsistent with the original set. More generally, we aim to understand more fully when the sequential recalibration approach proposed here that builds on existing posterior samples does not converge to the correct posterior distribution, or is not more efficient than starting over from scratch.
Conclusion
We proposed a method for recalibrating a MSM that can be used when new information becomes available that requires the original calibration targets to be supplemented with new targets. We applied the algorithm to recalibrate a natural history model for CRC to include additional targets based on larger studies because model examination and validation indicated that more information was needed to inform sojourn time. We found that the sequential approach to calibration was more efficient than starting over from scratch. The new targets resulted in changes in the posterior distribution for mean sojourn time, suggesting longer sojourn times than previously estimated.
Availability of data and materials
The datasets used and/or analysed during the current study available from the corresponding author on reasonable request.
Abbreviations
 ABC:

Approximate Bayesian computation
 CMS:

Centers for Medicare and Medicaid Services
 CRC:

Colorectal cancer
 CRCSPIN:

ColoRectal Cancer Simulated Population model for Incidence and Natural history
 IMABC:

Incremental Mixture Approximate Bayesian Computation
 MCMC:

Markov chain Monte Carlo
 MSMs:

Microsimulation models
 MST:

Mean sojourn time
 SEER:

Surveillance, epidemiology, and end results
 SMD:

Standardized difference in means
 UKFSS:

United Kingdom Flexible Sigmoidoscopy Screening
References
Beaumont MA, Corneaut J, Marin JM, Robert CP. Adaptive approximate Bayesian computation. Biometrika. 2009;96(4):983–90.
Church JM. Clinical significance of small colorectal polyps. Dis Colon Rectum. 2004;47:481–5.
Corley DA, Jensen CD, Marks AR, Zhao WK, de Boer J, Levin TR, Doubeni C, Fireman BH, Quesenberry CP. Variation of adenoma prevalence by age, sex, race, and colon location in a large population: implications for screening and quality programs. Clin Gastroenterol Hepatol. 2013;11(2):172–80.
de Koning HJ, Meza R, Plevritis SK, Ten Haaf K, Munshi VN, Jeon J, Erdogan SA, Kong CY, Han SS, van Rosmalen J, et al. Benefits and harms of computed tomography lung cancer screening strategies: a comparative modeling study for the US Preventive Services Task Force. Ann Intern Med. 2014;160(5):311–20.
Drechsler J, Reiter J. Disclosure risk and data utility for partially synthetic data: an empirical study using the German IAB Establishment Survey. J Off Stat. 2009;25:589–603.
Garrett ES, Zeger SL. Latent class model diagnostics. Biometrics. 2000;56:1055–67.
Gramacy RB. Surrogates: Gaussian process modeling, design and optimization for the applied science. Boca Raton: CRC; 2000.
Gramacy RB. Surrogates: Gaussian process modeling, design and optimization for the applied sciences. Boca Raton: CRC; 2020. http://bobby.gramacy.com/surrogates/.
Hanushek E, Citro C. Improving information for social policy decisions: the uses of microsimulation modeling. Washington: National Academy Press; 1991.
Hixson L, Fennerty M, Sampliner R, McGee D, Garewal H. Prospective study of the frequency and size distribution of polyps missed by colonoscopy. J Natl Cancer Inst. 1990;82:1769–72.
Imperiale TF, Wagner DR, Lin CY, Larkin GN, Rogge JD, Ransohoff DF. Risk of advanced proximal neoplasms in asymptomatic adults according to the distal colorectal findings. N Engl J Med. 2000;343:169–74.
Investigators UFSST, et al. Single flexible sigmoidoscopy screening to prevent colorectal cancer: baseline findings of a UK multicentre randomised trial. Lancet. 2002;359(9314):1291–300.
Karr A, Kohnen C, Organian A, Reiter J, Saniel A. A framework for evaluating utility of data altered to protect confidentiality. Am Stat. 2006;60:224–32.
Kim JJ, Burger EA, Regan C, Sy S. Screening for cervical cancer in primary care: a decision analysis for the U.S. Preventive Services Task Force. Technical report, Agency for Healthcare Research and Quality. Contract No. HHSA290201200015I. 2017.
Kim J, Kuntz K, Stout N, Mahmud S, Villa L, Franco E, Goldie S. Multiparameter calibration of a natural history model of cervical cancer. Am J Epidemiol. 2007;166:137–50.
Knudsen AB, Hur C, Gazelle GS, Schrag D, McFarland EG, Kuntz KM. Rescreening of persons with a negative colonoscopy result: results from a microsimulation model. Ann Intern Med. 2012;157(9):611–20.
Knudsen AB, Zauber AG, Rutter CM, Naber SK, DoriaRose VP, Pabiniak C, Johanson C, Fischer SE, LansdorpVogelaar I, Kuntz KM. Estimation of benefits, burden, and harms of colorectal cancer screening strategies: modeling study for the US Preventive Services Task Force. J Am Med Assoc. 2016;315(23):2595–609.
Koh KJ, Lin LH, Huang SH, Wong JU. CARE—pediatric colon adenocarcinoma: a case report and literature review comparing differences in clinical features between children and adult patients. Medicine. 2015;94(6):e503.
Kong CY, McMahon PM, Gazelle GS. Calibration of disease simulation model using an engineering approach. Value Health. 2009;12(4):521–9.
Leslie A, Carey FA, Pratt NR, Steele RJC. The colorectal adenomacarcinoma sequence. Br J Surg. 2002;89:845–60.
Lieberman D, Moravec M, Holub J, Michaels L, Eisen G. Polyp size and advanced histology in patients undergoing colonoscopy screening: implications for CT colonography. Gastroenterology. 2008;135(4):1100–5.
Mandelblatt JS, Stout NK, Schechter CB, Van Den Broek JJ, Miglioretti DL, Krapcho M, TrenthamDietz A, Munoz D, Lee SJ, Berry DA, et al. Collaborative modeling of the benefits and harms associated with different US breast cancer screening strategies. Ann Intern Med. 2016;164(4):215–25.
Marin JM, Pudlo P, Robert CP, Ryder RJ. Approximate Bayesian computational methods. Stat Comput. 2012;22(6):1167–80.
McKinley T, Vernon I, Andrianakis I, McCreesh N, Oakley J, Nsubuga R, Goldstein M, White R. Approximate Bayesian computation and simulationbased inference for complex stochastic epidemic models. Stat Sci. 2018;33(1):4–18.
Menzies N, Soeteman D, Pandya A, Kim J. Bayesian methods for calibrating health policy models: a tutorial. Pharmacoeconomics. 2017;35(6):613–24.
Muto T, Bussey HJR, Morson BC. The evolution of cancer in the colon and rectum. Cancer. 1975;36:2251–70.
National Cancer Institute. Surveillance, epidemiology, and end results (SEER) program. Technical report, National Cancer Institute, DCCPS, Surveillance Research Program, Cancer Statistics Branch. released April 2004, based on the November 2003 submission. 2004. www.seer.cancer.gov.
National Cancer Institute. Cancer INtervention and Surveillance Modeling Network (CISNET). Technical report. 2018. https://cisnet.cancer.gov.
National Center for Health Statistics. US life tables. 2000. http://www.cdc.gov/nchs/products/pubs/pubd/lftbls/life/1966.htm.
Nelder JA, Mead R. A simplex method for function minimization. Comput J. 1965;7:308–13.
Ozik J, Collier NT, Wozniak JM, Spagnuolo C. From desktop to largescale model exploration with Swift/T. In: Proceedings of the 2016 Winter Simulation Conference (WSC). IEEE Press; 2016. p. 206–20.
Pastore M, Calcagnì A. Measuring distribution similarities between samples: a distributionfree overlapping index. Front Psychol. 2019;10:1089. https://doi.org/10.3389/fpsyg.2019.01089.
Pickhardt PJ, Choi R, Hwang I, Butler JA, Puckett ML, Hildebrandt HA, Wong RK, Nugent PA, Mysliwiec PA, Schindler WR. Computed tomographic virtual colonoscopy to screen for colorectal neoplasia in asymptomatic adults. N Engl J Med. 2003;349:2191–200.
Poncet P, The R Core Team. statip: statistical functions for probability distributions and regression. R package version 0.2.3. 2019. https://CRAN.Rproject.org/package=statip.
Press B, Teukolsky S, Vatterby W, Flannery B. Numerous receipes in Fortran—the art of scientific computing. New York: Cambridge University Press; 1992.
Raftery A, Bao L. Estimating and projecting trends in HIV/AIDS generalized epidemics using incremental mixture importance sampling. Biometrics. 2010;66:1162–73.
Ramsey S, McIntosh M, Etzioni R, Urban N. Simulation modeling of outcomes and cost effectiveness. Hematol Oncol Clin N Am. 2000;14:925–38.
Rex D, Cutler C, Lemmel G, Rahmani E, Clark D, Helper D, Lehman G, Mark D. Colonoscopic miss rates of adenomas determined by backtoback colonoscopies. Gastroenterology. 1997;112:24–8.
Rutter CM, Johnson EA, Feuer EJ, Knudsen AB, Kuntz KM, Schrag D. Secular trends in colon and rectal cancer relative survival. J Natl Cancer Inst. 2013;105(23):1806–13.
Rutter CM, Knudsen AB, Marsh TL, DoriaRose VP, Johnson E, Pabiniak C, Kuntz KM, van Ballegooijen M, Zauber AG, LansdorpVogelaar I. Validation of models used to inform colorectal cancer screening guidelines: accuracy and implications. Med Decis Mak. 2016;36:604–14.
Rutter CM, Miglioretti DL, Savarino JE. Bayesian calibration of microsimulation models. J Am Stat Assoc. 2009;104(488):1338–50.
Rutter CM, Savarino JE. An evidencebased microsimulation model for colorectal cancer: validation and application. Cancer Epidemiol Prev Biomarkers. 2010;19:1055–9965.
Rutter C, Ozik J, DeYoreo M, Collier N. Microsimulation model calibration using incremental mixture approximate Bayesian computation. Ann Appl Stat. 2019;13:2189–212.
Rutter C, Zaslavsky A, Feuer E. Dynamic microsimulation models for health outcomes: a review. Med Decis Mak. 2011;31:10–8.
Sacks J, Welch W, Mitchell T, Winn H. Design and analysis of computer experiments. Stat Sci. 1989;4:409–35.
Sisson SA, Fan Y, Tanaka MM. Sequential Monte Carlo without likelihoods. Proc Natl Acad Sci. 2007;104(6):1760–5.
Steele R, Raftery A, Edmond M. Computing normalizing constants for finite mixture models via incremental mixture importance sampling (IMIS). J Comput Graph Stat. 2006;15:712–34.
Tjørve E, Tjørve KM. A unified approach to the Richardsmodel family for use in growth analyses: why we need only two model forms. J Theor Biol. 2010;267(3):417–25.
van Hees F, Habbema JD, Meester RG, LansdorpVogelaar I, van Ballegooijen M, Zauber AG. Should colorectal cancer screening be considered in elderly persons without previous screening? A costeffectiveness analysis. Ann Intern Med. 2014;160(11):750–9.
Whyte S, Walsh C, Chilcott J. Bayesian calibration of a natural history model with application to a population model for colorectal cancer. Med Decis Mak. 2011;31:625–41.
Zauber AG, Knudsen AB, Rutter CM, LansdorpVogelaar I, Savarino JE, van Ballegooijen M, Kuntz KM. Costeffectiveness of CT colonography to screen for colorectal cancer: report to the Agency for Healthcare Research and Quality from the Cancer Intervention and Surveillance Modeling Network (CISNET) for MISCAN, SimCRC, and CRCSPIN Models. Technical report. Project ID: CTCC0608. 2009. https://www.cms.gov/medicarecoveragedatabase/details/technologyassessmentsdetails.aspx?TAId=58.
Zhao S, Wang S, Pan P, Xia T, Chang X, Yang X, Guo L, Meng Q, Yang F, Qian W, et al. Magnitude, risk factors, and factors associated with adenoma miss rate of tandem colonoscopy: a systematic review and metaanalysis. Gastroenterology. 2019;156:1661–74.
Acknowledgements
Not applicable.
Funding
This publication was made possible by Grant Numbers U01CA199335 and U01CACA253913 from the National Cancer Institute as part of the Cancer Intervention and Surveillance Modeling Network (CISNET), which supported Drs. DeYoreo and Rutter. Its contents are solely the responsibility of the authors and do not necessarily represent the official views of the National Cancer Institute. Drs. Ozik and Collier were also supported by the NIH (grant R01GM121600), the NCIDOE Joint Design of Advanced Computing Solutions for Cancer program, and through resources provided by the University of Chicago Research Computing Center, and the Laboratory Computing Resource Center at Argonne National Laboratory. This material is based upon work supported by the U.S. Department of Energy, Office of Science, under contract number DEAC0206CH11357.
Author information
Authors and Affiliations
Contributions
All authors contributed to this manuscript. M.D. and C.M.R conceived of the idea and M.D. led the research and wrote the manuscript. J.O. and N.C. implemented the algorithm runs and contributed to computational aspects of the manuscript. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
The study was approved by the Human Subjects Protection Committee at the RAND Corporation. Informed consent was not applicable.
Consent for publication
Not applicable.
Competing interests
The authors have no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Additional file 1:
Description of the Incremental Mixture Approximate Bayesian Computation Algorithm.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
DeYoreo, M., Rutter, C.M., Ozik, J. et al. Sequentially calibrating a Bayesian microsimulation model to incorporate new information and assumptions. BMC Med Inform Decis Mak 22, 12 (2022). https://doi.org/10.1186/s12911021017260
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s12911021017260