Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

calibmsm: An R package for calibration plots of the transition probabilities in a multistate model

  • Alexander Pate ,

    Roles Conceptualization, Formal analysis, Methodology, Software, Writing – original draft

    alexander.pate@manchester.ac.uk

    Affiliation Division of Imaging, Informatics and Data Science, University of Manchester, Manchester, United Kingdom

  • Matthew Sperrin,

    Roles Methodology, Writing – review & editing

    Affiliation Division of Imaging, Informatics and Data Science, University of Manchester, Manchester, United Kingdom

  • Richard D. Riley,

    Roles Methodology, Writing – review & editing

    Affiliations Department of Applied Health Sciences, School of Health Sciences, College of Medicine and Health, University of Birmingham, Birmingham, United Kingdom, National Institute for Health and Care Research (NIHR) Birmingham Biomedical Research Centre, United Kingdom

  • Ben Van Calster,

    Roles Methodology, Writing – review & editing

    Affiliations Department of Development and Regeneration, KU Leuven, Leuven, Belgium, Leuven Unit for Health Technology Assessment Research (LUHTAR), KU Leuven, Leuven, Belgium

  • Glen P. Martin

    Roles Conceptualization, Funding acquisition, Methodology, Supervision, Writing – review & editing

    Affiliation Division of Imaging, Informatics and Data Science, University of Manchester, Manchester, United Kingdom

Abstract

Background and objective

Multistate models, which allow the prediction of complex multistate survival processes such as multimorbidity, or recovery, relapse and death following treatment for cancer, are being used for clinical prediction. It is paramount to evaluate the calibration (as well as other metrics) of a risk prediction model before implementation of the model. While there are a number of software applications available for developing multistate models, currently no software exists to aid in assessing the calibration of a multistate model, and as a result evaluation of model performance is uncommon. calibmsm has been developed to fill this gap.

Methods

Assessing the calibration of predicted transition probabilities between any two states is made possible through three approaches. The first two utilise calibration techniques for binary and multinomial logistic regression models in combination with inverse probability of censoring weights, whereas the third utilises pseudo-values. All methods are implemented in conjunction with landmarking to allow calibration assessment of predictions made at any time beyond the start of follow up. This study focuses on calibration curves, but the methodological framework also allows estimation of calibration slopes and intercepts.

Results

This article serves as a guide on how to use calibmsm to assess the calibration of any multistate model, via a comprehensive example evaluating a model developed to predict recovery, adverse events, relapse and survival in patients with blood cancer after a transplantation. The calibration plots indicate that predictions of relapse made at the time of transplant are poorly calibrated, however predictions of death are well calibrated. The calibration of all predictions made at 100 days post transplant appear to be poor, although a larger validation sample is required to make stronger conclusions.

Conclusions

calibmsm is an R package which allows users to assess the calibration of predicted transition probabilities from a multistate model. Evaluation of model performance is a key step in the pathway to model implementation, yet evaluation of the performance of predictions from multistate models is not common. We hope availability of software will help model developers evaluate the calibration of models being developed.

1. Introduction

Risk prediction models enable the prediction of clinical events in either diagnostic or prognostic settings [1] and are used widely to inform clinical practice. A multistate model [2] may be used when there are multiple outcomes of interest, or when a single outcome of interest may be reached via intermediate states. For example, prediction of death after local recurrence or distant metastasis in patients with breast cancer following surgery [3]; prediction of death following progression of chronic kidney disease [4]; prediction of non-AIDS events and death in individuals living with HIV [5]. Using a multistate model for prediction is important when the development of an intermediate condition occurring post index date may have an impact on the risk of future outcomes of interest. Risk prediction models developed for use in clinical practice should be evaluated in a relevant cohort, or preferably multiple settings/cohorts, prior to implementation [6]. If the intended use of this model is known, targeted validation in a specific setting may be preferred [7]. A key part of the validation process is assessment of the calibration of the model [8]. Calibration assesses whether the predicted risks match the observed event rates in the cohort of interest. Ideally calibration curves should be produced to estimate observed event rates as a function the predicted risks over the entire distribution of predicted risk. This corresponds to a moderate assessment of calibration [9]. Methodology on this topic is well developed for binary outcomes [9], survival outcomes [10,11], and survival outcomes in the presence of competing risks [12,13], however less so for multistate models, where there is often interest in prediction of more than one outcome state, and in predictions made at landmark times.

The R [14] package mstate provides a comprehensive set of tools to develop a multistate model and estimate patient-specific predictions for a continuously observed multistate survival process [15]. mstate focuses on non-parametric and semi-parametric multistate models where the cause-specific hazards have been fitted using cox-proportional hazards models. The flexsurv package [16] builds on the functionality of mstate, allowing users to fit parametric multistate models (still using the cause-specific hazards approach), as well as an approach that uses mixture models. Both mstate and flexsurv allow fitting of clock-forward (Markov) and clock reset (Semi-Markov) models. The SemiMarkov package [17] contains functions specifically for fitting semi-Markov models. The msm package [18] focuses on fitting multistate models to continuous time processes that are observed at arbitrary times (panel data). The flexmsm package [19] provides a general estimation framework for multistate Markov processes, with flexible specification of the transition intensities. Transition intensities can be specified through Generalised Additive Models, and allows models with forward and backward transitions to be fitted. The Lexis functions from the Epi package [20] provide a way to represent and manipulate data from multistate models, and provides an interface to the mstate. For a full list of packages available for fitting multistate models, see https://cran.r-project.org/web/views/Survival.html.

There are a number of R packages which enable the estimation of calibration curves for models predicting continuous, binary, or survival outcomes [2123]. However, despite a wide range of packages for developing multistate models, currently no software exists to aid researchers in assessing the calibration of a multistate model that has been developed for the purposes of individual risk prediction. The R package calibmsm has been developed to enable researchers to estimate calibration curves and scatter plots for the transition probabilities between any two states. The methodology is based on three approaches outlined previously [24], which focused on assessing the calibration of the transition probabilities out of the starting state. The work in this paper extends the framework to assess the calibration of transition probabilities out of any state at any time using landmarking [25,26], provides more details on estimation of the inverse-probability of censoring weights (where relevant), and demonstrates the process for estimating confidence intervals. calibmsm is available from the Comprehensive R Archive Network at https://CRAN.R-project.org/package=calibmsm, or can be installed from the following GitHub repository [27].

De Wreede et al. [15], used data from the European Society for Blood and Marrow Transplantation [28] to showcase how to develop a multistate model for clinical prediction of outcomes after bone morrow transplantation in leukemia patients (Fig 1). In this study, we show how to assess the calibration of a model developed on the same EBMT data as a way of illustrating the syntax and workflows of calibmsm. This clinical example also highlights some important differences between the methods in how they deal with informative censoring and computational feasibility, which may impact future uptake of the methods. Details on the methodology are given in section 2. The clinical example, including steps for data preparation and production of calibration plots are given in section 3. Section 4 contains a discussion and summary.

thumbnail
Fig 1. A six-state model for leukemia patients after bone marrow transplantation.

Figure taken from De Wreede et al [15].

https://doi.org/10.1371/journal.pone.0320504.g001

2. Methods and theory

2.1 Setup

Let be a multistate survival process with states. We assume a multistate model has already been developed and we want to assess the calibration of the predicted transition probabilities, , in a cohort of interest. The transition probabilities are the probability of being in state at time , if in state at time , where . To assess the calibration of the multistate model, we must estimate observed event probabilities:

In a well calibrated model, the transition probabilities will be equal to the observed event probabilities.

In the absence of censoring, can be estimated using cross sectional calibration techniques in a landmark [25,26] cohort of individuals who are in state at time (i.e., methods to assess the calibration of models predicting binary or multinomial outcomes). In the presence of censoring, calibration must be assessed in this landmark cohort of individuals either using cross sectional techniques in combination with inverse probability of censoring weights, or through pseudo-values. These approaches were previously proposed and evaluated in a simulation study [24], but were restricted to assessing calibration out of the starting state at time . The theory is summarised and revised to allow assessment of calibration out of any state at any time in sections 2.2–2.6.

2.2 Binary logistic regression with inverse probability of censoring weights (BLR-IPCW) calibration curves

The first approach produces calibration curves using a framework for binary logistic regression models in conjunction with inverse probability of censoring weights to account for informative censoring (BLR-IPCW). Let be an indicator for whether an individual is in state at time . is then modeled using a flexible approach with as the sole predictor. This model is fit in the landmark cohort (in state at time ) of individuals who are also still uncensored at time . This cohort is weighted using inverse probability of censoring weights (see section 2.4). We suggest using a loess smoother [29]:

(1)

or a logistic regression model with restricted cubic splines [30]:

(2)

Any flexible model for binary outcomes could be used, but these are the most common and are implemented in this package. Observed event probabilities are then estimated as fitted values from these models. These are commonly referred to as ‘predicted-observed risk’ values. The calibration curve is plotted using the set of points . To obtain unbiased calibration curves, the assumption that each outcome is independent from the censoring mechanism in the reweighted population must hold.

2.3 Multinomial logistic regression with inverse probability of censoring weights (MLR-IPCW) calibration scatter plots

The second approach produces calibration scatter plots using a framework for multinomial logistic regression models with inverse probability of censoring weights (MLR-IPCW). Let be a multinomial indicator variable taking values such that if an individual is in state at time . The nominal recalibration framework of Van Hoorde et al. [31,32], is then applied in the landmark cohort of individuals uncensored at time , weighted using inverse probability of censoring weights (section 2.4). First calculate the log-ratios of the predicted transition probabilities:

Then fit the following multinomial logistic regression model:

where is an arbitrary reference category which can be reached from state , takes values in the set of states that can be reached from state , and where is a vector spline smoother [33]. Observed event probabilities are then estimated as fitted values from this model. This results in a calibration scatter plot rather than a curve due to all states being modeled simultaneously, as opposed to BLR-IPCW, which is a “one vs all” approach. The scatter occurs because the observed event probabilities for state vary depending on the predicted transition probabilities of the other states. This is a stronger [9] form of calibration than that evaluated by BLR-IPCW, and will also result in observed event probabilities which sum to 1. In future iterations of calibmsm functionality will be added to produce smoothed curves estimated from these scatter plots. To obtain unbiased calibration curves, the assumption that the outcome is independent from the censoring mechanism in the reweighted population must hold.

2.4 Estimation of the inverse probability of censoring weights

The estimand for the weights is , the inverse of the probability of being uncensored at time if in state at time :

(3)

where denotes the history of the multistate survival process up to time , including the transition times, and is a set of baseline predictor variables believed to be predictive of the censoring mechanism. Note that may be the same as, but is not restricted to, the variables used for prediction when developing the multistate model. First the estimator is calculated by developing an appropriate survival model. The outcome in this model is the time until censoring occurs. Moving into an absorbing state prevents censoring from happening and is treated as a censoring mechanism in this model (i.e., a competing risks approach is not taken when fitting this model). is explicitly conditioned on when defining because the weights must reflect that censoring can no longer be observed for an individual if they enter an absorbing state at some time . Therefore, in calibmsm, the weights are estimated as:

Note that if the censoring mechanism is not conditionally independent from the outcome process given , i.e., the rate of censoring changes depending on outcome state occupancy, then this approach will be invalid. Instead, the outcome history up until time must be conditioned on when estimating the weights, as specified in equation (3). By default, is estimated in calibmsm using a cox proportional hazards model where all predictors are assumed to have a linear effect on the log-hazard. These conditions are highly restrictive. Users can therefore also input their own vector of weights which is strongly recommended. calibmsm has been developed to provide a framework for assessing calibration, and the packages’ primary purpose is not estimation of inverse probability of censoring weights, however, future versions may allow for more flexible models to estimate the weights. Given the BLR-IPCW and MLR-IPCW approaches are both reliant on correct estimation of the weights, we encourage users to take the time to carefully estimate these using a well specified model. The limitations of using the calibmsm internal functions for estimating the weights in this clinical example (section 4) are discussed in more detail later, and explored in vignette Sensitivity-analysis-for-IPCWs [27].

Stabilised weights can be estimated by multiplying by the weights by the mean probability of being uncensored:

The numerator can be estimated using an intercept only model, and note there is no dependence on .

Another option is to estimate , which is the inverse of the probability of being uncensored at time if uncensored at time :

This can be estimated using the same approach as for , except there is no requirement to be in state when landmarking at time . If the censoring mechanism is conditionally independent from after conditioning on , then , and any consistent estimator for will be a consistent estimator of . The advantage is that is calculated by developing a model in the cohort of individuals uncensored at time , which is a larger cohort than those uncensored and in state at time . Therefore will be a more precise estimator than . On the contrary, if the rate of differs depending on , there is a risk of bias in estimation of the weights. We therefore recommend using the estimator unless sample size (number of individuals in state at time ) is low, which may be assessed using sample size formula for prediction models with time-to-event outcomes [34]. If the sample size is deemed insufficient, one may consider using , but the risk of bias associated with this estimator must be carefully considered.

We note the importance of using inverse probability of censoring weights, even if the censoring mechanism is believed to be completely at random. When a multistate model has an absorbing state (which is the case for most), entry into this state will prevent censoring from happening. This induces a dependence between the outcome and the censoring mechanism which must be adjusted for using inverse probability of censoring weights. This issue was highlighted in the supplementary material of previous work [24].

2.5 Pseudo-value calibration plots

The third approach produces calibration curves using pseudo-values [35,36]. Pseudo-values can be used in place of the outcome of interest in a regression model if some outcomes are not observed due to right censoring. This is the case in models (1) and (2). For certain estimators (where estimates the expectation of the outcome it is replacing), the pseudo-value for individual is defined as:

,

where is equal to estimated in a cohort without individual . One such estimator for the outcomes in models (1) and (2) given the underlying multistate survival process, is the Landmark Aalen-Johansen estimator [37], which estimates the expectation of in the landmark cohort of individuals in which calibration is being assessed. The resulting pseudo-values are a vector with elements, one for each possible transition, for every individual . These pseudo-values can replace the outcome in equations (1) and (2) in order to estimate .

Pseudo-values are based on the same assumptions as the underlying estimator . The Landmark Aalen-Johansen estimator is valid for both Markov and non-Markov multistate models. However, it does make the assumption of random censoring. According to Kleinbaum and Klein [38], this means that “subjects who are censored at time should be representative of all the study subjects who remained at risk at time with respect to their survival experience”. The approach to alleviate this is to estimate the pseudo-values within sub-groups of individuals, now making the assumption of random censoring within the specified subgroups. This can be done by calculating the pseudo-values within subgroups defined by baseline predictors, or subgroups defined by the predicted transition probabilities . Both options are implemented in this package. When pseudo-values are calculated within subgroups, they are still used as the outcome in models (1) and (2) in the same way. Note that the pseudo-values are continuous, as opposed to binary , but the link function in model (1) remains the same to ensure are between zero and one. Here, are commonly referred to as ‘pseudo-observed risk’ values.

2.6 Estimation of confidence intervals

Confidence intervals for both BLR-IPCW and pseudo-value calibration curves can be estimated using bootstrapping. While theoretically feasible, it is currently unclear how to present confidence intervals for each data point in the calibration scatter plots produced by MLR-IPCW, and therefore these are omitted. A process for estimating the confidence intervals around the BLR-IPCW calibration curves is as follows:

  1. Resample validation dataset with replacement
  2. Landmark the dataset for assessment of calibration
  3. Calculate inverse probability of censoring weights
  4. Fit the preferred calibration model in the landmarked dataset (restricted cubic splines or loess smoother)
  5. Generate observed event probabilities for a fixed vector of predicted transition probabilities (specifically the predicted transition probabilities from the non-bootstrapped landmark validation dataset)

This will produce a number of bootstrapped calibration curves, all plotted over the same vectors of predicted transition probabilities. Taking the and percentiles of the observed event probabilities for each predicted transition probability gives the required confidence interval around the estimated calibration curve. To estimate confidence intervals for the pseudo-value calibration curves using bootstrapping, the same procedure is applied except the third step is replaced with ‘calculate the pseudo-values within the landmarked bootstrapped dataset’. This will be highly computationally demanding as the pseudo-values must be estimated in every bootstrap dataset. This process is integrated into calibmsm internally, however code for how to implement these steps manually is also provided in the vignette BLR-IPCW-manual-boostrap [27].

If using the pseudo-value method, confidence intervals can however be calculated using closed form estimates of the standard error when making predictions of the observed event probabilities (i.e., when obtaining fitted values from models (1) and (2)). We recommended this due to the computational burden of bootstrapping the confidence intervals around the pseudo-value calibration curves. There are a number of issues with estimating parametric confidence intervals for the BLR-IPCW calibration curves. Firstly, a robust sandwich-type estimator should be used to estimate the standard error [39], which are known to result in conservative confidence intervals. On the contrary, the size of the confidence interval will be underestimated as uncertainty in estimation of the weights is not considered. Due to the impact of these two factors, we recommend using bootstrapping to estimate the confidence intervals for BLR-IPCW calibration curves.

3. Description of package functions and interface

The methods for estimating the calibration curves are implemented using the calib_msm function. A step-by-step guide to this process is given in Fig 2.

thumbnail
Fig 2. Flowchart detailing steps for estimating calibration plots using calibmsm::calib_msm.

https://doi.org/10.1371/journal.pone.0320504.g002

Step 1 is to obtain access to the multistate model that is being evaluated. calibmsm is designed to assess the calibration of a pre-existing model. Details on development of multistate models is given elsewhere [2,15,16].

Step 2 is to derive the validation dataset that calibration will be assessed in. This must be formatted in two ways, and will be used as the inputs for the calib_msm arguments data_raw and data_ms. The data_raw argument requires the validation cohort in a data.frame format with one row per individual. For methods BLR-IPCW and MLR-IPCW, data_raw must contain variables dtcens (censoring time) and dtcens_s (censoring indicator), where dtcens_s = 1 if the individual is censored at time dtcens, dtcens_s = 0 otherwise, plus any baseline predictors used to estimate the weights. For the pseudo-value approach, data_raw must contain any baseline predictors being used to group individuals before calculating the pseudo-values. The data_ms argument requires a dataset of class msdata. A dataset of this class can be derived using the package mstate [15], and will contain a separate row for every transition for each individual. Both data_ms and data_raw should contain a corresponding patient ID variable id. The data_raw and data_ms arguments contain the same information about individuals in the validation cohort in different formats, and data_ms can be derived from data_raw using the function msprep from mstate [15]. The data_raw argument is used to build the calibration models and estimate the calibration curves, while the data_ms argument is used to implement the landmarking for all methods, and estimate the Aalen-Johansen estimator for the pseudo-value approach. They are specified separately to reduce computation time when bootstrapping. See section 4.1 for an example of datasets meeting these criteria and details on how they are derived.

Step 3 is to choose the state and time at which calibration will be assessed. The predicted transition probabilities out of state at time for individuals in the validation cohort must then be estimated using the multistate model that is being evaluated. We defer to de Wreede et al [15] and Jackson [16] for details on how to do this. The estimated transition probabilities are then specified in calib_msm through the tp_pred argument, which must contain a column for each transition , even if the transition from to has zero probability. The rows in tp_pred must be ordered in the same way as those in data_raw.

Step 4 is to choose a method for assessing calibration. When running calib_msm, the user can specify the method for assessing calibration through the calib_type argument, and the type of non-linear curve through curve_type argument. It is also recommended to specify other relevant input parameters, such as the variables for grouping individuals if using the pseudo-value approach, or the number of knots if using restricted cubic splines. A list of the function arguments and their required inputs is given in Table 1. For full details on function inputs we refer to the package documentation, obtained through help(package = “calibmsm”). For users new to R, we recommend starting with the book ‘R for Data Science’ [41].

thumbnail
Table 1. Arguments for the function calibmsm::calib_msm and required input.

https://doi.org/10.1371/journal.pone.0320504.t001

The methods in calibmsm require continuously observed data, however are agnostic to the type of multistate model used to estimate the transition probabilities. This includes Markov, Semi-Markov or non-Markov models, and non-parametric, semi-parametric or parametric models. A dataset of class msdata from mstate is required as input, however this is only required to apply landmarking, and determine the occupied state for each individual at time . The estimated transition probabilities, supplied through tp_pred can be estimated using any statistical software.

Once the data for the calibration curves or scatter plots has been estimated using calib_msm, they can be plotted using the S3 generic, plot. Plot methods have been written for three output classes of calib_msm: calib_blr, calib_mlr and calib_pv. Separating these processes allows users to manually estimate bootstrapped calibration curves (see vignette BLR-IPCW-manual-bootstrap) [27] using the output from calib_msm. It also allows users the flexibility of producing their own plots utilising the full functionality of ggplot2, rather than being reliant on the S3 generics provided.

4. Clinical example and typical program run

4.1 Clinical setting and data preperation

We utilise data from the European Society for Blood and Marrow Transplantation [28], containing multistate survival data after a transplant for patients with blood cancer. The start of follow up is the day of the transplant and the initial state is alive and in remission. There are three intermediate events (: recovery, : adverse event, or : recovery + adverse event), and two absorbing states (: relapse and : death). This data is available from the mstate package. We assume the user of calibmsm has experience with handling the type of data used to develop a multistate model as outlined by De Wreede et al [15].

Four datasets are provided to enable assessment of a multistate model fitted to these data. The code for deriving all these datasets is provided in the source code for calibmsm [27]. The first is ebmtcal, which is the input for argument data_raw. This data is the same as the ebmt dataset provided in mstate, with two extra variables derived: time until censoring (dtcens) and an indicator for whether censoring was observed (dtcens_s = 1) or an absorbing state was entered (dtcens_s = 0). This dataset contains baseline information on year of transplant (year), age at transplant (age), prophylaxis given (proph), and whether the donor was gender matched (match), and information on time until entry into the different outcome states. Such a dataset would need to be constructed by the user after prospective data collection or manipulation of observational data.

The second dataset provided is msebmtcal, which is the input for argument data_ms. This is a dataset of class msdata, and has been derived from the ebmt dataset by applying the processes and functions from the package mstate [15]. It contains all transition times, an event indicator for each transition, as well as a trans attribute containing the transition matrix. Note that while the data on transition times is not required to be present in ebmtcal in order to run calib_msm, it is required in order to derive the dataset msebmtcal from ebmtcal.

In the work of De Wreede et al. [15], the focus is on predicting transition probabilities made at times and days, across a range of follow up times , and comparing prognosis for patients in different states . In this study we also focus on assessing the calibration of the transition probabilities made at these times. We assess calibration of the transition probabilities at years, a common follow up time for cancer prognosis, but calibration of the model may vary for other values of . We estimate transition probabilities for each individual by developing a model as demonstrated in de Wreede et al. [15], following the theory of Putter et al [2].

The predicted transitions probabilities from each state at times and are contained in stacked datasets tps0 and tps100 respectively. A leave-one-out approach was used when estimating these transition probabilities. This means each individual was removed from the development dataset when fitting the multistate model to estimate their transition probabilities. This approach allows validation to be assessed in the same dataset that the model was developed with minimal levels of in-sample optimism. Note that for tps100 the predicted probabilities for some states are equal to . This is because no individuals in state at time transition into states or . This may be due to the definition of an adverse event having to occur within a certain number of days post transplant.

4.2 Calibration plots for the transition probabilities out of state at time

We start by producing calibration curves for the predicted transition probabilities out of state at time . Given all individuals start in state , there is no need to consider the transition probabilities out of states at . Calibration is assessed at follow up time ( days). We start by extracting the predicted transition probabilities from state at time from the object tps0. These are the transition probabilities we aim to assess the calibration of.

We first evaluate calibration using the BLR-IPCW approach by specifying calib_type = “blr”. We choose to estimate the calibration curves using restricted cubic splines, although the use of loess smoothers would be equally valid. When using restricted cubic splines, the number of knots must always be specified by the user, and 3 knots are chosen here given the reasonably small size of the dataset. Weights are estimated using the internal estimation procedure with baseline predictor variables year, agecl, proph and match. The w_max_follow=t_eval argument censors individuals at t_eval before fitting the model used to estimate the weights, i.e., a “stopped cox” approach [42]. This decision was made to help meet the proportional hazards assumption as there is differential follow up for individuals in different year groups (see vignette Sensitivity-analysis-for-IPCWs [27] for more details). The w_landmark_type argument assigns whether weights are estimated using all individuals uncensored at time , or only those uncensored and in state at time , as discussed in section 2.4 The maximum weight (w_max = 10) and stabilisation of weights (w_stabilised = TRUE) are left as default. Weights can also be manually specified using the weights argument. We request 95% confidence intervals for the calibration curves calculated through bootstrapping with 200 bootstrap replicates.

The first element of dat_calib_blr (named plotdata) contains 6 data frames. One for the calibration curves of the transition probabilities into each of the six states, Each data frame contains five columns, id: the identifier of each individual; pred: the predicted transition probabilities; obs: the observed event probabilities; obs_lower and obs_upper: the confidence interval for the observed event probabilities. The second element (named metadata) is a metadata argument containing information about the data and chosen calibration analysis. The plot data and metadata can be viewed using the print and metadata commands respectively. However, it is recommended to get acquainted with the underlying object structure, as accessing the plot data will be useful if wanting to customise plots or apply bootstrapping manually.

The BLR-IPCW calibration curves can then be generated by applying the plot function to the output from calib_msm. Note, when marginal density curves are requested (default), grid::grid_draw is the preferred approach to display this plot in the viewer. The calibration curves (Fig 3) indicate the level of calibration is different for the transition probabilities into each of the different states. The calibration into states and 6 looks the best. State has good calibration over the majority of the predicted risks but over predicts for individuals with the highest predicted risks. Transition probabilities into states and are over and under predicted respectively over most of the range of predicted risks. Importantly the calibration of the transition probabilities into state (Relapse), a key clinical outcome in this clinical setting, is extremely poor.

thumbnail
Fig 3. BLR-IPCW calibration curves out of state j= 1 at time s = 0.

https://doi.org/10.1371/journal.pone.0320504.g003

Next we use the pseudo-value approach to assess calibration by specifying calib_type = “pv”. Instead of specifying how the weights are estimated, we now specify variables to define groups within which pseudo-values will be calculated (see section 2.5). The goal is to induce uninformative censoring within the chosen subgroups. We chose to calculate pseudo-values in individuals with the same year of transplant (pv_group_vars = c(“year”)), and then split individuals into a further three groups defined by their predicted risk (pv_n_pctls = 3). The number of percentiles should be increased in bigger validation datasets, although guidance on specific numbers is currently lacking. Year of transplant was identified as a subgrouping variable because a later transplant resulted in a shorter possible follow up, an earlier administrative censoring time, and it was therefore highly predictive of being censored. Your data should be explored to identify appropriate variables for subgrouping (see vignette Evaluation-of-estimation-of-IPCWs). A parametric confidence interval is estimated as recommended in section 2.6.

The pseudo-value calibration curves (Fig 4) are largely similar to the BLR-IPCW calibration curves (Fig 3). The agreement in the calibration curves from two completely distinct methods provides some reassurance on the validity of the assessment of calibration. This is with the exception of state , where the pseudo-value calibration plot indicates the transition probabilities are well calibrated, but the BLR-IPCW calibration plot indicates the transition probabilities under predict. In a situation like this, we recommend testing the assumptions made by each of the methods to try and diagnose which are most likely to hold, and what may be driving the difference. In this particular example, we hypothesised that the cox-proportional hazards model for estimating the inverse probability of censoring weights may be misspecified, in particular the proportional hazards assumption, due to the strong effect of year of transplant on the censoring mechanism. We explored this theory in more detail (see vignette Sensitivity-analysis-for-IPCWs), but found little change when estimating the weights using a flexible parametric survival model. Instead, we identified that this may be caused by a difference in the censoring mechanism for individuals in the adverse event state, as it appeared these individuals were less likely to be censored. This will bias the results from the BLR-IPCW and MLR-IPCW methods unless the weights are conditional on the amount of time spent in each outcome state, something which calibmsm is not currently set up to do. Although it’s not possible to be certain that the individuals in the adverse event state were less likely to be censored purely from looking at the data, we concluded it was a strong possibility, and that the BLR-IPCW calibration curves may be biased in this particular clinical example.

thumbnail
Fig 4. Pseudo-value calibration curves out of state j= 1 at time s = 0.

https://doi.org/10.1371/journal.pone.0320504.g004

Next we use the MLR-IPCW to evaluate calibration which produces a calibration scatter plot by specifying calib_type = “mlr”. The inputs for calculating the weights are the same as for the BLR-IPCW approach, but a confidence interval is no longer requested which is not possible for the MLR-IPCW approach.

The MLR-IPCW calibration scatter plots, produced using plot are contained in Fig 5. Within each plot for state , there is a large amount of variation in calibration of the transition probabilities depending on the predicted transition probabilities into states . One valuable insight from these plots is that the variance in the calibration of the transition probabilities into state , is considerably smaller than that of state , despite these two states both having good calibration according to the BLR-IPCW plots (arguably state looked better calibrated). This means the calibration of the transition probabilities into state remains reasonably consistent, irrespective of the risks of the other states. On the contrary, the calibration of the predicted transition probabilities into state is more dependent on the predicted transition probabilities of the other states. These plots can also be used to identify groups of individuals where calibration is poor. For example, there is a cluster of individuals with a predicted-observed risk > 17.5% for state 3, where the model is under-predicting. These are all individuals aged 20–40, year group 1995–1998, no gender mismatch and no prophylaxis.

In practice, we recommend researchers produce calibration curves using all three methods. The MLR-IPCW approach is a stronger [9] form of calibration assessment than the BLR-IPCW and pseudo-value approaches, but all three provide important contextual information. Note that we hypothesized that the inverse probability of censoring weights may not be accurate in this example due to a censoring mechanism which changes depending on outcome state occupancy. Using the current approach for estimating the weights, this will result in biased calibration plots for the BLR-IPCW and MLR-IPCW methods, having a particular effect on the plots for state 3 (adverse event). For the following section, we therefore proceed using only the pseudo-value method.

thumbnail
Fig 5. MLR-IPCW calibration scatter plots out of state j= 1 at time s = 0.

https://doi.org/10.1371/journal.pone.0320504.g005

4.3 Calibration plots for the transition probabilities out of states and at time

In the work of De Wreede et al. [15], focus then shifts to comparing transition probabilities when depending on whether an individual has had an adverse event (state $3$) or remains in state (post transplant). Our focus therefore now shifts to assessing the calibration of these transition probabilities. This is done through landmarking as described in section 2. We start by extracting the predicted transition probabilities from state and at time from the object tps100. These are the transition probabilities we aim to assess the calibration of.

The process for estimating the calibration curves remains the same, changing the inputted values j and s, and specifying the appropriate predicted transition probabilities to the argument tp_pred. We start by producing the calibration plots for and using pseudo-value (Fig 6) method.

thumbnail
Fig 6. Pseudo-value calibration curves out of state j= 1 at time s = 100.

https://doi.org/10.1371/journal.pone.0320504.g006

There are only four calibration plots because no individuals in state at time are in states (adverse event) or (recovery + adverse event) after days. We believe this is due to the definition of an adverse event occuring within 100 days, but as secondary users of the data, cannot be sure about this. The calibration of the predicted transition probabilities according is very poor. Only for state is the observed risk a monotonically increasing function of the predicted transition probabilities. The confidence intervals are very large. For states and , we cannot rule out that the poor calibration is a result of sampling variation as opposed to a poorly performing prediction model. A larger validation dataset would be required to get to the bottom of this. There is a major issue with the calibration of the transition probabilities of staying in state , as the predicted risk is inversely proportional to the observed event rate.

Next we produce calibration plots for and (Fig 7).

thumbnail
Fig 7. Pseudo-value calibration curves out of state j= 3 at time s = 100.

https://doi.org/10.1371/journal.pone.0320504.g007

Again there are only four possible states that an individual may transition into, although this includes states (adverse event) and (recovery + adverse event), instead of (post transplant) and (recovery). This is because once an individual has entered state , they cannot move backwards into states or . The calibration plots are better than for . For transitions into states and , the calibration curves are monotonically increasing and comparatively close to the line of perfect calibration, although the confidence intervals are still quite large. Again the calibration of state is very poor. This makes it difficult to base any clinical decisions on the predicted transition probabilities for relapse out of states or at time , whereas making clinical decisions based on the risk of death () after survival for days is more viable, as this was well calibrated for both and .

We provide an overview of interpretating the calibration curves for this clinical example in Box 1.

Interpretation of the calibration results.

Please note, the model being evaluated was originally developed to showcase the functionality of the mstate package and was never intended for clinical use. This is a hypothetical discussion to illustrate how the calibration plots may be interpreted.

Assessing the calibration of multi-state clinical prediction models requires consideration of each of the states of the model, with a requirement for the re to be good calibration across all states before the model could be used in clinical practice. We have provided three methods to assess calibration (see Section 2), and we recommend assessing calibration using each, so that the results can be compared.

The calibration curves shown in Figs 35, which consider predictions out of the recovery state a time 0, show that there is good agreement between the observed and predicted risks for some – but not all – of the states. The results tell us that the transition probabilities of remaining in state (transplant) are pre-dominantly over-predicting. Specifically, the model is over-estimating the predicted risk of someone not recovering, having an adverse event, experiencing relapse or death following transplant. The transition probabilities of being in state 2 (recovery) or state 5 (adverse event + recovery) are either under or over-predicting depending on the predicted risk value. A key clinical outcome in this clinical setting is the risk of relapse (state 5), with these results showing that the model should not be used to inform risk estimation for this state, since the calibration of state 5 is extremely poor.

On the contrary, the calibration of transition probabilities for state 4 (adverse event + recovery) and state 6 (death) are reasonably well calibrated. Checking for consistency in conclusions across the three calibration methods is always recommended as it may reveal important insights from the analysis. Indeed, we found differences in the calibration results of state 3 across the three methods (as discussed in the main paper). This led to further investigation, which concluded that the calibration plots from the BLR-IPCW and MLR-IPCW approaches may be biased in this setting, in particular state 3 (adverse event). This led us to focus on the pseudo-value calibration plots which indicated the transition probabilities into the adverse event state were well calibrated.

The pseudo-value calibration curves in Figs 6 and 7, which consider predictions out of the recovery and adverse event states at time 100, show very poor agreement between the observed and predicted risks.

In our opinion, finding that there are some states with miscalibrated transition probabilities informs us that the predicted risks from the model should not be used to inform clinical decision-making. For example, it is clear that the model should not be used to aid clinical decision-making around relapse risk following transplant, especially when making predictions 100 days post-transplant. On the contrary, one may argue that if using the model to inform a clinical decision based solely around the risk of death (state 6), the model is appropriate. However, if the goal is not to make a clinical decision informed by the risk of all the states simultaneously, a multistate model is not necessary in the first place.

From this, one would conclude revisions to the model are needed prior to model implementation. For example, one could test the inclusion of interaction terms between predictors, or one could reduce the number of predictors in specific transition models to reduce overfitting. However, diagnosing which part of the multi-state model is causing certain transition probabilities to be miscalibrated is multifaceted. Indeed, miscalibrated transition probabilities could be driven by errors in any of the intermediate competing risks models into that state. Further research is needed around methods to help such identification.

5. Discussion

Multistate models are a unique tool for prediction, handling both competing risks and the occurrence of intermediate health states in the same model. Development of multistate models for prediction is becoming more common, yet validation of such models is still very uncommon. A major barrier to the implementation of statistical techniques is often the availability of software [43]. calibmsm has been developed to aid in the implementation of techniques to assess the calibration of the transition probabilities from a multistate model. This paper has extended previously proposed methods for assessing the calibration of the transition probabilities out of the initial state [24], to the transition probabilities out of any state at any time . While package development has focused on multistate models, calibmsm could, in theory, be used to assess the calibration of predicted risks from a range of other models, including: competing risks models [2], standard single outcome survival models, where predictions can be made at any landmark time (note these are both special cases of a multistate model), dynamic models [26,44], and any model which utilises information post baseline to update predictions [45].

All three methods (BLR-IPCW, MLR-IPCW and pseudo-value) have been shown to give an unbiased assessment of calibration under random censoring mechanisms, and a predominately unbiased assessment of calibration when there is a strong association between the outcome and censoring mechanisms that can be explained by baseline covariates [24]. This paper found broadly similar evaluation of calibration when using the BLR-IPCW and pseudo-value methods, however there were discrepancies in the evaluation of calibration of the transition probabilities into state . This is an indicator that the assumptions underpinning either one of the methods could be violated. This was explored in detail (see vignette Sensitivity-analysis-for-IPCWs [27]) and led to the conclusion that the BLR-IPCW and MLR-IPCW plots are likely unreliable, in particular for the adverse event state. We hypothesised this was driven by a differential censoring mechanism/observation process for individuals in the adverse event state. Simulations studies are required to 1) quantify this type of bias, and 2) explore whether this can be accounted for by estimating the inverse probability of censoring weights using approaches which are conditional on the time spent in each outcome state (for example a latent-class model). If such a study could be undertaken this would be highly valuable [46,47]. For now, we reiterate the importance of implementing these methods in settings where the observation process/censoring mechanism does not change depending on the outcome state an individual is in. It has previously been suggested to evaluate calibration using MLR-IPCW and one of the BLR-IPCW or pseudo-value approaches because MLR-IPCW provides a stronger assessment of calibration [24]. However, we now suggest to evaluate calibration using all three methods, and a comparison between the BLR-IPCW and pseudo-value approaches can be used to help assess whether the assumptions of either method may be violated.

Given it is possible to use calibmsm to validate a standard competing risks model [12,13], we carried out a sensitivity analysis to compare the approaches described in this paper with the ‘graphical calibration curves’ of Austin et al. [12], (see vignette Comparison-with-graphical-calibration-curves-in-competing-risks-setting [27]). BLR-IPCW, pseudo-values, and graphical calibration curves (MLR-IPCW excluded for not producing a calibration curve) all resulted in similar calibration curves. This is with the exception of BLR-IPCW for state , which has been previously discussed. The three methods take completely different approaches to assessing the calibration of a competing risks model. Therefore finding agreement between these assessments of calibration can provide reassurance that the calibration plots are correct, and is an exercise that could be repeated in practice. Despite this, the relative performance of each method in a wider range of competing risks scenarios remains unknown. A comparison of these methods in a simulation when the assumptions of each method do and do not hold, and under a range of sample sizes and multistate model structures, would be therefore valuable [46].

The BLR-IPCW, MLR-IPCW and pseudo-value approaches have different computational burdens. A calibration curve can be obtained reasonably quickly using the BLR-IPCW or MLR-IPCW approaches, however estimation of confidence intervals for BLR-IPCW using bootstrapping (the recommend method in section 2.6) will result in a high computation time in large validation datasets. On the contrary, obtaining the calibration curve itself using the pseudo-value approach has a high computational burden due to estimation of the pseudo-values. Once these have been calculated, a calibration curve and confidence interval can be estimated quickly using parametric techniques, meaning estimation of the confidence interval adds minimal computational burden. We plan to extend the package to allow users to estimate the pseudo-values for each individual seperately before estimating the calibration curve. This will allow the first part of the process to be parallelised and will make estimation of calibration curves using the pseudo-value approach more feasible in large datasets.

Estimation of the weights is clearly of high importance for the BLR-IPCW and MLR-IPCW approaches. If the model to do so is misspecified, this could lead to incorrect evaluation of the calibration. It is possible this is what is causing the difference between the BLR-IPCW and pseudo-value approaches for the calibration of transition probabilities from state at time into state , as was explored in vignette Sensitivity-analysis-for-IPCWs [27]. This package is focused on creation of calibration curves, but is not a dedicated package for estimating inverse probability of censoring weights. We encourage users to create a well specified model for the weights (see Robins and Hernan [39]) if using the BLR-IPCW or MLR-IPCW approaches. Custom functions for estimating the weights can be specified through the w.function argument in calib_msm. Alternatively, weights can be estimated externally and then specified though the weights argument. In this latter case, the internal bootstrapping procedure will not work, as the weights need to be re-estimated in each bootstrap dataset. We have provided a more detailed vignette about how to estimate calibration curves and confidence intervals using bootstrapping when defining your own function to estimate the weights (see vignette BLR-IPCW-manual-bootstrap) [27].

Despite sample size formulae being available for clinical prediction models predicting continuous [48], binary [34,49], time-to-event [34] and multinomial outcomes [50]; sample size formulae do not currently exist for when developing a multistate clinical prediction model. Given the combinatorial issues with multistate models, overfitting is of particular concern as the number of individuals passing through some transitions may be small. Future work in this area is therefore paramount. A multistate model, at its core, is a network of cause-specific hazards models [2], which are no different to a normal time-to-event model. We hypothesise that existing sample size formula could be applied to each model in isolation in order to get a minimum sample size per transition, which could then be divided by the proportion of individuals expected to reach the starting state for that transition in order to derive the total number of individuals required to satisfy that transitions target sample size. The maximum across all transitions would then be taken. For clock-forward models, this may be complicated by the fact that each cause-specific model is an interval censored model, and it is currently unclear how to apply existing sample size formula [34] to interval censored data.

In summary, calibmsm provides tools to assess the calibration of the transition probabilities of a multistate model or competing risks model using three approaches (BLR-IPCW, MLR-IPCW and pseudo-values). Further comparison of these approaches in targeted simulations to establish their performance under different censoring mechanisms and assumptions would be valuable. Future work will aim to develop methodology for other model evaluation metrics and incorporate these into calibmsm.

Supporting information

S1 File. R: a replication script for the analyses.

https://doi.org/10.1371/journal.pone.0320504.s001

(R)

References

  1. 1. van Smeden M, Reitsma JB, Riley RD, Collins GS, Moons KG. Clinical prediction models: diagnosis versus prognosis. J Clin Epidemiol. 2021;132:142–5. pmid:33775387
  2. 2. Putter H, Fiocco M, Geskus RB. Tutorial in biostatistics: competing risks and multi-state models. Stat Med. 2007;26(11):2389–430. pmid:17031868
  3. 3. Putter H, van der Hage J, de Bock GH, Elgalta R, van de Velde CJH. Estimation and prediction in a multi-state model for breast cancer. Biom J. 2006;48(3):366–80. pmid:16845902
  4. 4. Lintu MK, Shreyas KM, Kamath A. A multi-state model for kidney disease progression. Clin Epidemiol Glob Health. 2022;13:100946.
  5. 5. Masiá M, Padilla S, Moreno S, Barber X, Iribarren JA, Del Romero J, et al. Prediction of long-term outcomes of HIV-infected patients developing non-AIDS events using a multistate approach. PLoS One. 2017;12(9):e0184329. pmid:28886092
  6. 6. Steyerberg EW, Harrell FE Jr. Prediction models need appropriate internal, internal-external, and external validation. J Clin Epidemiol. 2016;69:245–7. pmid:25981519
  7. 7. Sperrin M, Riley RD, Collins GS, Martin GP. Targeted validation: validating clinical prediction models in their intended population and setting. Diagn Progn Res. 2022;6(1):24. pmid:36550534
  8. 8. Van Calster B, McLernon DJ, van Smeden M, Wynants L, Steyerberg EW, Topic Group ‘Evaluating diagnostic tests and prediction models’ of the STRATOS initiative. Calibration: the Achilles heel of predictive analytics. BMC Med. 2019;17(1):230. pmid:31842878
  9. 9. Van Calster B, Nieboer D, Vergouwe Y, De Cock B, Pencina MJ, Steyerberg EW. A calibration hierarchy for risk models was defined: from utopia to empirical data. J Clin Epidemiol. 2016;74:167–76. pmid:26772608
  10. 10. Crowson CS, Atkinson EJ, Therneau TM. Assessing calibration of prognostic risk scores. Stat Methods Med Res. 2016;25(4):1692–706. pmid:23907781
  11. 11. Austin PC, Harrell FE Jr, van Klaveren D. Graphical calibration curves and the integrated calibration index (ICI) for survival models. Stat Med. 2020;39(21):2714–42. pmid:32548928
  12. 12. Austin PC, Putter H, Giardiello D, van Klaveren D. Graphical calibration curves and the integrated calibration index (ICI) for competing risk models. Diagn Progn Res. 2022;6(1):2. pmid:35039069
  13. 13. Gerds TA, Andersen PK, Kattan MW. Calibration plots for risk prediction models in the presence of competing risks. Stat Med. 2014;33(18):3191–203. pmid:24668611
  14. 14. R Core Team. R: a language and environment for statistical computing [Online]. Vienna, Austria: R Foundation for Statistical Computing; 2023. Available from: https://www.r-project.org/
  15. 15. de Wreede LC, Fiocco M, Putter H. mstate: An R package for the analysis of competing risks and multi-state models. J Stat Softw. 2011;38(7).
  16. 16. Jackson CH. flexsurv: A platform for parametric survival modeling in R. J Stat Softw. 2016;70:i08. pmid:29593450
  17. 17. Król A, Saint-Pierre P. SemiMarkov: an R package for parametric estimation in multi-state semi-markov models. J Stat Soft. 2015;66(6).
  18. 18. Jackson CH. Multi-state models for panel data: the msm package for R. J Stat Softw. 2011;38(8):128–129.
  19. 19. Eletti A, Marra G, Radice R. flexmsm: A general framework for flexible multi-state survival modelling [Online]; 2024. Available from: https://cran.r-project.org/package=flexmsm
  20. 20. Cartensen B, Plummer M, Laara E, Hills M. Epi: a package for statistical analysis in epidemiology [Online]; 2024. Available from: https://cran.r-project.org/package=Ep
  21. 21. Martin GP, Jenkins D, Sperrin M. predRupdate: Prediction model validation and updating [Online]; 2024. Available from: https://github.com/GlenMartin31/predRupdate
  22. 22. Harrell FE Jr. Hmisc: harrell miscellaneous [Online]; 2023. Available from: https://hbiostat.org/R/Hmisc/
  23. 23. De Cock B, Nieboer D, Van Calster B, Steyerberg EW, Vergouwe Y. The CalibrationCurves package: assessing the agreement between observed outcomes and predictions. 2023.
  24. 24. Pate A, Sperrin M, Riley RD, Peek N, Van Staa T, Sergeant JC, et al. Calibration plots for multistate risk predictions models. Stat Med. 2024;43(14):2830–52. pmid:38720592
  25. 25. Dafni U. Landmark analysis at the 25-year landmark point. Circ Cardiovasc Qual Outcomes. 2011;4(3):363–71. pmid:21586725
  26. 26. Van Houwelingen HC. Dynamic prediction by landmarking in event history analysis. Scand J Stat. 2006;34(1):70–85.
  27. 27. Pate A, Martin GP. calibmsm [Online]; 2024. Available from: https://alexpate30.github.io/calibmsm/
  28. 28. EBMT. Data from the European Society for blood and marrow transplantation [Online]; 2023 [cited 2023 Apr 11. ]. Available from: https://search.r-project.org/CRAN/refmans/mstate/html/EBMT-data.html
  29. 29. Austin PC, Steyerberg EW. Graphical assessment of internal and external calibration of logistic regression models by using loess smoothers. Stat Med. 2014;33(3):517–35. pmid:24002997
  30. 30. Harrell FE. Regression modeling strategies, Springer S. Cham, Switzerland: Springer; 2015.
  31. 31. Van Hoorde K, Vergouwe Y, Timmerman D, Van Huffel S, Steyerberg EW, Van Calster B. Assessing calibration of multinomial risk prediction models. Stat Med. 2014;33(15):2585–96. pmid:24549725
  32. 32. Van Hoorde K, Van Huffel S, Timmerman D, Bourne T, Van Calster B. A spline-based tool to assess and visualize the calibration of multiclass risk predictions. J Biomed Inform. 2015;54:283–93. pmid:25579635
  33. 33. Yee TW. Vector generalized linear and additive models. 1st ed. New York (NY): Springer; 2015.
  34. 34. Riley RD, Snell KI, Ensor J, Burke DL, Harrell FE Jr, Moons KG, et al. Minimum sample size for developing a multivariable prediction model: PART II - binary and time-to-event outcomes. Stat Med. 2019;38(7):1276–96. pmid:30357870
  35. 35. Andersen PK, Wandall ENS, Pohar Perme M. Inference for transition probabilities in non-Markov multi-state models. Lifetime Data Anal. 2022;28(4):585–604. pmid:35764854
  36. 36. Andersen PK, Perme MP. Pseudo-observations in survival analysis. Stat Methods Med Res. 2010;19(1):71–99. pmid:19654170
  37. 37. Putter H, Spitoni C. Non-parametric estimation of transition probabilities in non-Markov multi-state models: the landmark Aalen-Johansen estimator. Stat Methods Med Res. 2018;27(7):2081–92. pmid:29846146
  38. 38. Kleinbaum DG, Klein M. Survival analysis: a self-learning text. 3rd ed. Springer; 2012.
  39. 39. Hernan M, Robins J. 12.2 Estimating IP weights via modeling. In: Causal inference: what if. Boca Raton: Chapman & Hall/CRC; 2020.
  40. 40. Yee TW. TheVGAMPackage for categorical data analysis. J Stat Softw. 2010;32(10).
  41. 41. Grolemund G, Cetinkaya-Rundel M, Wickham H. R for data science. 2nd ed. O’Reilly Media; 2023.
  42. 42. van Houwelingen HC, Putter H. Comparison of stopped Cox regression with direct methods such as pseudo-values and binomial regression. Lifetime Data Anal. 2015;21(2):180–96. pmid:25084763
  43. 43. Pullenayegum EM, Platt RW, Barwick M, Feldman BM, Offringa M, Thabane L. Knowledge translation in biostatistics: a survey of current practices, preferences, and barriers to the dissemination and uptake of new statistical methods. Stat Med. 2016;35(6):805–18. pmid:26307183
  44. 44. Grand MK, de Witte TJM, Putter H. Dynamic prediction of cumulative incidence functions by direct binomial regression. Biom J. 2018;60(4):734–47. pmid:29577376
  45. 45. Bull LM, Lunt M, Martin GP, Hyrich K, Sergeant JC. Harnessing repeated measurements of predictor variables for clinical risk prediction: a review of existing methods. Diagn Progn Res. 2020;4:9. pmid:32671229
  46. 46. Heinze G, Boulesteix A-L, Kammer M, Morris TP, White IR, Simulation Panel of the STRATOS initiative. Phases of methodological research in biostatistics-Building the evidence base for new methods. Biom J. 2024;66(1):e2200222. pmid:36737675
  47. 47. Boulesteix A-L, Lauer S, Eugster MJA. A plea for neutral comparison studies in computational sciences. PLoS One. 2013;8(4):e61562. pmid:23637855
  48. 48. Riley RD, et al. Minimum sample size for developing a multivariable prediction model: part I – continuous outcomes. Stat Med. 2019;38(7):1262–1275,
  49. 49. van Smeden M, Moons KG, de Groot JA, Collins GS, Altman DG, Eijkemans MJ, et al. Sample size for binary logistic prediction models: beyond events per variable criteria. Stat Methods Med Res. 2019;28(8):2455–74. pmid:29966490
  50. 50. Pate A, Riley RD, Collins GS, van Smeden M, Van Calster B, Ensor J, et al. Minimum sample size for developing a multivariable prediction model using multinomial logistic regression. Stat Methods Med Res. 2023;32(3):555–71. pmid:36660777