## Figures

## Abstract

We present a novel quasi-Bayesian method to weight multiple dynamical models by their skill at capturing both potentially non-linear trends and first-order autocorrelated variability of the underlying process, and to make weighted probabilistic projections. We validate the method using a suite of one-at-a-time cross-validation experiments involving Atlantic meridional overturning circulation (AMOC), its temperature-based index, as well as Korean summer mean maximum temperature. In these experiments the method tends to exhibit superior skill over a trend-only Bayesian model averaging weighting method in terms of weight assignment and probabilistic forecasts. Specifically, mean credible interval width, and mean absolute error of the projections tend to improve. We apply the method to a problem of projecting summer mean maximum temperature change over Korea by the end of the 21^{st} century using a multi-model ensemble. Compared to the trend-only method, the new method appreciably sharpens the probability distribution function (pdf) and increases future most likely, median, and mean warming in Korea. The method is flexible, with a potential to improve forecasts in geosciences and other fields.

**Citation: **Olson R, An S-I, Fan Y, Evans JP (2019) Accounting for skill in trend, variability, and autocorrelation facilitates better multi-model projections: Application to the AMOC and temperature time series. PLoS ONE 14(4):
e0214535.
https://doi.org/10.1371/journal.pone.0214535

**Editor: **Juan A. Añel, Universidade de Vigo, SPAIN

**Received: **November 15, 2018; **Accepted: **March 14, 2019; **Published: ** April 10, 2019

**Copyright: ** © 2019 Olson et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

**Data Availability: **The relevant data and code implementing the methodology and used to produce the results of the paper (including programs to make the figures) are available as a supplementary file S2_File.zip. Observational data used in this paper are freely available from third party websites, with details provided in the references. The Climate Model Intercomparison Project phase 5 model output is available from https://esgf-node.llnl.gov/projects/esgf-llnl/. Data access requires filling in a registration form at https://esgf-node.llnl.gov/user/add/. The results of the cross-validation and projection experiments, including the weights, the Markov chain Monte Carlo chains, the R programming language workspaces, and the control input files are available from the https://www.climatelab.yonsei.ac.kr/forum.

**Funding: **RO and SIA were supported by Basic Science Research Program through National Research Foundation of Korea (NRF-2017K1A3A7A03087790, NRF-2018R1A5A1024958) (http://www.nrf.re.kr/eng/main). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

**Competing interests: ** The authors have declared that no competing interests exist.

## 1 Introduction

A common forecasting problem is one of probabilistic multi-model forecasts of a stochastic dynamical system [1–18]. Sometimes, when a collection of complex dynamical models is used to provide multi-model forecasts, these forecasts are weighted according to model performance compared to observations [1,5,10,19–23]. The Bayesian approach to this problem assumes that associated with *k* dynamical models are *k* competing statistical models *M*_{i} for vector of observations ** y**. These statistical models result in a conditional probability density function (pdf) for

**given that**

*y**M*

_{i}is reasonable,

*p*(

**|**

*y**M*

_{i}). Typically, in a multi-model evaluation context, the pdf

*p*(

**|**

*y**M*

_{i}) is a multivariate statistical distribution centered on

*i*th dynamical model trend

*x*_{i}. Each model is associated with a prior belief in its adequacy (“prior”)

*p*(

*M*

_{i}), which can be derived from previous work, or may be more subjective. The posterior probability, or weight, for each model

*i*given the observations is then found using Bayes theorem [24]: (1)

Specifically, the posterior probability of each statistical (and corresponding dynamical) model is the likelihood of observations ** y** coming from the model (given by the pdf

*p*(

**|**

*y**M*

_{i})), multiplied by the model prior.

In ensemble modelling, models are usually judged on how well they represent the mean state of the system, its trend, or spatio-temporal fields [1,3,6,14,22,23]. However, it is increasingly being recognized that variability is of utmost importance for future prediction. Specifically, for some systems (stochastic dynamical systems) the stationary pdf of the equilibrium solution is directly affected by system dynamics (i.e., the nonlinear operator in the ordinary differential equations) through the Fokker-Planck (Kolmogorov forward) equation. Recent climate science work identifies variability as a key factor impacting climate projections [6,25]. Furthermore, variability has been used as a novel and effective constraint for climate sensitivity [26]. In addition, variability also has major relevance for forewarning of critical thresholds (i.e., a forcing value above which the underlying system shifts to a new equilibrium; [27]). Specifically, an increase in variance or lag-1 autocorrelation with time, as well as skewness and kurtosis, have been used as such early warning indicators [28–31]. This motivates using variability properties of the system as a novel metric to assess performance of multiple dynamical system models.

Several new studies break important new ground by incorporating variability into the weighting [17,32–34], but they typically assume stationarity of the pdf of the system [17,32,33], or cannot work with complex dynamical models [34]. Some previous work does explicitly weight dynamical models by performance in variability and trends in a statistically-sound way [35]. However, the method in its current form works only for linear trends (as a function of time) and does not account for autocorrelation in the variability.

Here we propose a novel method to weight models of complex dynamical systems by their performance in autocorrelation, variability, and a potentially nonlinear trend (i.e., nonlinear with time) compared to observations, and to make probabilistic forecasts. The method is based on Bayesian Model Averaging (BMA) [20,21]. While the framework is Bayesian, it deviates from traditional Bayesian theory in some steps of the estimation process. We highlight these deviations where they arise in more detail in later sections. Consequently, we call our approach “quasi-Bayesian”. Using several simulated and observed datasets (involving AMOC, its temperature-based index, and summer mean maximum temperature over Korea) we show that the new method results in better weighting and tends to improve forecasts of system mean change under new conditions compared to when trend-only BMA weighting is used. Thus, this work has implications for improving projections of many environmental systems. The approach is not restricted to linear trends, making it relatively easy to apply to new datasets. Finally, we apply the method to a real case problem of projecting future summer mean temperature changes over Korea.

The rest of the paper is structured as follows. Section 2 describes the novel methodology to weight models by trend and variability performance, to combine those weights, to make multi-model weighted projections, as well as the computational details. The main interest here is not the procedure for obtaining the trend and variability components, but the algorithm for model weighting. In Section 3 we describe leave-one-out cross-validation experiments to test method performance against a trend-only BMA method. Here we also provide the specific details on how the trend and variability components were extracted from the data. Section 4 describes the results of these experiments. Section 5 discusses the application of the method to make multi-model probabilistic projections of Korean summer mean maximum temperature change. Section 6 briefly discusses the main findings of the study and places it in context of prior work. Section 7 discusses the limitations of the work, and Section 8 presents conclusions.

## 2 Materials and methods

### 2.1 Overview of the method

At the start of the analysis, we assume that we have a collection of dynamical model time series outputs, and that these outputs can be decomposed into long-term trend and variability components. The details of this decomposition are not critical for this study, as we focus on the statistical methodology for the weighting. The weights (or probabilities) for the two submodels are calculated separately, using the Bayesian statistical paradigm, and then combined. The combined weights can then be used to make predictions (Fig 1).

### 2.2 Notation and decomposition of model output

Consider that *k* models are available. We postulate that each dynamical model is associated with a statistical model *M*_{i} for the observations. *M*_{i} can be thought of as a statistical event, which when true indicates that *i*th dynamical model is a reasonable representation of real system. *M*_{i} consists of two submodels: a trend submodel *M*_{T,i} (related to the trend in the system), and a variability submodel *M*_{V,i} (modelling internal fluctuations in the system). When *M*_{T,i} is true, the *i*th dynamical model correctly captures the trend of the system. Likewise, when *M*_{V,i} is true, the *i*th dynamical model correctly captures the variability of the system. Alternatively, we can consider the model for anomalies scaled by the mean (*M*_{V0,i}). Each model produces time series output of a physical quantity during the period when observations are available (“calibration period”), as well as under new forcing conditions, usually associated with future system projections (“projection period”). We are interested in finding the probability distribution of a change of the system mean Δ between a “projection reference period” (typically the same as the calibration period) and the projection period. We denote the raw calibration period model output from the *i*th dynamical model by vector where superscript “*'*” indicates that the output is raw (un-smoothed), and *n* is the length of the record. The model output is a regularly spaced time series. We consider decomposition of the form:
(2)

We will use the term “anomalies” to refer to the variability component of the time series. The trend *x*_{i} can be either a linear trend, or a more flexible nonlinear trend obtained, for example, from robust locally weighted regression [36]. We assume that this decomposition is deterministic, unique, and is performed before the start of the main analysis. We also assume that the estimate of the trend is a reasonable proxy for the true unknown trend. While it may be possible to also incorporate the uncertainty in this decomposition, we leave it to future work. The focus here is not on how to properly decompose a time series into a long-term trend and variability, but on the novel methodology for weighting by performance in both. See [18] for an example of use of an alternative methodology to decompose the data. The use of alternative methods for data decomposition is subject of future research. We describe the decomposition method we use for each dataset in Section 3. The same decomposition is also applied to the observed time series ** y'**:
(3)

Another option is relative decomposition. It takes the following form:
(4)
where is the deterministic sample mean of the *i*th dynamical model output, and are normalized anomalies; and similarly for the observations:
(5)
where is the observed mean.

Next subsections contain the following: subsection 2.3 discusses the trend submodel weighting (which largely follows previous work), subsection 2.4 centers on the variability submodel weighting, section 2.5 discusses combining the component weights for each model, section 2.6 is dedicated to procedure for making weighted multi-model projections, and section 2.7 presents computational details on the implementation of the method.

### 2.3 Weighting the trend submodels

The trend submodel weighting is implemented following prior work, and full details are provided there [9]. Essentially, this method is BMA that also considers the uncertainty due to model error, and uncertainty in statistical properties of data-model residuals. Here, we consider *k* competing statistical models *M*_{T,i} for raw observations ** y’**. We stress that statistical and dynamical models are conceptually related: i.e., if the statistical model

*M*

_{T,i}is true, it implies that the associated

*i*th dynamical model correctly represents the trend in the system. Each

*M*

_{T,i}is a hierarchical statistical model that connects modelled deterministic trend from the

*i*th model during the calibration period

*x*_{i}to real system trend

**, and then the system trend to actual observations**

*y***(Eq 6): (6) where**

*y'**f*

*ε*_{D}is random discrepancy (long-term model error), and

*is random internal variability (as well as short-term observational error).*

**ε**_{NV}Here we deviate somewhat from orthodox Bayesian practice. A typical Bayesian approach would assume a distributional form for the discrepancy vector *f**ε*_{D}. However, because this error is likely long-term dependent, and the probability distributions for its components are not necessarily normal, finding and justifying a proper parametric model for it is non-trivial. To deal with this conundrum, we adopt an approach inspired by prior work [37]. We postulate that model error can be derived from inter-model trend differences. The reasoning for this implementation is as follows. Imagine a particular trend submodel *M*_{T,i} represents the “true” system. Associated with this system is trend *x*_{i} and pseudo-observations **. If only the rest of the models are available to the researcher, then the best-fit model ***j* to these pseudo-observations is associated with trend *x*_{j}. The difference between the best model and the pseudo-observed trends is then the unscaled error of the *j*th model. Thus, we obtain samples for unscaled discrepancy *ε*_{D} directly from the differences between each model’s trend and the next-closest model trend (see [9] for details). We acknowledge that this parameterization is simplified; model error is an emergent research topic [37]. We thus hope this work can galvanize more research on parametrizing model error.

The second non-orthodox idea, is related to the deterministic *f* factor (“error expansion factor”, Eq 6). This factor is a new addition to the model presented in previous work [9]. *f* is a parameter that scales *ε*_{D} to account for potential overconfidence. The non-orthodox idea relates to the procedure for selecting *f*. Specifically, we do not estimate *f* from present-day observations as a strict Bayesian would do, but rather we select *f* that results in correct coverage of the 90% posterior credible intervals during cross-validation experiments (different *f* for each dataset). The reason for this is as follows. Using just present-day observations to estimate *f* may produce small *f* values that result in overconfident future projections. This is because models have been developed so that they match observed data. Philosophically, present-day model-data agreement may be due to overfitting, and may not be reflective of the actual amount of error in the models.

The internal variability *ε*_{NV} (Eq 6) is modelled as an AR(1) process with random parameters ** θ** = (

*σ*,

*ρ*), where

*σ*is innovation standard deviation and

*ρ*is autocorrelation. Following Bayes theorem, and marginalization theorem, the trend model weights are then calculated as: (7)

Here, *p*(*M*_{T,i}) denotes the prior for the *i*th trend model, *p*(*y**'*|** y**,

**,**

*θ**M*

_{T,i}) is the AR1 likelihood resulting from the bottom line of Eq (6),

*p*(

**) denotes the prior for the AR1 parameters, and**

*θ**p*(

**|**

*y**M*

_{T,i}) is obtained according to the top line of Eq (6) using samples from

*f*

*ε*_{D}as discussed above. Unlike the previous work [9], here we assume uniform prior probabilities for trend models

*p*(

*M*

_{T,i}). The integral is evaluated using Monte Carlo integration, which is simpler to implement than Markov chain Monte Carlo methods used in some studies [2]. For the relative low dimension parameter space that we deal with here, simple Monte Carlo is adequate. Additional experiments suggest the sample size we use for the Monte Carlo integration is reasonable to minimize Monte Carlo error (Text A in S1 File). Once calculated, the weights are normalized to sum to 1 to facilitate interpretation as probabilities. We provide technical details in Text A in S1 File.

### 2.4 Weighting the variability submodels

Variability models are weighted using similar ideas to the ones used in trend weight estimation. We consider *k* competing statistical models for calibration period anomalies observations Δ* y* = (Δ

*y*

_{1}, Δ

*y*

_{2}, …, Δ

*y*

_{n}) (see Eq (3)). Each

*i*th variability model

*M*

_{V,i}models the anomalies hierarchically in the following form: (8) where , are autocorrelation and innovation standard deviation of the real climate, are summary statistics of autocorrelation and innovation standard deviation from

*i*th model anomalies,

*f*

*ε*^{V}is model error (where

*ε*^{V}= (

*ε*

_{σ},

*ε*

_{ρ}), and

*f*is a deterministic scaling factor to widen the distribution to correct for potential overconfidence), and . The top line of Eq (8) connects real system anomaly properties to model summary statistics, and the bottom line shows that observed anomalies are modelled as red noise with parameters (

*σ*

_{y},

*ρ*

_{y}) of the real system.

Thus, in the top line of Eq (8) instead of performing full posterior sampling to obtain samples for real system autocorrelation and innovation standard deviation parameters we assume they are centered around summary statistics of *i*th physical model anomalies with an additive error *f ε^{V}*. Each model’s summary statistics are taken as the corresponding MLE estimates. Again, we refrain from assuming any parametric form for

*. Similar to the error for the trend model, here we also assume samples for*

**ε**^{V}*are obtained from differences between each model MLE summary statistics and the next-closest model summary statistics . The next-closest model is found as follows: for each model*

**ε**^{V}*i*we compare the conditional likelihood of

*i*th model anomalies given AR(1) parameters of other variability submodels ,

*j*≠

*i*under the AR(1) statistical model, and find a model

*j*that maximizes this likelihood. We also add a sample of zero vector (0,0) to

*for computational stability. We post-multiply these samples by a scaling factor*

**ε**^{V}*f*to obtain samples for

*f*.

**ε**^{V}*f*is the same parameter that is used to scale trend model discrepancy (Section 2.3).

This approach gives us only *k*+1 samples from *f ε^{V}*. To obtain a larger number of samples which are well-dispersed, we add to

*f*realizations from an independent bivariate normal distribution with standard deviations in each dimension set to 1/5 of the original

**ε**^{V}*k*+1 sample ranges. We use the value of 1/5 because it results in samples with a reasonably smooth density that preserves large scale cross-correlation structure between the original

*k*+1 samples of

*ε*

_{σ}and

*ε*

_{ρ}, and provides a decent approximation to the underlying pdf for

*f*(Fig A in S1 File). Sensitivity tests indicate that using lower standard deviations can degrade the smoothness of the pdf (not shown).

**ε**^{V}Then, the posterior probability of the variability model *i* is, using Bayes rule [24] and probability rules:
(9)
where is an AR1 likelihood function, is sampled using the top line of Eq (8) using bootstrapping from *f ε^{V}* as described above, and

*p*(

*M*

_{V,i}) is the prior probability (“prior”) for the

*i*th variability submodel. We assume equal priors for all submodels. This integral is also evaluated using Monte Carlo integration. Specifically, we sample from the conditional pdf of real system summary statistics given each variability model as described above, and for each sample we calculate the conditional likelihood for the observed anomalies . The integral is approximated as a simple mean of the conditional likelihoods across the samples. Probabilities are calculated for each submodel and are normalized to sum up to 1. The implementation using relative variability

*M*

_{V0}is identical except the residuals

*Δx*_{i}and

**are normalized by the respective model and observational means prior to the analysis. We provide technical details on the implementation in Text B in S1 File.**

*Δy*### 2.5 Combined weights and Bayesian model averaging

In the next step, the weights for the two submodels are put together to form a single combined model weight. Using probability laws: (10)

We make two simplifying assumptions. First, we observe that in the datasets described in Section 3 typically the relationships between the variability summary statistics and on one hand, and trend model probability on the other hand, appear to be weak (Figs B-K in S1 File). In addition, the corresponding linear coefficients are almost always weak (weak is defined as the absolute values less than 0.5). Assuming that the relationships based on the sample summary statistics are a good proxy for those based on the population properties, we make an assumption that the probability of the trend model is independent of the variability model: (11) which allows us to directly plug in trend model weights obtained using the method in Section 2.3. Second, since only anomalies are used to weight the variability model: (12)

This quantity is obtained following Section 2.4. As a result, the combined weights can be expressed as a product of the trend and variability submodel weights: (13)

We stress that even though the independence assumption generally appears reasonable here, it may not always apply. Hence, it is recommended to check it when applying the methodology to new datasets. Incorporating the potential dependence between the trend and variability submodels into our framework is the subject of future research. Once calculated, the probabilities are normalized to sum up to 1, meaning that we restrict our probability space to the union of available models *M*_{i}.

### 2.6 Future Projections

Future model projections are implemented largely following previous work [9]. Once the weights are obtained, the statistical model for system change between projection reference and projection periods Δ follows the BMA formula [20,21]:
(14)
where *D* = (** y**,

**) is collection of all available observations,**

*Δy**p*(Δ|

*M*

_{i},

*D*) is conditional probability for the change given than

*i*th dynamical model is correct, and

*w*

_{i}=

*p*(

*M*

_{i}|

*D*) is the probability for the

*i*th model (i.e., model weight) found earlier (Eq (13)) as the product of the trend and variability model probabilities. This represents a skill-weighted mixture of pdfs from individual models. Here we consider Δ to be a simple difference between projection period mean and forecast reference period mean. Future predictions are largely modelled following prior work [9]. Just as for the calibration period, we assume a deterministic decomposition of projection period output into trend and anomalies: (15)

The exact decomposition method for each dataset is listed in Section 3. Next, we consider the following statistical model for dynamical system time-series projections (all quantities are vectors):
(16)
where **y***'*^{(f)} is the projection time series, is *i*th model trend output from Eq (15), *b*^{(f)} = *b*^{(f)}**1** is random time-constant bias, and is random short-term internal variability in each model. Thus, we assume fthat if *i*th model is correct, the vector projection is the sum of *i*th model trend, a time constant bias, and internal variability. Here we again deviate somewhat from the traditional Bayesian theory in that the components of this model are partially informed by inter-model differences, and by model output during cross-validation experiments. Such steps are necessitated by the absence of actual system observations over the projection period to inform us about these components. We model the bias parameter as where is sample standard deviation of future period-mean next-closest model differences (where next-best is used in the *l*_{1} distance sense), and *f* is the deterministic model error expansion factor (the same factor that is used for model weighting). Two different formulations are implemented for internal variability. In the first formulation (“boot”; [9]) we use simple bootstrapping from to generate internal variability samples. In the alternative formulation (“ar1”) we sample as a red noise process with parameters , the sample innovation standard deviation and autocorrelation of future anomalies. An improvement would be to consider the uncertainty in the AR1 parameters; we do not do this here to simplify the method. To obtain projection period mean changes from the reference period, we take weighted samples of future projections using Eqs (14) and (16), and simply subtract projection reference period mean modeled value for each model. As in previous work [9], we use 100,000 samples for all experiments.

The overall algorithm for the method is illustrated in Fig 2. The method estimates model weights from calibration period observations, and has one fixed parameter *f*, quantifying model error. Larger *f* values lead to higher model errors, and as a result broader projections with higher coverage of the 90% posterior credible intervals. Unlike standard Bayesian analysis, we first choose *f* to obtain approximately correct empirical coverage of the 90% posterior credible intervals during cross-validation. For the cross-validation, each model is selected as the “truth” one-at-a-time. Models are weighted using the output from the “true” model. The “true” model is then excluded from the model set, and the future weighted projections from the remaining models are compared to the output from the “true” model. Once *f* achieves approximately correct empirical coverage, the method is used for actual projections constrained by real observations. If there are many replicates (or regions) of the system, cross-validation can also be performed by splitting the calibration period into two subperiods. In step 1, observations during the first subperiod in each region/replicate can be used to assign replicate/region-specific weights. In step 2, observations during the second subperiod can test the empirical coverage of the posterior credible intervals. Here, however, we focus on the one-at-a-time cross-validation using future model output. This is because (i) the length of historical record for which high-quality observations are available is too short for most of the experiments [38,39], (ii) observational records suffer from observational errors, and (iii) climate signal (e.g., the magnitude of climate changes) is quite low in the historical period. We choose various variables and periods to test the method under different conditions.

### 2.7. Computational details

All experiments have been performed on an Intel Xeon CPU X5650 @ 2.67GHz GNU/Linux 2.6.18-164.el5 supercomputer, using R programming language version 3.3.3. For other required packages the following versions were used: mblm 0.12 and KernSmooth 2.23–15. We provide the R code as S2 File. This code is provided under the GNU general public license v3.

In the next section we describe several cross-validation experiments for our method and compare the performance of the method (which we call hereafter “trend+var”) with a BMA method where all variability submodel weights are set to equal (termed hereafter “trend”). Note that “trend” method is BMA which still weights models by their performance in terms of trend.

## 3 Leave-one-out cross-validation experiments to test method skill

### 3.1 Overview of leave-one-out cross-validation experiments

To evaluate method performance, we carry out leave-one-out cross-validation experiments with several simulated and observed datasets: (i) Atlantic meridional overturning circulation (AMOC) strength [Sv] from 13 global climate models (GCMs) (AMOC experiment), (ii) Korean summer mean maximum temperatures from 29 GCMs (Korea_temp), (iii) Korean temperatures with an extended calibration period (Korea_temp_long), (iv) winter East Sea surface temperatures (SSTs) (Winter SST Experiment), (v) temperature-based AMOC Index (temperature in northern North Atlantic “gyre” minus Northern Hemisphere temperature) from 13 GCMs (AMOCIndex), and (vi) the same as (v) but also considering information from climate observations (AMOCIndex_obs). We discuss each experiment in greater detail in the following subsections. The cases differ in terms of the calibration, projection, and projection reference periods (Table 1). In experiments involving model output only, each of the models is selected as “truth” one at time, and its output is used to weight the models. Then, during the validation period, the projected pdfs of changes using the remaining models are compared to the “true” model output. The set-up for the AMOCIndex_obs is slightly different: both calibration and validation periods have available instrumental observations. Here, instead of selecting each model output as pseudo-observations one-at-a-time, we simply use actual observations to both weight the climate models, and to evaluate the projections. All experiments are performed with both “trend” and “trend+var” methods. Both methods have been calibrated for each experiment to have approximately correct coverage (correct % of cases where the “truth” is outside the 90% posterior credible intervals) by adjusting the model error expansion factor *f* (Table 1). The calibrated values of *f* for the AMOCIndex experiments are also used for the corresponding AMOCIndex_obs experiments. We focus on the Winter_SST experiment here, however summary results for all experiments are also provided.

**Bold font** indicates improvement of the “trend+var” method, compared to the “trend” method. *k* is the number of models in the ensemble; MCIW is mean 90% credible interval width; MAE is mean absolute bias of the mean; CIW is 90% credible interval width; AB is absolute bias of the mean.

### 3.2 AMOC experiment

For the AMOC experiment (Table 1), data extraction and processing largely follow previous work [9]. The Climate Model Intercomparison Project phase 5 (CMIP5; [40]) model output for this (and other) experiments has been obtained from the ESGF LLNL portal [41]. Future forecasts use the RCP8.5 emissions scenario [42]. We use robust locally-weighted “lowess” regression [36] to obtain the trend model component during the calibration period, and Theil-Sen slopes [43]–in the validation period. We set the “lowess” smoother span parameter to 0.8 during the smoothing. We use this span value because it appears effective at removing interdecadal variability. The smoothed model output is illustrated in Fig 3. Importantly, we see nonlinearities in the modeled trends. Previous variability weighting work does not account for such nonlinearities [35]. During the trend weighting we use smoothed output as anomalies with respect to the entire calibration period. We use normalized (by the absolute AMOC) anomalies to weight the variability models. Future projections use the “boot” variant of the method.

### 3.3 Korea_temp and Korea_temp_long experiments

Korea_temp and Korea_temp_long differ only in the calibration periods and the error expansion factors *f*, with Korea_temp_long using a longer calibration period. These experiments use output from historical and future RCP8.5 runs of 29 CMIP5 model runs (Table 1, Table A in S1 File). First, Korean daily maximum temperatures are calculated as spatial averages over land grid cells (cells with more than 80% land) between 34–40°N and 125–130°E [38]. The JJA (June, July, August) means are then obtained for each year. Theil-Sen slopes are used for obtaining model trends during the model output decomposition. During the weighting, smoothed output is used as anomalies with respect to the entire calibration period. Future projections use the “boot” variant of the method. Note that the Korea_temp “trend+var” experiment has a slightly elevated coverage of 93%. Decreasing *f* to obtain approximately 90% coverage is expected to improve performance metrics, but also to make probability densities too discontinuous. Hence, we use the value of *f* = 0.75.

### 3.4 Winter_SST experiment

Winter_SST experiment uses winter sea surface temperatures from the East Sea from historical and future RCP8.5 runs of 26 CMIP5 climate models (Table 1, Table B in S1 File). We select this dataset because we find considerable relationships between present-day internal variability properties and future SST change in this region and season (Fig 4; for model number corresponding to each model see Fig L in S1 File). We define the East Sea as the area between 35 °N and 42°N, and between 130 °E and 139 °E. We use a simple average of all ocean points in this region. During the weighting we use the output as anomalies with respect to the calibration period. Furthermore, we use Theil-Sen slopes to obtain model output trends. Future projections use the “ar1” variant of the method, since we detect a considerable autocorrelation in the model output anomalies.

### 3.5 AMOCIndex experiment

AMOCIndex experiment (Table 1) relies on historical output from the same 13 CMIP5 models used for the AMOC experiment. AMOC Index is defined as sea surface temperature in northern North Atlantic “gyre” minus Northern Hemisphere temperature. It is physically linked to northward heat transport by the AMOC, and hence can be used as a proxy for AMOC [9,44]. Data extraction and processing follow [9], with a few changes. The Index is used as an anomaly with respect to the entire historical period 1880–2004. We then use a portion of the historical period (1880–1945) for calibration, and another portion (1965–2004) for projections. Smoothing is performed using Theil-Sen slopes. Projections use the “ar1” variant of the method.

### 3.6 AMOCIndex_obs experiment

For the AMOCIndex_obs we use actual observations both to weight the models, and to validate the probabilistic projections. Otherwise, the experiment relies on the same model output as AMOCIndex experiment. The observations are a simple average of two AMOC Index versions: one calculated with ERSSTv4 SSTs [45–47], and one with COBE-SST2 SSTs [48]. ERSSTv4 data is publicly curated by National Oceanic and Atmospheric Administration [49], while COBE-SST2 observations are provided by M. Ishii on the servers of Hokkaido University, Japan [50]. Both versions use GISTEMP Northern Hemisphere temperatures [51]. GISTEMP observations are maintained by NASA Goddard Institute for Space Studies [52]. For comparison with model output the COBE-SST2 SSTs are first interpolated to a 2×2° grid using bilinear interpolation, while the ERSSTv4 observations are already on such a grid. For both “trend” and “trend+var” experiments, *f* is taken from corresponding AMOCIndex experiments.

## 4 Results of leave-one-out cross-validation experiments

The new method tends to be better able to correctly identify the “true” model from pseudo-observations (Fig 5, Fig M in S1 File). This is not surprising since it uses extra variability information that is not available to the “trend” method. This extra information can provide a powerful constraint because models differ considerably in their representation of internal variability, based on sample estimates of the variability properties (Fig 4). The most striking improvement is obtained for the AMOC experiment while arguably the least improvement–for the AMOCIndex (Fig M in S1 File).

Rows represent different “true” models. Color represents model weight.

Another important metric is the factor *f* that provides calibrated projections. This factor can be interpreted as a rough measure of model error relative to the next-closest inter-model differences in output space. The values feature a substantial range from 0.75 to 3.75 (Table 1). For experiments AMOC, AMOCIndex, and Korea_temp_long, *f* is the same or similar for both methods. Thus, under our statistical model, the best dynamical models for both “trend” and “trend+var” experiments are approximately equally close to the “true” unobserved trends in the real system, in both calibration and projection periods. However, for the rest of the simulated data experiments the new method achieves a lower *f*. Here, the best model for the “trend+var” method is closer (more than twice as close for Korea_temp) to the “true” trend of the system, compared to the best model under the “trend” experiment, both in calibration *and* projection periods.

We now turn our attention to the question of future prediction. First, it is worth noting that we do not find a significant bias between projections and “true” model output in any of our leave-one-out cross-validation experiments. The new method tends to improve in terms of the mean 90% credible interval width as well as mean absolute bias of the mean (Table 1, Figs 6 and 7, Figs N-Q in S1 File). Specifically, in the Korea_temp experiments, the forecast 90% credible intervals on average sharpen by about 25%. For some cases (e.g., models 3 and 22 of the Korea_temp experiment), the improvements are particularly dramatic, featuring a drastic sharpening of the pdf and a strong reduction in the 90% credible intervals, with a low bias; Figs P and Q in S1 File). The only cases with no improvement are AMOCIndex, and corresponding AMOCIndex_obs (Table 1). We note that these experiments rely on the same model output. They also use a weaker historical climate forcing during the projection period, whereas other experiments use stronger RCP8.5 future forcing. It is worthwhile noting that the experiments with the improvement boast a visual relationship between sample estimates of variability properties and future changes (Fig 4). Specifically, models with higher innovation standard deviation tend to produce higher summer mean maximum temperature warming in the Korean temperature experiments. A positive relationship between standard deviation and future temperature change has been previously found in previous work for many regions [6]. The relationships for the AMOC experiment are different: future AMOC slowdown appears to be stronger for models with higher autocorrelation and low normalized innovation standard deviation. In the Winter_SST experiment, the relationships also involve both variability properties: higher and low in the models are associated with smallest future warming. Thus, we speculate that the degree of improvement may be related to the strength of statistical relationships between the variability parameters and future change. Testing this hypothesis is left to future work. There can be considerable shifts in the pdf between the “trend” and “trend+var” method (Figs N-Q in S1 File). This is consistent with the fact that additional fluctuation data can provide a relatively independent constraint on the model weights.

Subplots differ in the assumed “true” model. Red circles are deterministic projections from each remaining model, red dotted lines are 90% posterior credible intervals. Black lines are changes from the “true” models.

We note that the improvement in performance by the “trend+var” method is not caused by any increase in number of parameters resulting in overfitting. The overall statistical model for the projections is the same in both cases: a weighted mixture of pdfs from individual models. The increase in skill is due to better estimation of individual model weights *w*_{i} in the “trend+var” model through using new variability data constraints on the models.

## 5 Real-Case application: Projecting korean summer mean maximum temperature

We now apply both the “trend” and “trend+var” methods to make projections of Korean summer mean maximum temperature. Specifically, we use 29 GCMs from Coupled Model Intercomparison Project phase 5 (CMIP5, [40]) model runs (the same model set as for the Korea_temp experiment). The models are weighted using 1973–2005 station observational data provided by Korean Meteorological Administration (KMA) weather stations [38,53]. We apply simple area average to daily mean maximum temperatures from the stations before calculating summer mean values. We use this short period because it has the best observational coverage, however to provide a liberal estimate of the uncertainty we take model error expansion factors *f* from the corresponding longer-period Korea_temp_long experiments. Future changes (2081–2100 minus 1973–2005) under the RCP8.5 emissions scenario [42] are presented in Fig 8. The results show (a) notably higher projected warming and (b) considerable reduction of the low-warming (< 2 K) tails after the variability weighting. Specifically, the mean increases from 4.9 K to 5.6 K, and the 5^{th} percentile from 1.8 K to 3.2 K. The new projection mode leaps from 5.3 K to 6.6 K (Table 2). In addition, the 90% credible interval shrinks from 5.5 K to 4.3 K (22% reduction).

Vertical lines are the means and the 90% posterior credible intervals.

## 6 Discussion

Here we present a novel method “trend+var” to weight models of complex dynamical systems by their skill at representing both autocorrelated variability and trend in observations. The key step is association of two statistical models with each dynamical model: a trend statistical model, and a variability statistical model. The component submodels are weighted separately using relevant observations, and then the weights are combined. The combined weights are used to make weighted probabilistic multi-model projections. In a series of cross-validation experiments, we show that the new method appears to better identify the “true” model compared to the trend-only weighting method (“trend”). The new method also tends to perform better in terms of mean 90% posterior credible interval and mean absolute bias. Our analysis deviates in some aspects from the traditional Bayesian framework, in order to avoid making difficult-to-justify parametric assumptions about model error, and to alleviate potential overconfidence in one-at-a-time cross-validation experiments.

Applying the new method to the real case of projecting Korean summer mean maximum temperature change by the end of this century considerably increases future projections. These projections are more informative than from the “trend” method because they use the additional variability and short-term memory (quantified by the lag-1 autocorrelation coefficient) information from both models and observations. Since the BMA predictive model is the same (Eq 14), the increase in skill is not due to an increased number of parameters, but is derived purely through better estimation of model weights. Recent work has found correlations with absolute values of up to approximately 0.8 between present-day interannual summer temperature sample standard deviation in global and regional climate models, and long-term future mean and/or variability changes for some regions [6,25]. This suggests that historical variability in those regions may provide a valuable constraint on the models. Applying the method to those regions should be considered for future work.

It is worth discussing differences between this study and previous Bayesian work. Here we for the first time implement a quasi-Bayesian statistical method that weights models by their performance in terms of both trend, variability, and short-term memory (as quantified by the lag-1 autocorrelation) for a relatively general case: arbitrary (potentially non-linear) trend function and red noise variability. The method can be extended to more complex variability structures. Model weights are obtained by constraining the method with calibration period observations, while a parameter controlling model error assumptions is calibrated using cross-validation experiments. Some prior work does also incorporate variability into model weights [35], however their method has so far been demonstrated on a simple case: serially uncorrelated variability, and a linear mean function. Other studies [3,4,6] also incorporates variability into the analyses. However, these studies do not actually use variability performance to weight the models and ignore autocorrelation skill. Unlike previous work, we do consider autocorrelation, which is a common feature of variability in many observed and modeled processes [54–56].

## 7 Caveats

Our study is subject to several caveats. First, the anomalies around the long-term trend, as well as model-observational residuals are assumed to be red noise processes. However, our framework can be extended to more general cases in the future. We compare the spectra of model anomalies (normalized in the AMOC experiment) for each model and experiment to the 90% confidence intervals for the corresponding AR1 process spectra, based on 1000 random realizations (Figs R-T in S1 File). Relevant comparison for the AMOCIndex experiment is shown in Fig 4 of a preceding study [9]. These results indicate that AR1 process is a reasonable approximation to the internal variability for these systems. Second, when combining the weights of the variability and trend submodels we are assuming independence. While this assumption appears to be generally reasonable here, it may not apply for other datasets. Incorporating dependence should be considered in future studies. Thus, our method is expected to be ideal for cases where there is at least some relationship between present-day variability and future changes, yet the relationship between present-day trends and variability in the models is sufficiently weak to justify the independence assumption we make here. Third, by using a common error expansion factor *f* for the internal variability, trend submodel errors, as well as for the forecasts, we are assuming the magnitudes of errors in these three components are linked. A way forward in subsequent work may be to assume different *f* for trend and variability. The best *f* values could then be found using constrained optimization (optimizing future performance metrics while constraining coverage to be correct). This is beyond the scope of this study. Fourth, when sampling future internal variability, we do not consider the uncertainty in the AR1 parameters of the anomalies. However, as explained in Section 3, we calibrate our method to account for potential overconfidence by scaling the magnitude of the model errors. Other caveats include the simplicity of the future model bias and of the cross-validation experiments, as well as no explicit representation of observational error. For the future Korean temperature projections, the high density of observational network mitigates some of these concerns, as random errors are expected to decrease after averaging across many stations. In addition, theoretically if modelled and observed data from multiple regions are used together in a cross-validation framework, the observational error will be implicitly incorporated into the analysis after nudging the *f* parameter. Nonetheless, an explicit representation of observed error should be considered in the future.

While the focus on this paper is on the statistical weighting methodology by trend and variability performance, the simplicity of the decomposition into trend and variability (e.g., lowess method or linear detrending) deserves mention. The nonlinear trends discussed here may include residual contributions from long-term internal climate variability. However, this can be handled by the trend-weighting part of the method since this part accounts for long-term model error [35]. The unfiltered long-term variability in each model can be simply considered as part of this long-term model error. Previous work provides examples of using a more sophisticated decomposition [18]. Improving the decomposition methodology is beyond the scope of the paper, and is subject of future work.

This work assumes stationarity of model weights: if a model is correct during the calibration period, it is also assumed to be correct in the validation period. This is a standard assumption of the BMA method [1,5,20,21,35].

Notably, this work does not properly confront the issue of model dependence (e.g., the fact that models coming from the same research group, or models with similar outputs are dependent in the general sense of the term) [12,57–60]. This needs to be addressed in future work.

The best new datasets to apply the method to are the ones either with many regions, or with repeated experiments, and where a long calibration period can be split into two subperiods. In this case method performance can be systematically assessed using real observations in cross-validation experiments, and *f* can be properly calibrated. However, any assumption about *f* under new conditions is inherently untestable. Hence, we recommend including equal weights projections along with projections from this (or any other) weighting scheme. In the absence of many regions, and with only short time series available, one has to resort to simulated cross-validation experiments using calibration, projection, and projection reference period model output to calibrate the method. In such cases, if models share common errors, the real value of *f* may be higher than estimated.

## 8 Conclusions

We present a statistically-rigorous novel method to weight multiple models of stochastic dynamical systems by their skill at representing both internal variability (including autocorrelation) and a nonlinear trend of a time series process, and to make predictions of system change under new conditions. The weight is interpreted as a likelihood of a dynamical model being adequate at capturing both trend and variability aspects of the process. This is a particularly important diagnostic given the broad relevance of variability (e.g., variability can affect extreme events such as heat waves and droughts in climate science). We show that the proposed method tends to better identify “true” models in a suite of leave-one-out cross-validation experiments compared to a typically-used trend-only BMA weighting method. The new method also tends to improve forecasts, as judged by the mean 90% credible interval width and mean absolute bias. This has important implications specifically for multi-model climate projections. Applying the method to project Korean summer mean maximum temperature changes over this century considerably increases future projections. Specifically, the mode of 1973–2005 to 2081–2100 warming under the RCP8.5 emissions scenario increases by 1.3 K to 6.6 K, while the mean shifts from 4.9 K to 5.6 K. Furthermore, the pdf becomes 22% sharper as measured by the 90% posterior credible interval.

## Supporting information

### S1 File. The combined supporting information file contains supplementary Texts A and B, supplementary Figs A to T, and supplementary Tables A and B.

https://doi.org/10.1371/journal.pone.0214535.s001

(PDF)

### S2 File. The file contains the R programming language computer code that implements the model weighting methodology.

https://doi.org/10.1371/journal.pone.0214535.s002

(ZIP)

## Acknowledgments

For their roles in producing, coordinating, and making available the CMIP5 model output, we acknowledge the climate modeling groups, the World Climate Research Programme’s (WCRP) Working Group on Coupled Modelling (WGCM), and the Global Organization for Earth System Science Portals (GO-ESSP). A portion of model outputs used here has been obtained from the German Climate Computing Centre (DKRZ), funded through the Federal Ministry for Education and Research. Jong-Soo Shin provided technical assistance with extracting Korean and East Sea temperature model output. We are thankful to Cameron Farnsworth for providing comments on the manuscript, and to Soong-Ki Kim for thought-provoking discussions.

## References

- 1. Bhat KS, Haran M, Terando A, Keller K. Climate Projections Using Bayesian Model Averaging and Space-Time Dependence. J Agric Biol Environ Stat. 2011;16(4):606–28.
- 2. Buser CM, Künsch HR, Lüthi D, Wild M, Schär C. Bayesian multi-model projection of climate: bias assumptions and interannual variability. Clim Dyn. 2009 Nov 1;33(6):849–68.
- 3. Buser CM, Kúnsch HR, Schär C. Bayesian multi-model projections of climate: Generalization and application to ENSEMBLES results. Clim Res. 2010;44(2–3):227–41.
- 4. Chandler RE. Exploiting strength, discounting weakness: combining information from multiple climate simulators. Philos Trans R Soc Lond Ser A. 2013 Apr 1;371:20120388–20120388.
- 5. Duan Q, Ajami NK, Gao X, Sorooshian S. Multi-model ensemble hydrologic prediction using Bayesian model averaging. Adv Water Resour. 2007 May;30(5):1371–86.
- 6. Huttunen JMJ, Räisänen J, Nissinen A, Lipponen A, Kolehmainen V. Cross-validation analysis of bias models in Bayesian multi-model projections of climate. Clim Dyn. 2017 Mar 1;48(5–6):1555–70.
- 7. Knutti R. Should we believe model predictions of future climate change? Philos Trans R Soc Lond Math Phys Eng Sci. 2008 Dec 28;366(1885):4647–64.
- 8. Olson R, Fan Y, Evans JP. A simple method for Bayesian model averaging of regional climate model projections: Application to southeast Australian temperatures. Geophys Res Lett. 2016 Jul 28;43(14):2016GL069704.
- 9. Olson R, An S-I, Fan Y, Evans JP, Caesar L. North Atlantic observations sharpen meridional overturning projections. Clim Dyn. 2017 Aug 23;1–18.
- 10. Raftery AE, Gneiting T, Balabdaoui F, Polakowski M. Using Bayesian model averaging to calibrate forecast ensembles. Mon Weather Rev. 2005;133(5):1155–74.
- 11. Rougier J, Goldstein M, House L. Second-order exchangeability analysis for multimodel ensembles. J Am Stat Assoc. 2013;108(503):852–63.
- 12. Steinschneider S, McCrary R, Mearns LO, Brown C. The effects of climate model similarity on probabilistic climate projections and the implications for local, risk-based adaptation planning. Geophys Res Lett. 2015 Jun 28;42(12):2015GL064529.
- 13.
Tebaldi C, Sanso B, Smith RL, Ferreira MAR. Characterizing Uncertainty of Future Climate Change Projections using Hierarchical Bayesian Models. In: Jose Bernardo, Bayarri MJ, Berger JO, Dawid AP, Heckerman D, Smith AFM, et al., editors. Bayesian Statistics 9. Oxford Scholarship Online; 2011.
- 14. Tebaldi C, Sansó B. Joint projections of temperature and precipitation change from multiple climate models: A hierarchical Bayesian approach. J R Stat Soc Ser A Stat Soc. 2009;172(1):83–106.
- 15. Terando A, Keller K, Easterling WE. Probabilistic projections of agro-climate indices in North America. J Geophys Res Atmospheres. 2012;117(8).
- 16. Wallach D, Mearns LO, Ruane AC, Rötter RP, Asseng S. Lessons from climate modeling on the design and use of ensembles for crop modeling. Clim Change. 2016;139(3–4):551–64.
- 17. Xu Y, Gao X, Giorgi F. Upgrades to the reliability ensemble averaging method for producing probabilistic climate-change projections. Clim Res. 2010;41(1):61–81.
- 18. Qi Y, Qian C, Yan Z. An alternative multi-model ensemble mean approach for near-term projection. Int J Climatol. 2017 Jan 1;37(1):109–22.
- 19. Acharya N, Chattopadhyay S, Mohanty UC, Ghosh K. Prediction of Indian summer monsoon rainfall: a weighted multi-model ensemble to enhance probabilistic forecast skills. Meteorol Appl. 2014 Jul 1;21(3):724–32.
- 20. Hoeting JA, Madigan D, Raftery AE, Volinsky CT. Bayesian Model Averaging: A Tutorial. Stat Sci. 1999;14(4):382–417.
- 21. Montgomery JM, Nyhan B. Bayesian model averaging: Theoretical developments and practical applications. Polit Anal. 2010;18(2):245–70.
- 22. Yun WT, Stefanova L, Krishnamurti TN. Improvement of the Multimodel Superensemble Technique for Seasonal Forecasts. J Clim. 2003 Nov 1;16(22):3834–40.
- 23. Yun WT, Stefanova L, Mitra AK, Kumar TSVV, Dewar W, Krishnamurti TN. A multi-model superensemble algorithm for seasonal climate prediction using DEMETER forecasts. Tellus A. 2005 May 1;57(3):280–9.
- 24. Bayes M, Price M. An Essay towards Solving a Problem in the Doctrine of Chances. By the Late Rev. Mr. Bayes, F. R. S. Communicated by Mr. Price, in a Letter to John Canton, A. M. F. R. S. Philos Trans. 1763 Jan 1;53:370–418.
- 25. Fischer EM, Rajczak J, Schär C. Changes in European summer temperature variability revisited. Geophys Res Lett. 2012 Oct 16;39(19):L19702.
- 26. Cox PM, Huntingford C, Williamson MS. Emergent constraint on equilibrium climate sensitivity from global temperature variability. Nature. 2018;553(7688):319–22. pmid:29345639
- 27. Kopp RE, Shwom RL, Wagner G, Yuan J. Tipping elements and climate–economic shocks: Pathways toward integrated assessment. Earths Future. 2016;4(8):346–72.
- 28. Feng QY, Viebahn JP, Dijkstra HA. Deep ocean early warning signals of an Atlantic MOC collapse. Geophys Res Lett. 2014;41(16):6008–14.
- 29. Kleinen T, Held H, Petschel-Held G. The potential role of spectral properties in detecting thresholds in the Earth system: Application to the thermohaline circulation. Ocean Dyn. 2003;53(2):53–63.
- 30. Lenton TM, Livina VN, Dakos V, Van Nes EH, Scheffer M. Early warning of climate tipping points from critical slowing down: Comparing methods to improve robustness. Philos Trans R Soc Math Phys Eng Sci. 2012;370(1962):1185–204.
- 31. Thomas ZA. Using natural archives to detect climate and environmental tipping points in the Earth System. Quat Sci Rev. 2016;152:60–71.
- 32. Braverman A, Cressie N, Teixeira J. A likelihood-based comparison of temporal models for physical processes. Stat Anal Data Min. 2011 Jun 1;4(3):247–58.
- 33. Kwasniok F. Analysis and modelling of glacial climate transitions using simple dynamical systems. Phil Trans R Soc A. 2013 May 28;371(1991):20110472. pmid:23588043
- 34. Peavoy D, Franzke C. Bayesian analysis of rapid climate change during the last glacial using Greenland δ18O data. Clim Past. 2010;6(6):787–94.
- 35. Fan Y, Olson R, Evans JP. A Bayesian posterior predictive framework for weighting ensemble regional climate models. Geosci Model Dev. 2017 Jun 23;10(6):2321–32.
- 36. Cleveland WS. Robust Locally Weighted Regression and Smoothing Scatterplots. J Am Stat Assoc. 1979 Dec 1;74(368):829–36.
- 37. Sexton DMH, Murphy JM, Collins M, Webb MJ. Multivariate probabilistic projections using imperfect climate models part I: Outline of methodology. Clim Dyn. 2012;38(11–12):2513–42.
- 38. Shin J, Olson R, An S-I. Projected Heat Wave Characteristics over the Korean Peninsula During the Twenty-First Century. Asia-Pac J Atmospheric Sci. 2018;54(1):53–61.
- 39. Kanzow T, Cunningham SA, Johns WE, Hirschi JJ-M, Marotzke J, Baringer MO, et al. Seasonal Variability of the Atlantic Meridional Overturning Circulation at 26.5°N. J Clim. 2010 Jun 11;23(21):5678–98.
- 40. Taylor KE, Stouffer RJ, Meehl GA. An Overview of CMIP5 and the Experiment Design. Bull Am Meteorol Soc. 2011 Oct 7;93(4):485–98.
- 41.
ESGF LLNL. ESGF @ DOE/LLNL [Internet]. 2016. https://esgf-node.llnl.gov/projects/esgf-llnl/
- 42. Moss RH, Edmonds JA, Hibbard KA, Manning MR, Rose SK, Vuuren DP, et al. The next generation of scenarios for climate change research and assessment. Nature. 2010 Feb 11;463(7282):747–56. pmid:20148028
- 43. Sen PK. Estimates of the Regression Coefficient Based on Kendall’s Tau. J Am Stat Assoc. 1968;63(324):1379–89.
- 44. Rahmstorf S, Box JE, Feulner G, Mann ME, Robinson A, Rutherford S, et al. Exceptional twentieth-century slowdown in Atlantic Ocean overturning circulation. Nat Clim Change. 2015 May;5(5):475–80.
- 45. Huang B, Banzon VF, Freeman E, Lawrimore J, Liu W, Peterson TC, et al. Extended Reconstructed Sea Surface Temperature Version 4 (ERSST.v4). Part I: Upgrades and Intercomparisons. J Clim. 2014 Oct 14;28(3):911–30.
- 46. Huang B, Thorne PW, Smith TM, Liu W, Lawrimore J, Banzon VF, et al. Further Exploring and Quantifying Uncertainties for Extended Reconstructed Sea Surface Temperature (ERSST) Version 4 (v4). J Clim. 2015 Dec 11;29(9):3119–42.
- 47. Liu W, Huang B, Thorne PW, Banzon VF, Zhang H-M, Freeman E, et al. Extended Reconstructed Sea Surface Temperature Version 4 (ERSST.v4): Part II. Parametric and Structural Uncertainty Estimations. J Clim. 2014 Nov 20;28(3):931–51.
- 48. Hirahara S, Ishii M, Fukuda Y. Centennial-Scale Sea Surface Temperature Analysis and Its Uncertainty. J Clim. 2013 Sep 3;27(1):57–75.
- 49.
Huang B, Banzon VF, Freeman E, Lawrimore J, Liu W, Peterson TC, et al. Extended Reconstructed Sea Surface Temperature (ERSST), Version 4, NOAA National Centers for Environmental Information [Internet]. 2016. https://data.nodc.noaa.gov/cgi-bin/iso?id=gov.noaa.ncdc:C00884
- 50.
Ishii M. COBE-SST2 Version 2.9.2 [Internet]. 2016. https://amaterasu.ees.hokudai.ac.jp/~ism/pub/cobe-sst2/
- 51. Hansen J, Ruedy R, Sato M, Lo K. Global Surface Temperature Change. Rev Geophys. 2010 Dec 1;48(4):RG4004.
- 52.
GISTEMP Team. GISS Surface Temperature Analysis (GISTEMP) [Internet]. NASA Goddard Institute for Space Studies; 2016 [cited 2016 Jun 15]. https://data.giss.nasa.gov/gistemp/
- 53.
Korea Meteorological Administration. Open Portal for Meteorological Data (in Korean) [Internet]. 2016. https://data.kma.go.kr
- 54. Hasselmann K. Stochastic climate models Part I. Theory. Tellus. 1976 Jan 1;28(6):473–85.
- 55. Keller K, McInerney D. The dynamics of learning about a climate threshold. Clim Dyn. 2008 Feb 1;30(2–3):321–32.
- 56. Olson R, Sriver R, Chang W, Haran M, Urban NM, Keller K. What is the effect of unresolved internal climate variability on climate sensitivity estimates? J Geophys Res Atmospheres. 2013;118(10):4348–58.
- 57. Annan JD, Hargreaves JC. On the meaning of independence in climate science. Earth Syst Dyn. 2017;8(1):211–24.
- 58. Bishop CH, Abramowitz G. Climate model dependence and the replicate Earth paradigm. Clim Dyn. 2013;41(3–4):885–900.
- 59. Haughton N, Abramowitz G, Pitman A, Phipps SJ. Weighting climate model ensembles for mean and variance estimates. Clim Dyn. 2015;45(11–12):3169–81.
- 60. Leduc M, Laprise R, de Elía R, Šeparović L. Is institutional democracy a good proxy for model independence? J Clim. 2016;29(23):8301–16.