Next Article in Journal
Modeling the Impacts of Climate Change on Crop Yield and Irrigation in the Monocacy River Watershed, USA
Previous Article in Journal
Climate Change and Its Possible Impact in Groundwater Resource of the Kankai River Basin, East Nepal Himalaya
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Effect of Statistical Downscaling on the Weighting of Multi-Model Ensembles of Precipitation

1
South Central Climate Adaptation Science Center, University of Oklahoma, Norman, OK 73019, USA
2
Jet Propulsion Laboratory, California Institute of Technology, Pasadena, CA 91109, USA
*
Author to whom correspondence should be addressed.
Submission received: 23 October 2020 / Revised: 19 November 2020 / Accepted: 20 November 2020 / Published: 25 November 2020
(This article belongs to the Section Climate Dynamics and Modelling)

Abstract

:
Recently, assessments of global climate model (GCM) ensembles have transitioned from using unweighted means to weighted means designed to account for skill and interdependence among models. Although ensemble-weighting schemes are typically derived using a GCM ensemble, statistically downscaled projections are used in climate change assessments. This study applies four ensemble-weighting schemes for model averaging to precipitation projections in the south-central United States. The weighting schemes are applied to (1) a 26-member GCM ensemble and (2) those 26 members downscaled using Localized Canonical Analogs (LOCA). This study is distinct from prior research because it compares the interactions of ensemble-weighting schemes with GCMs and statistical downscaling to produce summarized climate projection products. The analysis indicates that statistical downscaling improves the ensemble accuracy (LOCA average root mean square error is 100 mm less than the CMIP5 average root mean square error) and reduces the uncertainty of the projected ensemble-mean change. Furthermore, averaging the LOCA ensemble using Bayesian Model Averaging reduces the uncertainty beyond any other combination of weighting schemes and ensemble (standard deviation of the mean projected change in the domain is reduced by 40–50 mm). The results also indicate that it is inappropriate to assume that a weighting scheme derived from a GCM ensemble matches the same weights derived using a downscaled ensemble.

1. Introduction

Climate modeling is traditionally directed toward research to improve our understanding of the climate system [1] and to develop climate projections that aid climate change decision-making [2]. However, regional and local scales are not well represented by global climate models GCMs, [3]. Downscaling techniques can be used to reduce the biases of GCMs, translate the GCM-simulated response to local scales, and provide added information for decision-making [4,5]. Downscaling techniques include regional climate modeling and statistical downscaling (SD) and bias correction methods. SD methods are both computationally efficient and flexible, which has led to their use for impact assessments, e.g., [6,7,8,9,10], including the National Climate Assessment NCA, [11].
Numerous assessments of GCMs currently exist, e.g., [12,13,14,15,16,17], and downscaling is being investigated in programs such as the Coordinated Regional Downscaling Experiments CORDEX, e.g., [18]. Recently, there has been a transition from using an unweighted, multi-model ensemble mean to more advanced methods that account for the skill and independence of models to inform the weighting strategy [15,16,19,20]. This transition has occurred, in part, from the recognition that some models can have more skill for certain variables and regions, but also because the use of models with common code bases can result in a lack of independence between GCMs [15,16,19,20,21,22]. In addition, it has been noted that the spread in the weighted multi-model mean can be interpreted more accurately as a measure of uncertainty while the spread of the unweighted average is not a measure of uncertainty [23,24]. Given this knowledge, the Fourth National Climate Assessment (NCA4) was the first of the NCA reports to use an ensemble-weighting scheme [25]. Its ensemble weights were calculated using an ensemble of GCMs, but the ensemble weights were applied to a statistically downscaled ensemble.
One reason to address model or statistical downscaling (SD) method independence with ensemble weighting is the replication of programming code across institutions [22,26]. While individual GCMs share parts of code from various modeling institutions, one can argue that applying an SD method to an ensemble of GCMs has two significant effects. First, it applies a single common code and statistical framework to all GCMs in the ensemble, which can cause the weighting scheme of a post-SD ensemble to force the weights of all members to be equal [24]. Second, SD methods correct GCM biases based on a single common training dataset which often has biases of its own that can influence the projections [27,28]. In essence, all the members of a post-SD ensemble of GCMs share the biases of a single statistical technique and training dataset, which may further reduce ensemble independence. In this study, our results will provide details on how the model independence is sensitive to the SD method and the weighting scheme used for the model averaging.
Sanderson et al. [19] present a strategy to weight a climate model ensemble that considers both the skill in the climatological performance of individual models as well as their interdependency. The purpose of this weighting strategy is to weigh GCMs to account for historical skill and model independence. This strategy has been implemented in several studies so far, e.g., [15,19,20,22] as well as in NCA4 [11]. Eyring et al. [29] thoroughly discuss the state of GCM evaluation and how others have constrained the spread of uncertainty in future projections. They show that there is a need to better understand advanced weighting methods to provide decision-makers a more tightly constrained set of climate projections. Although it is important to consider a range of future projections in risk assessment, e.g., [30], stakeholders searching for projections express a need for assistance to select useful projections and assess uncertainty, e.g., [31]. Therefore, the scientific community needs to apply model-averaging techniques to more judiciously combine multi-model climate projections while considering the interdependency of the GCMs. One technique to do so is the Bayesian Model Averaging (BMA) method [15,16,32]. BMA produces a multi-model average created from optimized model weights such that the BMA-weighted model ensemble average for the historical simulation closely matches the observation and its uncertainty. In this study, we apply both the BMA weighting strategy and the weighting scheme introduced by Sanderson et al. [19] to study the post-SD effect on the weighting of multi-model ensembles of precipitation for the south-central United States.
To our knowledge, the influence of statistical downscaling on weighted-model projections and the independence of the ensemble has not been examined. In addition, BMA has not yet been implemented with a statistically downscaled ensemble, although examples of recent work include the implementation of BMA in a hybrid approach with quantile mapping [24] in the context of global atmospheric rivers [15] and for precipitation over the contiguous United States [16]. This study will focus on three questions. First, what are the sensitivities of the GCM and statistically downscaled ensembles to the various weighting schemes? Second, will a BMA approach together with statistical downscaling increase the historical ensemble accuracy and improve confidence in the ensemble-mean projected change? Finally, is it appropriate to assume that ensemble weighting derived from GCMs will produce similar results when applied to a GCM and statistically downscaled ensemble? In this manuscript, Section 2 overviews the methods and data used for this analysis. Section 3 presents our results and discussion. Section 4 concludes this work and recommends future work to further this study.

2. Materials and Methods

2.1. Study Domain and Variables

The south-central United States (from about 26° N 108.5° W to 40° N 91° W) has a varied topography with a sharp gradient in mean annual precipitation from the east (humid) to the west (arid). The region includes the Mississippi River Valley and the Ozark Mountains in the east (elevations of 200–800 m), the Rocky Mountains in the west (1500–4400 m), and the Gulf of Mexico in the southeast (near sea level). Precipitation in the southeast portion of the domain can be eight times higher than drier western locations (Figure 1).
Our study domain excludes Mexico and the Gulf of Mexico, as gridded observations are limited over these areas. The precipitation gradient is challenging for GCMs to represent in this region [28]. Precipitation is also more difficult to simulate in climate models and to downscale, e.g., [14,33,34,35,36]. The complex topography and climatological differences in precipitation provide an opportunity to assess the weighting schemes and statistical downscaling, helping generalize the results of this study to other regions.

2.2. Climate Projection Datasets

We use one member each from 26 GCMs in from the Coupled Model Intercomparison Projection Phase 5 (CMIP5 [40]) archive to form the GCM multi-model ensemble. To form the downscaled ensemble, the same 26 GCMs are used from the downscaled projections created with the Localized Constructed Analogs (LOCA) method [41]. The LOCA-downscaled projections have been used in other studies and the NCA4 [11]. Table S1 lists the GCMs used for both the GCM ensemble (hereafter CMIP5 ensemble) and downscaled ensemble (hereafter LOCA ensemble).
To facilitate analysis, the data for each ensemble member are interpolated from their native resolution to a common 10 km grid across the south-central U.S. using a bi-linear interpolation similar to that described in Wootten et al. [28]. We examine projected precipitation changes from 1981–2005 to 2070–2099 driven by RCP 8.5, which ramps the anthropogenic radiative forcing to 8.5 W/m2 by 2100 [42,43]. We chose RCP 8.5 to maximize the change signals and allow us to analyze greater differences between weight schemes and downscaling techniques. The historical period (1981–2005) is used for both the historical simulations and historical observations (next section) to facilitate comparisons with other efforts in this region [28] and because the historical period of the CMIP5 archive (and SD projections) also ends in 2005 [40]. This historical period is also of similar length to other similar studies, e.g., [13,14,44,45].

2.3. Observation Dataset

Although weather stations can be used for training or calibration purposes in statistical downscaling, many publicly available downscaled projections (including LOCA) are created using gridded observation-based data for training. Gridded observations are based largely on station data that are adjusted and interpolated to a grid in a manner that attempts to account for data biases, temporal/spatial incoherence, and missing station data [27,28,46,47]. In this study, we use Livneh version 1.2 (hereafter Livneh [37]) as the gridded observation data used for comparison to the ensembles. The Livneh data served as the training data for the LOCA ensemble, so it is expected that LOCA will be more accurate than the GCMs when compared to the Livneh dataset. The Livneh observations are also interpolated to the same 10 km grid using bilinear interpolation to facilitate analysis. While we recognize that different gridded observations and downscaling techniques influence projections of precipitation, the effect is minimal on mean annual precipitation [28]. Therefore, we hold that it is appropriate to make use of only one SD method and one gridded observation dataset.

2.4. Weighting Schemes

In this analysis, we apply an unweighted multi-model mean and four weighting schemes to both the CMIP5 and LOCA ensembles, similar to the experimental design of Massoud et al. [15]. These four weighting schemes are detailed briefly below along with a brief discussion of differences between the Sanderson et al. [20] method and BMA. Figure 2 shows a schematic flowchart of the analysis with the weighting schemes.
The CMIP5 and LOCA ensembles are supplied to each weighting scheme. All the weighting schemes used in this analysis derive model weights using precipitation in the full domain separately for both the CMIP5 and LOCA ensembles. The ensemble mean for each combination of weighting scheme and ensemble (CMIP5 and LOCA) is calculated at each grid cell in the domain.

2.4.1. Historical Skill Weighting

The first weighting scheme (hereafter Skill) is a simple skill weighting based on the root mean square error (RMSE) of each ensemble member compared to the Livneh gridded observations. The weights are normalized so that the resulting weights of the 26 ensemble members sum to one.

2.4.2. Historical Skill and Historical Independence Weighting (SI-h)

The second weighting method is described by Sanderson et al. hereafter SI-h, [20]. The SI-h uses a normalized area-weighted RMSE matrix that compares the RMSE of each model against observations (skill) and each model against all other models (independence). In a multivariate context, a normalized matrix for each variable that is linearly combined using equal weights for each variable. In this study, we use a single normalized RMSE matrix given our focus on precipitation only. A set of skill and independence weights are calculated using the normalized RMSE matrix. For the independence weights, the effective repetition of each model is calculated using the normalized matrix of RMSE. The independence weight of a model is the inverse of the effective repetition. Models that are more alike to other models in the ensemble have a higher (lower) effective repetition (independence weight). The skill weights compare the normalized RMSE of each model against the observations. Models with lower RMSE are given a higher skill weight. Both the independence and skill weights are calculated using a radius of similarity and quality which determines the degree to which a model should be down-weighted for independence and skill. For this analysis we set the radius of similarity and quality to 0.48 and 0.8, respectively, matching Sanderson et al. [20]. The resulting skill and independence weights for the ensemble members are multiplied together to produce the model weights. These weights also are normalized (i.e., final weights sum to one). SI-h weights are calculated only for our study region.

2.4.3. Historical Skill and Future Independence Weighting (SI-c)

The third weighting method is a variation of the SI-h method described in Sanderson et al. [20]. One can argue that statistical downscaling improves the skill of the ensemble in the historical period by training with gridded observations, simultaneously nudging every model toward these observations during the historical period (i.e., bias correction). As such, one would expect that the skill (independence) weights of a downscaled ensemble would be high (low). Still, a primary goal of statistical downscaling is to retain the future-change signal of the GCMs in the downscaled output [48]. Therefore, it is appropriate to examine the effect of downscaling on the independence of the change signals in the multi-model ensemble. To do so, we alter SI-h by calculating the normalized area-weighted RMSE matrix using the projected change in precipitation from each model in the ensemble for the independence weighting. That is, the independence weighting in this adjusted scheme (hereafter SI-c) focuses on the models’ effective repetition of the change signal instead of the historical climatology. The SI-c skill weights are identical to those in SI-h, and the skill and independence weights are combined and normalized in the same manner as the SI-h scheme to determine the SI-c final model weights.

2.4.4. Bayesian Model Averaging

Hoeting et al. [32] provide a comprehensive overview of Bayesian Model Averaging (BMA) and its variants, which have been widely used in previous literature. BMA differs from other model-averaging methods by explicitly estimating the weights and associated uncertainties of the weights for each model through maximizing a specified likelihood function. In other words, BMA obtains model weights that produce model combinations that have the maximum likelihood of matching historical observations compared to other model combinations. Using these optimized weights, BMA constructs the mean and uncertainty distribution of the performance metrics (or objective function) of interest. Applications of BMA have been described in numerous works [15,16,49,50,51,52,53,54,55,56,57].
The BMA method offers an alternative to the selection of a single model from several candidate models, by weighting each candidate model according to its statistical evidence, which is proportional to the model’s skill and independence. Since the BMA method estimates a distribution of model weights, various model combinations become possible, which implicitly addresses the model dependence issue. In other words, consider that in the BMA framework there is a hypothetical ‘Model A’ and a ‘Model B’ that are similar and therefore not independent; ‘Model A’ may have higher weights in some combinations, and conversely, ‘Model B’ might have higher weights in other combinations. Consequently, if both models are rewarded in the same set of weights, each model likely receives a reduced weight because both models are providing information to the model average. Therefore, model dependence can play a role in the BMA scheme since both of the dependent models can affect each other’s weights, which can be portrayed in the posterior samples. See Massoud et al. [15] and the supplementary information of Massoud et al. [16] for more information on how BMA is implemented.
Successful use of the BMA/Markov Chain Monte Carlo (MCMC) application depends on many input factors, such as the number of chains, the prior used for the parameters, the number of generations to sample, the convergence criteria, among other things [58]. For our application, we used C = 10 chains, the prior was a uniform distribution from 0–1 for each model weight, each sampled set of weights was normalized so that the sum of weights equaled 1, the number of generations was set at G = 5000 for each metric being fit, and the convergence of the chains relied on the Gelman and Rubin [59] diagnostic, where we applied the commonly used convergence threshold of R = 1.2.

2.4.5. Differences between SI-h and BMA

Similar to how other studies have defined dependence [15,19,20], we define it here as the ‘distance’ of the models in the ‘model space’. This definition indicates that a metric like RMSE describes dependence. The closer two models are (i.e., the lower the RMSE between them), the more dependent they are and therefore provide similar information to the model ensemble. Conversely, the larger the RMSE (i.e., the larger the ‘distance’) between two models, the more independent they are. In the Sanderson approach, shown in Sanderson et al. [19,20] and Massoud et al. [15], there is an explicit consideration of both the skill and the dependence of the model ensemble when calculating the model weights. This explicit consideration only covers the first-order effect, however, by slightly nudging model weights based on a quantitative assessment of model skill and dependence. This quantitative assessment is subjective because the user can choose whether to apply higher weights for dependent models or higher weights for skillful models. In comparison, the BMA method is objective, through the use of the objective function (or likelihood function), and is a method that captures the total order of the model weight combinations. The field of climate research still warrants proper model averaging methods see [29]. We believe that the combination of methods described in our paper (i.e., Sanderson approach and the BMA method) offers a way to infer the total distributions of these model combinations based on statistics that the user desires.

3. Results and Discussion

To begin, we explore the influence of statistical downscaling on the ensemble weighting followed by an examination of the historical accuracy and projected changes resulting from each ensemble.

3.1. Ensemble Weights

Applying the SI-h and SI-c approaches, one can examine the behavior of the skill and independence weights separately. The SI-h weighted CMIP5 ensemble has both smaller skill weights and greater independence weights than the SI-h weighted LOCA ensemble (Figure 3).
This result is to be expected given the nature of statistical downscaling with GCMs. Each member of the ensemble downscaled with LOCA is bias-corrected to closely resemble the training data in the historical period. This connection, in turn, makes the LOCA downscaled ensemble more skillful compared to the observations but also makes each ensemble member in the historical period alike to the training data used for downscaling. That is, each downscaled ensemble member now shares in common the errors of the training data used. As a result, the weights from the CMIP5 and LOCA ensembles can be quite different, though the method used to craft the ensemble weights is the same. The average skill weights for the CMIP5 and LOCA ensembles using SI-h are 0.44 and 0.95, respectively. The average independence weights for the CMIP5 and LOCA ensembles using SI-h are 0.13 and 0.040, respectively.
Recall that the SI-c method examines the independence of the ensemble change signal alongside the historical skill to craft the associated ensemble weights. The SI-c weighted ensembles have higher independence weights than the SI-h ensembles, but they retain the same skill weights. As such, the LOCA SI-c ensemble has a higher skill weight than the CMIP5 SI-c ensemble but also retains a similar range of independence weights as compared to the CMIP5 SI-c ensemble. The average independence weight for the LOCA SI-c ensemble (0.69) is slightly lower than that of the CMIP5 SI-c ensemble (0.72). This result suggests that the statistical downscaling technique may have little influence on the independence of the ensemble. In addition, precipitation is one of the most challenging variables to represent in climate modeling, e.g., [34] and this study developed the ensemble weights for each scheme using precipitation only. Using a multivariate approach to developing the ensemble weights may further decrease the independence weights of the downscaled ensemble.
Figure 4 shows the model weights estimated using the various methods. The difference between the SI-h and SI-c approaches to independence is shown by the resulting overall weights of each approach (Figure 4).
The focus only on the historical period causes the SI-h and skill weights to be nearly identical in the LOCA ensemble (where each member has a nearly equivalent weight). However, the SI-c weights are more variable than the SI-h weight and have unique weights for each ensemble member. That said, for both CMIP5 and LOCA ensembles, the weights of the Skill, SI-h, and SI-c approaches generally cluster together (Figure 4).
We then look at the BMA weights for the CMIP5 and LOCA ensembles. The BMA method provides a distribution of weights for each model, and Figure 4 shows this distribution compared to other weighting schemes. For both the CMIP5 and LOCA ensembles, the BMA distributions of weights (shown in the blue box-and-whisker plots in Figure 4) tend to span the values estimated with the Skill, SI-h, and SI-c approaches, but also cover a broader region of the model weight space, since many model combinations are possible with BMA. Furthermore, the best combination estimated with BMA, marked as ‘BMA best’ in Figure 4, is significantly different from the other weighting schemes.
While formulated slightly differently, the Skill, SI-h, and SI-c weighting schemes look only at one moment of the distribution, essentially using only one sample to estimate the model weights. In contrast, the BMA approach examines multiple moments of the distribution of the model weights using MCMC sampling. The BMA computes a distribution of weights that gives a similar fit to the observations and results in thousands of samples that can be used to propagate the uncertainty from the multi-model ensemble [15,16]. In addition, the BMA approach rewards skillful models while also penalizing co-dependence, but it does so using a greater number of samples compared to the SI-h and SI-c weightings. As such, it is likely the multiple moments used in BMA drive the weighting for models to be significantly different than the SI-c, SI-h, and skill weights.
In addition, one also can observe that the BMA weights (and their distributions) are not the same as those of the CMIP5 and LOCA downscaled ensembles. The top three models in the CMIP5 ensemble based on the BMA best weights are ACCESS 1-3, EC-EARTH, and CMCC-CM (Figure 4a). The top three models in the LOCA ensemble based on the BMA best weights are CanESM2, CESM1-BGC, and MIROC-ESM (Figure 4b). Statistical downscaling bias corrects each GCM, inherently improving the historical skill. However, given the common modeling framework that statistical downscaling applies to each GCM, there is an additional level of co-dependence applied to each GCM. In the SI-h and SI-c, the resulting pattern of the weights is similar across both ensembles, though the magnitude of the weights’ changes. Given that the BMA weights are quite different between ensembles, it is plausible that the BMA weighting trained with the LOCA ensemble responds to and captures the changes to skill and co-dependence in the ensemble more effectively than the SI-c and SI-h.

3.2. Historical Biases and RMSE

In the historical period (1981–2005), the variability of the within-ensemble members is different pre and post-SD. The LOCA ensemble members are inherently less variable than those of the CMIP5 ensemble (Figure 5, left) as a result of the bias correction performed during the downscaling process.
For each ensemble, the variabilities of the BMA-simulated means also are lower than their corresponding raw ensemble (Figure 5, right). This increased confidence in the ensemble mean is reflected in the associated standard deviations (Table 1).
In addition, the best BMA weighted ensemble means for the entire domain are consistently closer to Livneh for both the CMIP5 and LOCA ensembles. This result also is reflected in the RMSE of the weighted ensemble means (Figure 6).
Although LOCA reduces the RMSE of each ensemble member, any weighting scheme tends to reduce the RMSE of the ensemble mean compared to the unweighted mean. In addition, the ensemble mean across the domain created using the best BMA weights has the lowest RMSE compared to the observations.
From Figure 5 and Figure 6, we note that the LOCA ensemble itself is more accurate than the CMIP5 ensemble for precipitation, but also tends to underestimate precipitation in the south-central United States. In addition, we also note that the BMA best weighting does not provide the most accurate ensemble mean of the BMA simulations for individual states (Figure 6). All the weighting schemes used in this analysis derive weights using precipitation in the full domain. The resulting weighting schemes may not produce the most accurate ensemble means in regional subsets. One could repeat the derivation of weights for individual states or smaller regions. Therefore, the results of the analyses presented in this section and the following sections are still valid for this and other regions.

3.3. Projected Changes

Figure 7 displays the mean projected change and also highlights unique differences between ensemble means.
Although the spatial pattern for the mean projected change is similar between the unweighted, Skill-weighted, SI-h weighting, and SI-c weighting, the ensemble mean based on the BMA best weighting has a slightly different pattern (Figure 7). For the CMIP5 ensemble, the BMA ensemble mean projects a significant increase in precipitation in the northeast portion of the domain with little to no decrease in the southwest portion of the domain. For the LOCA ensemble, the BMA ensemble mean projects an increase in precipitation across much of the northern portion of the domain, with significant drying in eastern Texas and Louisiana. For the SI-h, SI-c, skill, and unweighted ensembles, the spatial patterns remain consistent across both weighting schemes and ensembles (though the magnitudes are different). Lastly, we notice in Figure 6 that the spatial pattern of the BMA mean projected change is different between the CMIP5 and LOCA ensembles.
The mean projected change across the region from the raw ensembles also is more variable, and thus more uncertain, than the corresponding BMA simulations (Figure 8).
However, the distribution of the mean projected change from the BMA is not identical between the LOCA and CMIP5 ensembles either across the domain or for individual states. In addition, the BMA best weighted mean does not share the same pattern between ensembles while the other weighting schemes demonstrate a similar pattern for mean projected change across the region across ensembles. This result reflects the differences in the weights between ensembles, as BMA accounts for the downscaling procedure. The BMA best weighting generally improves the confidence in the mean projected change over the other weighting schemes or unweighted mean (Table 2). Finally, the combination of the LOCA and BMA results in the smallest uncertainty associated with this region and three of four states, as compared to all other combinations tested in this analysis.

3.4. Implications of Results

To the best of our knowledge, this study is the first to comparatively analyze multiple ensemble weighting schemes and the interaction with statistical downscaling. In addition, it appears to be the first study using BMA with a set of statistically downscaled projections to determine ensemble weighting used for both historic and future model simulations of precipitation. In prior efforts to derive ensemble weighting, including the NCA4 [25], the implicit assumption has been that the ensemble weighting derived using a GCM ensemble will match the weights derived using a downscaled ensemble. The previous work of Sanderson et al. [20] shows the common biases and co-dependence associated with shared code between GCMs, and we have extended this logic to the added process of statistical downscaling. To create a statistically downscaled set of projections, a single statistical modeling process is applied to all GCMs used to improve the spatial resolution and correct biases toward a single common set of training data. Therefore, each member of a downscaled ensemble now shares the biases of the statistical downscaling technique and the training data used. That is, using statistical downscaling (applying similar coding to all ensemble members) introduces additional co-dependences into the ensemble in a similar fashion to what is already known for GCMs [20]. This result is confirmed in our paper, as displayed in Figure 3 which shows how the LOCA historical independence weights are relatively low for all models in the ensemble, compared to those of the historical CMIP5 model ensemble.
While somewhat evident when using the modified Sanderson approach (SI-c), the influence of downscaling on the ensemble is most evident in the BMA results. The BMA is designed to determine the appropriate weighting based on thousands of simulations and the multiple moments of the resulting distribution [15]. BMA rewards the skillful model while penalizing co-dependence. While the Sanderson approach (both SI-h and SI-c) examines only a single moment of the distribution, BMA’s incorporation of multiple moments of the distribution allows it to more thoroughly address both concerns of co-dependence and historical skill. As a result, the BMA more capably captures the additional co-dependence and skill introduced by statistically downscaling an ensemble. The consideration of the additional moments of the distribution also allows BMA to reduce the uncertainty of the mean projected changes beyond the Sanderson approaches or skill weighting. These results are also in line with the work of Massoud et al. [15,16].
The BMA generally reduces the uncertainty of the mean projected change of the ensemble beyond any of the other weighting schemes used in this study (which is in line with results from [15] and [16]); it is also apparent that using BMA with the LOCA downscaled projections reduces the uncertainty even further than BMA with the CMIP5 ensemble alone (which is similar to the results suggested in [60] which combined multiple statistically downscaled projections). This result stems from the significant amount of bias correction in the LOCA downscaling method, with the BMA framework further reducing biases from the multi-model ensemble. Given the results of this study, we find that it is not appropriate to assume that the weighting scheme derived from a GCM ensemble would match those derived using a downscaled ensemble (particularly when using BMA). Hence, we recommend the use of multiple weighting methods, if possible, when combining multi-model ensembles to ensure that the model averaging does not solely rely on a few models and to maintain the total order of the distributions of the variables of interest. We also recommend that multiple weighting schemes be considered with downscaled projections for future assessments.
Precipitation remains among the most challenging variables to simulate in global climate models, e.g., [14,33,34,61]. Although this study focused solely on mean annual precipitation, we speculate that our conclusions will remain valid in multivariate analysis incorporating variables that are less challenging to model or thresholds of variables (such as the number of days with rainfall over 1 inch). Such an analysis is planned for our future work. Although this analysis focused on the south-central United States, the various weighting schemes have been used individually in multiple other regions around the world, e.g., [15,16]. Therefore, we also speculate that the results of this analysis can be generalized to other regions. In addition, all of the metrics for the weighting schemes were calculated using the entire south-central U.S., but the uncertainty associated with each state in the region likely can be reduced further by training weighting schemes with a downscaled ensemble for individual states. As such, for individual states or adaptation decision making in smaller regions, it is likely beneficial to consider using weighting schemes tailored for those regions.

4. Conclusions

In this study, we have compared four ensemble weighting schemes on mean annual precipitation for both GCM output and the same output after downscaling with a single statistical downscaling method. Traditionally, prior efforts involving ensemble weighting calculate the weights using the raw GCM ensemble and implicitly assume that the same weights apply to a statistically downscaled ensemble (i.e., the weight-transfer assumption). The weight-transfer assumption ignores the additional co-dependence and improved skill introduced by statistical downscaling to all members of the ensemble. The effect of these attributes is most strongly reflected by the BMA weighting scheme, whose weights change dramatically between the CMIP5 and LOCA ensemble. The use of statistical downscaling generally reduces the uncertainty and improves the accuracy associated with the historical mean and mean projected change in annual precipitation (Table 1 and Table 2, Figure 5, Figure 6 and Figure 8). However, the BMA weighted LOCA ensemble also reduces the uncertainty of the mean projected change in annual precipitation above and beyond the other three weighting schemes (Table 2).
While we recommend using multiple weighting schemes, we recognize this may be impractical for large-scale assessments (such as the NCA). However, it may be beneficial to compute the weights for smaller regions when conducting state or local impact assessments or adaptation decisions. Alternatively, one also may consider using a BMA process to select a subset of a multi-model ensemble specific to the needs of stakeholders, for their domain and variables of interest. The development of the associated evaluation and selection framework for these products for smaller domains will greatly benefit local climate change impact assessments and is also the subject of our future work on this topic.

Supplementary Materials

The following are available online at https://0-www-mdpi-com.brum.beds.ac.uk/2225-1154/8/12/138/s1, Table S1: Global Climate Models used to create both the CMIP5 and LOCA ensembles.

Author Contributions

Conceptualization, A.M.W.; methodology, A.M.W. and E.C.M.; formal analysis, investigation, A.M.W. and E.C.M.; writing—original draft preparation, A.M.W.; writing—review and editing, A.M.W. and E.C.M.; visualization, A.M.W., E.C.M., and A.S.; supervision, D.E.W. and H.L.; All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by National Aeronautics and Space Administration. This material is based upon work supported in part by the National Aeronautics and Space Administration under Grant No. NNX15AK02H issued through the NASA Oklahoma Space Grant Consortium.

Acknowledgments

The authors thank the reviewers for their comments and critiques to strengthen this article. A portion of this research was carried out at the Jet Propulsion Laboratory, California Institute of Technology, under a contract with the National Aeronautics and Space Administration. Copyright 2020.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Weart, S. The development of general circulation models of climate. Stud. Hist. Philos. Sci. Part B Stud. Hist. Philos. Mod. Phys. 2010, 41, 208–217. [Google Scholar] [CrossRef]
  2. Smith, L.; Stern, N. Uncertainty in science and its role in climate policy. Philos. Trans. R. Soc. A 2011, 369, 1–24. [Google Scholar] [CrossRef] [Green Version]
  3. Rummukainen, M. State-of-the-art with regional climate models. WIREs Clim. Chang. 2010, 1. [Google Scholar] [CrossRef]
  4. Rummukainen, M. Added value in regional climate models. WIREs Clim. Chang. 2016, 7. [Google Scholar] [CrossRef]
  5. Tabari, H.; DeTroch, R.; Giot, O.; Hamdi, R.; Termonia, P.; Saeed, S.; Brisson, E.; Van Lipzig, N.; Williams, P. Local impact analysis of climate change on precipitation extremes: Are high-resolution climate models needed for realistic simulations? Hydrol. Earth Syst. Sci. 2016, 20, 3843–3857. [Google Scholar] [CrossRef] [Green Version]
  6. Gergel, D.R.; Nijssen, B.; Abatzoglou, J.T.; Lettenmaier, D.P.; Stumbaugh, M.R. Effect of climate change on snowpack and fire potential in the western USA. Clim. Chang. 2017, 141, 287–299. [Google Scholar] [CrossRef]
  7. Allstadt, A.J.; Vavrus, S.J.; Heglund, P.J.; Pidgeon, A.M.; Thogmartin, W.E.; Radeloff, V.C. Spring plant phenology and false springs in the conterminous US during the 21st century. Environ. Res. Lett. 2015, 10, 104008. [Google Scholar] [CrossRef]
  8. Basso, B.; Hyndman, D.W.; Kendall, A.D.; Grace, P.R.; Roberston, G.P. Can impacts of climate change agricultural adaptation strategies be accurately quantified if crop models are annually re-initialized? PLoS ONE 2015, 10, e0127333. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Fatichi, S.; Ivanov, V.Y.; Caporali, E. Simulation of future climate scenarios with a weather generator. Adv. Water Resour. 2011, 34, 448–467. [Google Scholar] [CrossRef]
  10. Li, X.; Babovic, V. Multi-site multivariate downscaling of global climate model outputs: An integrated framework combining quantile mapping, stochastic weather generator and Empirical Copula approaches. Clim. Dyn. 2019, 52, 5775–5799. [Google Scholar] [CrossRef]
  11. USGCRP. Climate Science Special Report: Fourth National Climate Assessment Volume I; U.S. Global Change Research Program: Washington, DC, USA, 2017; 470p. [CrossRef] [Green Version]
  12. Eyring, V.; Bony, S.; Meehl, G.A.; Senior, C.A.; Stevens, B.; Souffer, R.J.; Taylor, K.E. Overview of the Coupled Model Intercomparison Project Phase 6 (CMIP6) experimental design and organization. Geosci. Model Dev. 2016, 9, 1937–1958. [Google Scholar] [CrossRef] [Green Version]
  13. Lee, H.; Goodman, A.; McGibbney, L.; Waliser, D.E.; Kim, J.; Loikith, P.C.; Gibson, P.B.; Massoud, E.C. Regional Climate Model Evaluation System powered by Apache Open Climate Workbench v1.3.0: An enabling tool for facilitating regional climate studies. Geosci. Model Dev. 2018. [Google Scholar] [CrossRef] [Green Version]
  14. Gibson, P.B.; Waliser, D.E.; Lee, H.; Tian, B.; Massoud, E. Climate model evaluation in the presence of observational uncertainty: Precipitation indices over the contiguous United States. J. Hydrometeorol. 2019, 20, 1339–1357. [Google Scholar] [CrossRef]
  15. Massoud, E.C.; Espinoza, V.; Guan, B.; Waliser, D.E. Global Climate Model Ensemble Approaches for Future Projections of Atmospheric Rivers. Earth’s Future 2019, 7, 1136–1151. [Google Scholar] [CrossRef] [Green Version]
  16. Massoud, E.C.; Lee, H.; Gibson, P.B.; Loikith, P.; Waliser, D.E. Bayesian Model Averaging of Climate Model Projections Constrained by Precipitation Observations over the Contiguous United States. J. Hydrometeorol. 2020, 21, 2401–2418. [Google Scholar] [CrossRef]
  17. Parding, K.M.; Dobler, A.; McSweeney, C.; Landgren, O.A.; Benestad, R.; Erlandsen, H.B.; Mezghani, A.; Gregow, H.; Raty, O.; Viktor, E. GCMeval—An interactive tool for evaluation and selection of climate model ensembles. Clim. Serv. 2020, 18, 100167. [Google Scholar] [CrossRef]
  18. Diaconescu, E.; Gachon, P.; Laprise, R.; Scinocca, J. Evaluation of Precipitation Indices over North America from Various Configurations of Regional Climate Models. Atmosphere-Ocean 2016, 54, 418–439. [Google Scholar] [CrossRef]
  19. Sanderson, B.M.; Knutti, R.; Caldwell, P. Addressing interdependency in a multimodel ensemble by interpolation of model properties. J. Clim. 2015, 13, 5150–5170. [Google Scholar] [CrossRef]
  20. Sanderson, B.M.; Wehner, M.; Knutti, R. Skill and independence weighting for multi-model assessments. Geosci. Model Dev. 2017, 10, 2379–2395. [Google Scholar] [CrossRef] [Green Version]
  21. Knutti, R. The end of model democracy? Clim. Chang. 2010, 102. [Google Scholar] [CrossRef]
  22. Knutti, R.; Sedlacek, J.; Sanderson, B.M.; Lorenz, R.; Fischer, E.M.; Eyring, V. A climate model weighting scheme accounting for performance and independence. Geophys. Res. Lett. 2017, 44. [Google Scholar] [CrossRef] [Green Version]
  23. Lorenz, R.; Herger, N.; Sedlacek, J.; Eyring, V.; Fischer, E.M.; Knutti, R. Prospects and Caveats of Weighting Climate Models for Summer Maximum Temperature Projections Over North America. J. Geophys. Res. Lett. Atmos. 2018, 123, 4509–4526. [Google Scholar] [CrossRef]
  24. Shin, Y.; Lee, Y.; Park, J. A weighting scheme in a multi-model ensemble for bias-corrected climate simulation. Atmosphere 2020, 11, 775. [Google Scholar] [CrossRef]
  25. Sanderson, B.M.; Wehner, M.F. Model Weighting Strategy. In Climate Science Special Report: Fourth National Climate Assessment Volume I; Wuebbles, D.J., Fahey, D.W., Hibbard, K.A., Dokken, D.J., Stewart, B.C., Maycock, T.K., Eds.; U.S. Global Change Research Program: Washington, DC, USA, 2017; pp. 436–442. [Google Scholar] [CrossRef] [Green Version]
  26. Knutti, R.; Masson, D.; Gettelman, A. Climate model genealogy: Generation CMIP5 and how we got there. Geophys. Res. Lett. 2013, 40, 1194–1199. [Google Scholar] [CrossRef]
  27. Behnke, R.; Vavrus, S.; Allstadt, A.; Thogmartin, W.; Radelhoff, V.C. Evaluation of downscaled gridded climate data for the conterminous United States. Ecol. Appl. 2016, 26, 1338–1351. [Google Scholar] [CrossRef]
  28. Wootten, A.M.; Dixon, K.W.; Adams-Smith, D.J.; McPherson, R.A. Statistically downscaled precipitation sensitivity to gridded observation data and downscaling technique. Int. J. Climatol. 2020. [Google Scholar] [CrossRef]
  29. Eyring, V.; Cox, P.M.; Flato, G.M.; Glecker, P.J.; Abramowitz, G.; Caldwell, P.; Collins, W.D. Taking climate model evaluation to the next level. Nat. Clim. Chang. 2019, 9, 102–110. [Google Scholar] [CrossRef] [Green Version]
  30. Kotamarthi, R.; Mearns, L.; Hayhoe, K.; Castro, C.L.; Wuebbles, D. Use of Climate Model Information for Decision-Making and Impacts Research: State of Our Understanding; Department of Defense, Strategic Environemental Research and Development Program: Washington, DC, USA, 2016; 55p.
  31. Dilling, L.; Berrgren, J. What do stakeholders need to manage for climate change and variability? A document-based analysis from three mountain states in the Western USA. Reg. Environ. Chang. 2014, 15, 657–667. [Google Scholar] [CrossRef]
  32. Hoeting, J.A.; Madigan, D.; Raftery, A.E.; Volinsky, C.T. Bayesian model averaging: A tutorial. Stat. Sci. 1999, 14, 382–401. [Google Scholar]
  33. Cesana, G.; Suselj, K.; Brient, F. On the Dependence of Cloud Feedback on Physical Parameterizations in WRF Aquaplanet Simulations. Geophys. Res. Lett. 2017, 44, 10762–10771. [Google Scholar] [CrossRef]
  34. Tapiador, F.J.; Roca, R.; Genio, A.D.; Dewitte, B.; Petersen, W.; Zhang, F. Is Precipitation a good metric for model performance? Bull. Am. Meteorol. Soc. 2019, 100, 223–233. [Google Scholar] [CrossRef] [PubMed]
  35. Paschalis, A.; Molnar, P.; Fatichi, S.; Burlando, P. A stochastic model for high-resolution space-time precipitation simulation. Water Resour. Res. 2013, 49, 8400–8417. [Google Scholar] [CrossRef]
  36. Li, X.; Meshgi, A.; Wang, X.; Zhang, J.; Tay, S.H.X.; Pijcke, G.; Manocha, N.; Ong, M.; Nguyen, M.T.; Babovic, V. Three resampling approaches based on method of fragments for daily-to-subdaily precipitation disaggregation. Int. J. Clim. 2018, 38, e1119–e1138. [Google Scholar] [CrossRef]
  37. Livneh, B.; Rosenberg, E.A.; Lin, C.; Nijissen, B.; Mishra, V.; Andreadis, K.M.; Maurer, E.P.; Lettenmaier, D.P. A long-term hydrologically based dataset of land surface fluxes and states for the conterminous United States: Updates and extensions. J. Clim. 2013, 26, 9384–9392. [Google Scholar] [CrossRef]
  38. Amante, C.; Eakins, B.W. ETOPO1 1 Arc-Minute Global Relief Model: Procedures, Data Sources and Analysis; NOAA Technical Memorandum NESDIS NGDC-24; National Geophysical Data Center, NOAA: Boulder, CO, USA, 2009.
  39. GRDC. Major River Basins of the World/Global Runoff Data Centre, GRDC, 2nd ed.; Federal Institute of Hydrology (BfG): Koblenz, Germany, 2020.
  40. Taylor, A.; Gregory, J.M.; Webb, M.J.; Taylor, K.E. Forcing, feedbacks and climate sensitivity in CMIP5 couple atmosphere-ocean climate models. Geophys. Res. Lett. 2012, 39. [Google Scholar] [CrossRef] [Green Version]
  41. Pierce, D.W.; Cayan, D.R.; Thrasher, B.L. Statistical downscaling using Localized Constructed Analogs (LOCA). J. Hydrometeorol. 2014, 15, 2558–2585. [Google Scholar] [CrossRef]
  42. Van Vuuren, D.P.; Edmonds, J.; Kainuma, M.; Riahi, K.; Thomson, A.; Hibbard, K.; Hurtt, G.C.; Kram, T.; Krey, V.; Lamarque, J.-F.; et al. The representative concentration pathways: An overview. Clim. Chang. 2011, 5, 5–31. [Google Scholar] [CrossRef]
  43. Riahi, K.; Grubler, A.; Nakicenovic, N. Scenarios of long-term socio-economic and environmental development under climate stabilization. Technol. Forecast. Soc. Chang. 2007, 74, 887–935. [Google Scholar] [CrossRef]
  44. Espinoza, V.; Waliser, D.E.; Guan, B.; Lavers, D.A.; Ralph, F.M. Global Analysis of Climate Change Projection Effects on Atmospheric Rivers. Geophys. Res. Lett. 2018, 45, 4299–4308. [Google Scholar] [CrossRef]
  45. Lavers, D.A.; Ralph, F.M.; Waliser, D.; Gershunov, A.; Dettinger, M.D. Climate change intensification of horizontal water vapor transport in CMIP5. Geophys. Res. Lett. 2015, 42, 5617–5625. [Google Scholar] [CrossRef] [Green Version]
  46. Karl, T.R.; Williams, C.N.; Young, P.J.; Wendland, W.M. A model to estimate the time of observation bias associated with monthly mean maximum, minimum, and mean temperatures for the United States. J. Clim. Appl. Meteorol. 1986, 25, 145–160. [Google Scholar] [CrossRef] [Green Version]
  47. Abatzoglou, J. Development of gridded surface meteorological data for ecological applications and modeling. Int. J. Climatol. 2013, 33, 121–131. [Google Scholar] [CrossRef]
  48. Schoof, J.T. Statistical downscaling in climatology. Geogr. Compass 2013, 7, 249–265. [Google Scholar] [CrossRef] [Green Version]
  49. Raftery, A.E.; Gneiting, T.; Balabdaoui, F.; Polakowski, M. Using Bayesian model averaging to calibrate forecast ensembles. Mon. Weather Rev. 2005, 133, 1155–1174. [Google Scholar] [CrossRef] [Green Version]
  50. Gneiting, T.; Raftery, A.E. Weather forecasting with ensemble methods. Science 2005, 310, 248–249. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  51. Duan, Q.; Newsha, K.A.; Gao, X.; Sorooshian, S. Multi-model ensemble hydrologic prediction using Bayesian model averaging. Adv. Water Resour. 2007, 30, 1371–1386. [Google Scholar] [CrossRef] [Green Version]
  52. Vrugt, J.A.; Robinson, B.A. Treatment of uncertainty using ensemble methods: Comparison of sequential data assimilation and Bayesian model averaging. Water Resour. Res. 2007, 43. [Google Scholar] [CrossRef] [Green Version]
  53. Vrugt, J.A.; Cajo, T.F.; Clark, M.P.; Hyman, J.M.; Robinson, B.A. Treatment of input uncertainty in hydrologic modeling: Doing hydrology backward with Markov chain Monte Carlo simulation. Water Resour. Res. 2008, 44. [Google Scholar] [CrossRef] [Green Version]
  54. Bishop, C.H.; Shanley, K.T. Bayesian model averaging’s problematic treatment of extreme weather and a paradigm shift that fixes it. Mon. Weather Rev. 2008, 136, 4641–4652. [Google Scholar] [CrossRef]
  55. Olson, R.; Fan, Y.; Evans, J.P. A simple method for Bayesian model averaging of regional climate model projections: Application to southeast Australian temperatures. Geophys. Res. Lett. 2016, 43, 7661–7669. [Google Scholar] [CrossRef]
  56. Olson, R.; An, S.; Fan, Y.; Evans, J.P. Accounting for skill in trend, variability, and autocorrelation facilitates better multi-model projections: Application to the AMOC and temperature time series. PLoS ONE 2019, 14, e0214535. [Google Scholar] [CrossRef] [PubMed]
  57. Fan, Y.; Olson, R.; Evans, J.P. A Bayesian posterior predictive frameworks for weighting ensemble regional climate models. Geosci. Model Dev. 2017, 10, 2321–2332. [Google Scholar] [CrossRef] [Green Version]
  58. Vrugt, J.A.; Massoud, E.C. Uncertainty quantification of complex system models: Bayesian analysis. In Handbook of Hydrometeorological Ensemble Forecasting; Duan, Q., Pappenberger, F., Thielen, J., Wood, A., Cloke, H.L., Schaake, J.C., Eds.; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
  59. Gelman, A.; Rubin, D.B. Inference from iterative simulation using multiple sequences. Stat. Sci. 1992, 7, 457–472. [Google Scholar] [CrossRef]
  60. Liu, J.; Yuan, D.; Zhang, L.; Zou, X.; Song, X. Comparison of Three Statistical Downscaling Methods and Ensemble Downscaling Method Based on Bayesian Model Averaging in Upper Hanjiang River Basin, China. Adv. Meteorol. 2016, 2016. [Google Scholar] [CrossRef]
  61. Sempreviva, A.M.; Schiano, M.E.; Pensieri, S.; Semedo, A.; Tomé, R.; Bozzano, R.; Borghini, M.; Grasso, F.; Soerensen, L.L.; Teixeira, J.; et al. Observed development of the vertical structure of the marine boundary layer during the LASIE experiment in the Ligurian Sea. Ann. Geophys. 2009, 28, 17–25. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Left—Study domain overlaid with annual average precipitation (mm) from Livneh v. 1.2 [37]. Right—Topographical map for the study domain: The elevation map of south-central United States with major rivers overlaid on it. Brown/green shading denotes elevation (in units of m), while the rivers are outlined in blue. Topography, bathymetry and shoreline data are obtained from the National Oceanic and Atmospheric Administration (NOAA) National Geophysical Data Center’s ETOPO1 Global Relief Model [38]. This is a 1 arc-minute model of the Earth’s surface developed from diverse global and regional digital datasets and then shifted to a common horizontal and vertical datum. River shapefiles are obtained from the Global Runoff Data Centre’s Major River Basins of the World [39].
Figure 1. Left—Study domain overlaid with annual average precipitation (mm) from Livneh v. 1.2 [37]. Right—Topographical map for the study domain: The elevation map of south-central United States with major rivers overlaid on it. Brown/green shading denotes elevation (in units of m), while the rivers are outlined in blue. Topography, bathymetry and shoreline data are obtained from the National Oceanic and Atmospheric Administration (NOAA) National Geophysical Data Center’s ETOPO1 Global Relief Model [38]. This is a 1 arc-minute model of the Earth’s surface developed from diverse global and regional digital datasets and then shifted to a common horizontal and vertical datum. River shapefiles are obtained from the Global Runoff Data Centre’s Major River Basins of the World [39].
Climate 08 00138 g001
Figure 2. Flowchart showing the process of analysis with weighting schemes. The skill weighting uses only the historical skill. The SI-h uses the historical skill and independence is used to determine model weights. The SI-c weighting uses the historical skill and the independence of the projected change signal to determine model weights. Bayesian Model Averaging (BMA) explicitly accounts for both skill and independence during the optimization process. Results from the weighting schemes are compared to the results from the unweighted LOCA and CMIP5 ensembles.
Figure 2. Flowchart showing the process of analysis with weighting schemes. The skill weighting uses only the historical skill. The SI-h uses the historical skill and independence is used to determine model weights. The SI-c weighting uses the historical skill and the independence of the projected change signal to determine model weights. Bayesian Model Averaging (BMA) explicitly accounts for both skill and independence during the optimization process. Results from the weighting schemes are compared to the results from the unweighted LOCA and CMIP5 ensembles.
Climate 08 00138 g002
Figure 3. Skill and Independence weights for the CMIP5 (black) and LOCA (blue) ensemble. The solid lines are for the ensembles using only historical precipitation information (SI-h). The dashed lines make use of the projected change in precipitation to derive independence weights and historical accuracy to derive skill weights (SI-c). The individual weights for each ensemble member are contained within each polygon.
Figure 3. Skill and Independence weights for the CMIP5 (black) and LOCA (blue) ensemble. The solid lines are for the ensembles using only historical precipitation information (SI-h). The dashed lines make use of the projected change in precipitation to derive independence weights and historical accuracy to derive skill weights (SI-c). The individual weights for each ensemble member are contained within each polygon.
Climate 08 00138 g003
Figure 4. (a) BMA best, SI-h, SI-c and Skill weights for the CMIP5 ensemble; (b) BMA best, SI-h, SI-c and Skill weights for the LOCA ensemble. Boxplots in the background are the weights across all BMA simulations.
Figure 4. (a) BMA best, SI-h, SI-c and Skill weights for the CMIP5 ensemble; (b) BMA best, SI-h, SI-c and Skill weights for the LOCA ensemble. Boxplots in the background are the weights across all BMA simulations.
Climate 08 00138 g004
Figure 5. Left—Historical values for CMIP5 and LOCA ensembles with the Livneh observed values averaged over the full domain and subdomains. Right—Weighted ensemble mean values with Livneh observed values and boxplot of BMA simulated ensemble means averaged over the full domain and subdomains.
Figure 5. Left—Historical values for CMIP5 and LOCA ensembles with the Livneh observed values averaged over the full domain and subdomains. Right—Weighted ensemble mean values with Livneh observed values and boxplot of BMA simulated ensemble means averaged over the full domain and subdomains.
Climate 08 00138 g005
Figure 6. Left—RMSE of CMIP5 and LOCA ensembles. Right—RMSE of weighted ensemble means with boxplot of BMA simulated RMSE of ensemble means.
Figure 6. Left—RMSE of CMIP5 and LOCA ensembles. Right—RMSE of weighted ensemble means with boxplot of BMA simulated RMSE of ensemble means.
Climate 08 00138 g006
Figure 7. Projected change (mm) in annual average precipitation from the unweighted ensembles (left hand column) and all eight weighted ensembles.
Figure 7. Projected change (mm) in annual average precipitation from the unweighted ensembles (left hand column) and all eight weighted ensembles.
Climate 08 00138 g007
Figure 8. Left—Projected change (mm) from CMIP5 and LOCA ensembles. Right—Mean projected change of weighted ensembles with boxplot of BMA simulated mean change.
Figure 8. Left—Projected change (mm) from CMIP5 and LOCA ensembles. Right—Mean projected change of weighted ensembles with boxplot of BMA simulated mean change.
Climate 08 00138 g008
Table 1. Standard Deviation (mm) of Historical Ensemble Means—Unweighted CMIP5 and LOCA ensembles vs. weighted CMIP5 and LOCA ensembles for the full domain and all subdomains.
Table 1. Standard Deviation (mm) of Historical Ensemble Means—Unweighted CMIP5 and LOCA ensembles vs. weighted CMIP5 and LOCA ensembles for the full domain and all subdomains.
Group—WeightingFullNew MexicoTexasOklahomaLouisiana
CMIP5—Unweighted119.96149.24133.61140.29216.42
CMIP5—Skill111.37139.51124.86127.60226.01
CMIP5—SI-h122.63139.91138.72137.11255.65
CMIP5—SI-c112.91134.55125.47132.19225.47
CMIP5—BMA30.6220.9638.5138.3862.52
LOCA—Unweighted13.969.6018.1922.3929.18
LOCA—Skill13.949.5918.1622.3629.15
LOCA—SI-h14.009.6218.2322.4229.28
LOCA—SI-c14.149.9918.3621.9029.74
LOCA—BMA1.852.423.143.9211.07
Table 2. Standard Deviation (mm) of Ensemble Mean Projected Change—Unweighted CMIP5 and LOCA ensembles vs. weighted CMIP5 and LOCA ensembles for the full domain and all subdomains.
Table 2. Standard Deviation (mm) of Ensemble Mean Projected Change—Unweighted CMIP5 and LOCA ensembles vs. weighted CMIP5 and LOCA ensembles for the full domain and all subdomains.
Group—WeightingFullNew MexicoTexasOklahomaLouisiana
CMIP5—Unweighted62.8260.5969.2877.19116.53
CMIP5—Skill64.1959.0373.1681.50111.35
CMIP5—SI-h68.1963.4076.2683.84116.78
CMIP5—SI-c67.4061.5177.1184.36118.63
CMIP5—BMA26.6224.6828.6234.5136.73
LOCA—Unweighted65.0646.5969.8582.75178.51
LOCA—Skill64.9946.5769.7982.61178.31
LOCA—SI-h64.8846.6069.6982.53178.26
LOCA—SI-c70.8752.3175.7287.78188.64
LOCA—BMA16.3214.8719.0521.5559.39
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wootten, A.M.; Massoud, E.C.; Sengupta, A.; Waliser, D.E.; Lee, H. The Effect of Statistical Downscaling on the Weighting of Multi-Model Ensembles of Precipitation. Climate 2020, 8, 138. https://0-doi-org.brum.beds.ac.uk/10.3390/cli8120138

AMA Style

Wootten AM, Massoud EC, Sengupta A, Waliser DE, Lee H. The Effect of Statistical Downscaling on the Weighting of Multi-Model Ensembles of Precipitation. Climate. 2020; 8(12):138. https://0-doi-org.brum.beds.ac.uk/10.3390/cli8120138

Chicago/Turabian Style

Wootten, Adrienne M., Elias C. Massoud, Agniv Sengupta, Duane E. Waliser, and Huikyo Lee. 2020. "The Effect of Statistical Downscaling on the Weighting of Multi-Model Ensembles of Precipitation" Climate 8, no. 12: 138. https://0-doi-org.brum.beds.ac.uk/10.3390/cli8120138

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop