Journal article
2013
APA
Click to copy
Schwalm, C., Huntinzger, D. N., Michalak, A., Fisher, J., Kimball, J., Mueller, B., … Zhang, Y. (2013). Sensitivity of inferred climate model skill to evaluation decisions: a case study using CMIP5 evapotranspiration.
Chicago/Turabian
Click to copy
Schwalm, C., D. N. Huntinzger, A. Michalak, J. Fisher, J. Kimball, B. Mueller, Kecheng Zhang, and Yongqiang Zhang. “Sensitivity of Inferred Climate Model Skill to Evaluation Decisions: a Case Study Using CMIP5 Evapotranspiration” (2013).
MLA
Click to copy
Schwalm, C., et al. Sensitivity of Inferred Climate Model Skill to Evaluation Decisions: a Case Study Using CMIP5 Evapotranspiration. 2013.
BibTeX Click to copy
@article{c2013a,
title = {Sensitivity of inferred climate model skill to evaluation decisions: a case study using CMIP5 evapotranspiration},
year = {2013},
author = {Schwalm, C. and Huntinzger, D. N. and Michalak, A. and Fisher, J. and Kimball, J. and Mueller, B. and Zhang, Kecheng and Zhang, Yongqiang}
}
Confrontation of climate models with observationally-based reference datasets is widespread and integral to model development. These comparisons yield skill metrics quantifying the mismatch between simulated and reference values and also involve analyst choices, or meta-parameters, in structuring the analysis. Here, we systematically vary five such meta-parameters (reference dataset, spatial resolution, regridding approach, land mask, and time period) in evaluating evapotranspiration (ET) from eight CMIP5 models in a factorial design that yields 68 700 intercomparisons. The results show that while model–data comparisons can provide some feedback on overall model performance, model ranks are ambiguous and inferred model skill and rank are highly sensitive to the choice of meta-parameters for all models. This suggests that model skill and rank are best represented probabilistically rather than as scalar values. For this case study, the choice of reference dataset is found to have a dominant influence on inferred model skill, even larger than the choice of model itself. This is primarily due to large differences between reference datasets, indicating that further work in developing a community-accepted standard ET reference dataset is crucial in order to decrease ambiguity in model skill.