dc.contributor.author | Siegert, S | |
dc.contributor.author | Bellprat, O | |
dc.contributor.author | Ménégoz, M | |
dc.contributor.author | Stephenson, DB | |
dc.contributor.author | Doblas-Reyes, FJ | |
dc.date.accessioned | 2018-11-21T10:39:51Z | |
dc.date.issued | 2017-01-16 | |
dc.description.abstract | The skill of weather and climate forecast systems is often assessed by calculating the correlation coefficient between past forecasts and their verifying observations. Improvements in forecast skill can thus be quantified by correlation differences. The uncertainty in the correlation difference needs to be assessed to judge whether the observed difference constitutes a genuine improvement, or is compatible with random sampling variations. A widely used statistical test for correlation difference is known to be unsuitable, because it assumes that the competing forecasting systems are independent. In this paper, appropriate statistical methods are reviewed to assess correlation differences when the competing forecasting systems are strongly correlated with one another. The methods are used to compare correlation skill between seasonal temperature forecasts that differ in initialization scheme and model resolution. A simple power analysis framework is proposed to estimate the probability of correctly detecting skill improvements, and to determine the minimum number of samples required to reliably detect improvements. The proposed statistical test has a higher power of detecting improvements than the traditional test. The main examples suggest that sample sizes of climate hindcasts should be increased to about 40 years to ensure sufficiently high power. It is found that seasonal temperature forecasts are significantly improved by using realistic land surface initial conditions. | en_GB |
dc.description.sponsorship | The authors acknowledge support by the European Union Program FP7/2007-13 under Grant Agreement 3038378 (SPECS). The work of O. Bellprat was funded by ESA under the Climate Change Initiative (CCI) Living Planet Fellowship VERITAS-CCI. | en_GB |
dc.identifier.citation | Vol. 145, pp. 437 - 450 | en_GB |
dc.identifier.doi | 10.1175/MWR-D-16-0037.1 | |
dc.identifier.uri | http://hdl.handle.net/10871/34834 | |
dc.language.iso | en | en_GB |
dc.publisher | American Meteorological Society | en_GB |
dc.rights | © 2017 American Meteorological Society | en_GB |
dc.subject | Statistical techniques | en_GB |
dc.subject | Statistics | en_GB |
dc.subject | Climate prediction | en_GB |
dc.subject | Forecast verification/skill | en_GB |
dc.subject | Hindcasts | en_GB |
dc.subject | Seasonal forecasting | en_GB |
dc.title | Detecting improvements in forecast correlation skill: Statistical testing and power analysis | en_GB |
dc.type | Article | en_GB |
dc.date.available | 2018-11-21T10:39:51Z | |
dc.identifier.issn | 0027-0644 | |
dc.description | This is the final version. Available from the American Meteorological Society via the DOI in this record | en_GB |
dc.identifier.journal | Monthly Weather Review | en_GB |