Show simple item record

contributor authorHamill, Thomas M.
date accessioned2017-06-09T14:57:03Z
date available2017-06-09T14:57:03Z
date copyright1999/04/01
date issued1999
identifier issn0882-8156
identifier otherams-3033.pdf
identifier urihttp://onlinelibrary.yabesh.ir/handle/yetl/4167657
description abstractWhen evaluating differences between competing precipitation forecasts, formal hypothesis testing is rarely performed. This may be due to the difficulty in applying common tests given the spatial correlation of and non-normality of errors. Possible ways around these difficulties are explored here. Two datasets of precipitation forecasts are evaluated, a set of two competing gridded precipitation forecasts from operational weather prediction models and sets of competing probabilistic quantitative precipitation forecasts from model output statistics and from an ensemble of forecasts. For each test, data from each competing forecast are collected into one sample for each case day to avoid problems with spatial correlation. Next, several possible hypothesis test methods are evaluated: the paired t test, the nonparametric Wilcoxon signed-rank test, and two resampling tests. The more involved resampling test methodology is the most appropriate when testing threat scores from nonprobabilistic forecasts. The simpler paired t test or Wilcoxon test is appropriate to use in testing the skill of probabilistic forecasts evaluated with the ranked probability score.
publisherAmerican Meteorological Society
titleHypothesis Tests for Evaluating Numerical Precipitation Forecasts
typeJournal Paper
journal volume14
journal issue2
journal titleWeather and Forecasting
identifier doi10.1175/1520-0434(1999)014<0155:HTFENP>2.0.CO;2
journal fristpage155
journal lastpage167
treeWeather and Forecasting:;1999:;volume( 014 ):;issue: 002
contenttypeFulltext


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record