What to look for when using forecasts from NWP models for streamflow forecasting?

What to look for when using forecasts from NWP models for streamflow forecasting?

By Durga Lal Shrestha, James Bennett, David Robertson and QJ Wang, members of the CSIRO Columnist Team

There have been a few posts on NWP performance lately, and so we thought we’d add our perspective. We’ve been working closely with the Bureau of Meteorology to extend their new 7-day deterministic streamflow forecasting service to an operational ensemble streamflow forecasting service. One of the fundamental choices we have to make is the source of quantitative precipitation forecasts (QPFs). This is not as straightforward as it may seem – QPFs that show strong performance in some ways may not be suitable in others, while even if a hypothetical ‘ideal’ QPF exists, it may not be useable e.g. because it is inconsistent with other forecasts the Bureau issues.

Here’re some of the things we look for in QPFs from Numerical Weather Prediction (NWP) models:

  • QPFs that perform best after post-processing. We use a post-processor to produce ensemble forecasts from deterministic QPFs (we call it the Rainfall Post-Processor or RPP; see Robertson et al., 2013). It’s a little similar to EMOS-type post-processors (see, e.g. Gneiting et. al; 2005) with which you may be familiar. Post-processors are able to correct many of the deficiencies of rainfall forecasts, so we are looking for QPFs that add extra information.
  • QPFs that produce skilful forecasts in a range of catchments. The goal is to apply a single system across all of Australia, so we need to handle all the conditions Australia throws at us: from tropical/monsoonal climates, through arid areas, to cool temperate climates.
  • QPFs that produce skilful forecasts of accumulated rainfalls. Yes, the timing of rain storms is very important, and we of course consider this when we evaluate QPFs. But for streamflow forecasting, the total amount of rainfall (irrespective of timing) forecasted for all lead-times can be crucial. Assessing the performance of accumulated rainfall forecasts – particularly for measures of reliability – is not commonly considered in literature published on evaluating QPFs.

In Australia, the Bureau of Meteorology uses the ACCESS suite of NWP models and the ‘Poor Man’s Ensemble’ (PME) to provide guidance to issue official weather forecasts. The PME is simply a collection of 8 deterministic NWP models. The global version of ACCESS (ACCESS-G) is developed and operated by the Bureau to produce 10-day deterministic rainfall forecasts, while the PME form the major basis of operational weather forecasting products issued by the Bureau. The Bureau asked us to evaluate if ACCESS or PME would be preferable for an ensemble streamflow forecasting system.

After post-processing there is little to separate the performance of PME from ACCESS-G across a range of catchments (Figure 1) – even where there are big differences in the performance (particularly bias) of the raw forecasts, such as for tropical Queensland. As you can see CRPS is almost identical for raw ACCESS-G and PME, but the bias is vastly different. The PME has a tendency to predict too many small rain events because of “smearing out” rainfall from component models in the ensemble mean. These small positive errors help cancel out a few significant negative errors (underestimation) which leads to smaller bias in PME. In contrast, the small positive errors add up with the negative errors when computing CRPS (absolute error for deterministic forecast).

D:\My Publications\Blogs\fig1_bias_crps\fig1_bias_crps.png
Figure 1. Catchment average rainfall and CRPS in the raw and calibrated rainfall forecasts in Tully (Queensland – Tropical) and North Esk (Tasmania – Temperate) catchments

Calibrated QPFs from ACCESS-G and PME are equally reliable at individual lead times (Figure 2). When we consider accumulated rainfall totals, the story becomes more interesting: reliability of calibrated PME QPFs is worse than calibrated ACCESS-G QPFs (thick lines in Figure 2). We suspect that this could be due to autocorrelation in raw PME forecasts. Taking the average of any set of deterministic forecasts will generally give a smoother (i.e., more autocorrelated) rainfall forecast than an individual NWP forecast. This autocorrelation isn’t compatible with the Schaake Shuffle (Clark et al., 2004), which we use to instil temporal and spatial correlations in the QPFs. (Though we’re not sure this is the cause – anyone else come across this problem?) Of course, the benefit of averaging many deterministic NWP models to produce the PME should be an increase in forecast accuracy – but in this case this advantage is only slight.

D:\My Publications\Blogs\fig2\fig2.png
Figure 2. Probability Integral Transform – uniform probability diagrams (PIT plots) for the catchment average calibrated QPFs in Tully and North Esk catchments. Thin lines are for individual and thick lines – cumulative, yellow lines = calibrated ACCESS-G, violet lines – calibrated PME

Of course, in a brief blog post it’s not possible to cover all the aspects of NWP QPFs that interest us: for example, we are also very interested in the streamflow forecasts that are produced by our calibrated QPFs. So we’ll leave you to fill in the gaps:

  • What makes a good QPF for you? What’s important to you when you assess QPFs for streamflow forecasts? Do you look for something others don’t?
North Esk River in Launceston, Tasmania (Temperate/Perennial) (Creative Commons License)
Tully River in Far North Queensland (Tropical/Perennial) (Creative Commons License)
Ord River downstream of Lake Argyle in North East of Western Australia (Grassland/Intermittent) (permission from https://austravelphotography.files.wordpress.com)

 

3 thoughts on “What to look for when using forecasts from NWP models for streamflow forecasting?

  1. From an operational perspective, accessibility is important (can we get it on time in a format we can use, updated frequently enough to be relevant?). Information about past performance (i.e. hydrologically relevant verification) is also surprisingly hard to come by.

    I suppose there’s the old joke about “This food tastes terrible. And the portions are too small.” High resolution (e.g. 2 km), long leadtime (e.g. 7 days), 10-50 member ensembles would be great, as long as they’re also high quality (skillful, bias free/bias corrected).

    1. Hah! Thanks for highlighting this – as you say, practical considerations like data formats and accessibility are hugely important, as well as readily accessible technical information about forecast performance.

    2. Thanks Tom for comments. In this blog we did not consider the practicability issue of the NWP forecasts. For example, which forecasts should be used if there are multiple forecasts issued in a day? ACCESS-G forecasts are issued twice a day at 00:00 UTC and 12:00 UTC. The Bureau of Meteorology runs the hydrologic models to issue streamflow forecasts at 09:00 Local time (23:00 UTC). The ACCESS-G forecasts which will be relevant for hydrological forecasting (to be issued at 23:00 UTC) are those issued at 12:00 UTC. The 11 hours difference between the ACCESS-G forecasts and hydrological forecasts is good because 1) the first few hours of NWP forecasts are not generally reliable because of spin-up time of NWP models 2) the ACCESS-G typically take about 5 hours to run.

      Another issue as you pointed out is that format of the NWP output. The NWP outputs are gridded, so this requires an interpolation of the NWP model precipitation forecasts to catchment subareas. That means NWP forecasts can not be input directly to the hydrological models for real time application.

Leave a Reply to Thomas Pagano Cancel reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.