Subseasonal streamflow forecasts inform a multitude of water management decisions, from early flood warning to reservoir operation. Seamless forecasts, i.e. forecasts that are reliable and sharp over a range of lead times (1–30 d) and aggregation timescales (e.g. daily to monthly) are of clear practical interest. However, existing forecast products are often non-seamless, i.e. developed and applied for a single timescale and lead time (e.g. 1 month ahead). If seamless forecasts are to be a viable replacement for existing non-seamless forecasts, it is important that they offer (at least) similar predictive performance at the timescale of the non-seamless forecast.
This study compares forecasts from two probabilistic streamflow post-processing (QPP) models, namely the recently developed seamless daily Multi-Temporal Hydrological Residual Error (MuTHRE) model and the more traditional (non-seamless) monthly QPP model used in the Australian Bureau of Meteorology's dynamic forecasting system. Streamflow forecasts from both post-processing models are generated for 11 Australian catchments, using the GR4J hydrological model and pre-processed rainfall forecasts from the Australian Community Climate and Earth System Simulator – Seasonal (ACCESS-S) numerical weather prediction model. Evaluating monthly forecasts with key performance metrics (reliability, sharpness, bias, and continuous ranked probability score skill score), we find that the seamless MuTHRE model achieves essentially the same performance as the non-seamless monthly QPP model for the vast majority of metrics and temporal stratifications (months and years). As such, MuTHRE provides the capability of seamless daily streamflow forecasts with no loss of performance at the monthly scale – the modeller can proverbially “have their cake and eat it too”. This finding demonstrates that seamless forecasting technologies, such as the MuTHRE post-processing model, are not only viable but also a preferred choice for future research development and practical adoption in streamflow forecasting.
Subseasonal streamflow forecasts (with lead times up to 30 d) can be used to inform a range of water management decisions, from flood warning and reservoir flood management at shorter lead times (e.g. up to a week) to river basin management at timescales up to a month. The uncertainty in these forecasts is often represented using ensemble and probabilistic methods. Probabilistic streamflow forecasts have traditionally been developed and applied at only a single lead time and timescale (e.g. Gibbs et al., 2018; Mendoza et al., 2017; Souza Filho and Lall, 2003; Pal et al., 2013; Hidalgo-Muñoz et al., 2015). However, since different applications require forecasts over a range of lead times and timescales, recent research has focussed on producing “seamless” forecasts, i.e. forecasts from a single product that are (statistically) reliable and sharp across multiple lead times and aggregation timescales (McInerney et al., 2020). For seamless forecasts to be a viable replacement for more traditional “non-seamless” forecasts (i.e. forecasts for a single lead time and timescale), it is important to establish that the performance of seamless forecasts is competitive with their non-seamless counterparts at the native timescale of the latter.
Recent research by McInerney et al. (2020) has shown that seamless subseasonal forecasting is achievable. McInerney et al. (2020) developed the Multi-Temporal Hydrological Residual Error (MuTHRE) model for post-processing daily streamflow forecasts in order to improve reliability across a range of timescales. Using a case study with 11 catchments in the Murray–Darling basin, Australia, it was concluded that subseasonal forecasts generated using the MuTHRE streamflow post-processing model are indeed seamless because daily forecasts are consistently reliable (i) for lead times between 1 and 30 d and (ii) when aggregated to the monthly scale.
Seamless subseasonal forecasts are reliable over a wide range of aggregation
timescales (e.g. daily to monthly) and lead times (1–30 d). In contrast,
non-seamless forecasts are either (i) only available at a single timescale
(e.g. a post-processing model developed directly at the monthly scale does
not generate daily forecasts) or (ii) cannot be reliably aggregated to longer timescales (e.g. from daily to monthly). The practical benefits of seamless forecasts are as follows.
flood warning, where short-term forecasts (up to 1 week) on individual days are of practical interest (Cloke and Pappenberger, 2009), managing hydropower systems, which can utilise forecasts of inflow between 7 and 15 d to increase production in the electricity grid (Boucher and Ramos, 2019), managing reservoirs for rural water supply, where forecast volumes over long aggregation scales (e.g. weeks/months), and at long lead times (up to 1 month), are required due to long travel times (Murray-Darling Basin Authority, 2019), and operating urban water supply systems, where monthly forecasts are of value (Zhao and Zhao, 2014).
These practical benefits of seamless forecasts provide a clear motivation for their development and use. However, for seamless forecasts to be a viable replacement for non-seamless forecasts, it is important that they do not come at the cost of a substantial loss of performance at the native timescale of the non-seamless forecast. For example, if aggregated forecasts from a seamless daily model were considerably worse than monthly forecasts from an existing non-seamless model, then users of the monthly forecasts would prefer to continue using forecasts from the non-seamless model. In general, one might expect forecasts from a non-seamless model, developed and calibrated at single timescale, to provide superior performance compared to forecasts from a seamless model calibrated at shorter timescale and then aggregated. While the non-seamless model has only one job to do, which is to provide quality forecasts at a single timescale, the seamless model is expected to produce good performance over a range of lead times and aggregation timescales. Herein lies a major challenge of seamless forecasting.
Our interest in comparing the performance of aggregated seamless forecasts with non-seamless forecasts at their native timescale has similarities to previous research in aggregating deterministic streamflow predictions. For example, Wang et al. (2011) found that the Wapaba monthly rainfall–runoff model produced similar/better performance than aggregated predictions from the SIMHYD and AWBM daily rainfall–runoff models, despite only using observed monthly forcing data. Yang et al. (2016) compared daily and sub-daily versions of the SWAT model (with daily and sub-daily observed rainfall inputs) and found large differences in the partitioning of baseflow and direct runoff. However, to the best of the authors' knowledge, no studies have compared aggregated probabilistic forecasts from a seamless model against probabilistic forecasts from a non-seamless model.
The aim of this study is to establish whether aggregated forecasts from a (probabilistic) seamless model achieve comparable performance to those from a non-seamless (probabilistic) model at its native timescale. This aim is achieved by comparing the monthly forecast performance of the seamless MuTHRE post-processing model (aggregated from daily to monthly) against the non-seamless monthly streamflow post-processing model used in the Australian Bureau of Meteorology's dynamic forecasting system (Woldemeskel et al., 2018).
The remainder of the paper is organised as follows. Section 2 describes the forecasting methods, with a focus on the streamflow post-processing models, Sect. 3 introduces the case study methods, Sects. 4 and 5 present and discuss case study results, and Sect. 6 provides concluding remarks.
The forecasting methods investigated in this study share a similar general structure but differ in the streamflow post-processing (QPP) model. To facilitate the presentation, this section is organised as follows. The general structure is outlined in Sect. 2.1. Common features of the post-processing models are described in Sect. 2.2. Specific details of the MuTHRE and monthly QPP models are described in Sects. 2.3 and 2.4.
The forecasting methods in this study employ a deterministic hydrological model forced with an ensemble of rainfall forecasts and combined with a streamflow post-processing model. This general structure is illustrated schematically in Fig. 1 and detailed next.
Illustration of general approach used to produce streamflow forecasts. Layers represent ensemble members.
The deterministic hydrological model,
The streamflow forecasts are obtained in two steps. First, an ensemble of
The streamflow post-processing models are constructed using the residual error modelling approach. They comprise a deterministic component and a residual error model. The residual error model employs a streamflow transformation to represent the heteroscedasticity and skew of the errors, an autoregressive term to represent error persistence, and components to capture other features of errors such as seasonality.
We consider two forecasting methods which differ in the structure and
details of the streamflow post-processing model as follows, with a schematic representation of these models shown in Fig. 2a:
The post-processing models also differ in their parameter estimation (calibration) procedure. Figure 2b shows that the MuTHRE model is calibrated using observed daily rainfall and observed daily streamflow, whereas the monthly QPP model is calibrated to forecast daily rainfall and observed monthly streamflow (see Sects. 2.3.4 and 2.4.4 for details).
Conceptual diagrams of the seamless MuTHRE model and the non-seamless monthly QPP model. Ensemble components are indicated with multiple layers. Panel
Figure 2c illustrates the key operational distinction between the models. The MuTHRE model produces seamless daily streamflow forecasts that can be used at a range of lead times and aggregation periods (e.g. daily, weekly, fortnightly, and monthly). In contrast, the monthly QPP model produces only 1-month-ahead non-seamless monthly forecasts.
The next section presents common features of the post-processing models before moving to specific model details.
The deterministic component
The residual error model describing the relationship between the
probabilistic streamflow estimate
The transformation
The residual error term
When generating forecasts, recent streamflow observations are used to update
errors via the AR1 model and reduce uncertainty in
The seamless MuTHRE post-processing model operates at the daily timescale.
Uncertainty due to forecast rainfall and hydrological errors is represented
using the ensemble dressing approach (Pagano et al., 2013). The ensemble of daily raw streamflow forecasts,
In the context of Eq. (1), the deterministic component in the MuTHRE
model at its daily time step
The MuTHRE model assumes that the mean of the residual error –
The scaling factor –
Innovations are modelled using a two-component mixed Gaussian distribution as follows:
The parameters of the residual error model daily hydrological model simulations daily observed streamflow
Seasonality (
The non-seamless monthly QPP model operates at the monthly timescale. The raw forecasts are aggregated from daily to monthly scale and collapsed to their median value yielding
The deterministic component in the non-seamless model at its monthly time step
The residual error model is applied at the monthly scale after collapsing the ensemble of raw forecasts to a single time series.
The monthly residual error model captures seasonality in residuals by
varying the mean
The parameters of the monthly residual error model monthly deterministic forecasts monthly observed streamflow
All parameters are calibrated using the method of moments. Full details are provided in Woldemeskel et al. (2018).
The case study uses a set of 11 catchments from the Murray–Darling basin in Australia, including four catchments on the Upper Murray River (New South Wales and Victoria) and seven catchments on the Goulburn River (Victoria). These catchments have winter-dominated rainfall which leads to higher streamflow between June and October (see Fig. 3) and have less than 5 % of days with no flow. Catchment properties are summarised in Table 1. This same set of catchments was used to extensively evaluate the MuTHRE model in McInerney et al. (2020).
Location of the 11 case study catchments
Properties of the 11 case study catchments.
Time series of daily observed streamflow over a 22-year period between 1991
and 2012 are obtained from the Hydrologic Reference Stations (HRS) dataset
(
Rainfall forecasts are provided by the Australian Community Climate and Earth System Simulator – Seasonal (ACCESS-S; Hudson et al., 2017). The ACCESS-S rainfall forecasts are pre-processed using the method of Schepen et al. (2018) in order to reduce biases and improve the reliability in comparison to observed rainfall. An ensemble of 100 pre-processed rainfall forecasts that begin on the first day of each month and extend out to a maximum lead time of 1 month are used.
The conceptual rainfall–runoff model GR4J (Perrin et al., 2003) is used as the deterministic hydrological model
Calibration of model parameters and evaluation of forecasts is performed
using a leave-1-year-out cross validation procedure (McInerney et al., 2020). For each calendar year
Note that, in this work, we do not consider parametric uncertainty (in the hydrological and residual error models), which is expected to be a (relatively) minor contributor to total forecast uncertainty, given the long data period used in the estimation; this simplification is common in contemporary forecasting implementations (e.g. Engeland and Steinsland, 2014; Verkade et al., 2017).
For each year
This calibration/forecasting process is repeated for all 22 years, resulting in 22 sets of 1-year forecasts, which are subsequently merged into a single 22-year forecast to facilitate evaluation against streamflow observations.
Streamflow forecasts are evaluated using numerical metrics for the following
attributes:
Reliability refers to the degree of statistical consistency between the forecast distribution and the observed data. It is evaluated using the reliability metric of Evin et al. (2014). Lower metric values are better, with 0 representing perfect reliability and 1 representing the worst reliability. Sharpness refers to the spread of the forecast distribution, with sharper forecasts being those with lower spread. We use the sharpness metric of McInerney et al. (2020), which is based on the ratio of the average 90 % interquantile range (IQR) of the forecasts and a climatological distribution (described below). Lower values are better, with 0 representing a deterministic forecast (with no spread) and 1 representing the same sharpness as climatology. In contrast to the other attributes considered here, sharpness is a property of the forecast only and does not depend on the observed data. Volumetric bias refers to the long-term water balance error. It is quantified using the metric of McInerney et al. (2017) as the relative absolute difference between total observed streamflow and the total forecast streamflow (averaged over the forecast ensemble). Lower values are better, with 0 representing unbiased forecasts. Combined performance is quantified using the continuous ranked probability score (CRPS). The CRPS is defined as the sum of squared differences between forecast cumulative distribution function (CDF) and the empirical CDF of the observation. Note that the CRPS can be decomposed into terms representing individual performance aspects, namely reliability, and uncertainty/resolution (related to sharpness; Hersbach, 2000). We express this metric as a skill score (CRPSS) relative to the climatological distribution. Higher CRPSS values are better, with a value of 1 representing a perfectly accurate deterministic forecast and 0 representing the same skill as the climatological distribution.
The climatological distribution represents the distribution of daily streamflow for a given time of the year based solely on previously observed streamflow at that time of the year. It is constructed using a 29 d moving window approach, which is described in detail in McInerney et al. (2020).
The study focuses on the performance of the streamflow post-processing models at the monthly scale. The monthly MuTHRE forecasts are obtained by aggregating daily forecasts to the monthly scale. The monthly QPP model generates monthly forecasts directly.
Overall evaluation of monthly forecasts is performed using data from the entire evaluation period, i.e. all months and years, with more detailed stratified performance evaluation performed for individual months and years.
We also demonstrate the ability of the MuTHRE model to produce seamless forecasts, which are reliable over a range of lead times and aggregation scales. This is achieved by evaluating both (i) daily forecasts stratified by lead times from 1 to 28 d and (ii) cumulative flow forecasts for periods of 1–28 d. The forecast is considered to be seamless if reliability metrics are similar across all lead times and aggregation scales. The evaluation of cumulative flow forecasts expands on the analysis of McInerney et al. (2020), who evaluated only daily and monthly forecasts, and provides and important demonstration of seamless forecasting over the entire range of timescales from 1 to 28 d. We note that cumulative flow forecasts over 1 month correspond to monthly forecasts.
Forecast performance of the two streamflow post-processing models is compared across multiple catchments using practical significance tests, as described next. For each combination of performance metric (e.g. reliability) and stratification (e.g. month), a statistical test is used to determine whether differences in metric values over the range of catchments exceed a predefined margin representing practical significance (relevance).
The statistical tests are performed using the paired Wilcoxon signed rank test (Bauer, 1972), with controls applied to reduce the false discovery rate to 5 %, corresponding to a confidence level of 95 % (Wilks, 2006; Benjamini and Hochberg, 1995). The practical significance margin is taken as 20 % of the median metric value for the non-seamless monthly QPP model (following McInerney et al., 2020).
Figure 4 illustrates the streamflow forecast time series in the Biggara catchment (catchment ID 401012; see Fig. 3). Daily forecasts from the seamless MuTHRE model for a representative time period beginning on 1 May 2002 are shown in Fig. 4a. The observed daily streamflow lies within the 90 % probability limits of the MuTHRE forecasts for each lead time. As expected, the probability limits are tight for short lead times (when forecast rainfall uncertainty and hydrological uncertainty are small) and widen for longer lead times.
Time series of daily and cumulative probabilistic forecasts from the seamless MuTHRE model for the Murray River at Biggara (401012; see Fig. 3) for May 2002. The non-seamless monthly QPP model does not have the capability to produce these forecasts.
Figure 5 (left column) shows the performance of daily forecasts from the MuTHRE model for lead times of 1 to 28 d, evaluated over all case study catchments. The key finding from this analysis is that reliability is relatively constant over all lead times, with median metric values lying in the tight range of 0.04–0.06 (Fig. 5a). We also note that forecasts are sharper and have better CRPSS at short lead times and that bias is relatively constant.
Performance of MuTHRE forecasts in terms of daily streamflow
Figure 4b shows cumulative flow forecasts out to 28 d in the Biggara catchment for the representative time period. The cumulative flows based on observed streamflow lie well within the 90 % probability limits of the MuTHRE forecasts for all lead times.
Figure 5 (right column) shows the performance of cumulative flow forecasts from the MuTHRE model for lead times of 1 to 28 d over all catchments. Again, we see that reliability is relatively constant over all lead times, with median metric values between 0.04 and 0.06 (Fig. 5b). We also note that sharpness, volumetric bias, and CRPSS metrics are typically better for cumulative forecasts than for daily forecasts (compare the left and right columns in Fig. 5).
In summary, the forecasts from the MuTHRE model are seamless because they are reliable over (a) the range of lead times and (b) multiple aggregation scales, from the shortest scale of 1 d to the longest scale of 1 month, and everything in between. This result confirms and extends previous findings in McInerney et al. (2020), who focused on daily and monthly scales only. In contrast to the seamless MuTHRE model, the non-seamless monthly QPP model does not have the capability to produce forecasts of daily streamflow and cumulative flows for time periods below 1 month.
Figure 6 compares monthly forecasts from the seamless MuTHRE model and non-seamless monthly QPP model for the Biggara catchment. While there are some minor differences between the two forecasts (e.g. the monthly QPP model produces larger spread than the MuTHRE model during 2010), the two forecasts are clearly very similar.
Time series of monthly probabilistic forecasts for Murray River at Biggara (401012; see Fig. 3), for the seamless MuTHRE model and non-seamless monthly QPP model. Results are shown between the years 2000 and 2011.
Figure 7 compares monthly forecasts from the MuTHRE and monthly QPP models in terms of overall performance (left column) and when stratified by month (middle column) and year (right column). The key findings are as follows.
Overall performance (all months and years;
In summary, aggregated forecasts from the seamless MuTHRE model offer similar (not practically significant) and, in some cases, superior performance to forecasts from the non-seamless monthly QPP model, for the vast majority of performance metrics and stratifications considered in this study.
The empirical results show that the seamless MuTHRE model achieves essentially the same performance as the non-seamless monthly QPP model at
the monthly timescale and even provides improvement in some aspects. At
first glance, this outcome may seem surprising for the following reasons:
The seamless MuTHRE model is required to produce reliable forecasts over a range of lead times and aggregations scales, whereas the non-seamless monthly QPP model is only required to produce reliable monthly streamflow forecasts. The seamless MuTHRE model is calibrated at the daily scale, using only observed daily streamflow during calibration, while the non-seamless monthly QPP model is calibrated to match the observed monthly streamflow. The seamless MuTHRE model does not see the forecast rainfall during calibration, whereas the non-seamless monthly QPP model does.
The subsections below describe how the seamless MuTHRE model is able to achieve comparable/better performance than the non-seamless monthly QPP model despite these apparent challenges.
The seamless MuTHRE model produces daily forecasts that can be aggregated from timescales of 1 d to 1 month, whereas the non-seamless monthly QPP model produces forecasts only at the monthly scale. One might expect the enhanced capability obtained from the seamless MuTHRE model to come at some cost in performance at the monthly scale. Encouragingly, this is not the case.
The ability to reliably aggregate daily forecasts to the monthly scale demonstrates that the seamless MuTHRE model is adequately capturing temporal persistence in daily forecasts. The MuTHRE model represents temporal persistence in hydrological errors using the daily Eq. (
The seamless MuTHRE model is not calibrated to optimise performance at the monthly scale, as it uses only observed daily streamflow during calibration. On the other hand, the non-seamless monthly QPP model is calibrated to match the observed monthly streamflow, which could lead to improved performance at the monthly scale compared to the seamless MuTHRE model. As such, the comparable performance of the MuTHRE model at the monthly scale is particularly encouraging given that monthly data are not used in its calibration.
Both approaches use the same deterministic hydrological model calibrated using observed rainfall and streamflow data. However, due to structural differences in their representation of residual errors, the seamless MuTHRE and non-seamless monthly QPP models differ in the approach used to calibrate the residual error model parameters. The residual error model in the non-seamless monthly QPP model represents combined rainfall and hydrological uncertainty. It uses forecast rainfall during calibration and can (in theory) correct for biases and under-/overdispersion in rainfall forecasts. In contrast, the seamless MuTHRE model represents only hydrological uncertainty and is calibrated using observed rainfall. Uncertainty due to forecast rainfall is represented by propagating rainfall forecasts through the hydrological model. Since the MuTHRE model does not correct for forecast rainfall errors, this approach requires rainfall forecasts to be reliable in order to produce reliable streamflow forecasts (Verkade et al., 2017).
In this study we have used rainfall forecasts from the ACCESS-S numerical weather prediction model (Hudson et al., 2017), which were pre-processed at the catchment scale with the aim of reducing biases and over-/underdispersion at the daily scale and capturing the temporal persistence in rainfall (Schepen et al., 2018). As a result, these pre-processed rainfall forecasts are reliable and sharp at both the daily and monthly scale and do not have a detrimental impact on the performance of the seamless MuTHRE model.
Since the seamless MuTHRE model uses only observed rainfall in calibration, it does not require recalibration if an improved rainfall forecast product becomes available, which is a useful advantage in operational settings. In contrast, the monthly QPP model is calibrated using forecast rainfall and must be recalibrated whenever a new rainfall forecast is to be used. Note that this is a benefit of the ensemble dressing approach used in the MuTHRE model, rather than of the forecasts being seamless.
The MuTHRE and monthly QPP models differ in their use of recently observed streamflow data. The MuTHRE model uses daily streamflow observations to update both (i) the dynamic bias component of the error model, to account for monthly errors, and (ii) the daily AR1 model to account for recent daily errors and improve sharpness of forecasts for short lead times. In contrast, the monthly model only uses monthly aggregated streamflow observations to update the monthly AR1 model. The ability to utilise the most recent time series of daily streamflow observations provides the MuTHRE model with a potential advantage over the monthly QPP model (which see only monthly totals) and may be another reason why the MuTHRE model performs so well compared with the monthly QPP model.
The key practical benefits that the seamless MuTHRE model provide over the
non-seamless monthly QPP model are summarised below:
Seamless forecasts can be used to inform decisions at a range of timescales.
Seamless daily forecasts are easily integrated into river system models used for real-time decision-making. The forecasting system is simplified as a single seamless product can serve a range of forecast requirements at different timescales. Improvements in rainfall forecasting are easily integrated into the forecasting system (as described in Sect. 5.1.2).
The competitive performance of the seamless MuTHRE model, even at the native scale of the non-seamless monthly QPP forecasts, is clearly encouraging – it indicates that seamless forecasts do not require a compromise between capability (range of available forecast timescales) and performance. Proverbially speaking, a user of seamless forecasts can “have their cake and eat it too”. This finding provides further motivation to adopt seamless forecasts in research and practical work.
Future work is recommended on the following aspects:
Subseasonal streamflow forecasts at timescales ranging from daily to monthly are of major interest in water management. This study compares two streamflow post-processing (QPP) models, namely the seamless daily Multi-Temporal Hydrological Residual Error (MuTHRE) model and the more traditional non-seamless monthly QPP model used in the Australian Bureau of Meteorology's dynamic forecasting system. The MuTHRE model is designed at the daily scale and can be aggregated up to the monthly scale, whereas the monthly QPP model is designed directly at the monthly scale and does not produce forecasts at the daily scale. A case study with 11 catchments in southeastern Australia, the GR4J conceptual rainfall–runoff model, and pre-processed ACCESS-S rainfall forecasts are reported.
The key finding is that the seamless MuTHRE model achieves essentially the same monthly scale performance as the non-seamless monthly QPP model for the majority of metrics (reliability, sharpness, bias, and CRPSS) and stratifications (monthly and yearly). Remarkably, the seamless post-processing model achieves high-quality forecasts (based on the metrics considered in this study) at its native daily scale and matches the performance of the non-seamless monthly model at the monthly scale, despite not being calibrated at that timescale.
Seamless subseasonal forecasts, which are reliable over a wide range of lead times (1–30 d) and timescales (daily–monthly), offer numerous practical benefits over non-seamless forecasts. For users, seamless subseasonal forecasts can inform a wide range of management decisions, from flood warning to water supply operation, while for service providers, seamless forecasts will reduce the number of forecast products that require development and operation. As such it represents a single modelling tool with great versatility. The encouraging results from this study help motivate broader adoption of seamless forecasts, as they offer additional capability without a loss in performance.
The code used to produce the MuTHRE and monthly QPP forecasts in this study are available upon reasonable request from the contact author.
Data used in the case studies are available at
DM, MT, DK, GK, and NT conceptualised the project. DM, MT, DK, and GK developed the methods, and DM performed analysis. RL and FW prepared the data. DM, MT, DK, and GK wrote the original draft. All authors reviewed and edited the paper.
The contact author has declared that none of the authors has any competing interests.
Publisher's note: Copernicus Publications remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Support with supercomputing resources was provided by the Phoenix HPC service at the University of Adelaide. We thank Surendra Rauniyar and Christopher Pickett-Heaps, for their insightful feedback during the Bureau review of this paper, and two anonymous reviewers, for their constructive comments, all of which helped improve the quality of this work.
The research presented in this paper has been funded by the Australian Bureau of Meteorology.
This paper was edited by Micha Werner and reviewed by Marie-Amélie Boucher and one anonymous referee.