Multistepahead daily inflow forecasting using the ERAInterim reanalysis data set based on gradientboosting regression trees
 ^{1}Institute of Hydropower System and Hydroinformatics, Dalian University of Technology, Dalian 116024, China
 ^{2}Key Laboratory of Ocean Energy Utilization and Energy Conservation of Ministry of Education, Dalian University of Technology, Dalian 116024, China
 ^{1}Institute of Hydropower System and Hydroinformatics, Dalian University of Technology, Dalian 116024, China
 ^{2}Key Laboratory of Ocean Energy Utilization and Energy Conservation of Ministry of Education, Dalian University of Technology, Dalian 116024, China
Correspondence: Zhanwei Liu (337891617@qq.com)
Hide author detailsCorrespondence: Zhanwei Liu (337891617@qq.com)
Inflow forecasting plays an essential role in reservoir management and operation. The impacts of climate change and human activities have made accurate inflow prediction increasingly difficult, especially for longer lead times. In this study, a new hybrid inflow forecast framework – using the ERAInterim reanalysis data set as input and adopting gradientboosting regression trees (GBRT) and the maximal information coefficient (MIC) – is developed for multistepahead daily inflow forecasting. Firstly, the ERAInterim reanalysis data set provides more information for the framework, allowing it to discover inflow for longer lead times. Secondly, MIC can identify an effective feature subset from massive features that significantly affects inflow; therefore, the framework can reduce computational burden, distinguish key attributes from unimportant ones and provide a concise understanding of inflow. Lastly, GBRT is a prediction model in the form of an ensemble of decision trees, and it has a strong ability to more fully capture nonlinear relationships between input and output at longer lead times. The Xiaowan hydropower station, located in Yunnan Province, China, was selected as the study area. Six evaluation criteria, namely the mean absolute error (MAE), the rootmeansquared error (RMSE), the Pearson correlation coefficient (CORR), Kling–Gupta efficiency (KGE) scores, the percent bias in the flow duration curve highsegment volume (BHV) and the index of agreement (IA) are used to evaluate the established models utilizing historical daily inflow data (1 January 2017–31 December 2018). The performance of the presented framework is compared to that of artificial neural network (ANN), support vector regression (SVR) and multiple linear regression (MLR) models. The results indicate that reanalysis data enhance the accuracy of inflow forecasting for all of the lead times studied (1–10 d), and the method developed generally performs better than other models, especially for extreme values and longer lead times (4–10 d).
Reliable and accurate inflow forecasting 1–10 d in advance is significant with respect to the efficient utilization of water resources, reservoir operation and flood control, especially in areas with concentrated rainfall. Rainfall in southern China is usually concentrated for several days at a time due to strong convective weather, such as typhoons. Low accuracy inflow predictions can consequently mean that power stations are unable to devise reasonable power generation plans 7–10 d ahead of disaster events, and this subsequently leads to unnecessary water abandonment and even substantial economic losses. Figure 1 shows the “losses of electric quantity due to discarded water” (LEQDW) in Yunnan and Sichuan provinces, China, from 2011 to 2016 (Sohu, 2017; Jiang, 2018). The total LEQDW in Yunnan and Sichuan provinces increased from 1.5 ×10^{9} to 45.6 ×10^{9} kWh during the period from 2011 to 2016, with an average annual growth rate of 98.0 %. In recent years, due to the increasing number of hydropower stations and improved hydropower capacity, the problem of discarding water due to inaccurate inflow forecasting is becoming increasingly serious and has had a negative impact on the development of hydropower in China.
The main challenge with respect to inflow forecasting at present, which is caused by climate change and human activities, is low accuracy, especially for longer lead times (Badrzadeh et al., 2013; ElShafie et al., 2007). Meanwhile, streamflow variation, which also stems from climate change and anthropogenic activities, means that inflow forecasting models often need to be rebuilt, and the model parameters need to be recalibrated according to the actual inflow and meteorological data within 1 or 2 years.
To address these problems, a variety of models and approaches have been developed. These approaches can be divided into three categories: statistical methods (Valipour et al., 2013), physical methods (Duan et al., 1992; Wang et al., 2011; Robertson et al., 2013) and machinelearning methods (Chau et al., 2005; Liu et al., 2015; Rajaee et al., 2019; Zhang et al., 2018; Yaseen et al., 2019; Fotovatikhah et al., 2018; Mosavi et al., 2018; Chau, 2017; Ghorbani et al., 2018). Each method has its own conditions and scope of application.
Statistical methods are usually based on historical inflow records and mainly include the autoregressive model, the autoregressive moving average (ARMA) model and the autoregressive integrated moving average (ARIMA) model (Lin et al., 2006). Statistical methods assume that the inflow series is stationary and the relationship between input and output is simple. However, real inflow series are complex, nonlinear and chaotic (Dhanya and Kumar, 2011), making it difficult to obtain highaccuracy predictions using statistical models.
Physical methods, which have clear mechanisms, are implemented using theories of inflow generation and confluence. These methods can reflect the characteristics of the catchment but are very strict with initial conditions and input data (Bennett et al., 2016). Meanwhile, these methods, which are used for flood forecasting, have a shorter lead time and cannot be utilized to acquire longterm forecasting results due to input uncertainty.
Machinelearning methods, which have a strong ability to handle the nonlinear relationship between input and output and have recently shown excellent performance with respect to inflow prediction, are widely used for medium and longterm inflow forecasts. In particular, several studies have shown that artificial neural networks (ANNs; Rasouli et al., 2012; Cheng et al., 2015; ElShafie and Noureldin, 2011) and support vector regression (SVR; Tongal and Booij, 2018; Luo et al., 2019; Moazenzadeh et al., 2018) are two powerful models for inflow prediction. However, these models still have some inherent disadvantages. For example, ANNs are prone to being trapped by local minima, and both ANN and SVR suffer from overfitting issues and reduced generalizing performance.
In recent years, gradientboosting regression trees (GBRT) (Fienen et al., 2018; Friedman, 2001), a nonparametric machinelearning method based on a boosting strategy and decision trees, has been developed and has been used to study traffic (Zhan et al., 2019) and environmental (Wei et al., 2019) issues, where it has proven to alleviate the abovementioned problems. Thus, GBRT is selected for daily inflow prediction with lead times of 1–10 d in this paper. Compared with ANN and SVR, GBRT also has two other advantages. Firstly, GBRT can rank features according to their contribution to model scores, which is of great significance for reducing the complexity of the model. Secondly, GBRT is a white box model and can be easily interpreted. To the best of our knowledge, GBRT has not previously been used for daily inflow prediction with lead times of 1–10 d. For comparison purposes, ANN, SVR and multiple linear regression (MLR) have been employed to forecast daily inflow and are considered to be benchmark models in this study.
In addition to forecasting models, a vital reason why many approaches cannot attain a higher inflow prediction accuracy is that inflow is influenced by various factors (Yang et al., 2019), such as rainfall, temperature and humidity. Thus, it is very difficult to select appropriate features for inflow forecasting. The current feature selection methods for inflow forecasting mainly include two methodologies. The first method is the modelfree method (Bowden, 2005; Snieder et al., 2020) which employs a measure of the correlation coefficient criterion (Badrzadeh et al., 2013; Siqueira et al., 2018; Pal et al., 2013) to characterize the correlation between a potential model input and the output variable. The second method is the modelbased method (Snieder et al., 2020) which usually utilizes the model and search strategies to determine the optimal input subset. Common search strategies include forward selection and backward elimination (May et al., 2011). The correlation coefficient has a limited ability to capture nonlinear relationships and exhaustive searches tend to increase the computational burden. Thus, in order to accurately and quickly select effective inputs, the maximal information coefficient (MIC; Reshef et al., 2011) is used to select input factors for inflow forecasting.
MIC is a robust measure of the degree of correlation between two variables and has attracted a lot attention from academia (Zhao et al., 2013; Ge et al., 2016; Lyu et al., 2017; Sun et al., 2018). In addition, sufficient potential input factors are a prerequisite for obtaining reliable and accurate prediction results, and it is not enough to use only antecedent inflow series as model input. To enhance the accuracy of inflow forecasting and acquire a longer lead time, increasing amounts of meteorological forecasting data have been used for inflow forecasting (Lima et al., 2017; Fan et al., 2015; Rasouli et al., 2012). However, with extended lead times, the errors of forecast data continuously increase, as the variables are obtained using a numerical weather prediction (NWP) system are also affected by complex factors (Mehr et al., 2019). Moreover, due to the continuous improvement of forecasting systems, it is difficult to obtain consistent and long series of forecasting data (Verkade et al., 2013).
To mitigate these problems, reanalysis data generated by the European Centre for MediumRange Weather Forecasts (ECMWF) Reanalysis Interim – ERAInterim (Dee et al., 2011), which employs one of the best methods for the reanalysis of data describing atmospheric circulation and elements (Kishore et al., 2011), have been used as model input. The reanalysis data have less error than observed data and forecast data, which is the result of assimilating observed data with forecast data. ERAInterim provides the results of a global climate reanalysis from 1979 to date, which have been produced using a fixed version of a NWP system. The fixed version ensures that there are no spurious trends caused by an evolving NWP system. Therefore, these meteorological reanalysis data satisfy the need for long sequences of consistent data and have been used for the prediction of wind speeds (Stopa and Cheung, 2014) and solar radiation (Ghimire et al., 2019; LinaresRodríguez et al., 2011) in the past.
This study aims to provide a reliable inflow forecasting framework with longer lead times for daily inflow forecasting. The framework adopts the ERAInterim reanalysis data as model input, which ensures that ample information is supplied to depict inflow. The MIC is used to select appropriate features to avoid overfitting and the waste of computing resources caused by feature redundancy. GBRT, which is robust to outliers and has a strong nonlinear fitting ability, is used as the prediction model to improve the inflow forecasting accuracy of longer lead times. This paper is organized as follows: Sect. 2 describes a case study and the data collected; Sect. 3 introduces the theory and processes of the methods used, including the MIC and GBRT; Sect. 4 presents the results and a discussion of the data; and the conclusions are given in Sect. 5.
2.1 The study area and the data utilized
The Xiaowan hydropower station in the lower reaches of the Lancang River was chosen as the study site (Fig. 2). The Xiaowan hydropower station is the main controlling hydropower station in the Lancang River; therefore, it is very meaningful to adopt it as the case study. The Lancang River, which is also known as the Mekong River, is approximately 2000 km long and has a drainage area of 113 300 km^{2} above the Xiaowan hydropower station. The river originates on the Tibetan Plateau and runs through China, Myanmar, Laos, Thailand, Cambodia and Vietnam. The major source of water flowing into the Lancang River in China comes from melting snow on the Tibetan Plateau (Mekong River Commission, 2005).
We collected the ERAInterim reanalysis data set and the observed daily inflow and rainfall data for Xiaowan for 8 years (January 2011 to December 2018). Figure 3 depicts the daily inflow series. The data from January 2011 to December 2014 (1461 d, accounting for approximately 50 % of the whole data set), from January 2015 to December 2016 (731 d, accounting for approximately 25 % of the whole data set) and from January 2017 to December 2018 (730 d, accounting for approximately 25 % of the whole data set) are used as the training, validation and testing data sets respectively. The reanalysis data set can be downloaded from https://apps.ecmwf.int/datasets/data/interimfulldaily/levtype=sfc/ (last access: 1 July 2019), and it is provided every 12 h on a 0.25^{∘} × 0.25^{∘} spatial grid. Based on expert knowledge and available literature, the 26 nearsurface variables (Table A1), which include the total precipitation (tp), the 2 m temperature (t2m) and the total column water (tcw), from the reanalysis data are considered as potential predictors for inflow forecasting. More details regarding the ERAInterim data set are presented in Appendix A.
2.2 Feature scaling and feature selection
Feature scaling is necessary for machinelearning methods, and all features are scaled to the range between zero and one, as shown in Eq. (1), before being included in the calculation.
where x_{scale} and x_{original} indicate the scaled and original data respectively; and x_{max} and x_{min} represent the maximum and minimum of inflow series respectively. The reasonable selection of input variables can reduce the computational burden and improve the prediction accuracy of the model by removing redundant feature information and reducing the dimensions of the features. If too many features are selected, the model will become very complex, which will cause trouble when adjusting parameters and subsequently result in overfitting and difficult convergence. Moreover, natural patterns in the data will be blurred by noise (Zhao et al., 2013). Conversely, if irrelevant features are chosen, noise will be added into the model and hinder the learning process. The MIC is employed to select input data from candidate predictors from the reanalysis data set. The lagged inflow and rainfall series are identified using the partial autocorrelation function (PACF) and the crosscorrelation function (CCF). The corresponding 95 % confidence interval is used to identify significant correlations. Furthermore, when the correlation coefficient slowly declines and cannot fall into the confidence interval, a trialanderror procedure is used to determine the optimum lag, i.e. starting from one lag and then modifying the external inputs by successively adding one more lagged time series into the input data (Amiri, 2015; Shoaib et al., 2015).
3.1 Feature selection via the maximal information coefficient
The calculation of the MIC is based on the concept of mutual information (MI; Kinney and Atwal, 2014). For a random variable X, such as observed inflow, the entropy of X is defined as
where p(x) is the probability density function of X=x. Furthermore, for another random variable Y, such as observed rainfall, the conditional entropy of X given Y may be evaluated using the following expression:
where H(XY) is the uncertainty of X given knowledge of Y, and p(x, y) and p(xy) are the joint probability density and the conditional probability of X=x and Y=y respectively. The reduction of the original uncertainty of X, due to the knowledge of Y, is called the MI (Amorocho and Espildora, 1973; Chapman, 1986) and is defined by
Considering a given data set D, including variable X and Y with a sample size n, the calculation of the MIC is divided into three steps. Firstly, scatter plots of X and Y as well as grids for partitioning, which are called “xbyy” grids, are drawn. Let DG denote the distribution of D divided by one of the xbyy grids as G. MI${}^{*}(D,x,y)=\mathrm{max}\phantom{\rule{0.25em}{0ex}}\mathrm{MI}\left(D\rightG)$, where MI(DG) is the mutual information of DG. Secondly, the characteristic matrix is defined as
Lastly, the MIC is introduced as the maximum value of the characteristic matrix: MIC$\left(D\right)=\underset{xy<B\left(n\right)}{max}M(D{)}_{x,y}$, where B(n) is the upper bound of the grid size and is a function of sample size, which is defined as B=n^{0.6}. We perform feature selection from the ERAInterim reanalysis data set in two steps via MIC. First, we compute the MIC of each of the reanalysis variables and observed inflow. Then, we sort features based on the MIC in descending order and determine the optimum inputs using a trialanderror procedure, i.e. starting from the top feature and then modifying the external inputs by successively adding one more feature into model input. The selected k features from the reanalysis data are used as part of the model input.
3.2 Gradientboosting regression trees
GBRT is an ensemble model that mainly includes two algorithms: the decision tree algorithm and the boosting algorithm. The decision tree is robust to outliers and is used as a primitive model, and the boosting algorithm, which is an integration rule, is used to improve inflow forecasting accuracy.
3.2.1 The decision tree
The decision tree in this paper refers to decision tree learning used in computer science, which is one of the predictive modelling approaches used in machine learning. A decision tree consists of branch nodes (the tree structure) and leaf nodes (the tree output). Assuming that a training data set is given in a feature space with N features and each feature has n samples, (X_{1}, y_{1}), (X_{2}, y_{2}), …, (X_{n}, y_{n}) (X_{i}=(x_{1}, x_{2}, …, x_{N}), i=1, 2, …, n). In the input space where the training set is located, each region is recursively divided into two subregions, and the output value of each subregion is used to construct a binary decision tree. The topdown cyclic branch learning of the decision tree adopts a greedy algorithm where each branch node only cares about its own objective function. By traversing all features and all segmentation points of each feature, the best feature j and segmentation points “s” can be found by minimizing squared loss:
where
Here, y_{i} is the observed value, and R_{1}(j, s) and R_{2}(j, s) are the results of partitioning. c_{1} and c_{2} are output values of R_{1}(j, s) and R_{2}(j, s) respectively. Figure 4 shows an example of a decision tree model with a max depth and number of leaf nodes of 3 and 5 respectively. If the threshold of loss is set as the stopping condition of the decision tree, it will easily lead to overfitting problems. Hence, we set the following parameters to alleviate the overfitting problem of the decision tree model: the maximum depth of the tree, the minimum number of samples required to split an internal node, the minimum number of samples required at a leaf node and the number of leaf nodes. These parameters are also those used for optimization when using the decision tree.
3.2.2 The boosting algorithm
The idea of gradient boosting originated from the observation by Breiman (Breiman, 1997) and can be interpreted as an optimization algorithm based on a suitable cost function. Explicit regression gradientboosting algorithms have been subsequently developed (Friedman, 2001; Mason et al., 1999). The boosting algorithm used is described in the following. Supposing a training data set with n samples (X_{1}, y_{1}), (X_{2}, y_{2}), …, (X_{n}, y_{n}), a squared loss function is used to train the decision tree:
The core of the GBRT algorithm is the iterative process of training the decision with a residual method. The iterative training process of GBRT with M decision trees is explained in the following.

Initialization ${f}_{\mathrm{0}}\left(x\right)=\mathrm{arg}\underset{c}{min}\sum _{i=\mathrm{1}}^{n}L({y}_{i},c)$;

for m (m=1, 2, …, M) decision trees:
 a.
operating i (i=1, 2, …, n) sample points, and using the negative gradient of the loss function to replace the residual in the current model ${r}_{{m}_{i}}={\left[\frac{\partial L\left({y}_{i},f\left({x}_{i}\right)\right)}{\partial f\left({x}_{i}\right)}\right]}_{f\left(x\right)={f}_{m\mathrm{1}}\left(x\right)}$;
 b.
fitting a regression tree with ${R}_{{m}_{t}}\left\{\left({x}_{i},{r}_{{m}_{i}}\right)\right\}$, the ith regression tree with ${R}_{{m}_{t}}(t=\mathrm{1}$, 2, …, T) as its corresponding leaf node region is obtained, where t is the number of leaf nodes of regression;
 c.
for each leaf region t=1, 2, …, T, the best fitting value is calculated by ${c}_{{m}_{t}}={\scriptscriptstyle \begin{array}{c}\mathrm{argmin}\\ c\end{array}}\sum _{{x}_{i}\in {R}_{{m}_{t}}}L\left({y}_{i},{f}_{m\mathrm{1}}\left({x}_{i}\right)+c\right)$;
 d.
the fitting results are updated by adding the fitting values obtained to the previous values using ${f}_{{m}_{t}}\left({x}_{i}\right)={f}_{m\mathrm{1}}\left({x}_{i}\right)+\sum _{t=\mathrm{1}}^{T}{c}_{{m}_{t}}I\left({x}_{i}\in {R}_{{m}_{t}}\right)$;
 a.

finally, a strong learning method is obtained $\widehat{f}\left({x}_{i}\right)={f}_{M}\left({x}_{i}\right)=\sum _{m=\mathrm{1}}^{M}\sum _{t=\mathrm{1}}^{T}{c}_{{m}_{t}}I\left({x}_{i}\in {R}_{{m}_{t}}\right)$.
According to the above introduction to GBRT, the parameters of the GBRT can be divided into two categories: boosting parameters and tree parameters. The boosting parameters include the learning rate and the number of weak learners (learning_rate and n_estimators). The learning rate setting is used to reduce the gradient step. The learning rate influences the overall time of training: the smaller the value, the more iterations are required for training. There are four tree parameters: max_leaf_nodes, min_samples_leaf, min_samples_split and max_depth. Hence, GBRT has six parameters that control model complexity (Fienen et al., 2018), which we adjusted for tuning using a trialanderror procedure.
3.3 Evaluation criteria of the models
It is critical to carefully define the meaning of performance and to evaluate the performance on the basis of the forecasting and fitted values of the model compared with historical data. The rootmeansquared error (RMSE) and mean absolute error (MAE) are the most commonly used criteria to assess model performance, and they are calculated using Eqs. (11) and (12) respectively:
where ${\widehat{Q}}_{i}$ and Q_{i} are the inflow estimation and observed value at time i respectively, and n is the number of samples. The RMSE is more sensitive to extremes in the sample sets, and it is consequently used to evaluate the model's ability to simulate flood peaks. The Pearson correlation coefficient (CORR) is a measure of the strength of the association between the observed inflow series and the forecasted inflow series, and it is calculated according to Eq. (13):
where $\stackrel{\mathrm{\u203e}}{\widehat{Q}}$ is the mean of the estimation series. The range of the CORR is between zero and one and values close to one demonstrate a perfect estimation result. The Kling–Gupta efficiency (KGE) score (Knoben et al., 2019) is also a widely used evaluation index. It can be provided following Eqs. (14) and (15):
where σ is the standard deviation of the observed values, $\widehat{\mathit{\sigma}}$ is the standard deviation of the inflow estimation, μ is the mean of the observed series and $\widehat{\mathit{\mu}}$ is the mean of the inflow estimation series. The percent bias in the flow duration curve highsegment volume (BHV; Yilmaz et al., 2008; Vogel and Fennessey, 1994) is presented to estimate the prediction performance of the model for extreme values. It can be provided following Eq. (16):
where h=1, 2, …, H is the inflow index for inflows with exceedance probabilities lower than 0.02. In this paper, the inflow threshold of exceedance probabilities equalling 0.02 is 1722 m^{3} s^{−1}. The index of agreement (IA; Willmott, 1981) plays a significant role in evaluating the degree of the agreement between observed series and inflow estimation series. Similar to CORR, it ranges between zero (no agreement at all) and one (perfect fit). It is given by
3.4 Overview of framework
Figure 5 illustrates the overall structure of the framework presented. This structure consists of two major models: GBRT and GBRTMIC. In GBRT, we measure the relevance of the different lags in observed inflow and rainfall with observed inflow at the time of forecast using the partial autocorrelation function (PACF) and the crosscorrelation function (CCF; Badrzadeh et al., 2013), and we select appropriate lags as predictors for the model using hypothesis testing and trialanderror procedures. Data preprocessing and feature scaling are then carried out for selected predictors. Next, the data set is divided into a training set, a validation set, and a testing set according to the length of each data set, which is specified in advance (in Sect. 2.2). A grid search algorithm, which is an exhaustive searchall candidate parameter combination method, is guided to the optimization model parameters by the evaluation of the validation set for each lead time (Chicco, 2017). Lastly, the prediction results are evaluated based on the testing set. Compared with GBRT, GBRTMIC adds reanalysis data which are selected via MIC (in Sect. 3.1) as the model input. Moreover, GBRTMIC also calculates the importance of features according to the prediction results and ranks the features (Louppe, 2014).
It is difficult to perform multistep forecasting due to the accumulation of errors, reduced accuracy and increased uncertainty. Therefore, the current state of multistepahead forecasting is reviewed. There are two main strategies that one can use for multistep forecasting for single output, namely, static (direct) multistep forecasting and recursive multistep forecasting (Bontempi et al., 2012; Taieb et al., 2012). The recursive forecasting strategy is biased when the underlying model is nonlinear, and it is sensitive to the estimation error as estimated values, instead of actual values, are used more often as the forecasts move further into the future (Bontempi et al., 2012). Thus, the static multistep forecasting strategy is employed in this paper. As the static strategy does not use any approximated values to compute the forecasts, it is not prone to the accumulation of errors. The model structures of GBRT and GBRTMIC are as follows:
where ${\widehat{Q}}_{t+T}^{\mathrm{I}}$ and ${\widehat{Q}}_{t+T}^{\mathrm{II}}$ are the forecasted values of GBRT and GBRTMIC at lead time T of current time t respectively; ${\mathit{\theta}}_{t}^{\mathrm{I}}$ and ${\mathit{\theta}}_{t}^{\mathrm{II}}$ are parameters of GBRT and GBRTMIC at lead time T of current time t respectively; p and q are lags of the observed inflow and rainfall determined using the PACF and CCF respectively; E_{t} represents the features from reanalysis data at the current time t; and k is the number of features from the reanalysis data determined via the MIC.
In order to compare them with GBRTMIC, the ANNMIC, SVRMIC and MLRMIC models, which were obtained by replacing GBRT in the framework with ANN, SVR and MLR respectively, are also employed for inflow forecasting with lead times of 1–10 d. As mentioned previously, six indices, the MAE, RMSE, CORR, KGE, BHV and IA, are calculated to evaluate the performance of models based on the testing set. We also explored the feature importance based on the GBRTMIC model (Louppe, 2014). All computations carried out in this paper were performed on a ThinkPad P1 workstation containing an Intel Core i79850H CPU with 2.60 GHz and 16.0 GB of RAM, using the version 3.7.10 of Python (Python Software Foundation, 2020), which is powerful, fast and open, and the scikitlearn package (Pedregosa et al., 2011).
Note: “obs” represents the selected observed optimal input set, obs = {Q_{t−1}, Q_{t−4}, R_{t−1}, R_{t−2}, R_{t−3}, R_{t−4}, R_{t−5}, R_{t−6}}. “rea” represents the selected input set from the reanalysis, rea = {stl3_{t−1}, d2m_{t−1}, tcwv_{t−1}, tcw_{t−1}, stl2_{t−1}, mn2t_{t−1}, tsn_{t−1}, stl4_{t−1}, stl1_{t−1}, ro_{t−1}, swvl1_{t−1}, swvl2_{t−1}, swvl3_{t−1}}.
4.1 Feature selection
Figure 6 shows the PACF, CCF and the corresponding 95 % confidence interval from lag 1 to lag 12. The PACF shows significant autocorrelation at lag 1 and lag 4 respectively (Fig. 6a); thus, inflow series 1 and 4 d lag are selected as the model inputs. The CCF between inflow and rainfall gradually decreases as the time lag increases (Fig. 6b) and cannot fall into the 95 % confidence interval. Therefore, a trialanderror procedure is used to determine the optimal selection of the lagged rainfall series. A total of 13 input structures are tried (Table 1), and the trial results are shown in Fig. A1. The results indicate that 7th input structure shows the best performance. Accordingly, rainfall series from 1 to 6 d lag are selected as the model input. As mentioned previously, based on the MIC between inflow and the reanalysis variable (Table A1), a trialanderror procedure is used to determine the optimal input subset. A total of 26 input structures are tried (Table 2), and the trial results are shown in Fig. A2. The results show that 8th input structure shows the best performance; thus, the predictors nos. 1–8 in Table A1 are selected as the model input. Finally, a total of 16 variables including 8 observed variables and 8 reanalysis variables are selected as the model input (Table 3). As shown in Table 3, nos. 9–18 are reanalysis variables, and the range of the MIC of the reanalysis variables selected is 0.643 to 0.847. Furthermore, no. 9 and nos. 13–16 are variables related to temperature. Soil temperature level 3 (no. 9) is the temperature of the soil in layer 3 (28–100 cm, where the surface is at 0 cm). The temperature of the snow layer (no. 13) gives the temperature of the snow layer from the ground to the snow–air interface. Nos. 10–12 are variables related to the water content of the atmosphere. The 2 m dewpoint temperature (no. 10) is a measure of the humidity of the air, and, when combined with temperature and pressure, it can be used to calculate the relative humidity. The total column water vapour (no. 11) is the total amount of water vapour, which is a fraction of the total column water. The total column water (no. 12) is the sum of the water vapour, liquid water, cloud ice, rain and snow in a column extending from the surface of the Earth to the top of the atmosphere. The volumetric soil water layer 1 (no. 19) is the volume of water in soil layer 1. In summary, all the selected predictors are interpretable and have a good physical connection to inflow.
4.2 Hyperparameter optimization
For machinelearning methods, hyperparameters are parameters that are set before training and cannot be directly learnt from the regular training process. In order to improve the performance of models, it is imperative to tune the hyperparameters of models. The grid search function is employed to tune the hyperparameters of GBRT, GBRTMIC, ANNMIC and SVRMIC. After a review of the available literature (Badrzadeh et al., 2013; Rasouli et al., 2012), an optimizer in the family of quasiNewton methods, namely LBFGS, is chosen as the training algorithm for the ANN, and the number of hidden layers is fixed to three. Another two parameters, namely the activation function and the number of nodes of the hidden layer, need to be adjusted. A range of 2–20 neurons and four commonly used activation functions (Table 4) are selected by grid search. To alleviate the influence of the random initialization of weights, 50 ANNMIC models are trained for each parameter combination. The optimal activation function and the number of nodes of the hidden layer are determined by selecting the minimal MAE of the validation set for each lead time. The results of the trials show that tanh and the logistic function are the two more robust activation functions (Fig. 7), and ANN, with fewer nodes, is inclined to obtain a lower error. The optimal parameter combination for each lead time is listed in Table 5. It can be seen that the optimal number of nodes is 2, 3 or 4, and the optimal activation function is either tanh or the logistic function.
For SVR, according to Lin et al. (2006) and Dibike et al. (2001), the radial basis function (RBF) outperforms other kernel functions for runoff modelling; thus, RBF is used as the kernel function in this study. There are three parameters that need to be adjusted. Firstly, an appropriate tuning range of each parameter is determined by a trialanderror procedure. Then, to reach at an optimal choice of these parameters, the MAE is used to optimize the parameters using grid search. The optimal tuning parameters of SVR are shown in Table 5. As mentioned earlier, for GBRT, there are six parameters need to be adjusted. In order to obtain an optimal parameter combination as soon as possible, we optimize all parameters in two steps. Firstly, n_estimators and learning_rate are fixed to 100 and 0.1 respectively. The max_leaf_nodes, min_samples_leaf, max_depth and min_samples_split tuning parameters generate 40 000 models at each lead time. Secondly, after the tree parameters are determined, learning_rate is modified to 0.01 and n_estimators is determined using grid search. To accommodate the computational burden, all models are distributed among about 12 central processing units (CPUs), and the total wall time for the runs is about 7 h for GBRT_MIC and GBRT. Table 6 lists the optimal tuning parameters for GBRT and GBRTMIC.
Note: the bold text, (min, max, step) represents $\left[\mathrm{min}+\frac{\mathrm{max}\mathrm{min}}{\mathrm{step}\mathrm{1}}\times \mathrm{0},\phantom{\rule{0.25em}{0ex}}\mathrm{min}+\frac{\mathrm{max}\mathrm{min}}{\mathrm{step}\mathrm{1}}\times \mathrm{1},\phantom{\rule{0.125em}{0ex}}\mathrm{\dots},\phantom{\rule{0.125em}{0ex}}\mathrm{min}+\frac{\mathrm{max}\mathrm{min}}{\mathrm{step}\mathrm{1}}\times (\mathrm{step}\mathrm{1})\right]$.
4.3 Input comparison
Figure 8 illustrates the performance indices of GBRT and GBRTMIC for the testing set (1 January 2017–31 December 2018) at lead times of 1–10 d. It is obvious that the reanalysis data selected by the MIC greatly improves upon the GBRT forecasting at both short and long lead times. In particular, for the longer lead times, GBRTMIC significantly outperforms GBRT. From Fig. 8a, it can be noted that the MAE of GBRTMIC decreases from 175 to 172, which is a decrease of 1.74 %, for 2 d ahead forecasting, and it decreases from 273 to 237, which is a decrease of 13.18 %, for 10 d ahead forecasting compared with GBRT. From Fig. 8b, it can be seen that the RMSE of GBRTMIC achieves a 1.4 % and 10.6 % reduction for 2 and 10 d ahead forecasting respectively compared with GBRT. Figure 8c, d and f show that the CORR, KGE and IA of GBRTMIC increase by 0.2 %, 2.2 %, 1.0 % for 2 d ahead forecasting and 3.4 %, 7.8 % and 2.2 % for 10 d ahead forecasting respectively. Figure 8e compares the BHV values for GBRT and GBRTMIC and indicates that the reanalysis data can enhance the forecasting of extreme values. Figure 9a shows the 5 d ahead forecasted inflow of GBRTMIC and GBRT versus the observed inflow in the testing set. The slopes of the fitting curves of GBRTMIC and GBRT are 0.89 and 0.81 respectively; this also demonstrates that GBRTMIC can obtain more accurate inflow forecasting than GBRT. Figure 9b illustrates the distribution of the forecast errors of GBRT and GBRTMIC. The results show that the prediction error of two models has an approximately normal distribution. This demonstrates that the prediction error contains information that is not extracted by the model and that more errors of the forecasted inflow concentrate at around zero for GBRTMIC than for GBRT. Figure 9c provides forecasted inflow time series (from the testing set) for GBRTMIC and GBRT at a lead time of 5 d. It can be seen that GBRTMIC shows great performance compared with GBRT, especially for extreme values. This reveals that the problem of inaccurate extreme value prediction that has arisen in areas with concentrated rainfall for the GBRT model could be mitigated by incorporating the reanalysis data identified by the MIC.
4.4 Model comparison
GBRTMIC, SVRMIC and ANNMIC with the optimal model parameters are employed for inflow forecasting of 1 to 10 d ahead. The summarized results for the training and testing set are presented in Tables 7 and 8 respectively. To avoid the problem of local minima, 50 ANNMIC models are trained for each lead time, and the median of the predictions of the 50 models is used as the final prediction. It is clear from Table 7 that the GBRTMIC model is more efficient in the training set than other models at lead times of 1–10 d, which demonstrates that GBRTMIC has a powerful fitting ability. Meanwhile, all machinelearning models obtain better forecasted results than MLRMIC, which cannot capture nonlinear relationships. It should be noted that ANNMIC has the best performance for extreme values in terms of the BHV in the training set. As shown in Table 8, GBRTMIC performs best for the testing set at lead times of 4–10 d in terms of the abovementioned six indices. At a lead time of 10 d, the KGE of GBRTMIC even reached 0.8317. At lead times of 1–3 d, three machinelearning models obtain good performance and outperform MLRMIC. The machinelearning models can acquire enough information to perform forecasting at short lead times (1–3 d). The performance indices of these four models in the testing set (2017–2018) at lead times of 1–10 d are presented in Fig. 10. The results indicate that the performance of these four models decreases (higher MAE, RMSE and BHV, and lower CORR, KGE and IA) as the lead time increases. As mentioned earlier, the four models perform equally well for 1 to 3 d ahead forecasting, whereas significant differences among their performance are found as lead times exceed 3 d. This clearly indicates that GBRT produces much higher CORR, KGE and IA values and lower MAE, RMSE and BHV values than the other three models for 4–10 d ahead forecasting, although ANNMIC performs almost as well as GBRTMIC for 10 d ahead forecasting. It should be noted that SVR shows the worst performance according to the BHV and KGE values, and this demonstrates that SVR cannot capture extreme values. On the contrary, GBRTMIC significantly outperforms other models in terms of the BHV at lead times of 1–10 d, which indicates that GBRTMIC is the most successful model with respect to obtaining extreme values among all of the models developed in this paper.
4.5 Feature importance
A benefit of using gradient boosting is that after the boosted trees are constructed, the relative importance scores for each feature can be acquired to estimate the contribution of each feature to inflow forecasting. Figure 11 shows the feature importance based on GBRTMIC for lead times of 1 and 10 d respectively. The 1 d lag observed time series (Q_{t−1}) is more important for shorter lead times (Fig. 11a), which demonstrates that the historical observed values are essential to inflow forecasting at shorter lead times. The features (e.g. stl3_{t−10} and d2m_{t−10}) from the reanalysis data have a high relative importance at longer lead times (Fig. 11b). Based on the analysis of the concepts of stl3_{t−10} and tcw_{t−10} (Sect. 4.1), we infer that the temperature near the ground impacts the inflow by affecting the melting of snow, which is consistent with the fact that the Lancang River is a snowmelt river. The 10 d lag observed time series (Q_{t−10}) is also very important, which indicates the long memory of inflow series (Salas, 1993). Meanwhile, it is found that the reanalysis data provide important information for inflow forecasting at longer lead times.
In this study, GBRTMIC is employed to make inflow forecasts for lead times of 1–10 d, and ANNMIC, SVRMIC and MLRMIC are developed to compare with GBRTMIC. The reanalysis data selected by the MIC and the antecedent inflow and the rainfall records selected by the PACF and CCF are used as predictors to drive the models. These models are compared using six evaluation criteria: the MAE, RMSE, CORR, KGE, BHV and IA. It is shown that GBRTMIC, ANNMIC and SVRMIC outperform MLRMIC at lead times of 1–10 d, and GBRTMIC performs best at lead times of 4–10 d, especially for the forecasting of extreme values. According to a comparison of the forecasted results of GBRT and GBRTMIC, we conclude that GBRTMIC can be used for more accurate and reliable inflow forecasting at lead times of 1–10 d and that reanalysis data selected by the MIC greatly improve upon the GBRT forecasting, especially for lead times of 4–10 d. In addition, the feature importance achieved by GBRTMIC demonstrates that soil temperature, the total amount of water vapour in a column and dewpoint temperature near the ground contribute to increasing the prediction accuracy of inflow at longer lead times. In summary, the framework developed integrates GBRT and reanalysis data selected by the MIC and can perform inflow forecasting well at lead times of 1–10 d. The results of this study are of significance as they can assist power stations in making power generation plans 7–10 d in advance in order to reduce LEQDW and flood disasters. Another possibility to improve the results may be the consideration of heuristic methods (e.g. the Grey Wolf algorithm) to optimize model parameters, which could search a wider range of hyperparameters and optimization parameters more quickly.
ERAInterim is a reanalysis product of global atmospheric forecasts at ECMWF that is produced using the Integrated Forecast System (IFS) data assimilation system. The system includes a fourdimensional variational analysis (4DVar) with a 12 h analysis window. The spatial resolution of the data set is approximately 80 km (0.72^{∘}) with vertical 60 levels from the surface up to 0.1 hPa (Berrisford et al., 2011). The 0.125 to 2.5^{∘} reanalysis meteorological products are generated by interpolation. Reanalysis meteorological products from the ERAInterim such as rainfall, maximum and minimum temperatures, and wind speed at a 0.25^{∘} × 0.25^{∘} (latitude × longitude) spatial and 12 h temporal resolution for the study period from 2011 to 2018 are downloaded from the ECMWF web page.
A total of 13 input structures from the observed data are tried, and 50 trials are performed for each input structure. The results (Fig. A1) show that the 7th input structure is the optimal input subset for GBRT.
A total of 26 input structures from the reanalysis data are tried, and 50 trials are performed for each input structure. The results (Fig. A2) show that the 8th input structure is the optimal input subset for GBRTMIC.
The code used for this study can be found in Liao (2020; https://doi.org/10.5281/zenodo.3701182).
The reanalysis data set can be downloaded from https://apps.ecmwf.int/datasets/data/interimfulldaily/levtype=sfc/ (last access: 1 July 2019) (European Centre for Medium Range Weather Forecasts, 2019).
SL carried out the study design, the analysis and interpretation of data, and drafted the paper. ZL and BL participated in the study design, data collection, analysis of data and preparation of the paper. CC and ZZ carried out the experimental work and the data collection and interpretation. XJ participated in the design and coordination of the experimental work and the acquisition of data. All authors read and approved the final paper.
The authors declare that they have no conflict of interest.
This research is supported by National Natural Science Foundation of China (grant nos. 51979023 and U1765103) and the Liaoning Province Natural Science Foundation of China (grant no. 20180550354). We are grateful for reanalysis data provided by European Centre for MediumRange Weather Forecasts.
This research has been supported by the National Natural Science Foundation of China (grant nos. 51979023 and U1765103) and the Liaoning Province Natural Science Foundation of China (grant no. 20180550354).
This paper was edited by Giuliano Di Baldassarre and reviewed by three anonymous referees.
Amiri, E.: Forecasting daily river flows using nonlinear time series models, J. Hydrol., 527, 1054–1072, https://doi.org/10.1016/j.jhydrol.2015.05.048, 2015. a
Amorocho, J. and Espildora, B.: Entropy in the assessment of uncertainty in hydrologic systems and models, Water Resour. Res., 9, 1511–1522, https://doi.org/10.1029/WR009i006p01511, 1973. a
Badrzadeh, H., Sarukkalige, R., and Jayawardena, A.: Impact of multiresolution analysis of artificial intelligence models inputs on multistep ahead river flow forecasting, J. Hydrol., 507, 75–85, https://doi.org/10.1016/j.jhydrol.2013.10.017, 2013. a, b, c, d
Bennett, J. C., Wang, Q. J., Li, M., Robertson, D. E., and Schepen, A.: Reliable longrange ensemble streamflow forecasts: Combining calibrated climate forecasts with a conceptual runoff model and a staged error model, Water Resour. Res., 52, 8238–8259, https://doi.org/10.1002/2016WR019193, 2016. a
Berrisford, P., Kållberg, P., Kobayashi, S., Dee, D., Uppala, S., Simmons, A. J., Poli, P., and Sato, H.: Atmospheric conservation properties in ERAInterim, Q. J. Roy. Meteorol. Soc., 137, 1381–1399, https://doi.org/10.1002/qj.864, 2011. a
Bontempi, G., Taieb, S. B., and Le Borgne, Y.A.: Machine learning strategies for time series forecasting, in: European business intelligence summer school, Springer, Switzerland, 62–77, https://doi.org/10.1007/9783319611648, 2012. a, b
Bowden, G. J. D.G. M. H.: Input determination for neural network models in water resources applications. Part 1  background and methodology, J. Hydrol., 301, 75–92, https://doi.org/10.1016/j.jhydrol.2004.06.021, 2005. a
Breiman, L.: Arcing the edge, Report, Statistics Department, Technical Report 486, University of California, available at: https://statistics.berkeley.edu/sites/default/files/techreports/486.pdf (last access: 1 February 2020), 1997. a
Chapman, T. G.: Entropy as a Measure of Hydrologic Data Uncertainty and Model Performance, J. Hydrol., 85, 111–126, https://doi.org/10.1016/00221694(86)90079X, 1986. a
Chau, K., Wu, C., and Li, Y.: Comparison of several flood forecasting models in Yangtze River, J. Hydrol. Eng., 10, 485–491, https://doi.org/10.1061/(ASCE)10840699(2005)10:6(485), 2005. a
Chau, K.W.: Use of metaheuristic techniques in rainfallrunoff modelling, Water, 9, 1–6, https://doi.org/10.3390/w9030186, 2017. a
Cheng, C. T., Feng, Z.K., Niu, W.J., and Liao, S.L.: Heuristic Methods for Reservoir Monthly Inflow Forecasting: A Case Study of Xinfengjiang Reservoir in Pearl River, China, Water, 7, 4477–4495, https://doi.org/10.3390/w7084477, 2015. a
Chicco, D.: Ten quick tips for machine learning in computational biology, Biodata Min., 10, 35, https://doi.org/10.1186/s1304001701553, 2017. a
Dee, D. P., Uppala, S., Simmons, A., Berrisford, P., Poli, P., Kobayashi, S., Andrae, U., Balmaseda, M., Balsamo, G., and Bauer, D. P.: The ERA‐Interim reanalysis: Configuration and performance of the data assimilation system, Q. J. Roy. Meteorol. Soc., 137, 553–597, https://doi.org/10.1002/qj.828, 2011. a
Dhanya, C. T. and Kumar, D. N.: Predictive uncertainty of chaotic daily streamflow using ensemble wavelet networks approach, Water Resour. Res., 47, W06507, https://doi.org/10.1029/2010WR010173, 2011. a
Dibike, Y. B., Velickov, S., Solomatine, D., and Abbott, M. B.: Model Induction with Support Vector Machines: Introduction and Applications, J. Comput. Civil Eng., 15, 208–216, https://doi.org/10.1061/(ASCE)08873801(2001)15:3(208), 2001. a
Duan, Q., Sorooshian, S., and Gupta, V.: Effective and efficient global optimization for conceptual rainfallrunoff models, Water Resour. Res., 28, 1015–1031, https://doi.org/10.1029/91WR02985, 1992. a
ElShafie, A. and Noureldin, A.: Generalized versus nongeneralized neural network model for multilead inflow forecasting at Aswan High Dam, Hydrol. Earth Syst. Sci., 15, 841–858, https://doi.org/10.5194/hess158412011, 2011. a
ElShafie, A., Taha, M. R., and Noureldin, A.: A neurofuzzy model for inflow forecasting of the Nile river at Aswan high dam, Water Resour. Manage., 21, 533–556, https://doi.org/10.1007/s1126900690271, 2007. a
European Centre for Medium Range Weather Forecasts: ERA Interim, available at: https://apps.ecmwf.int/datasets/data/interimfulldaily/levtype=sfc/, last access: 1 July 2019. a
Fan, F. M., Schwanenberg, D., Collischonn, W., and Weerts, A.: Verification of inflow into hydropower reservoirs using ensemble forecasts of the TIGGE database for large scale basins in Brazil, J. Hydrol., 4, 196–227, https://doi.org/10.1016/j.ejrh.2015.05.012, 2015. a
Fienen, M. N., Nolan, B. T., Kauffman, L. J., and Feinstein, D. T.: Metamodeling for groundwater age forecasting in the Lake Michigan Basin, Water Resour. Res., 54, 4750–4766, https://doi.org/10.1029/2017WR022387, 2018. a, b
Fotovatikhah, F., Herrera, M., Shamshirband, S., Chau, K.W., Faizollahzadeh Ardabili, S., and Piran, M. J.: Survey of computational intelligence as basis to big flood management: Challenges, research directions and future work, Eng. Appl. Comp. Fluid., 12, 411–437, https://doi.org/10.1080/19942060.2018.1448896, 2018. a
Friedman, J. H.: Greedy function approximation: a gradient boosting machine, Ann. Stat., 29, 1189–1232, 2001. a, b
Ge, R., Zhou, M., Luo, Y., Meng, Q., Mai, G., Ma, D., Wang, G., and Zhou, F.: McTwo: a twostep feature selection algorithm based on maximal information coefficient, Bmc Bioinform., 17, 142–155, https://doi.org/10.1186/s1285901609900, 2016. a
Ghimire, S., Deo, R. C., Downs, N. J., and Raj, N.: Global solar radiation prediction by ANN integrated with European Centre for medium range weather forecast fields in solar rich cites of queensland Australia, J. Clean. Prod., 216, 288–310, https://doi.org/10.1016/j.jclepro.2019.01.158, 2019. a
Ghorbani, M. A., Kazempour, R., Chau, K.W., Shamshirband, S., and Ghazvinei, P. T.: Forecasting pan evaporation with an integrated artificial neural network quantumbehaved particle swarm optimization model: a case study in Talesh, Northern Iran, Eng. Appl. Comp. Fluid., 12, 724–737, https://doi.org/10.1080/19942060.2018.1517052, 2018. a
Jiang, R.: Focus on the focus of the “Three abandoned” electricity nearly 110 billion kWh, the loss of 48.7 billion yuan! How much electricity will be discarded in 2017 by water, fire, wind, light, and nuclear?, available at: https://www.inen.com/article/html/energy2266458.shtml (last access: 9 February 2020), 2018. a
Kinney, J. B. and Atwal, G. S.: Equitability, mutual information, and the maximal information coefficient, P. Natl. Acad. Sci. USA, 111, 3354–3359, https://doi.org/10.1073/pnas.1309933111, 2014. a
Kishore, P., Ratnam, M. V., Namboothiri, S., Velicogna, I., Basha, G., Jiang, J., Igarashi, K., Rao, S., and Sivakumar, V.: Global (50^{∘} S–50^{∘} N) distribution of water vapor observed by COSMIC GPS RO: Comparison with GPS radiosonde, NCEP, ERAInterim, and JRA25 reanalysis data sets, J. Atmos. SolTerr. Phy., 73, 1849–1860, https://doi.org/10.1016/j.jastp.2011.04.017, 2011. a
Knoben, W. J. M., Freer, J. E., and Woods, R. A.: Technical note: Inherent benchmark or not? Comparing Nash–Sutcliffe and Kling–Gupta efficiency scores, Hydrol. Earth Syst. Sci., 23, 4323–4331, https://doi.org/10.5194/hess2343232019, 2019. a
Lima, A. R., Hsieh, W. W., and Cannon, A. J.: Variable complexity online sequential extreme learning machine, with applications to streamflow prediction, J. Hydrol., 555, 983–994, https://doi.org/10.1016/j.jhydrol.2017.10.037, 2017. a
Lin, J. Y., Cheng, C.T., and Chau, K.W.: Using support vector machines for longterm discharge prediction, Hydrolog. Sci. J., 51, 599–612, https://doi.org/10.1623/hysj.51.4.599, 2006. a, b
LinaresRodríguez, A., RuizArias, J. A., PozoVázquez, D., and TovarPescador, J.: Generation of synthetic daily global solar radiation data based on ERAInterim reanalysis and artificial neural networks, Energy, 36, 5356–5365, https://doi.org/10.1016/j.energy.2011.06.044, 2011. a
Liu, Z., Zhou, P., and Zhang, Y.: A Probabilistic Wavelet–Support Vector Regression Model for Streamflow Forecasting with Rainfall and Climate Information Input, J. Hydrometeorol., 16, 2209–2229, https://doi.org/10.1175/JHMD140210.1, 2015. a
Louppe, G.: Understanding Random Forests: From Theory to Practice, PhD thesis, University of Liège, Liège, Belgium, 2014. a, b
Luo, X., Yuan, X., Zhu, S., Xu, Z., Meng, L., and Peng, J.: A hybrid support vector regression framework for streamflow forecast, J. Hydrol., 568, 184–193, https://doi.org/10.1016/j.jhydrol.2018.10.064, 2019. a
Lyu, H., Wan, M., Han, J., Liu, R., and Cheng, W.: A filter feature selection method based on the Maximal Information Coefficient and GramSchmidt Orthogonalization for biomedical data mining, Comput. Biol. Med., 89, 264–274, https://doi.org/10.1016/j.compbiomed.2017.08.021, 2017. a
Mason, L., Baxter, J., Bartlett, P. L., and Frean, M. R.: Boosting algorithms as gradient descent, in: Advances in Neural Information Processing Systems 12, Neural Information Processing Systems, 29 November–4 December 1999, Colorado, USA, 512–518, 1999. a
May, R., Dandy, G., and Maier, H.: Review of Input Variable Selection Methods for Artificial Neural Networks, in: Artificial Neural Networks, chap. 2, edited by: Suzuki, K., IntechOpen, Rijeka, https://doi.org/10.5772/16004, 2011. a
Mehr, A. D., Jabarnejad, M., and Nourani, V.: Paretooptimal MPSAMGGP: A new geneannealing model for monthly rainfall forecasting, J. Hydrol., 571, 406–415, https://doi.org/10.1016/j.jhydrol.2019.02.003, 2019. a
Mekong River Commission: Overview of the Hydrology of the Mekong Basin, Report, Mekong River Commission, Vientiane, available at: http://www.mekonginfo.org/assets/midocs/0001968inlandwatersoverviewofhydrologyofthemekongbasin.pdf (last access: 1 February 2020), 2005. a
Moazenzadeh, R., Mohammadi, B., Shamshirband, S., and Chau, K.W.: Coupling a firefly algorithm with support vector regression to predict evaporation in northern Iran, Eng. Appl. Comp. Fluid., 12, 584–597, https://doi.org/10.1080/19942060.2018.1482476, 2018. a
Mosavi, A., Ozturk, P., and Chau, K.W.: Flood prediction using machine learning models: Literature review, Water, 10, 1536, https://doi.org/10.3390/w10111536, 2018. a
Pal, I., Lall, U., Robertson, A. W., Cane, M. A., and Bansal, R.: Predictability of Western Himalayan river flow: melt seasonal inflow into Bhakra Reservoir in northern India, Hydrol. Earth Syst. Sci., 17, 2131–2146, https://doi.org/10.5194/hess1721312013, 2013. a
Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P., Weiss, R., and Dubourg, V.: Scikitlearn: Machine learning in Python, J. Mach. Learn Res., 12, 2825–2830, 2011. a
Python Software Foundation: Python Language Reference, version 3.7, available at: http://www.python.org (last access: 1 July 2019), 2020. a
Rajaee, T., Ebrahimi, H., and Nourani, V.: A review of the artificial intelligence methods in groundwater level modeling, J. Hydrol., 572, 336–351, https://doi.org/10.1016/j.jhydrol.2018.12.037, 2019. a
Rasouli, K., Hsieh, W. W., and Cannon, A. J.: Daily streamflow forecasting by machine learning methods with weather and climate inputs, J. Hydrol., 414, 284–293, https://doi.org/10.1016/j.jhydrol.2011.10.039, 2012. a, b, c
Reshef, D. N., Reshef, Y. A., Finucane, H. K., Grossman, S. R., McVean, G., Turnbaugh, P. J., Lander, E. S., Mitzenmacher, M., and Sabeti, P. C.: Detecting novel associations in large data sets, Science, 334, 1518–1524, https://doi.org/10.1126/science.1205438, 2011. a
Robertson, D., Pokhrel, P., and Wang, Q.: Improving statistical forecasts of seasonal streamflows using hydrological model output, Q. J. Roy. Meteorol. Soc., 17, 579–593, https://doi.org/10.5194/hess175792013, 2013. a
Salas, J. D.: Analysis and modelling of hydrological time series, in: Stochastic Water Resources Technology, Springer, Switzerland, 20–66, https://doi.org/10.1007/9781349034673_2, 1993. a
Shoaib, M., Shamseldin, A. Y., Melville, B. W., and Khan, M. M.: Runoff forecasting using hybrid Wavelet Gene Expression Programming (WGEP) approach, J. Hydrol., 527, 326–344, https://doi.org/10.1016/j.jhydrol.2015.04.072, 2015. a
Siqueira, H., Boccato, L., Luna, I., Attux, R., and Lyra, C.: Performance analysis of unorganized machines in streamflow forecasting of Brazilian plants, Appl. Soft. Comput., 68, 494–506, https://doi.org/10.1016/j.asoc.2018.04.007, 2018. a
Snieder, E., Shakir, R., and Khan, U.: A comprehensive comparison of four input variable selection methods for artificial neural network flow forecasting models, J. Hydrol., 583, 124299, https://doi.org/10.1016/j.jhydrol.2019.124299, 2020. a, b
Sohu: The causes and Countermeasures of hydropower waste water in Sichuan and Yunnan, available at: https://www.sohu.com/a/209379703_357198 (last access: 9 February 2020), 2017. a
Stopa, J. E. and Cheung, K. F.: Intercomparison of wind and wave data from the ECMWF Reanalysis Interim and the NCEP Climate Forecast System Reanalysis, Ocean Model., 75, 65–83, https://doi.org/10.1016/j.ocemod.2013.12.006, 2014. a
Sun, G., Li, J., Dai, J., Song, Z., and Lang, F.: Feature selection for IoT based on maximal information coefficient, Future Gener. Comp. Sy., 89, 606–616, https://doi.org/10.1016/j.future.2018.05.060, 2018. a
Taieb, S. B., Bontempi, G., Atiya, A. F., and Sorjamaa, A.: A review and comparison of strategies for multistep ahead time series forecasting based on the NN5 forecasting competition, Expert Syst. Appl., 39, 7067–7083, https://doi.org/10.1016/j.eswa.2012.01.039, 2012. a
Tongal, H. and Booij, M. J.: Simulation and forecasting of streamflows using machine learning models coupled with base flow separation, J. Hydrol., 564, 266–282, https://doi.org/10.1016/j.jhydrol.2018.07.004, 2018. a
Valipour, M., Banihabib, M. E., and Behbahani, S. M. R.: Comparison of the ARMA, ARIMA, and the autoregressive artificial neural network models in forecasting the monthly inflow of Dez dam reservoir, J. Hydrol., 476, 433–441, https://doi.org/10.1016/j.jhydrol.2012.11.017, 2013. a
Verkade, J., Brown, J., Reggiani, P., and Weerts, A.: Postprocessing ECMWF precipitation and temperature ensemble reforecasts for operational hydrologic forecasting at various spatial scales, J. Hydrol., 501, 73–91, https://doi.org/10.1016/j.jhydrol.2013.07.039, 2013. a
Vogel, R. M. and Fennessey, N. M.: FlowDuration Curves. I: New Interpretation and Confidence Intervals, J. Water Res. Pl., 120, 485–504, https://doi.org/10.1061/(ASCE)07339496(1994)120:4(485), 1994. a
Wang, E., Zhang, Y., Luo, J., Chiew, F. H., and Wang, Q.: Monthly and seasonal streamflow forecasts using rainfall‐runoff modeling and historical weather data, Water Resour. Res., 47, W05516, https://doi.org/10.1029/2010WR009922, 2011. a
Wei, Z., Meng, Y., Zhang, W., Peng, J., and Meng, L.: Downscaling SMAP soil moisture estimation with gradient boosting decision tree regression over the Tibetan Plateau, Remote Sens. Environ., 225, 30–44, https://doi.org/10.1016/j.rse.2019.02.022, 2019. a
Willmott, C. J.: On the validation of models, Phys. Geogr., 2, 184–194, https://doi.org/10.1080/02723646.1981.10642213, 1981. a
Yang, Q., Zhang, H., Wang, G., Luo, S., Chen, D., Peng, W., and Shao, J.: Dynamic runoff simulation in a changing environment: A data stream approach, Environ. Model. Softw., 112, 157–165, https://doi.org/10.1016/j.envsoft.2018.11.007, 2019. a
Yaseen, Z. M., Sulaiman, S. O., Deo, R. C., and Chau, K.W.: An enhanced extreme learning machine model for river flow forecasting: stateoftheart, practical applications in water resource engineering area and future research direction, J. Hydrol., 569, 387–408, https://doi.org/10.1016/j.jhydrol.2018.11.069, 2019. a
Yilmaz, K. K., Gupta, H. V., and Wagener, T.: A processbased diagnostic approach to model evaluation: Application to the NWS distributed hydrologic model, Water Resour. Res., 44, W09417, https://doi.org/10.1029/2007WR006716, 2008. a
Zhan, X., Zhang, S., Szeto, W. Y., and Chen, X.: Multistepahead traffic speed forecasting using multioutput gradient boosting regression tree, J. Intell Transport. S., 24, 1547–2442, https://doi.org/10.1080/15472450.2019.1582950, 2019. a
Zhang, D., Lin, J., Peng, Q., Wang, D., Yang, T., Sorooshian, S., Liu, X., and Zhuang, J.: Modeling and simulating of reservoir operation using the artificial neural network, support vector regression, deep learning algorithm, J. Hydrol., 565, 720–736, https://doi.org/10.1016/j.jhydrol.2018.08.050, 2018. a
Zhao, X., Deng, W., and Shi, Y.: Feature selection with attributes clustering by maximal information coefficient, Procedia Comput. Sci., 17, 70–79, https://doi.org/10.1016/j.procs.2013.05.011, 2013. a, b