Have a personal or library account? Click to login
Hybrid wavelet transform – MLR and ANN models for river flow prediction: Case study of Brahmaputra river (Pancharatna station) Cover

Hybrid wavelet transform – MLR and ANN models for river flow prediction: Case study of Brahmaputra river (Pancharatna station)

Open Access
|Feb 2024

Full Article

Introduction

For water authorities to effectively manage water reserves for different water users like hydropower generation, agricultural, domestic, flood management, etc., it is necessary to forecast river flow hours, days, months, or possibly longer in advance. Large volumes of dynamic, non-linear, and noisy data can be handled effectively using soft computing approaches, especially when the underlying physical relations are not understood.

For rainfall-runoff modeling and runoff forecasting, neural network models are successfully applied and can be found in the literature (Dawson & Wilby, 1998; Jain & Chalisgaonkar, 2000; Raghuwanshi, Singh & Reddy, 2006; Tayfur & Singh, 2006; Besaw, Rizzo, Bierman & Hackett, 2010; Rezaeian Zadeh, Amin, Khalili & Singh, 2010). The ASCE Task Committee (2000) reviews hydrologic applications of artificial neural network (ANN). Solaimani (2009) utilized ANN for modeling the rainfall runoff relationship in a catchment area located in a semiarid region of Iran by adopting feed forward back propagation for the rainfall forecasting with various algorithms with performance of multi-layer perceptions. The monthly stream flow of Jarahi watershed was analyzed to calibrate the given models. The monthly hydrometric and climatic were ranged from 1969 to 2000. The results extracted from the comparative study indicated that ANN was more appropriate and efficient to predict the river runoff than the classic regression model. Using the feed-forward back propagation neural network (FFBPNN) and the cascade forward back propagation neural network (CFBPNN) models, Mohseni and Muskula (2023) created rainfall-runoff-based models in the Yerli sub-catchment of the upper Tapi basin using data from 36 years from 1981 to 2016 and concluded that the developed ANN model was capable of predicting runoff quite accurately.

Wavelet transform (WT) has gained popularity recently as a helpful method for examining trends, periodicities, and variations in time series. Similar to the Fourier transform and short-time Fourier transform, a WT is a powerful mathematical tool for signal processing that can analyze both stationary and non-stationary data and produce time and frequency information at a higher resolution that is not possible with the former. The WT offers a multiresolution analysis; that is, at low scales (high frequency), it provides poor frequency resolution and better time resolution; at high scales (low frequency), it provides poor frequency resolution and higher time resolution. Such information is significant in real practice for all time-series signals. The WT breaks down a non-stationary time series into a specific number of stationary time series. The WT is then used to integrate other single prediction techniques to increase prediction accuracy. Another trustworthy hybrid model for application in time series forecasting issues is the wavelet-ANN (WANN). To estimate runoff discharge for the Ligvanchai watershed in Tabriz, Iran, Nourani, Komasi and Mano (2009) explored the rainfall-runoff modeling utilizing the wavelet-ANN technique. The time series were decomposed up to four levels using Haar, Daubechies (db2), Symlet (sym3), and Coiflet (coif1) wavelets. The model's outcomes demonstrated the great quality of the Haar wavelet when compared to the others. Kisi (2009b) proposed neuro-wavelet model for forecasting daily intermittent stream flow. Forecasting accuracy of neuro-wavelet model was better than single ANN model. To predict the flow of the Malbrabha river in India, Nayak, Venkatesh, Krishna and Jain (2013) developed a wavelet neural network (WNN) hybrid model employing the db5 wavelet for one time step ahead forecasting. According to the results, the WANN model outperformed the ANN model. Nourani, Baghanam, Adamowski and Gebremichael (2013) utilized feed forward neural networks (FFNN) to model the rainfall-runoff process on a daily and multi-step (two days, three days and four days) ahead time scale. Authors used db4 and Haar wavelets to remove noise from runoff time series and discovered that the performance of the FFNN was improved by applying WT to the raw runoff data. Also, it was concluded that db4 wavelet exhibits superior results as compared to Haar wavelet. Shafaei and Kisi (2017) compared the WANN model with ANN and SVM for prediction of short-term daily river flow of Ajichai river of Iran. A db5 mother wavelet was used to decompose the raw data and concluded that the WANN model performed better than ANN and SVM. Wang et al. (2022) developed hybrid models by combining the wavelet theory with five diverse types of machine learning models such as support vector machine (SVM)-radial basis function, SVM-polynomial, decision tree, gradient boosting, random forest, and long- or short-term memory. The db4 mother wavelet was used in the study. A comparison revealed that hybrid models exhibit better estimates than the stand-alone ones. To estimate monthly stream flows in Amasya, Turkey, Katipoǧlu (2023a) combined a discrete wavelet transform and a feedforward backpropagation neural network. The author decomposed various meteorological data using different wavelets such as Haar, Daubechies 2, Daubechies 4, Discrete Meyer, Coiflet 3, Coiflet 5, Symlet 3, and Symlet 5, and found that the results of Coiflet 5 was superior. Katipoǧlu (2023b) developed wavelet-ANFIS (W-ANFIS) model using db10 mother wavelet for predicting monthly Bitlis river flows in Turkey and concluded that the W-ANFIS model proved successful. Other studies on runoff prediction using wavelet-ANN are found in literature (Rao & Krishna, 2009; Adamowski & Sun, 2010; Linh et al., 2021; Kumar, Kumar, Kumar, Elbeltagi & Kuriqi, 2022). Nourani, Baghanam, Adamowski and Kisi (2014), Khandekar and Deka (2016) review the application of WT in hydrology. Application of WT in drought prediction, groundwater prediction, evaporation prediction can be found in the literature (Djerbouai & Souag-Gamane, 2016; Patil & Deka, 2017; Araghi, Adamowski & Martinez, 2020; Katipoǧlu, 2023c; Katipoǧlu, 2023d; Katipoǧlu, 2023e).

In time series forecasting issues, the wavelet-MLR (WMLR) is another trustworthy hybrid model. Kisi (2010) combined discrete wavelet transform and linear regression (WR) for short-term stream flow forecasting of two stations in Turkey. In comparison to ANN and ARMA models, WR models are found to be more superior. Kisi (2011) proposed a wavelet regression (WR) model for daily river stage forecasting of two stations on the Schuylkill river in Philadelphia. The result of a WR model was superior to the ANN models. Zhang, Zhang and Sing (2018) developed four models – MLR, ANN, wavelet coupled with MLR (W-MLR) and ANN (W-ANN) – for stream flow forecasting at four stations in the East River basin in China. All models showed similar performance in forecasting stream flow one-day ahead, while W-MLR and W-ANN performed better in five-day ahead forecasting. Other studies on runoff prediction using wavelet-MLR is found in literature (Kisi, 2009a; Budu, 2013; Shoaib et al., 2018; Khazaee Poul, Shourian & Ebrahimi, 2019).

Most previous investigations employed a selected mother wavelet type. Also, more research is required to fully understand the potential of Daubechies wavelets of different orders in studying hydrologic time series behavior. So, in this study, it is suggested to create hybrid models by coupling wavelet transform with ANN and MLR, with the following objectives:

  • To compare the results of all hybrid models (WANN, WMLR) developed using db1, db2, db3, db8 and db10 Daubechies wavelets for multiple lead times (two days, four days, seven days, 14 days). Also, to compare results of hybrid models with standalone ANN and MLR models.

  • To investigate the effect of Daubechies wavelets db1, db2, db3, db8, and db10 on forecasting accuracy.

  • To investigate how decomposition level affects model effectiveness.

Study area and data collection

With an average discharge at its mouth of 19,830 cumec, Brahmaputra is the fourth-largest river in the world (Goswami, 1985). Brahmaputra river Pancharatna station is selected for the study. Ten-year (Jan 1990 – Dec 1999) daily flow data were collected from Water Resources Department, Assam, India. The catchment area up to Pancharatna station is 532,000 km2. The seasonal monsoon rhythm and the Himalayan snow's freeze-thaw cycle influence the river's hydrologic regime. During the flood season, there are noticeable large changes in discharge over a short period of time. A maximum differential of roughly 17,000 m3·s−1 in 24 h (June 7–8, 1990) and 24,000 m3·s−1 in 48 h (June 7–9, 1990) was observed in rising limb (Sarma, 2005). The location of Pancharatna station is shown in Figure 1.

FIGURE 1.

Location of the gauging sites

Source: Khandekar (2014)

Figure 2 displays the observed time series discharge at Pancharatna station. Figure 2 demonstrates that the discharge is remarkably non-stationary, particularly during monsoon season (from June to September, each year). The contribution of flow from snow melting, during February to April, causes the discharge in the rising limb to fluctuate as well.

FIGURE 2.

Observed flow series

Source: Khandekar (2014)

Table 1 displays the statistical properties of flow data and demonstrates substantial variability. The terms Qmean, Qmax, Qmin, Sd, and Cx in the table stand for the mean, maximum, minimum, standard deviation, and skewness, respectively.

TABLE 1.

Statistical properties of flow data

Statistical parameterTrainingTestingAll
Qmean [m3·s−1]16 15916 23616 161
Qmax [m3·s−1]59 83276 23676 236
Qmin [m3·s−1]2 6281 7231 723
Sd [m3·s−1]11 78312 38811 965
Cx0.7260.9680.809

The discharge throughout the study period exhibits significant fluctuations, as shown in Table 1 (minimum = 1,723 cumec, the highest = 76,236 cumec). The standard deviation is determined to be 11,965 cumec, showing a significant dispersion of values from the mean. Additionally, the observed flows show significant positive coefficient of skewness (Cx = 0.809), indicating that the data has a more scattered distribution about the mean.

Methodology
Wavelet transform (WT)

Wavelet theory is discussed thoroughly in works by Mallat (1998), and Labat, Ababou and Mangin (2000). Wavelet transform of a raw signal has the ability of providing time-frequency, using a range of window sizes. It divides the input signal into wavelets (small waves), which are scaled and shifted versions of the original wavelet, called mother wavelet. There are two forms of wavelet transformation: the continuous wavelet transform (CWT) and the discrete wavelet transform (DWT). The CWT of a signal x(t) is described: (1) CWT(a,b)=1ax(t)ψ*tbadt. CWT(a,b) = {1 \over {\sqrt a}}\mathop \smallint \nolimits_{- \infty}^\infty x(t) \cdot {\psi ^*}\left({{{t - b} \over a}} \right) \cdot {\rm{d}}t.

In Eq. (1), the transformed signal is a function of two variables, a and b, which represent the scale and translation factor of the function ψ(t), respectively; * corresponds to complex conjugate (Mallat, 1998).

The ψ(t) is the transforming function, called as mother wavelet, is mathematically defined as: (2) ψtdt=0. \mathop \smallint \nolimits_{- \infty}^\infty \psi \left(t \right){\rm{d}}t = 0.

From a data set of length (L), CWT produces L2 coefficients. Therefore, duplicate data is trapped within the coefficients, which may or may not be a desired quality (Nourani et al., 2009; Rajaee, Nourani, Zounemat-Kermani & Kisi, 2011). The analysis will be significantly more precise and efficient, resulting in N transform coefficients, if scales and positions are chosen based on the powers of two. The resulting transform is called DWT, and has the form: (3) ψm,nt=1aomψtnboaomaom, {\psi _{m,n}}\left(t \right) = {1 \over {\sqrt {a_o^m}}}\psi \left({{{t - n{b_o}a_o^m} \over {a_o^m}}} \right), where the wavelet translation and dilation are each controlled by an integer, m, and the other by n; bo which must be larger than 0, is the location parameter; ao is a specified fixed dilation step greater than 1. The most common and simplest choice for parameters ao and bo are 2 and 1 (time steps), respectively. This power of two logarithmic scaling of the translations and dilations is known as the dyadic grid arrangement (Mallat, 1989). For parameters ao and bo, 2 and 1 (time steps), respectively, are the most typical options. According to Mallat (1989), this power-of-two logarithmic scaling of the translations and dilations is referred to as the dyadic grid layout.

Two sets of functions, known as high-pass (wavelet function) and low-pass (scaling function) filters, are operated by DWT. The original time series undergo processing via high-pass and low-pass filters (as shown in Fig. 3), followed by a down-sampled process that discards every second data point (Deka & Prahlada, 2012). The detailed (D1, D2, ..., Dn) and approximation (A1, A2, ..., An) coefficients, which represent the low frequency and high frequency components of the original signal, respectively, are derived after the signal is passed through high pass and low pass filters. There will be a total of n + 1 coefficients at every given nth decomposition level, where there will be one series of approximation coefficients at the nth level (An) and n series of detailed coefficients (D1, D2, ..., Dn). The sum of An + D1 + D2 +... + Dn is equal to the original signal.

FIGURE 3.

Wavelet decomposition tree

Source: Khandekar (2014).

Artificial Neural Network (ANN)

Neural networks are interconnected groups of artificial neurons that may be utilized as a computational model for information processing based on the connectionist approach to computing. An ANN may be thought of mathematically as a universal approximator with the capacity to learn from instances without the need of explicit physics.

Three-layer-feedforward artificial neural networks are most frequently utilized in hydrologic time series modeling (Jain & Chalisgaonkar, 2000; Raghuwanshi, Singh & Reddy, 2006; Tayfur & Singh, 2006; Bajirao, Kumar, Kumar, Elbeltagi & Kuriqi, 2021; Katipoǧlu, 2023a). In the current study, the Lavenberg–Marquardt (LM) learning function and tangent sigmoid as the transfer function were employed in the feedforward ANN. The LM technique was used to train the ANN since it is more effective and faster than the traditional gradient descent technique (Kisi, 2009b; Mohseni & Muskula, 2023). In three-layer-feedforward ANN, information flow takes place from the input to the output side. Weights regulate the strength of a signal that passes across connections between neurons, altering the information that is transferred between them. While not connected to nodes in the same layer, nodes in one layer are connected to those in the next one. As a result, a node's output in a layer depends exclusively on the inputs and weights it receives from the previous layers. Every node multiplies each input by its weight, adds up the result, and then runs the total through a transfer function to get the desired outcome. Typically, this transfer function is represented as a sigmoid function, which is an S-shaped curve that increases continuously. The “S” upper and lower limb attenuation keeps the raw sums smoothly within predetermined bounds. The transfer function additionally adds a nonlinearity that improves the network's capacity to represent complicated functions (Jain & Chalisgaonkar, 2000). The sigmoid function is monotonically growing, continuous, and differentiable everywhere.

Multiple linear regression (MLR)

The MLR is one of the statistical methods which represents a mathematical equation expressing the relation between two or more explanatory variable and a response variable. The MLR attempts to model the link between two or more explanatory variables and a response variable by fitting a linear equation to the observed data (Sharma, Isik, Srivastava & Kalin, 2013). The relationship between the dependent variable y and the independent variable x is described by the MLR equation as follows: (4) y=A0+A1x1+A2x2++Anxn y = {A_0} + {A_1}{x_1} + {A_2}{x_2} + \cdots + {A_n}{x_n} where: xn is the value of nth predictor, A0 is the regression constant, Ai is the coefficient of nth predictor.

Model development
Input selection in model

For deciding the number of input parameters, no fixed rule is available. The values of the future time step (Qt+n) in any time-series forecasting are inevitably reliant on the antecedent values Qt, Qt−1, Qt−2, …, Qt−j (Table 2). But because the value of j (lag) is not known beforehand, it is difficult to determine how many lags in the past will lead to higher efficiency. In hydrologic time series forecasting, determining j is crucial since it can help to reduce information loss and the omission of critical input variables that could interfere with training.

TABLE 2.

Correlation coefficients for the flow series

Output* (Qt+n)Input (Qt−j)*
QtQt−1Qt−2Qt−3Qt−4Qt−5Qt−6Qt−7Qt−8Qt−9
Qt+20.9830.9710.9570.9420.9290.9150.9030.8910.8800.871
Qt+40.9570.9420.9280.9150.9020.8910.8800.8700.8620.854
Qt+70.9150.9020.8910.8800.8700.8620.8540.8460.8390.832
Qt+140.8460.8390.8320.8250.8180.8120.8050.7990.7920.784
*

n – lead time; j – lag.

Numerous researchers (Sudheer, Gosain & Ramsastri, 2002; Budu, 2013; Nayak et al., 2013; Khazaee Poul et. al., 2019; Sun, Niu & Sivakumar, 2019; Katipoǧlu, 2023a, 2023b, 2023c) have used the technique based on the statistical features such as cross-correlation, autocorrelation and partial correlation of the data series in order to identify a distinct input vector. In the current research, based on the autocorrelation coefficient between the relevant variables (shown in Table 2), the input vectors to the models are chosen. To forecast the value of discharge, based on the following various input combinations were taken into account (Khandekar, 2014):

  • Qt.

  • Qt, Q(t−1).

  • Qt, Q(t−1), Q(t−2).

  • Qt, Q(t−1), Q(t−2), Q(t−3).

  • Qt, Q(t−1), Q(t−2), Q(t−3), Q(t−4).

  • Qt, Q(t−1), Q(t−2), Q(t−3), Q(t−4), Q(t−5).

ANN and MLR model development

At the first stage, to forecast discharge, ANN and MLR models without data preprocessing were created. In the present study it was decided to predict discharge for multiple lead times such as two days, four days, seven days and 14 days because the Brahmaputra river is one of the major rivers in India which carries heavy flood during monsoon season, so it is necessary to design flood warning system well in advance. First 70% data was used to train the model and last 30% data was used for testing the model. Initially, a three-layer feedforward backpropagation ANN were applied in the study. For each lead time, an optimal input combination is obtained by providing each of one to six input combination obtained through autocorrelation as input to ANN and by varying number of neurons in the hidden layer from two to 10 through trial and error (Budu, 2013; Moosavi, Vafakhah, Shirmohammadi & Behnia, 2013). The optimal input combination that gave the minimum root mean square error (RMSE) during testing period for each lead time is shown in Table 3. The output layer has only one neuron which is the discharge value for the given lead time. From viewpoint of comparing all models, same input combination is employed for the MLR and all hybrid models. The time series data normalized between zero and one (Nourani et al., 2009) by dividing the discharge value by the maximum one.

TABLE 3.

Optimal input combination

Lead time [day]Input parameter*Output parameter** Q(t+n)
2Qt, Q(t−1)Q(t+2)
4Qt, Q(t−1), Q(t−2), Q(t−3)Q(t+4)
7Qt, Q(t−1), Q(t−2), Q(t−3)Q(t+7)
14Qt, Q(t−1), Q(t−2), Q(t−3)Q(t+14)
*

Qt – current discharge value; Q(t−1), Q(t−2), Q(t−3) – one-, two- and three-time step past discharge values;

**

n – lead time.

Development of hybrid models (WANN and WMLR)

After developing the ANN and MLR models, hybrid models were developed. First of all, using DWT, for developing hybrid models, the normalized input data were decomposed into approximation and detail coefficients. Since all hydrological data are observed at discrete time intervals, all hybrid models used DWT to process time series data in the form of approximations and details at various levels so that gross and small features of a signal can be separated (Deka & Prahlada, 2012). Then, to obtain the output at a predetermined lead time, the approximation and detail coefficient were fed as input to ANN and MLR. Without decomposition, the output signals were preserved as normalized original series. The complete flow chart of model development is shown in Figure 4.

FIGURE 4.

Flow chart of model development

Source: own elaboration.

Selection of mother wavelets and decomposition level

The mother wavelet that will be used depends on the data to be analyzed (Nayak et al., 2013). As per authors' knowledge, there is no comparative study on prediction of hydrological parameter using group of Daubechies wavelets. So, in this research, an irregular wavelet, Daubechies (db) of orders 1 (db1), 2 (db2), 3 (db3), 8 (db8), and 10 (db10), illustrated in Figure 5, was chosen to deal with very irregular signal form. Daubechies wavelets of order N (dbN) are all asymmetric, orthogonal, and biorthogonal. They are compactly supported wavelets with extremal phase and highest number of vanishing moments for a given support width (Misiti, Misiti, Oppenheim & Poggi, 2010). Daubechies wavelet has a support width of 2N – 1. Except for db1 (haar), no Daubechies wavelet has an explicit expression. Wavelets with compact support or a narrow window function are appropriate for a local analysis of the signal.

FIGURE 5.

Daubechies wavelets

Source: Khandekar (2014).

The optimal decomposition level was obtained using the formula l = int [log (L)], (Nourani et al., 2014; Khazaee et al., 2019; Tarate et al., 2021; Katipoǧlu, 2023a, 2023b, 2023c), where l is the decomposition level, L represents the number of time series data, int represents the integer part function, and log represents logarithm with base 10. In the present study, L is 3,650 so l is approximately four. However, in our study, we decomposed raw signal up to fifth level. At any lth decomposition level, DWT produces l detailed coefficients and one approximation coefficient at lth level.

Performance criteria

The following evaluation measures were used to compare model performance. (5) RMSE=i=1L(QobsQcom)2L, RMSE = \sqrt {{{\sum\nolimits_{i = 1}^L {{{({Q_{{\rm{obs}}}} - {Q_{{\rm{com}}}})}^2}}} \over L}}, (6) R2=1i=1L(QobsQcom)2i=1L(QobsQobs¯)2, {R^2} = 1 - {{\sum\nolimits_{i = 1}^L {{{({Q_{{\rm{obs}}}} - {Q_{{\rm{com}}}})}^2}}} \over {\sum\nolimits_{i = 1}^L {{{({Q_{{\rm{obs}}}} - Q\overline {_{{\rm{obs}}}})}^2}}}}, (7) MAE=1ni=1LQobsQcom, MAE = {1 \over n}\sum\nolimits_{i = 1}^L {\left| {{Q_{{\rm{obs}}}} - {Q_{{\rm{com}}}}} \right|}, (8) B=i=1LQcompi=1LQobs, B = {{\sum\nolimits_{i = 1}^L {{Q_{{\rm{comp}}}}}} \over {\sum\nolimits_{i = 1}^L {{Q_{{\rm{obs}}}}}}}, (9) SI=RMSEi=1LQobs¯, SI = {{RMSE} \over {\sum\nolimits_{i = 1}^L {Q\overline {_{{\rm{obs}}}}}}}, where: RMSE, R2, MAE, B, SI, L, Qobs, Qcom, and Qobs¯ Q\overline {_{{\rm{obs}}}} are root-mean squared error (RMSE), determination coefficient (R2), mean absolute error (MAE), bias (B), scatter index (SI), number of observations (L), observed data (Qobs), computed values (Qcom), mean of observed data ( Qobs¯ Q\overline {_{{\rm{obs}}}} ), respectively. A R2 of 0.9 or more is considered very satisfactory, 0.8 to 0.9 represents a fairly good model, and less than 0.8 is considered unsatisfactory (Dawson & Wilby, 2001).

Results and discussion
ANN and MLR model results

The results of all models are presented in Tables 47. It can be seen from Tables 47 that in comparison with MLR models, the ANN model has performed better for all lead times, except for the fourteen-day one. It is observed that the effectiveness of both models decreases while increasing lead time. This could be due to considerable fluctuations in the data around mean values, such as a high standard deviation. Tables 47 also indicates the optimal ANN structure (e.g. for a two-day lead time, 2-8-1 means two neurons in the input layer, eight neurons in the hidden layer, and one neuron in the output layer).

TABLE 4.

Values of statistical parameters for two-day lead time

Model typeTraining periodTesting periodOptimum ANN structure
RMSER2MAEBIASSIRMSER2MAEBIASSI
ANN1 764.660.9771 044.281.0020.1092 463.330.9601 401.831.0140.1522-8-1
MLR1 828.260.9761 058.121.0000.1132 535.960.9581 293.191.0000.156
WANN-db1l51 199.430.989757.471.0080.0741 795.990.9791 144.511.0180.11012-2-1
WMLR-db1l51 228.250.989680.191.0000.0761 758.640.980913.411.0000.109
WANN-db2l4960.300.993540.141.0010.0591 415.360.987796.431.0070.08710-2-1
WMLR-db2l51 093.660.991595.741.0000.0681 478.740.986769.430.9990.091
WANN-db3l4813.840.995469.561.0010.0501 202.330.990698.701.0070.07410-3-1
WMLR-db3l5909.810.994510.231.0000.0561 311.380.989629.760.9990.081
WANN-db8l5553.070.998368.931.0000.0341 054.100.992543.921.0000.06512-2-1
WMLR-db8l5526.830.998296.781.0000.033775.110.996406.200.9990.048
WANN-db10l5481.400.998298.510.9990.030933.060.994436.490.9990.05712-2-1
WMLR-db10l5471.990.998261.791.0000.029751.870.996369.691.0000.046

Note: MAE and RMSE are m3·s−1 unit (Results of only optimum decomposition level are shown).

Source: own elaboration.

TABLE 5.

Values of statistical parameters for four-day lead time

Model typeTraining periodTesting periodOptimum ANN structure
RMSER2MAEBIASSIRMSER2MAEBIASSI
ANN3 133.090.9292 067.411.0050.1943 890.350.9012 566.071.0310.2394-8-1
MLR3 136.830.9291 988.841.0000.1943 932.060.8992 274.721.0000.243
WANN-db1l51 918.220.9731 229.161.0080.1182 590.840.9561 717.581.0180.15924-3-1
WMLR-db1l52 088.670.9681 268.591.0000.1292 909.440.9451 618.861.0000.179
WANN-db2l51 431.040.985868.661.0010.0881 986.380.9741 137.201.0090.12224-2-1
WMLR-db2l41 922.490.9731 160.031.0000.1192 342.130.9641 354.140.9990.144
WANN-db3l51 280.060.988790.251.0010.0791 683.590.9821 034.971.0040.10424-2-1
WMLR-db3l51 528.750.983924.381.0000.0942 067.670.9721 084.630.9990.127
WANN-db8l4962.640.993601.761.0010.0591 465.350.986774.111.0000.09020-2-1
WMLR-db8l5940.030.994556.621.0000.0581 213.320.990664.470.9990.075
WANN-db10l5822.670.995527.851.0030.0511 477.120.986755.191.0040.09124-2-1
WMLR-db10l5785.370.995460.221.0000.0481 174.800.991612.691.0000.072

Note: MAE and RMSE are in m3·s−1 unit (Results of only optimum decomposition level are shown).

Source: own elaboration.

TABLE 6.

Values of statistical parameters for seven-day lead time

Model typeTraining periodTesting periodOptimum ANN structure
RMSER2MAEBIASSIRMSER2MAEBIASSI
ANN4 308.720.8662 801.051.0020.2665 185.330.8253 441.991.0320.3184-3-1
MLR4 490.920.8542 986.841.0000.2775 376.870.8113 429.750.9990.331
WANN-db1l52 830.190.9421 961.301.0040.1754 131.690.8892 942.381.0160.25424-3-1
WMLR-db1l53 122.620.9291 942.731.0000.1934 123.650.8892 415.440.9990.254
WANNdb2l52 038.560.9701 408.021.0040.1262 894.020.9451 927.131.0240.17824-2-1
WMLR-db2l52 623.550.9501 617.631.0000.1623 193.400.9331 937.760.9990.196
WANN-db3l51 683.520.9791 151.421.0030.1042 525.320.9581 639.741.0130.15524-3-1
WMLR-db3l52 129.550.9671 336.341.0000.1322 987.570.9421 640.060.9990.184
WANN-db8l51 313.630.987871.910.9990.0812 038.330.9731 233.970.9970.12524-2-1
WMLR-db8l51 318.740.987837.081.0000.0821 702.960.9811 042.410.9990.105
WANN-db10l51 187.510.989792.060.9990.0731 929.580.9751 060.011.0050.11924-2-1
WMLR-db10l51 167.650.990733.491.0000.0721 585.020.984881.761.0000.097

Note: MAE and RMSE are in m3·s−1 unit (Results of only optimum decomposition level are shown).

Source: own elaboration.

TABLE 7.

Values of statistical parameters for 14-day lead time

Model typeTraining periodTesting periodOptimum ANN structure
RMSER2MAEBIASSIRMSER2MAEBIASSI
ANN5 704.970.7654 035.671.0140.3537 084.210.6735 116.371.0680.4334-2-1
MLR4 584.370.8483 204.371.0000.2835 415.770.8093 630.940.9990.333
WANN-db1l54 173.090.8742 765.271.0070.2584 913.550.8433 425.921.0360.30024-4-1
WMLR-db1l54 761.190.8363 215.261.0000.2946 050.480.7623 886.880.9990.371
WANN-db2l53 071.960.9322 062.781.0040.1904 716.300.8553 017.381.0380.28824-3-1
WMLR-db2l54 037.440.8822 571.671.0000.2494 876.840.8453 109.410.9990.299
WANN-db3l52 420.950.9581 737.071.0140.1493 848.560.9042 663.881.0310.23524-2-1
WMLR-db3l53 172.170.9272 083.461.0000.1964 770.310.8522 694.260.9990.292––
WANN-db8l51 922.740.9731 353.081.0120.1193 746.150.9082 190.531.0200.22924-2-1
WMLR-db8l51 985.330.9711 325.191.0000.1232 612.090.9551 731.460.9990.160
WANN-db10l52 225.740.9641 600.630.9950.1383 539.600.9182 287.830.9860.21724-3-1
WMLR-db10l51 670.310.9791 137.671.0000.1032 196.460.9681 435.571.0000.134

Note: MAE and RMSE are in m3·s−1 unit (Results of only optimum decomposition level are shown).

Source: own elaboration.

WANN and WMLR model results

The normalized observed data was decomposed using Daubechies wavelets of order 1 (db1), 2 (db2), 3(db3), 8 (db8) and 10 (db10) up to fifth level decomposition, which were fed as input to ANN and MLR, making the models as WANN(dbNli) and WMLR(dbNli), respectively, where N is the order of Daubechies wavelet and i is decomposition level. The performances of these hybrid models only for best decomposition level (low RMSE) are presented in Tables 47. Results of Tables 47 show that wavelet-based hybrid models perform significantly better than the standalone model.

Effect of Daubechies wavelet order on model efficiency

After comparing the results of hybrid models with respect to wavelet order, it was observed that all models' efficiency is increasing with wavelet order, highest being at 10th order. For example, for four-day lead time (during testing period), from the Table 4 it was found that for WANN model the value of R2 increased from 0.956 (WANN-db1l5) to 0.986 (WANN-db10l5) and for WMLR model the increase is from 0.945 (WMLR-db1l5) to 0.991 (WMLR-db10l5). Similar trend was observed for all lead times. Table 8 shows percent improvement in RMSE values with increase in Daubechies wavelet order from db1 to db10. The average percent improvement in RMSE is found to be 57.19%. Figure 6 shows a sample plot of effect of Daubechies wavelet order on RMSE for four-day lead time.

TABLE 8.

Percent improvement in RMSE with increase in wavelet order from db1 to db10

Lead time [day]Improvement in RMSE [%]
257.25
454.66
761.56
1455.30

Source: own elaboration.

FIGURE 6.

Effect of Daubechies wavelet order on RMSE (four-day lead time)

Source: own elaboration.

In general, for all lead times, from the analysis it was found that for lower order wavelets (db1, db2, db3), WANN model performance was better as compared to the WMLR model, while for higher order wavelets (db8, db10), WMLR was found to be superior as compared to WANN. A careful study of Tables 47 reveals that the forecasting ability of each hybrid model improves with increasing wavelet order. According to the observed time series flow data (Fig. 2), it appears to occur in a high frequency during the monsoon season (June to September, i.e. four months out of the year) and in a low frequency during the other non-monsoon eight months. Wider-support wavelets can capture low frequencies, while wavelets with a smaller support can capture high frequencies. The support width of a Daubechies wavelet of order N is equal to 2N – 1 (Misiti et al., 2010), hence the support widths of db1, db2, db3, db8, and db10, wavelets are one, three, five, 15, and 19, respectively. In brief, compared to db1, db2, db3, and db8 wavelet-based forecast models, the db10 wavelet has a reasonable support and good time-frequency localization properties, which together allow the model to capture both the underlying trend and the short-term variations in the time series. This conclusion is consistent with the findings of (Nourani et al., 2013; Katipoǧlu, 2023b, 2023e), who demonstrated that higher order mother wavelet (db4 and db10) offered substantially better results than lower order Haar (db1) wavelet. Scatter plots for all lead times for best model [WMLR(db10)] are shown in Figure 7. The scatter plots showed that the majority of the points were quite close to the 45° line, with only a few having greater magnitudes of observed flow on the lower side, indicating model underestimate (BIAS slightly below 1.0). WMLR(db10) models performance was very satisfactory (R2 > 0.9) (Dawson & Wilby, 2001) for all lead times. Time series plots are shown in Figure 8 for all lead times.

FIGURE 7.

Scatter plot for lead time during different testing period

Source: own elaboration.

FIGURE 8.

Time series plot for lead time during different testing period influence of decomposition level on model performance

Source: own elaboration.

As mentioned earlier, the optimum decomposition level was determined using the formula l = int [log (L)]. In our study, we decomposed raw signal up to fifth level. For the best model WMLR(db10), the effect of decomposition level on R2 is presented in Table 9. A careful examination of Table 9 demonstrates that model efficiency enhances with decomposition level. The initial resolution level of the original time series captures the high frequency components, and as the decomposition level (scale) is increased, the signal becomes smoother and more stationary. As a result, prediction errors were not affected by scale. Figure 9 shows effect of decomposition level on R2 for WMLR(db10) model for all lead times.

TABLE 9.

Effect of decomposition level on determination coefficient (R2) for WMLR-db10 model (testing period)

Two-day lead timeFour-day lead timeSeven-day lead time14-day lead time
Model typeR2Model typeR2Model typeR2Model typeR2
WMLR-db10l10.976WMLR-db10l10.914WMLR-db10l10.824WMLR-db10l10.656
WMLR-db10l20.995WMLR-db10l20.966WMLR-db10l20.861WMLR-db10l20.684
WMLR-db10l30.996WMLR-db10l30.990WMLR-db10l30.967WMLR-db10l30.760
WMLR-db10l40.996WMLR-db10l40.991WMLR-db10l40.983WMLR-db10l40.953
WMLR-db10l50.996WMLR-db10l50.991WMLR-db10l50.984WMLR-db10l50.968

Source: own elaboration.

FIGURE 9.

Effect of decomposition level on determination coefficient (R2) for WMLR(db10) model

Source: own elaboration.

Analysis of results for monsoon season during testing period

Finally, because the Brahmaputra river carries substantial flood during the monsoon season (June to September), an attempt was made to assess the model's accuracy during the monsoon period (for three years in testing from 1997 to 1999) for the WMLR(db10) model. The statistical parameter values for the WMLR(db10) model during testing period of monsoon season are shown in Table 10. As shown in Table 10, the WMLR(db10) model demonstrated good performance during monsoon season producing extremely satisfactory results for all lead times, despite substantial non-stationarity. Figure 10 compares flow series between observed, WMLR(db10) and WANN(db10) modelled flow for monsoon season for two-day lead time. From Figure 10 it is clear that the WMLR(db10) model has captured well almost all peaks, except the highest.

TABLE 10.

Values of statistical parameters for WMLR(db10) model for monsoon season (June to September) in testing period

YearRMSER2MAEBIASS.I.
Two-day lead time
1997890.550.995596.9661.0000.034
19981 448.880.995767.8020.9990.042
19991 070.200.996629.8990.9990.034
Four-day lead time
19971 572.650.9841 072.411.0000.059
19982 188.390.9891 153.710.9990.063
19991514.680.992968.890.9980.048
Seven-day lead time
19972 243.790.9671 667.900.9980.084
19982 971.270.9801 677.601.0010.086
19991 797.460.9891 209.730.9990.058
14-day lead time
19973 035.860.9382 292.431.0020.115
19983 722.210.9672 295.271.0060.111
19992 588.930.9761 972.250.9980.084

Note: RMSE and MAE are in cumec unit.

Source: own elaboration.

FIGURE 10.

Flow series comparison between observed, WMLR(db10) and WANN(db10) modelled flow for monsoon season

Source: own elaboration.

Conclusions

In this study, hybrid models were created by coupling wavelet transform with ANN and MLR to predict the Brahmaputra River flow at Pancharatna station for two-, four-, seven-, and 14-day lead time using 10-year daily flow data. Daubechies wavelets db1, db2, db3, db8, and db10 were used to decompose the observed raw flow data up to fifth level, which were used as input to ANN and MLR. After comparing the results of WANN and WMLR models with respect to wavelet order, it was found that both hybrid models' efficiency is increasing with wavelet order, highest being at 10th order. Comparing the hybrid models' outcomes, it was discovered for all lead times that the WMLR-db10 model had produced more reliable and superior results than the WANN model. The average percent improvement in RMSE with increase in wavelet order from one to 10 was found to be 57.19%. Also, it was concluded that with increase in the decomposition level the model's efficiency was increased. Finally, it was concluded that the wavelet transform as a preprocessing tool has proved to be best for mapping the relation between input and output as compared to a single model. To evaluate the forecasting effectiveness of suggested wavelet coupled hybrid models, other hydrological time series variables, such as rainfall, temperature, and evapotranspiration, can be employed as model inputs. To investigate more accuracy, the work can be expanded with additional wavelet types. Study can further be applied at other stations with highly non-stationary data.

DOI: https://doi.org/10.22630/srees.5258 | Journal eISSN: 2543-7496 | Journal ISSN: 1732-9353
Language: English
Page range: 69 - 94
Submitted on: Sep 2, 2023
Accepted on: Jan 19, 2024
Published on: Feb 28, 2024
Published by: Warsaw University of Life Sciences - SGGW Press
In partnership with: Paradigm Publishing Services

© 2024 Sachin Dadu Khandekar, Dinesh Shrikrishna Aswar, Pandurang Digamber Sabale, Varsha Sachin Khandekar, Mohankumar Namdeorao Bajad, Shivakumar Khaple, published by Warsaw University of Life Sciences - SGGW Press
This work is licensed under the Creative Commons Attribution-NonCommercial 4.0 License.