Journal
JOURNAL OF HYDROLOGY
Volume 589, Issue -, Pages -Publisher
ELSEVIER
DOI: 10.1016/j.jhydrol.2020.125359
Keywords
Long short-term memory; Runoff forecast; Attention; Self-attentive long short-term memory
Funding
- National Key Research and Development Project of China [2017YFE0100700]
- National Natural Science Foundation of China [41871340, 41801036, 41911530191]
- Shanghai Key Laboratory of Multidimensional Information Processing, East China Normal University [2020KEY003]
- Institute of Eco-Chongming [ECNU-IEC-201902]
- Fundamental Research Funds for the Central Universities
Ask authors/readers for more resources
It is still very challenging to enhance the accuracy and stability of daily runoff forecasts, especially several days ahead, owing to the non-linearity of the forecasted processes. Here, we hypothesize that short lag-time has a significant impact on forecasting results. Thus, we incorporate short previous time steps into long short-term memory (LSTM) and develop the Self-Attentive Long Short-Term Memory (SA-LSTM). In SA-LSTM, the self-attention mechanism is used to model interdependencies within short previous time steps. SA-LSTM is evaluated at eight runoff datasets. The experimental results demonstrate that, compared with state-of-art benchmark models, SA-LSTM achieves the best performance. The RMSEs of SA-LSTM are at least 2.3% smaller than that of the second best model at the seventh day. The NSEs and NSE_In of SA-LSTM are at least 4.6% and 6.4% higher than those of the second best model at the seventh day. Furthermore, SA-LSTM also surpasses the baseline methods for base, mean and peak flows. The superiority of SA-LSTM can be attributed to its exploitation of information in short lag-time.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available