Hydrological concept formation inside long short-term memory (LSTM) networks

被引:58
|
作者
Lees, Thomas [1 ,6 ]
Reece, Steven [2 ]
Kratzert, Frederik [7 ]
Klotz, Daniel [3 ,4 ]
Gauch, Martin [3 ,4 ]
De Bruijn, Jens [6 ,8 ]
Kumar Sahu, Reetik [6 ]
Greve, Peter [6 ]
Slater, Louise [1 ]
Dadson, Simon J. [1 ,5 ]
机构
[1] Univ Oxford, Sch Geog & Environm, South Parks Rd, Oxford OX1 3QY, England
[2] Univ Oxford, Dept Engn, Oxford, England
[3] Johannes Kepler Univ Linz, LIT AI Lab, Linz, Austria
[4] Johannes Kepler Univ Linz, Inst Machine Learning, Linz, Austria
[5] UK Ctr Ecol & Hydrol, Maclean Bldg, Wallingford OX10 8BB, Oxon, England
[6] Int Inst Appl Syst Anal IIASA, Laxenburg, Austria
[7] Google Res, Vienna, Austria
[8] Vrije Univ Amsterdam, Inst Environm Studies, Boelelaan 1087, NL-1081 HV Amsterdam, Netherlands
基金
英国自然环境研究理事会;
关键词
MODELS;
D O I
10.5194/hess-26-3079-2022
中图分类号
P [天文学、地球科学];
学科分类号
07 ;
摘要
Neural networks have been shown to be extremely effective rainfall-runoff models, where the river discharge is predicted from meteorological inputs. However, the question remains: what have these models learned? Is it possible to extract information about the learned relationships that map inputs to outputs, and do these mappings represent known hydrological concepts? Small-scale experiments have demonstrated that the internal states of long short-term memory networks (LSTMs), a particular neural network architecture predisposed to hydrological modelling, can be interpreted. By extracting the tensors which represent the learned translation from inputs (precipitation, temperature, and potential evapotranspiration) to outputs (discharge), this research seeks to understand what information the LSTM captures about the hydrological system. We assess the hypothesis that the LSTM replicates real-world processes and that we can extract information about these processes from the internal states of the LSTM. We examine the cell-state vector, which represents the memory of the LSTM, and explore the ways in which the LSTM learns to reproduce stores of water, such as soil moisture and snow cover. We use a simple regression approach to map the LSTM state vector to our target stores (soil moisture and snow). Good correlations (R-2 > 0.8) between the probe outputs and the target variables of interest provide evidence that the LSTM contains information that reflects known hydrological processes comparable with the concept of variable-capacity soil moisture stores. The implications of this study are threefold: (1) LSTMs reproduce known hydrological processes. (2) While conceptual models have theoretical assumptions embedded in the model a priori, the LSTM derives these from the data. These learned representations are interpretable by scientists. (3) LSTMs can be used to gain an estimate of intermediate stores of water such as soil moisture. While machine learning interpretability is still a nascent field and our approach reflects a simple technique for exploring what the model has learned, the results are robust to different initial conditions and to a variety of benchmarking experiments. We therefore argue that deep learning approaches can be used to advance our scientific goals as well as our predictive goals.
引用
收藏
页码:3079 / 3101
页数:23
相关论文
共 50 条
  • [1] Niederschlags-Abfluss-Modellierung mit Long Short-Term Memory (LSTM)Rainfall-Runoff modeling with Long Short-Term Memory Networks (LSTM)—an overview
    Frederik Kratzert
    Martin Gauch
    Grey Nearing
    Sepp Hochreiter
    Daniel Klotz
    [J]. Österreichische Wasser- und Abfallwirtschaft, 2021, 73 (7-8) : 270 - 280
  • [2] Using Long Short-Term Memory (LSTM) networks with the toy model concept for compressible pulsatile flow metering
    Brahma, Indranil
    [J]. MEASUREMENT, 2023, 223
  • [3] Simplified Gating in Long Short-term Memory (LSTM) Recurrent Neural Networks
    Lu, Yuzhen
    Salem, Fathi M.
    [J]. 2017 IEEE 60TH INTERNATIONAL MIDWEST SYMPOSIUM ON CIRCUITS AND SYSTEMS (MWSCAS), 2017, : 1601 - 1604
  • [4] Long Short-Term Memory (LSTM) Neural Networks Applied to Energy Disaggregation
    Tongta, Anawat
    Chooruang, Komkrit
    [J]. 2020 8TH INTERNATIONAL ELECTRICAL ENGINEERING CONGRESS (IEECON), 2020,
  • [5] Economic Nowcasting with Long Short-Term Memory Artificial Neural Networks (LSTM)
    Hopp, Daniel
    [J]. JOURNAL OF OFFICIAL STATISTICS, 2022, 38 (03) : 847 - 873
  • [6] Applications of Long Short-Term Memory (LSTM) Networks in Polymeric Sciences: A Review
    Malashin, Ivan
    Tynchenko, Vadim
    Gantimurov, Andrei
    Nelyub, Vladimir
    Borodulin, Aleksei
    [J]. Polymers, 2024, 16 (18)
  • [7] Long Short-Term Memory (LSTM) Deep Neural Networks in Energy Appliances Prediction
    Kouziokas, Georgios N.
    [J]. 2019 PANHELLENIC CONFERENCE ON ELECTRONICS AND TELECOMMUNICATIONS (PACET2019), 2019, : 162 - 166
  • [8] Rainfall-runoff modelling using Long Short-Term Memory (LSTM) networks
    Kratzert, Frederik
    Klotz, Daniel
    Brenner, Claire
    Schulz, Karsten
    Herrnegger, Mathew
    [J]. HYDROLOGY AND EARTH SYSTEM SCIENCES, 2018, 22 (11) : 6005 - 6022
  • [9] Multilayer Long Short-Term Memory (LSTM) Neural Networks in Time Series Analysis
    Malinovic, Nemanja S.
    Predic, Bratislav B.
    Roganovic, Milos
    [J]. 2020 55TH INTERNATIONAL SCIENTIFIC CONFERENCE ON INFORMATION, COMMUNICATION AND ENERGY SYSTEMS AND TECHNOLOGIES (IEEE ICEST 2020), 2020, : 11 - 14
  • [10] Wind Speed Prediction and Visualization Using Long Short-Term Memory Networks (LSTM)
    Ehsan, Amimul
    Shahirinia, Amir
    Zhang, Nian
    Oladunni, Timothy
    [J]. 2020 10TH INTERNATIONAL CONFERENCE ON INFORMATION SCIENCE AND TECHNOLOGY (ICIST), 2020, : 234 - 240