Hydrological concept formation inside Long Short-Term Memory (LSTM) networks

Neural networks have been shown to be extremely effective rainfall-runoff models, where the river discharge is predicted from meteorological inputs. However, the question remains: what have these models learned? Is it possible to extract information about the learned relationships that map inputs to...

Full description

Bibliographic Details
Main Authors: Lees, T, Reece, S, Kratzert, F, Klotz, D, Gauch, M, De Bruijn, J, Sahu, R, Greve, P, Slater, LJE, Dadson, S
Format: Journal article
Language:English
Published: European Geosciences Union 2022
_version_ 1797108018791317504
author Lees, T
Reece, S
Kratzert, F
Klotz, D
Gauch, M
De Bruijn, J
Sahu, R
Greve, P
Slater, LJE
Dadson, S
author_facet Lees, T
Reece, S
Kratzert, F
Klotz, D
Gauch, M
De Bruijn, J
Sahu, R
Greve, P
Slater, LJE
Dadson, S
author_sort Lees, T
collection OXFORD
description Neural networks have been shown to be extremely effective rainfall-runoff models, where the river discharge is predicted from meteorological inputs. However, the question remains: what have these models learned? Is it possible to extract information about the learned relationships that map inputs to outputs, and do these mappings represent known hydrological concepts? Small-scale experiments have demonstrated that the internal states of long short-term memory networks (LSTMs), a particular neural network architecture predisposed to hydrological modelling, can be interpreted. By extracting the tensors which represent the learned translation from inputs (precipitation, temperature, and potential evapotranspiration) to outputs (discharge), this research seeks to understand what information the LSTM captures about the hydrological system. We assess the hypothesis that the LSTM replicates real-world processes and that we can extract information about these processes from the internal states of the LSTM. We examine the cell-state vector, which represents the memory of the LSTM, and explore the ways in which the LSTM learns to reproduce stores of water, such as soil moisture and snow cover. We use a simple regression approach to map the LSTM state vector to our target stores (soil moisture and snow). Good correlations (R2>0.8) between the probe outputs and the target variables of interest provide evidence that the LSTM contains information that reflects known hydrological processes comparable with the concept of variable-capacity soil moisture stores. <br><br> The implications of this study are threefold: (1) LSTMs reproduce known hydrological processes. (2) While conceptual models have theoretical assumptions embedded in the model a priori, the LSTM derives these from the data. These learned representations are interpretable by scientists. (3) LSTMs can be used to gain an estimate of intermediate stores of water such as soil moisture. While machine learning interpretability is still a nascent field and our approach reflects a simple technique for exploring what the model has learned, the results are robust to different initial conditions and to a variety of benchmarking experiments. We therefore argue that deep learning approaches can be used to advance our scientific goals as well as our predictive goals.
first_indexed 2024-03-07T07:23:38Z
format Journal article
id oxford-uuid:f2de3e77-0c6a-4d7a-9fcb-f9a863d48e38
institution University of Oxford
language English
last_indexed 2024-03-07T07:23:38Z
publishDate 2022
publisher European Geosciences Union
record_format dspace
spelling oxford-uuid:f2de3e77-0c6a-4d7a-9fcb-f9a863d48e382022-10-26T14:49:18ZHydrological concept formation inside Long Short-Term Memory (LSTM) networksJournal articlehttp://purl.org/coar/resource_type/c_dcae04bcuuid:f2de3e77-0c6a-4d7a-9fcb-f9a863d48e38EnglishSymplectic ElementsEuropean Geosciences Union2022Lees, TReece, SKratzert, FKlotz, DGauch, MDe Bruijn, JSahu, RGreve, PSlater, LJEDadson, SNeural networks have been shown to be extremely effective rainfall-runoff models, where the river discharge is predicted from meteorological inputs. However, the question remains: what have these models learned? Is it possible to extract information about the learned relationships that map inputs to outputs, and do these mappings represent known hydrological concepts? Small-scale experiments have demonstrated that the internal states of long short-term memory networks (LSTMs), a particular neural network architecture predisposed to hydrological modelling, can be interpreted. By extracting the tensors which represent the learned translation from inputs (precipitation, temperature, and potential evapotranspiration) to outputs (discharge), this research seeks to understand what information the LSTM captures about the hydrological system. We assess the hypothesis that the LSTM replicates real-world processes and that we can extract information about these processes from the internal states of the LSTM. We examine the cell-state vector, which represents the memory of the LSTM, and explore the ways in which the LSTM learns to reproduce stores of water, such as soil moisture and snow cover. We use a simple regression approach to map the LSTM state vector to our target stores (soil moisture and snow). Good correlations (R2>0.8) between the probe outputs and the target variables of interest provide evidence that the LSTM contains information that reflects known hydrological processes comparable with the concept of variable-capacity soil moisture stores. <br><br> The implications of this study are threefold: (1) LSTMs reproduce known hydrological processes. (2) While conceptual models have theoretical assumptions embedded in the model a priori, the LSTM derives these from the data. These learned representations are interpretable by scientists. (3) LSTMs can be used to gain an estimate of intermediate stores of water such as soil moisture. While machine learning interpretability is still a nascent field and our approach reflects a simple technique for exploring what the model has learned, the results are robust to different initial conditions and to a variety of benchmarking experiments. We therefore argue that deep learning approaches can be used to advance our scientific goals as well as our predictive goals.
spellingShingle Lees, T
Reece, S
Kratzert, F
Klotz, D
Gauch, M
De Bruijn, J
Sahu, R
Greve, P
Slater, LJE
Dadson, S
Hydrological concept formation inside Long Short-Term Memory (LSTM) networks
title Hydrological concept formation inside Long Short-Term Memory (LSTM) networks
title_full Hydrological concept formation inside Long Short-Term Memory (LSTM) networks
title_fullStr Hydrological concept formation inside Long Short-Term Memory (LSTM) networks
title_full_unstemmed Hydrological concept formation inside Long Short-Term Memory (LSTM) networks
title_short Hydrological concept formation inside Long Short-Term Memory (LSTM) networks
title_sort hydrological concept formation inside long short term memory lstm networks
work_keys_str_mv AT leest hydrologicalconceptformationinsidelongshorttermmemorylstmnetworks
AT reeces hydrologicalconceptformationinsidelongshorttermmemorylstmnetworks
AT kratzertf hydrologicalconceptformationinsidelongshorttermmemorylstmnetworks
AT klotzd hydrologicalconceptformationinsidelongshorttermmemorylstmnetworks
AT gauchm hydrologicalconceptformationinsidelongshorttermmemorylstmnetworks
AT debruijnj hydrologicalconceptformationinsidelongshorttermmemorylstmnetworks
AT sahur hydrologicalconceptformationinsidelongshorttermmemorylstmnetworks
AT grevep hydrologicalconceptformationinsidelongshorttermmemorylstmnetworks
AT slaterlje hydrologicalconceptformationinsidelongshorttermmemorylstmnetworks
AT dadsons hydrologicalconceptformationinsidelongshorttermmemorylstmnetworks