Temporal Fusion Transformers for Streamflow Prediction: Value of Combining Attention with Recurrence

05/21/2023
by   Sinan Rasiya Koya, et al.
0

Over the past few decades, the hydrology community has witnessed notable advancements in streamflow prediction, particularly with the introduction of cutting-edge machine-learning algorithms. Recurrent neural networks, especially Long Short-Term Memory (LSTM) networks, have become popular due to their capacity to create precise forecasts and realistically mimic the system dynamics. Attention-based models, such as Transformers, can learn from the entire data sequence concurrently, a feature that LSTM does not have. This work tests the hypothesis that combining recurrence with attention can improve streamflow prediction. We set up the Temporal Fusion Transformer (TFT) architecture, a model that combines both of these aspects and has never been applied in hydrology before. We compare the performance of LSTM, Transformers, and TFT over 2,610 globally distributed catchments from the recently available Caravan dataset. Our results demonstrate that TFT indeed exceeds the performance benchmark set by the LSTM and Transformers for streamflow prediction. Additionally, being an explainable AI method, TFT helps in gaining insights into the streamflow generation processes.

READ FULL TEXT

page 3

page 6

page 7

page 11

page 12

research
06/21/2023

Probing the limit of hydrologic predictability with the Transformer network

For a number of years since its introduction to hydrology, recurrent neu...
research
07/07/2021

BumbleBee: A Transformer for Music

We will introduce BumbleBee, a transformer model that will generate MIDI...
research
04/30/2019

Very Deep Self-Attention Networks for End-to-End Speech Recognition

Recently, end-to-end sequence-to-sequence models for speech recognition ...
research
11/13/2022

GC-GRU-N for Traffic Prediction using Loop Detector Data

Because traffic characteristics display stochastic nonlinear spatiotempo...
research
05/19/2020

Should we hard-code the recurrence concept or learn it instead ? Exploring the Transformer architecture for Audio-Visual Speech Recognition

The audio-visual speech fusion strategy AV Align has shown significant p...
research
12/13/2020

Radial Deformation Emplacement in Power Transformers Using Long Short-Term Memory Networks

A power transformer winding is usually subject to mechanical stress and ...
research
12/29/2022

Unsupervised construction of representations for oil wells via Transformers

Determining and predicting reservoir formation properties for newly dril...

Please sign up or login with your details

Forgot password? Click here to reset