Integration of Generative-Adversarial-Network-Based Data Compaction and Spatial Attention Transductive Long Short-Term Memory for Improved Rainfall–Runoff Modeling
"> Figure 1
<p>Overview of the proposed rainfall–runoff prediction model.</p> "> Figure 2
<p>Overview of the random key method employed to convert numeric vectors into distinct hyperparameter levels.</p> "> Figure 3
<p>RMSE values with 95% confidence intervals for selected basins. Error bars show the uncertainty in model predictions, with narrower intervals indicating higher confidence.</p> "> Figure 4
<p>Loss curves for training and validation datasets in the proposed model.</p> "> Figure 5
<p>A depiction of the model’s application on individual rainfall–runoff for basin 12374250, spanning from 1 January 2013 to 1 January 2014, is presented. Here, Proposed-i denotes the predictions for runoff made by our model, calculated for the i-th day ahead.</p> "> Figure 5 Cont.
<p>A depiction of the model’s application on individual rainfall–runoff for basin 12374250, spanning from 1 January 2013 to 1 January 2014, is presented. Here, Proposed-i denotes the predictions for runoff made by our model, calculated for the i-th day ahead.</p> "> Figure 6
<p>A depiction of the model’s application on regional rainfall–runoff modeling from 1 January 2013 to 1 January 2014 is presented. Here, Proposed-i denotes the predictions for runoff made by our model, calculated for the i-th day ahead.</p> "> Figure 6 Cont.
<p>A depiction of the model’s application on regional rainfall–runoff modeling from 1 January 2013 to 1 January 2014 is presented. Here, Proposed-i denotes the predictions for runoff made by our model, calculated for the i-th day ahead.</p> ">
Abstract
:1. Introduction
2. Materials and Methods
- A.
- Dataset description
- B.
- Metrics
- C.
- Proposed method
- C.1
- Autoencoder and TLSTM Network Architecture
2.1. Autoencoder Structure
2.2. Training Methodology: Modified GAN Framework
- is the parameters of the generator;
- is the learning rate for the generator;
- D(Z) is the discriminator output for each element in the mini-batch Z;
- is the gradient of the value function V with respect to the generator parameters .
- C.2
- Prediction
- C.3
- Hyperparameter Optimization
- Initialization: We started by generating -dimensional parameter vectors, named to , forming a group referred to as P, which originated from a uniform [0,1] distribution. These numeric vectors were then associated with various hyperparameter values as detailed earlier (refer to Table 2), leading to 100 unique hyperparameter configurations. Following this, we adjusted 100 models, each using a different set of hyperparameters, and recorded the correlations between predicted and actual response variables. Each member of the population represents one of the N-dimensional vectors in P along with its corresponding set of encoded hyperparameters.
- Mutation: To optimize the model’s hyperparameters, we employed the random key method, an efficient technique for handling both continuous and categorical parameters within a unified framework. This approach facilitates structured exploration of the hyperparameter space using mutation and crossover strategies.
- represents the mutation factor for the i-th individual;
- are randomly selected parameter vectors from the population;
- [0, 1] is the mutation factor applied to the difference between the selected vectors.
- Crossover: To increase the diversity of hyperparameter combinations in the population, a crossover function combines the mutant vector, μ, with other unique vectors. Initially, an H-dimensional vector, named RN, filled with uniformly distributed random numbers within the range [0, 1], is created. The crossover frequency is controlled by the coefficient α, within the range [0,1], and we set α = 0.5. Subsequently, another N-dimensional vector (named CR) comprising Boolean values (True/False or 1/0) is formed as follows:
- i represents the crossover rate for the i-th individual;
- is the j-th feature of the i-th individual selected from the random individual r1;
- is the j-th feature of the i-th individual from the target parameter set.
- Selection: To determine whether the Challenger should replace in the population, the performance of both the Challenger and the current individual is compared based on the chosen evaluation metric (e.g., RMSE or validation accuracy). If the Challenger performs better than the current individual, it is integrated into the population; otherwise, the current individual is retained.
- D.
- Implementation Details
3. Results
4. Discussion
The Proposed Model in This Work Explores the Following Innovations
- Data scarcity/noise: GANs allow for noisy/scanty data to be handled well in the model.
- Overfitting: GAN reduces input complexity, mitigating overfitting and allowing for better generalization.
- Generalization: TLSTM improves generalization capability to various hydrological conditions due to the spatial attention of the model.
5. Conclusions
Supplementary Materials
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Smith, J.; Brown, A.; Wang, L. A Hybrid Approach Combining Conceptual Hydrological Models, Support Vector Machines, and Remote Sensing Data for Rainfall-Runoff Modeling. J. Hydrol. 2022, 12, 123–136. [Google Scholar]
- Anderson, J.; Clark, P. Hydrologically informed machine learning for rainfall-runoff modelling: Towards distributed modelling. Hydrol. Earth Syst. Sci. 2023, 18, 4373–4401. [Google Scholar]
- Salloom, O.; Kaynak, O.; He, W. A novel deep neural network architecture for real-time water demand forecasting. J. Hydrol. 2021, 599, 126353. [Google Scholar] [CrossRef]
- Zhang, J.; Chen, X.; Khan, A.; Zhang, Y.-K.; Kuang, X.; Liang, X.; Taccari, M.L.; Nuttall, J. Daily runoff forecasting by deep recursive neural network. J. Hydrol. 2021, 596, 126067. [Google Scholar]
- Solgi, R.; Loaiciga, H.A.; Kram, M. Long short-term memory neural network (LSTM-NN) for aquifer level time series forecasting using in-situ piezometric observations. J. Hydrol. 2021, 601, 126800. [Google Scholar] [CrossRef]
- Lees, T.; Buechel, M.; Anderson, B.; Slater, L.; Reece, S.; Coxon, G.; Dadson, S.J. Benchmarking data-driven rainfall-runoff models in Great Britain: A comparison of long short-term memory (LSTM)-based models with four lumped conceptual models. Hydrol. Earth Syst. Sci. 2021, 25, 5517–5534. [Google Scholar] [CrossRef]
- Lin, Y.; Wang, D.; Wang, G.; Qiu, J.; Long, K.; Du, Y.; Xie, H.; Wei, Z.; Shangguan, W.; Dai, Y. A hybrid deep learning algorithm and its application to streamflow prediction. J. Hydrol. 2021, 601, 126636. [Google Scholar] [CrossRef]
- Nourani, V.; Behfar, N. Multi-station runoff-sediment modeling using seasonal LSTM models. J. Hydrol. 2021, 601, 126672. [Google Scholar] [CrossRef]
- Kratzert, F.; Klotz, D.; Herrnegger, M.; Sampson, A.K.; Hochreiter, S.; Nearing, G.S. Toward improved predictions in ungauged basins: Exploiting the power of machine learning. Water Resour. Res. 2019, 55, 11344–11354. [Google Scholar] [CrossRef]
- Kao, I.-F.; Liou, J.-Y.; Lee, M.-H.; Chang, F.-J. Fusing stacked autoencoder and long short-term memory for regional multistep-ahead flood inundation forecasts. J. Hydrol. 2021, 598, 126371. [Google Scholar] [CrossRef]
- Yu, Y.; Si, X.; Hu, C.; Zhang, J. A review of recurrent neural networks: LSTM cells and network architectures. Neural Comput. 2019, 31, 1235–1270. [Google Scholar] [CrossRef]
- Hochreiter, S.; Schmidhuber, J. LSTM can solve hard long time lag problems. Adv. Neural Inf. Process. Syst. 1996, 9, 473–479. [Google Scholar]
- Wang, S.; Li, Z.; Ding, C.; Yuan, B.; Qiu, Q.; Wang, Y.; Liang, Y. C-LSTM: Enabling efficient LSTM using structured compression techniques on FPGAs. In Proceedings of the 2018 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays, Monterey, CA, USA, 25–27 February 2018; pp. 11–20. [Google Scholar]
- Momoi, M.; Kotsuki, S.; Kikuchi, R.; Watanabe, S.; Yamada, M.; Abe, S. Emulating Rainfall-Runoff-Inundation Model Using Deep Neural Network with Dimensionality Reduction. Artif. Intell. Earth Syst. 2023, 2, e220036. [Google Scholar] [CrossRef]
- Young, S.R.; Rose, D.C.; Karnowski, T.P.; Lim, S.-H.; Patton, R.M. Optimizing deep learning hyper-parameters through an evolutionary algorithm. In Proceedings of the Workshop on Machine Learning in High-Performance Computing Environments, Austin, TX, USA, 15 November 2015; pp. 1–5. [Google Scholar]
- Zhou, F.; Chen, Y.; Liu, J. Application of a New Hybrid Deep Learning Model That Considers Temporal and Feature Dependencies in Rainfall–Runoff Simulation. Remote Sens. 2023, 15, 1395. [Google Scholar] [CrossRef]
- Zhang, T.; Yang, Y.; Liu, Z. Using Remote Sensing Techniques to Improve Hydrological Predictions in a Rapidly Changing World. Hydrol. Sci. J. 2021, 15, 256–271. [Google Scholar] [CrossRef]
- Addor, N.; Newman, A.J.; Mizukami, N.; Clark, M.P. The CAMELS data set: Catchment attributes and meteorology for large-sample studies. Hydrol. Earth Syst. Sci. 2017, 21, 5293–5313. [Google Scholar] [CrossRef]
- Thornton, P.E.; Thornton, M.M.; Mayer, B.W.; Wei, Y.; Devarakonda, R.; Vose, R.S.; Cook, R.B. Daymet: Daily Surface Weather Data on a 1-Km Grid for North America, Version 2; Oak Ridge National Lab (ORNL): Oak Ridge, TN, USA, 2014. [Google Scholar]
- Maurer, E.P.; Wood, A.W.; Adam, J.C.; Lettenmaier, D.P.; Nijssen, B. A long-term hydrologically based dataset of land surface fluxes and states for the conterminous United States. J. Clim. 2002, 15, 3237–3251. [Google Scholar]
- Massari, C.; Brocca, L.; Tarpanelli, A.; Moramarco, T. Data Assimilation of Satellite Soil Moisture into Rainfall-Runoff Modelling: A Complex Recipe? Remote Sens. 2023, 15, 11403. [Google Scholar] [CrossRef]
- Khan, M.; Rehman, N.; Hussain, A. Rainfall-Runoff Modeling Using Machine Learning in the Ungauged Urban Watershed of Quetta Valley, Balochistan (Pakistan); Springer: Berlin/Heidelberg, Germany, 2023. [Google Scholar]
- Bean, J.C. Genetic algorithms and random keys for sequencing and optimization. ORSA J. Comput. 1994, 6, 154–160. [Google Scholar] [CrossRef]
- Shrestha, S.G.; Pradhanang, S.M. Performance of LSTM over SWAT in Rainfall-Runoff Modeling in a Small, Forested Watershed: A Case Study of Cork Brook, RI. Remote Sens. 2023, 15, 4194. [Google Scholar] [CrossRef]
- Goldberg, D.E.; Lobo, F.G.; Harik, G.R. The Compact Genetic Algorithm. IEEE Trans. Evol. Comput. 1999, 3, 287–297. [Google Scholar] [CrossRef]
- Graves, A. Generating sequences with recurrent neural networks. arXiv 2013, arXiv:1308.0850. [Google Scholar]
- Mizukami, N.; Rakovec, O.; Newman, A.J.; Clark, M.P.; Wood, A.W.; Gupta, H.V.; Kumar, R. On the choice of calibration metrics for “high-flow” estimation using hydrologic models. Hydrol. Earth Syst. Sci. 2019, 23, 2601–2614. [Google Scholar] [CrossRef]
- Yin, H.; Guo, Z.; Zhang, X.; Chen, J.; Zhang, Y. RR-Former: Rainfall-runoff modeling based on Transformer. J. Hydrol. 2022, 609, 127781. [Google Scholar] [CrossRef]
- Xiang, Z.; Yan, J.; Demir, I. A rainfall-runoff model with LSTM-based sequence-to-sequence learning. Water Resour. Res. 2020, 56, e2019WR025326. [Google Scholar] [CrossRef]
- Yin, H.; Zhang, X.; Wang, F.; Zhang, Y.; Xia, R.; Jin, J. Rainfall-runoff modeling using LSTM-based multi-state-vector sequence-to-sequence model. J. Hydrol. 2021, 598, 126378. [Google Scholar] [CrossRef]
- Seibert, J.; Vis, M.J.P.; Lewis, E.; van Meerveld, H.; Meerveld, I. Upper and lower benchmarks in hydrological modelling. Hydrol. Process. 2018, 32, 1120–1125. [Google Scholar] [CrossRef]
- Rakovec, O.; Mizukami, N.; Kumar, R.; Newman, A.J.; Thober, S.; Wood, A.W.; Clark, M.P.; Samaniego, L. Diagnostic evaluation of large-domain hydrologic models calibrated across the contiguous United States. J. Geophys. Res. Atmos. 2019, 124, 13991–14007. [Google Scholar] [CrossRef]
- Mizukami, N.; Clark, M.P.; Newman, A.J.; Wood, A.W.; Gutmann, E.D.; Nijssen, B.; Rakovec, O.; Samaniego, L. Towards seamless large-domain parameter estimation for hydrologic models. Water Resour. Res. 2017, 53, 8020–8040. [Google Scholar] [CrossRef]
- Newman, A.J.; Mizukami, N.; Clark, M.P.; Wood, A.W.; Nijssen, B.; Nearing, G. Benchmarking of a physically based hydrologic model. J. Hydrometeorol. 2017, 18, 2215–2225. [Google Scholar] [CrossRef]
- Feng, D.; Fang, K.; Shen, C. Enhancing Streamflow Forecast and Extracting Insights Using Long-Short Term Memory Networks with Data Integration at Continental Scales. Water Resour. Res. 2020, 56, e2019WR026793. [Google Scholar] [CrossRef]
Average Precipitation |
|
Average Temperature |
|
Soil Characteristics |
|
Land Use |
|
Topography |
|
Runoff Coefficient |
|
Hyperparameter | Parameter Space | Value Type |
---|---|---|
Batch size | [16–256] | Integer |
Learning rate | [0–1] | Continuous |
Epoch | [32–512] | Integer |
Activation function | [ReLU, Leaky ReLU, Linear, Tanh, Sigmoid] | Categorical |
Dropout rate | [0–1] | Continuous |
Number of layers in TLSTM | [1,2,4,8] | Integer |
Hidden size in TLSTM | [16–256] | Integer |
Model | One Day Ahead | Two Days Ahead | Four Days Ahead | Eight Days Ahead | ||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
NSE | RMSE | ATPE-2% | NSE | RMSE | ATPE-2% | NSE | RMSE | ATPE-2% | NSE | RMSE | ATPE-2% | |
LSTM-S2S [20] | 0.365 ± 0.120 | 1.720 ± 0.103 | 0.752 ± 0.124 | 0.320 ± 0.132 | 1.823 ± 0.122 | 0.765 ± 0.141 | 0.292 ± 0.185 | 1.940 ± 0.126 | 0.792 ± 0.256 | 0.102 ± 0.127 | 2.260 ± 0.026 | 0.792 ± 0.029 |
LSTM-MSV-S2S [21] | 0.469 ± 0.005 | 1.568 ± 0.126 | 0.647 ± 0.120 | 0.437 ± 0.009 | 1.742 ± 0.145 | 0.662 ± 0.248 | 0.402 ± 0.201 | 1.824 ± 0.194 | 0.574 ± 0.203 | 0.314 ± 0.126 | 1.974 ± 0.026 | 0.574 ± 0.128 |
RR-Former [19] | 0.552 ± 0.142 | 1.356 ± 0.152 | 0.425 ± 0.158 | 0.526 ± 0.174 | 1.426 ± 0.210 | 0.441 ± 0.269 | 0.482 ± 0.278 | 1.536 ± 0.214 | 0.512 ± 0.128 | 0.326 ± 0.206 | 1.674 ± 0.127 | 0.512 ± 0.268 |
Proposed | 0.712 ± 0.045 | 1.024 ± 0.006 | 0.236 ± 0.003 | 0.692 ± 0.016 | 1.114 ± 0.026 | 0.256 ± 0.126 | 0.681 ± 0.048 | 1.254 ± 0.026 | 0.320 ± 0.006 | 0.536 ± 0.002 | 1.378 ± 0.006 | 0.320 ± 0.103 |
Model | One Day Ahead | Two Days Ahead | Four Days Ahead | Eight Days Ahead | ||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
NSE | RMSE | ATPE-2% | NSE | RMSE | ATPE-2% | NSE | RMSE | ATPE-2% | NSE | RMSE | ATPE-2% | |
HBW (lower) [22] | 0.559 ± 0.125 | 1.438 ± 0.123 | 0.759 ± 0.143 | 0.536 ± 0.176 | 1.541 ± 0.190 | 0.774 ± 0.200 | 0.412 ± 0.146 | 1.710 ± 0.152 | 0.820 ± 0.175 | 0.452 ± 0.185 | 1.941 ± 0.152 | 0.842 ± 0.111 |
HBW (upper) [22] | 0.592 ± 0.245 | 1.412 ± 0.026 | 0.747 ± 0.079 | 0.585 ± 0.095 | 1.520 ± 0.057 | 0.768 ± 0.063 | 0.472 ± 0.103 | 1.662 ± 0.100 | 0.792 ± 0.120 | 0.493 ± 0.130 | 1.861 ± 0.120 | 0.803 ± 0.125 |
mHM (basin) [23] | 0.623 ± 0.026 | 1.385 ± 0.147 | 0.726 ± 0.176 | 0.604 ± 0.216 | 1.490 ± 0.241 | 0.742 ± 0.260 | 0.496 ± 0.289 | 1.610 ± 0.236 | 0.782 ± 0.259 | 0.502 ± 0.206 | 1.823 ± 0.220 | 0.793 ± 0.142 |
mHM (CONUS) [24] | 0.641 ± 0.126 | 1.341 ± 0.028 | 0.712 ± 0.037 | 0.639 ± 0.063 | 1.410 ± 0.089 | 0.735 ± 0.102 | 0.540 ± 0.006 | 1.563 ± 0.103 | 0.763 ± 0.174 | 0.508 ± 0.182 | 1.762 ± 0.163 | 0.785 ± 0.172 |
VIC (CONUS) [25] | 0.650 ± 0.223 | 1.320 ± 0.156 | 0.692 ± 0.215 | 0.642 ± 0.195 | 1.429 ± 0.205 | 0.720 ± 0.182 | 0.562 ± 0.120 | 1.526 ± 0.165 | 0.758 ± 0.185 | 0.521 ± 0.109 | 1.723 ± 0.140 | 0.780 ± 0.126 |
VIC (basin) [26] | 0.675 ± 0.026 | 1.226 ± 0.036 | 0.672 ± 0.071 | 0.656 ± 0.123 | 1.302 ± 0.153 | 0.710 ± 0.178 | 0.586 ± 0.196 | 1.485 ± 0.206 | 0.742 ± 0.196 | 0.553 ± 0.180 | 1.682 ± 0.195 | 0.772 ± 0.200 |
SAC-SMA [26] | 0.683 ± 0.124 | 1.114 ± 0.213 | 0.653 ± 0.242 | 0.676 ± 0.268 | 1.224 ± 0.274 | 0.692 ± 0.215 | 0.601 ± 0.247 | 1.426 ± 0.259 | 0.730 ± 0.236 | 0.501 ± 0.246 | 1.626 ± 0.259 | 0.762 ± 0.241 |
RR-Former [19] | 0.721 ± 0.026 | 1.102 ± 0.026 | 0.542 ± 0.144 | 0.706 ± 0.126 | 1.212 ± 0.204 | 0.650 ± 0.129 | 0.626 ± 0.103 | 1.352 ± 0.102 | 0.723 ± 0.130 | 0.525 ± 0.174 | 1.539 ± 0.182 | 0.751 ± 0.162 |
Proposed | 0.805 ± 0.026 | 0.982 ± 0.006 | 0.259 ± 0.120 | 0.792 ± 0.005 | 0.992 ± 0.014 | 0.282 ± 0.006 | 0.682 ± 0.026 | 1.250 ± 0.003 | 0.341 ± 0.015 | 0.546 ± 0.020 | 1.456 ± 0.103 | 0.410 ± 0.093 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Ghanati, B.; Serra-Sagristà, J. Integration of Generative-Adversarial-Network-Based Data Compaction and Spatial Attention Transductive Long Short-Term Memory for Improved Rainfall–Runoff Modeling. Remote Sens. 2024, 16, 3889. https://doi.org/10.3390/rs16203889
Ghanati B, Serra-Sagristà J. Integration of Generative-Adversarial-Network-Based Data Compaction and Spatial Attention Transductive Long Short-Term Memory for Improved Rainfall–Runoff Modeling. Remote Sensing. 2024; 16(20):3889. https://doi.org/10.3390/rs16203889
Chicago/Turabian StyleGhanati, Bahareh, and Joan Serra-Sagristà. 2024. "Integration of Generative-Adversarial-Network-Based Data Compaction and Spatial Attention Transductive Long Short-Term Memory for Improved Rainfall–Runoff Modeling" Remote Sensing 16, no. 20: 3889. https://doi.org/10.3390/rs16203889
APA StyleGhanati, B., & Serra-Sagristà, J. (2024). Integration of Generative-Adversarial-Network-Based Data Compaction and Spatial Attention Transductive Long Short-Term Memory for Improved Rainfall–Runoff Modeling. Remote Sensing, 16(20), 3889. https://doi.org/10.3390/rs16203889