Predictive Analytics of Air Temperature in Alaskan Permafrost Terrain Leveraging Two-Level Signal Decomposition and Deep Learning
<p>Data from the SNAP dataset [<a href="#B36-forecasting-06-00004" class="html-bibr">36</a>] from three field sites in Nome, Bethel, and Utqiagvik in Alaska. (<b>a</b>) Daily temperature time series over time with distinctions between freezing and thawing temperatures; (<b>b</b>) Map of the three field sites in Alaska.</p> "> Figure 2
<p>Multi-horizon forecasting of the temperature using VMD-WT-InceptionTime.</p> "> Figure 3
<p>Moving window schematic to construct training and testing sets. The testing set period spans between 21 May 2008 and 29 October 2015.</p> "> Figure 4
<p>Spearman correlation results between different inputs and the target output for every location. Different input lengths were considered for each input feature spanning nine weeks.</p> "> Figure 5
<p>Example of temperature sequences sequentially processed using VMD (<math display="inline"><semantics> <mrow> <mi>M</mi> <mo>=</mo> <mn>3</mn> </mrow> </semantics></math>) and WT (<math display="inline"><semantics> <mrow> <mi>I</mi> <mo>=</mo> <mn>3</mn> </mrow> </semantics></math>). (<b>a</b>) Raw temperature sequence; (<b>b</b>) Resulting VMD decomposed sequences from the temperature sequence; (<b>c</b>) Single-sided amplitude spectrum of the VMD decomposed sequences; (<b>d</b>) Resulting WT decomposed sequences from a single IMF; (<b>e</b>) Single-sided amplitude spectrum of WT decomposed sequences.</p> "> Figure 6
<p>Scatterplots comparing observed and forecast air temperatures in the three field sites using InceptionTime under four approaches: no decomposition, WT decomposition only, VMD only, and the proposed hybrid VMD-WT technique. (<b>a</b>) Nome; (<b>b</b>) Bethel; (<b>c</b>) Utqiagvik.</p> "> Figure 7
<p>Impact of different VMD decomposition levels on the average performance of the proposed hybrid model on the test set in terms of RMSE and <math display="inline"><semantics> <msup> <mi>R</mi> <mn>2</mn> </msup> </semantics></math>. (<b>a</b>) Nome; (<b>b</b>) Bethel; (<b>c</b>) Utqiagvik.</p> "> Figure 8
<p>Examples of air temperature forecasts using the optimized proposed forecasting technique (<math display="inline"><semantics> <mrow> <mi>M</mi> <mo>=</mo> <mn>30</mn> </mrow> </semantics></math> for Nome and Bethel and <math display="inline"><semantics> <mrow> <mi>M</mi> <mo>=</mo> <mn>39</mn> </mrow> </semantics></math> for Utqiavik) compared with actual measurements from the test set. Additional forecasts using the proposed technique under a sub-optimal decomposition level (<span class="html-italic">M</span> = 3), no decomposition using InceptionTime, and the historical means are shown for reference. All plots share the same vertical axis limits for comparison reasons. (<b>a</b>) randomly selected sequences. (<b>b</b>) randomly selected sequences. (<b>c</b>) the worst performance.</p> "> Figure 9
<p>Boxplot of per-horizon errors found using the optimized proposed technique under <math display="inline"><semantics> <mrow> <mi>M</mi> <mo>=</mo> <mn>30</mn> </mrow> </semantics></math> for Nome and Bethel and under <math display="inline"><semantics> <mrow> <mi>M</mi> <mo>=</mo> <mn>39</mn> </mrow> </semantics></math> for Utqiagvik on the testing sets.</p> "> Figure 10
<p>RMSE performance distribution of the optimized forecasting technique, segmented by the range of daily air temperature changes. The forecast and observation sequences were binned into intervals of <math display="inline"><semantics> <mrow> <mn>2</mn> <mi>K</mi> </mrow> </semantics></math> by computing the amplitude change between consecutive pairs of days (K/day). It is noteworthy that even under rapid air temperature fluctuations, the technique is capable of producing forecasts with low RMSE values at all three locations.</p> ">
Abstract
:1. Introduction
- Proposal of VMD-WT-InceptionTime for short-term multi-step air temperature forecasting. This hybrid technique is based on consecutive variational mode decomposition (VMD) and wavelet transform (WT) decompositions aiming to uncover hidden patterns and to reduce the complexity in past temperature and specific humidity sequences. These processed features are fed into a deep convolutional neural network forecasting model (InceptionTime).
- Comparison of the performance gains achieved through combined VMD and WT decompositions against using no decomposition or single decomposition techniques. To the best of the authors’ knowledge, the use of VMD and WT has not yet been investigated for the forecasting task at hand.
- Examination of the effects of VMD decomposition levels on the performance of the proposed forecasting technique and identification of the optimal level of decomposition.
- Assessment and validation of the technical experiments using multiple forecasting metrics and daily historical temperature data from three field sites in Alaska spanning 35+ years.
2. Background
2.1. Signal Decomposition Techniques
2.1.1. Variational Mode Decomposition
2.1.2. Wavelet Decomposition
2.2. InceptionTime
2.3. Dataset Description and Study Locations
- Nome is located on the coast of the Bering Sea on the Seward Peninsula at a latitude of 64.5 N and is in the discontinuous (50–90% permafrost coverage) permafrost zone [38]. Nome experiences a mean annual temperature of −2.2 , yet its position on the Bering Sea moderates the temperatures since the nearby large water mass provides thermal insulation from extreme air masses.
- Bethel is located on the Kuskokwim river in western Alaska 95 km inland from the mouth of the Kuskokwim River on the Bering Sea at a latitude of 60.8 N. The mean annual air temperature of Bethel is °C and it receives less temperature moderating influence from the Bering Sea than Nome due to its location nearly 100 km inland. The city is situated in the discontinuous permafrost zone [39] and is surrounded by many thermokarst lakes.
- Utqiagvik is the most northern study location at 71 N and is situated on the Arctic Ocean in the continuous permafrost zone (>90% permafrost coverage) [40] with a mean annual air temperature of .
3. Technical Implementation
3.1. Temperature Forecasting Using the Proposed Technique
- Temperature at two meters of height : that provides direct historical temperature data.
- Humidity at two meters of height : to provide direct historical humidity data.
- Precipitation : to provide direct historical precipitation data.
- Day of month: ranging from 1 to 31.
- Month of year: ranging from 1 to 12.
- Year ranging from 1979 to 2015.
- Season ranging from 1 to 4, with 1 referring to winter, 2 to spring, 3 to summer, and 4 to fall.
3.2. Performance Evaluation
3.2.1. Baseline Models
- Historical mean: A simple baseline model relying on the values from the previous month (i.e., the last 30 elements) to provide short-term forecasts. Other variations of this model were considered (e.g., same week averaged over the past three months, same month averaged over the past three years), but the proposed historical model proved the best one.
- RF: It is an ensemble classifier that uses multiple decision trees to obtain a better prediction performance. A bootstrap technique is used to train each tree from the set of training data [43].
- GBDT [44]: It is an iterative ensemble model of multiple decision trees. In each iteration, GBDT learns the decision trees by fitting the negative gradients (also known as residual errors). The output of the GBDT is the accumulation of the outputs of all its component decision trees.
- SVR [45]: It is a popular conventional machine learning model for regression. In this work, we employ SVR with the sigmoid kernel.
- ARIMA [46]: It is a well-known statistical model for forecasting. ARIMA is generally applicable to non-stationary time series data. Its difference transformation can effectively transform non-stationary data into stationary data. In this work, we employ the Seasonal ARIMA model.
- TST [47]: It is a recent deep neural network that handles long-term dependencies while tracking relationships in sequential input to learn context and, subsequently, meaning. This model was initially proposed in 2017 for translation tasks for natural language processing in [48], but has now become a state-of-the-art model for various tasks in that field. Multihead-self-attention is the core component of TST that makes it suitable for processing time series data. This mechanism identifies multiple dynamic contextual information (i.e., past values, future values) of every element in a sequence, with every attention head. In the recent literature, attention-based deep learning has been effectively employed for uni- and multivariate time series forecasting problems [49,50,51].
- XCM [52]: It is a recently developed convolutional neural network that efficiently captures information related to both the observed variables and the timing of events directly from the input data, enabling it to have more robust and generalized learning on smaller and larger datasets.
- LSTM/GRU: LSTM is an RNN that is capable of learning long-term dependencies, especially in sequence prediction problems. It does this by introducing three gates known as the input gate, the forget gate, and the output gate that cooperate to control the information flow [53]. GRU was later introduced as a simpler alternative to LSTM, having its gating signal reduced to two (i.e., an update gate and a reset gate) and eliminating the need for distinguishing between memory cells and hidden states [54].
- MiniRocket [55]: It is a recent and computationally efficient alternative to the original ROCKET model while still achieving high performance on time series classification tasks. MiniRocket incorporates a series of random convolutional kernels to transform the input data into a high-dimensional feature space before feeding them to the classification or regression layer.
3.2.2. Model Hyper-Parameter Tuning
3.2.3. Evaluation Metrics
4. Results
4.1. Input Correlation Analysis
4.2. Decomposition Analysis
4.3. Performance Benchmark
4.4. Impact of VMD Decomposition Level
5. Discussion
5.1. Preliminary Comparison with NOAA’s GFSv16
5.2. Computational Costs
5.3. Forecasting Using Different Combinations of Inputs
5.4. Forecasting the Air Temperatures under Rare Conditions
5.5. The Most Effective Use of the Findings
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Lawrence, D.M.; Slater, A.G. A projection of severe near-surface permafrost degradation during the 21st century. Geophys. Res. Lett. 2005, 32, L24401. [Google Scholar] [CrossRef]
- Shen, X.; Wang, J.; Li, Z.; Chen, D.; Gong, J. Research and operational development of numerical weather prediction in China. J. Meteorol. Res. 2020, 34, 675–698. [Google Scholar] [CrossRef]
- Thomas, G.; Mahfouf, J.F.; Montmerle, T. Toward a variational assimilation of polarimetric radar observations in a convective-scale numerical weather prediction (NWP) model. Atmos. Meas. Tech. 2020, 13, 2279–2298. [Google Scholar] [CrossRef]
- Shin, Y.; Yi, C. Statistical downscaling of urban-scale air temperatures using an analog model output statistics technique. Atmosphere 2019, 10, 427. [Google Scholar] [CrossRef]
- Cho, D.; Yoo, C.; Son, B.; Im, J.; Yoon, D.; Cha, D.H. A novel ensemble learning for post-processing of NWP Model’s next-day maximum air temperature forecast in summer using deep learning and statistical approaches. Weather Clim. Extrem. 2022, 35, 100410. [Google Scholar] [CrossRef]
- DelSole, T.; Zhao, M.; Dirmeyer, P.A.; Kirtman, B.P. Empirical correction of a coupled land–atmosphere model. Mon. Weather Rev. 2008, 136, 4063–4076. [Google Scholar] [CrossRef]
- Hou, Z.; Li, J.; Wang, L.; Zhang, Y.; Liu, T. Improving the forecast accuracy of ECMWF 2-m air temperature using a historical dataset. Atmos. Res. 2022, 273, 106177. [Google Scholar] [CrossRef]
- Chen, K.; Wang, P.; Yang, X.; Zhang, N.; Wang, D. A model output deep learning method for grid temperature forecasts in Tianjin area. Appl. Sci. 2020, 10, 5808. [Google Scholar] [CrossRef]
- Schultz, M.G.; Betancourt, C.; Gong, B.; Kleinert, F.; Langguth, M.; Leufen, L.H.; Mozaffari, A.; Stadtler, S. Can deep learning beat numerical weather prediction? Philos. Trans. R. Soc. A Math. Phys. Eng. Sci. 2021, 379, 20200097. [Google Scholar] [CrossRef]
- Ayad, S. Modeling and forecasting air temperature in Tetouan (Morocco) using SARIMA model. J. Earth Sci. Geotech. Eng. 2022, 12, 1–13. [Google Scholar] [CrossRef]
- Curceac, S.; Ternynck, C.; Ouarda, T.B.; Chebana, F.; Niang, S.D. Short-term air temperature forecasting using Nonparametric Functional Data Analysis and SARMA models. Environ. Model. Softw. 2019, 111, 394–408. [Google Scholar] [CrossRef]
- Abhishek, K.; Singh, M.; Ghosh, S.; Anand, A. Weather forecasting model using artificial neural network. Procedia Technol. 2012, 4, 311–318. [Google Scholar] [CrossRef]
- Ustaoglu, B.; Cigizoglu, H.; Karaca, M. Forecast of daily mean, maximum and minimum temperature time series by three artificial neural network methods. Meteorol. Appl. J. Forecast. Pract. Appl. Train. Tech. Model. 2008, 15, 431–445. [Google Scholar] [CrossRef]
- Afzali, M.; Afzali, A.; Zahedi, G. The potential of artificial neural network technique in daily and monthly ambient air temperature prediction. Int. J. Environ. Sci. Dev. 2012, 3, 33. [Google Scholar] [CrossRef]
- Fahimi Nezhad, E.; Fallah Ghalhari, G.; Bayatani, F. Forecasting maximum seasonal temperature using artificial neural networks “Tehran case study”. Asia-Pac. J. Atmos. Sci. 2019, 55, 145–153. [Google Scholar] [CrossRef]
- Alomar, M.K.; Khaleel, F.; Aljumaily, M.M.; Masood, A.; Razali, S.F.M.; AlSaadi, M.A.; Al-Ansari, N.; Hameed, M.M. Data-driven models for atmospheric air temperature forecasting at a continental climate region. PLoS ONE 2022, 17, e0277079. [Google Scholar] [CrossRef]
- Roy, D.S. Forecasting the air temperature at a weather station using deep neural networks. Procedia Comput. Sci. 2020, 178, 38–46. [Google Scholar] [CrossRef]
- Nandi, A.; De, A.; Mallick, A.; Middya, A.I.; Roy, S. Attention based long-term air temperature forecasting network: ALTF Net. Knowl. Based Syst. 2022, 252, 109442. [Google Scholar] [CrossRef]
- Huang, N.E.; Shen, Z.; Long, S.R.; Wu, M.C.; Shih, H.H.; Zheng, Q.; Yen, N.C.; Tung, C.C.; Liu, H.H. The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis. Proc. R. Soc. Lond. Ser. A Math. Phys. Eng. Sci. 1998, 454, 903–995. [Google Scholar] [CrossRef]
- Wu, Z.; Huang, N.E. Ensemble empirical mode decomposition: A noise-assisted data analysis method. Adv. Adapt. Data Anal. 2009, 1, 1–41. [Google Scholar] [CrossRef]
- Dragomiretskiy, K.; Zosso, D. Variational Mode Decomposition. IEEE Trans. Signal Process. 2014, 62, 531–544. [Google Scholar] [CrossRef]
- Kulaglic, A.; Üstündağ, B.B. Stock price forecast using wavelet transformations in multiple time windows and neural networks. In Proceedings of the 2018 3rd IEEE International Conference on Computer Science and Engineering (UBMK), Sarajevo, Bosnia and Herzegovina, 20–23 September 2018; pp. 518–521. [Google Scholar]
- Peng, L.; Chen, K.; Li, N. Predicting stock movements: Using multiresolution wavelet reconstruction and deep learning in neural networks. Information 2021, 12, 388. [Google Scholar] [CrossRef]
- Ali, M.; Khan, D.M.; Alshanbari, H.M.; El-Bagoury, A.A.A.H. Prediction of complex stock market data using an improved hybrid emd-lstm model. Appl. Sci. 2023, 13, 1429. [Google Scholar] [CrossRef]
- Xing, Y.; Yue, J.; Chen, C.; Cong, K.; Zhu, S.; Bian, Y. Dynamic displacement forecasting of dashuitian landslide in China using variational mode decomposition and stack long short-term memory network. Appl. Sci. 2019, 9, 2951. [Google Scholar] [CrossRef]
- Li, G.; Ma, X.; Yang, H. A hybrid model for monthly precipitation time series forecasting based on variational mode decomposition with extreme learning machine. Information 2018, 9, 177. [Google Scholar] [CrossRef]
- Duarte, D.; Walshaw, C.; Ramesh, N. A comparison of time-series predictions for healthcare emergency department indicators and the impact of COVID-19. Appl. Sci. 2021, 11, 3561. [Google Scholar] [CrossRef]
- Chen, P.; Deng, Y.; Zhang, X.; Ma, L.; Yan, Y.; Wu, Y.; Li, C. Degradation trend prediction of pumped storage unit based on MIC-LGBM and VMD-GRU combined model. Energies 2022, 15, 605. [Google Scholar] [CrossRef]
- Akin, M. Comparison of wavelet transform and FFT methods in the analysis of EEG signals. J. Med Syst. 2002, 26, 241–247. [Google Scholar] [CrossRef]
- Boyd, S.; Parikh, N.; Chu, E.; Peleato, B.; Eckstein, J. Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers. Found. Trends Mach. Learn. 2011, 3, 1–122. [Google Scholar] [CrossRef]
- Addison, P.S. Wavelet transforms and the ECG: A review. Physiol. Meas. 2005, 26, R155. [Google Scholar] [CrossRef]
- Ismail Fawaz, H.; Lucas, B.; Forestier, G.; Pelletier, C.; Schmidt, D.F.; Weber, J.; Webb, G.I.; Idoumghar, L.; Muller, P.A.; Petitjean, F. InceptionTime: Finding AlexNet for time series classification. Data Min. Knowl. Discov. 2020, 34, 1936–1962. [Google Scholar] [CrossRef]
- Zhou, K.; Wang, W.; Hu, T.; Deng, K. Time series forecasting and classification models based on recurrent with attention mechanism and generative adversarial networks. Sensors 2020, 20, 7211. [Google Scholar] [CrossRef] [PubMed]
- Guo, F.; Li, W.; Jiang, P.; Chen, F.; Liu, Y. Deep learning approach for damage classification based on acoustic emission data in composite materials. Materials 2022, 15, 4270. [Google Scholar] [CrossRef]
- Wang, P.; Wu, J.; Wei, Y.; Li, T. CEEMD-MultiRocket: Integrating CEEMD with Improved MultiRocket for Time Series Classification. Electronics 2023, 12, 1188. [Google Scholar] [CrossRef]
- Bieniek, P.A. Downscaled Climate Data for Alaska; Technical Report; International Arctic Research Center, University of Alaska Fairbanks: Fairbanks, AK, USA, 2019; Available online: https://registry.opendata.aws/wrf-alaska-snap (accessed on 26 June 2023).
- Mittal, A.K. Alaska Native Villages: Limited Progress Has Been Made on Relocating Villages Threatened by Flooding and Erosion; DIANE Publishing: Collingdale, PA, USA, 2009. [Google Scholar]
- Debolskiy, M.; Nicolsky, D.; Hock, R.; Romanovsky, V. Modeling present and future permafrost distribution at the Seward Peninsula, Alaska. J. Geophys. Res. Earth Surf. 2020, 125, e2019JF005355. [Google Scholar] [CrossRef]
- Zhao, Y.; Yang, Z.J. Historical and long-term climate trends in warm permafrost regions: A case study of Bethel, AK. Cold Reg. Sci. Technol. 2022, 204, 103677. [Google Scholar] [CrossRef]
- Brewer, M.C. Some results of geothermal investigations of permafrost in northern Alaska. Eos Trans. Am. Geophys. Union 1958, 39, 19–26. [Google Scholar]
- SAID, S.E.; DICKEY, D.A. Testing for unit roots in autoregressive-moving average models of unknown order. Biometrika 1984, 71, 599–607. [Google Scholar] [CrossRef]
- Kwiatkowski, D.; Phillips, P.C.; Schmidt, P.; Shin, Y. Testing the null hypothesis of stationarity against the alternative of a unit root: How sure are we that economic time series have a unit root? J. Econom. 1992, 54, 159–178. [Google Scholar] [CrossRef]
- Ho, T.K. Random decision forests. In Proceedings of the 3rd IEEE International Conference on Document Analysis and Recognition, Montreal, QC, Canada, 14–16 August 1995; Volume 1, pp. 278–282. [Google Scholar]
- Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
- Drucker, H.; Burges, C.J.; Kaufman, L.; Smola, A.; Vapnik, V. Support vector regression machines. In Proceedings of the NIPS’96: Proceedings of the 9th International Conference on Neural Information Processing Systems, Denver, CO, USA, 2–5 December 1996. [Google Scholar]
- Newbold, P. ARIMA model building and the time series analysis approach to forecasting. J. Forecast. 1983, 2, 23–35. [Google Scholar] [CrossRef]
- Zerveas, G.; Jayaraman, S.; Patel, D.; Bhamidipaty, A.; Eickhoff, C. A Transformer-Based Framework for Multivariate Time Series Representation Learning. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining (KDD ’21), New York, NY, USA, 14–18 August 2021; pp. 2114–2124. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
- Lara-Benítez, P.; Gallego-Ledesma, L.; Carranza-García, M.; Luna-Romera, J.M. Evaluation of the Transformer Architecture for Univariate Time Series Forecasting. In Proceedings of the Conference of the Spanish Association for Artificial Intelligence; Springer: Cham, Switzerland, 2021; pp. 106–115. [Google Scholar]
- Hu, Y.; Xiao, F. Network self attention for forecasting time series. Appl. Soft Comput. 2022, 124, 109092. [Google Scholar] [CrossRef]
- He, X.; Shi, S.; Geng, X.; Xu, L. Dynamic Co-Attention Networks for multi-horizon forecasting in multivariate time series. Future Gener. Comput. Syst. 2022, 135, 72–84. [Google Scholar] [CrossRef]
- Fauvel, K.; Lin, T.; Masson, V.; Fromont, É.; Termier, A. Xcm: An explainable convolutional neural network for multivariate time series classification. Mathematics 2021, 9, 3137. [Google Scholar] [CrossRef]
- Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
- Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
- Dempster, A.; Schmidt, D.F.; Webb, G.I. MiniRocket: A Very Fast (Almost) Deterministic Transform for Time Series Classification. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 14–18 August 2021; pp. 248–257. [Google Scholar]
- Akiba, T.; Sano, S.; Yanase, T.; Ohta, T.; Koyama, M. Optuna: A Next-generation Hyperparameter Optimization Framework. In Proceedings of the 25rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 2623–2631. [Google Scholar]
- Pincus, S.M.; Goldberger, A.L. Physiological time-series analysis: What does regularity quantify? Am. J. Physiol. Heart Circ. Physiol. 1994, 266, H1643–H1656. [Google Scholar] [CrossRef]
- Levitt, J.; Row, M.; Tallapragada, V.; Adimi, F.; Yang, F.; Treadon, R. GFSv16 Verification. National Oceanic and Atmospheric Administration; 2021. Available online: https://emc.ncep.noaa.gov/users/verification/global/gfs/gfsv16/grid2obs/rmse (accessed on 1 January 2024).
- Parker, C.L.; Mooney, P.A.; Webster, M.A.; Boisvert, L.N. The influence of recent and future climate change on Spring Arctic Cyclones. Nat. Commun. 2022, 13, 6514. [Google Scholar] [CrossRef]
- Arcomano, T.; Szunyogh, I.; Wikner, A.; Pathak, J.; Hunt, B.R.; Ott, E. A hybrid approach to atmospheric modeling that combines machine learning with a physics-based numerical model. J. Adv. Model. Earth Syst. 2022, 14, e2021MS002712. [Google Scholar] [CrossRef]
Parameter | Value |
---|---|
Dataset | SNAP [36] |
Temp. resolution | Daily |
Spatial resolution | 20 km |
Time range | 2 January 1979–29 October 2015 |
Number of samples | 262 × 262 × 13,450 |
Features [Units] | Air temperature (T2) [K] |
Specific humidity (Q2) [kg/kg] | |
Precipitation (PCPT) [mm] |
Feature | Location | Stats | ADF | KPSS | ||||||
---|---|---|---|---|---|---|---|---|---|---|
Mean | StD | Entropy | FF | FF | Test Stat | cValue | Test Stat | cValue | ||
T2 | Nome | 270.27 | 12.13 | 3.43 | 32.29% | 1.29% | −0.080 | −2.643 | 0.269 | 0.216 |
Bethel | 273.09 | 12.22 | 3.20 | 33.44% | 2.39% | −0.073 | −2.643 | 0.259 | 0.216 | |
Utqiagvik | 262.34 | 13.75 | 3.49 | 32.21% | 1.16% | −0.054 | −2.643 | 0.251 | 0.216 | |
Q2 | Nome | 0.003 | 0.002 | 2.929 | - | - | −0.637 | −2.643 | 0.253 | 0.216 |
Bethel | 0.004 | 0.002 | 3.128 | - | - | −0.636 | −2.643 | 0.235 | 0.216 | |
Utqiagvik | 0.002 | 0.002 | 2.697 | - | - | −0.694 | −2.643 | 0.248 | 0.216 | |
PCPT | Nome | 1.486 | 3.427 | 0.609 | - | - | −3.576 | −2.643 | 0.122 | 0.216 |
Bethel | 1.841 | 3.309 | 0.982 | - | - | −3.474 | −2.643 | 0.108 | 0.216 | |
Utqiagvik | 0.599 | 1.453 | 0.773 | - | - | −3.680 | −2.643 | 0.112 | 0.216 |
Hyper-Parameter | Nome | Bethel | Utqiagvik | ||
---|---|---|---|---|---|
Inception modules | Module number | 6 | 6 | 6 | |
Bottleneck layer | filters | 64 | 48 | 16 | |
kernel sizes | 1 | 1 | 1 | ||
stride | 1 | 1 | 1 | ||
Convolution layers | number of layers | 3 | 3 | 3 | |
filters | 64,64,64 | 48,48,48 | 16,16,16 | ||
kernel sizes | 39,19,9 | 19,9,5 | 79,39,19 | ||
stride | 1,1,1 | 1,1,1 | 1,1,1 | ||
padding | 19,9,4 | 9,4,2 | 39,19,9 | ||
Max-pooling layer | kernel size | 3 | 3 | 3 | |
stride | 1 | 1 | 1 | ||
padding | 1 | 1 | 1 | ||
Convolution layer | filters | 64 | 48 | 16 | |
kernel size | 1 | 1 | 1 | ||
stride | 1 | 1 | 1 | ||
Concatenation layer | dimension | 1 | 1 | 1 | |
Batch normalization | features | 256 | 192 | 64 | |
momentum | 0.1 | 0.1 | 0.1 | ||
Dropout | p | 32.3E-3 | 19.56E-5 | 7.6E-3 | |
Activation layer | function | ReLu | ReLu | ReLu | |
Optimizer | Adam | Adam | Adam | ||
Loss function | Flattened MSE | Flattened MSE | Flattened MSE | ||
Outputs | 7 | 7 | 7 | ||
Epochs | 30 | 30 | 30 | ||
Batch size | 1024 | 1024 | 1024 | ||
Learning rate | 30 | 30 | 30 |
Input | Seq. | T2 | Q2 | ||||
---|---|---|---|---|---|---|---|
Nome | Bethel | Utqiagvik | Nome | Bethel | Utqiagvik | ||
Raw TS | - | 3.434 | 3.202 | 3.498 | 2.929 | 3.128 | 2.697 |
VMD(TS) | IMF1 | 0.873 | 1.009 | 1.004 | 0.99 | 1.109 | 1.092 |
IMF2 | 0.973 | 0.964 | 0.987 | 1.001 | 0.999 | 1.005 | |
IMF3 | 0.652 | 0.653 | 0.651 | 0.667 | 0.676 | 0.654 | |
IMF4 | 0.449 | 0.443 | 0.435 | 0.437 | 0.42 | 0.431 | |
IMF5 | 0.289 | 0.266 | 0.278 | 0.282 | 0.279 | 0.3 | |
IMF6 | 0.203 | 0.212 | 0.189 | 0.205 | 0.214 | 0.222 | |
IMF7 | 0.174 | 0.173 | 0.171 | 0.174 | 0.18 | 0.186 | |
IMF8 | 0.163 | 0.18 | 0.169 | 0.16 | 0.157 | 0.171 | |
IMF9 | 0.168 | 0.163 | 0.177 | 0.179 | 0.181 | 0.174 | |
IMF10 | 0.164 | 0.164 | 0.188 | 0.19 | 0.171 | 0.189 | |
IMF11 | 0.165 | 0.17 | 0.186 | 0.175 | 0.175 | 0.175 | |
IMF12 | 0.16 | 0.157 | 0.163 | 0.17 | 0.193 | 0.175 | |
IMF13 | 0.103 | 0.112 | 0.095 | 0.11 | 0.116 | 0.11 | |
IMF14 | 0.024 | 0.025 | 0.02 | 0.027 | 0.023 | 0.021 | |
IMF15 | 0.053 | 0.052 | 0.055 | 0.018 | 0.021 | 0.018 | |
res. | 0.283 | 0.271 | 0.279 | 0.267 | 0.286 | 0.275 | |
WT(IMFx) | D1 | 0.446 | 0.4515 | 0.4575 | 0.4657 | 0.4726 | 0.4701 |
D2 | 0.5371 | 0.5344 | 0.5446 | 0.5488 | 0.5579 | 0.555 | |
D3 | 0.5538 | 0.5569 | 0.5624 | 0.5797 | 0.5778 | 0.5698 | |
A4 | 0.5612 | 0.5648 | 0.565 | 0.5797 | 0.5854 | 0.5834 |
Case | Model | Nome | Bethel | Utqiagvik | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
RMSE | MAPE | DA | RMSE | MAPE | DA | RMSE | MAPE | DA | |||||
No decomposition | InceptionTime | 5.528 | 1.541 | 0.790 | 63.121 | 6.574 | 1.749 | 0.701 | 62.306 | 4.817 | 1.356 | 0.866 | 63.4 |
Single decomposition | WT + InceptionTime | 5.286 | 1.486 | 0.808 | 64.192 | 5.6 | 1.513 | 0.783 | 64.699 | 4.6 | 1.307 | 0.878 | 63.9 |
VMD + InceptionTime | 0.798 | 0.214 | 0.995 | 90.648 | 0.809 | 0.205 | 0.995 | 91.261 | 0.739 | 0.202 | 0.996 | 91.2 | |
Proposed technique | VMD + WT + InceptionTime | 0.751 | 0.197 | 0.996 | 91.581 | 0.795 | 0.198 | 0.996 | 91.709 | 0.673 | 0.176 | 0.997 | 92.2 |
Model | Nome | Bethel | Utqiagvik | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
RMSE | MAPE | DA | RMSE | MAPE | DA | RMSE | MAPE | DA | ||||
Historical Mean | 7.334 | 2.063 | 0.63 | 70.347 | 7.868 | 2.098 | 0.571 | 69.302 | 6.343 | 1.826 | 0.768 | 73.368 |
SVR-Sigmoid | 12.083 | 3.79 | −0.004 | 55.39 | 12.038 | 3.68 | −0.004 | 54.122 | 13.25 | 4.487 | −0.014 | 52.853 |
ExtraTrees | 13.417 | 4.092 | −0.238 | 55.129 | 16.88 | 4.845 | −0.973 | 56.919 | 17.057 | 5.168 | −0.68 | 55.017 |
GBDT | 17.212 | 5.148 | −1.037 | 58.896 | 16.907 | 4.855 | −0.98 | 56.994 | 19.155 | 5.889 | −1.119 | 58.448 |
RF | 14.993 | 4.485 | −0.546 | 55.054 | 16.963 | 4.876 | −0.993 | 56.882 | 18.434 | 5.613 | −0.962 | 57.367 |
SARIMAX | 12.924 | 4.007 | −0.149 | 54.569 | 11.928 | 3.638 | 0.015 | 55.651 | 13.142 | 4.396 | 0.003 | 53.115 |
Model | Nome | Bethel | Utqiagvik | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
RMSE | MAPE | DA | RMSE | MAPE | DA | RMSE | MAPE | DA | ||||
TST | 5.837 | 1.682 | 0.766 | 61.9 | 6.019 | 1.691 | 0.749 | 62.67 | 5.054 | 1.466 | 0.852 | 61.9 |
XCM | 5.664 | 1.625 | 0.78 | 61.8 | 6.266 | 1.764 | 0.728 | 62.51 | 5 | 1.462 | 0.856 | 62.1 |
LSTM | 6.061 | 1.757 | 0.747 | 60 | 6.539 | 1.857 | 0.704 | 61.38 | 5.656 | 1.671 | 0.815 | 59.8 |
GRU | 6.112 | 1.761 | 0.743 | 60.2 | 6.285 | 1.773 | 0.727 | 62.41 | 5.42 | 1.595 | 0.83 | 60.1 |
MiniRocket | 7.566 | 2.25 | 0.607 | 57 | 7.641 | 2.243 | 0.596 | 58.46 | 6.383 | 1.924 | 0.765 | 57.1 |
Proposed technique | 0.751 | 0.197 | 0.997 | 91.6 | 0.795 | 0.198 | 0.996 | 91.71 | 0.673 | 0.176 | 0.997 | 92.2 |
Location | # | Error | Horiz. | Target Temperature Sequence | ||||
---|---|---|---|---|---|---|---|---|
Value | Range | Min. | Max. | Month | ||||
Nome | 1st | −7.19 | 7 | 283.02 | 7.48 | 283.02 | 290.5 | June |
2nd | 5.16 | 6 | 288.34 | 9.37 | 280.01 | 289.38 | June | |
3rd | 4.66 | 4 | 290.5 | 7.48 | 283.02 | 290.5 | June | |
Bethel | 1st | −5.3 | 7 | 268.33 | 15.22 | 257.69 | 272.91 | March |
2nd | −4.26 | 1 | 265.19 | 11.57 | 258.25 | 269.82 | February | |
3rd | −3.81 | 4 | 250.63 | 22.88 | 247.64 | 270.52 | January | |
Utqiagvik | 1st | 6.11 | 7 | 284.59 | 11.23 | 275.44 | 286.67 | July |
2nd | 5.3 | 6 | 265.03 | 20.32 | 244.72 | 265.03 | November | |
3rd | 3.87 | 3 | 283.82 | 11.23 | 275.44 | 286.67 | July |
Location | # | Error | Horiz. | Target Temperature Sequence | ||||
---|---|---|---|---|---|---|---|---|
Value | Range | Min. | Max. | Month | ||||
Nome | 1st | −0.0003 | 4 | 284.31 | 5.9 | 282.81 | 288.71 | May |
2nd | −0.0002 | 7 | 276.5 | 5.16 | 276.5 | 281.66 | April | |
3rd | 0.0001 | 6 | 281.33 | 3.83 | 279.21 | 283.04 | April | |
Bethel | 1st | 0.0004 | 6 | 258.25 | 12.14 | 257.68 | 269.82 | February |
2nd | −0.0001 | 4 | 270.62 | 8.14 | 268.5 | 276.64 | February | |
3rd | 0.0001 | 7 | 282.33 | 6.52 | 282.33 | 288.85 | July | |
Utqiagvik | 1st | −0.0001 | 7 | 285.03 | 8.86 | 276.17 | 285.03 | May |
2nd | −0.0001 | 5 | 266.95 | 5.96 | 265.26 | 271.21 | September | |
3rd | 0.0001 | 3 | 232.97 | 11.49 | 232.97 | 244.46 | December |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Ahajjam, A.; Putkonen, J.; Chukwuemeka, E.; Chance, R.; Pasch, T.J. Predictive Analytics of Air Temperature in Alaskan Permafrost Terrain Leveraging Two-Level Signal Decomposition and Deep Learning. Forecasting 2024, 6, 55-80. https://doi.org/10.3390/forecast6010004
Ahajjam A, Putkonen J, Chukwuemeka E, Chance R, Pasch TJ. Predictive Analytics of Air Temperature in Alaskan Permafrost Terrain Leveraging Two-Level Signal Decomposition and Deep Learning. Forecasting. 2024; 6(1):55-80. https://doi.org/10.3390/forecast6010004
Chicago/Turabian StyleAhajjam, Aymane, Jaakko Putkonen, Emmanuel Chukwuemeka, Robert Chance, and Timothy J. Pasch. 2024. "Predictive Analytics of Air Temperature in Alaskan Permafrost Terrain Leveraging Two-Level Signal Decomposition and Deep Learning" Forecasting 6, no. 1: 55-80. https://doi.org/10.3390/forecast6010004
APA StyleAhajjam, A., Putkonen, J., Chukwuemeka, E., Chance, R., & Pasch, T. J. (2024). Predictive Analytics of Air Temperature in Alaskan Permafrost Terrain Leveraging Two-Level Signal Decomposition and Deep Learning. Forecasting, 6(1), 55-80. https://doi.org/10.3390/forecast6010004