Intelligent Financial Forecasting with Granger Causality and Correlation Analysis Using Bayesian Optimization and Long Short-Term Memory
<p>The proposed system architecture. The architecture combines Bayesian optimization and LSTM networks to enhance the accuracy and robustness of financial predictions, offering improved insights and performance in volatile markets.</p> "> Figure 2
<p>Correlation matrix. Granger causality analysis showing significant predictive relationships among OHLC prices and Volume, highlighting the dynamic interactions and justifying the inclusion of these variables in the model.</p> "> Figure 3
<p>Granger causality analysis. The Augmented Dickey–Fuller (ADF) test indicates that the <span class="html-italic">Close</span> price series is non-stationary, with a high <span class="html-italic">p</span>-value of 0.6557, suggesting varying statistical properties over time. In contrast, the <span class="html-italic">Volume</span> series is stationary, with a very low <span class="html-italic">p</span>-value of 0.0006, confirming consistent statistical behavior. <a href="#electronics-13-04408-t003" class="html-table">Table 3</a> provides detailed results.</p> "> Figure 4
<p>Predicted vs. actual stock prices. It compares actual and predicted stock prices, showcasing the hybrid model’s ability to capture data patterns accurately and demonstrating the effectiveness of combining Bayesian optimization with LSTM networks for enhanced convergence and minimized prediction errors in financial forecasting.</p> "> Figure 5
<p>Training and validation MAEs. It illustrates the Mean Absolute Error (MAE) trends over training epochs, comparing the model’s performance on training and validation datasets. Initially, the training MAE is high, decreasing as the model learns, indicating improved accuracy. The validation MAE reflects how well the model generalizes to new data, with a stable or slightly increasing trend suggesting overfitting. This comparison helps assess the model’s effectiveness and informs training adjustments to optimize performance on unseen data.</p> "> Figure 6
<p>Residual plot. It displays the residual plot, showing the differences between observed and predicted values. Ideally, residuals should be randomly scattered around y = 0, indicating unbiased predictions. In this case, most data points align closely along the y = 0 line, suggesting that the model accurately captures the data patterns. The lack of visible trends in the residuals confirms the model’s robustness and reliability in making accurate predictions.</p> "> Figure 7
<p>Training and validation losses. It shows the training and validation loss curves across epochs, illustrating the model’s learning progress, where decreasing and stabilizing losses indicate effective learning and good generalization, while divergence suggests overfitting.</p> "> Figure 8
<p>Plots comparing BO-LSTM and LSTM-only.</p> ">
Abstract
:1. Introduction
2. Theoretical Background
3. Materials and Methods
3.1. Materials
3.1.1. System Architecture
3.1.2. Dataset
3.1.3. Data Preprocessing
Algorithm 1: Data Preprocessing |
Load the dataset: Input: Path to the stock data file (china_stock_data.csv) Output: data Check for missing values: If missing values are found, fill them using the forward fill method. Convert the ‘Date’ column to a date–time format. Calculate daily price changes: Compute the difference between the ‘Close’ and ‘Open’ prices and store as ‘Price Change’. Calculate the daily percentage change: \text{Price Change %} = \frac{\text{Close} − \text{Open}}{\text{Open}} \times 100 Normalize feature values: Use Min–Max Scaling to transform the following columns: ‘Open’, ‘High’, Low’, ‘Close’, ‘Volume’. Prepare training and test datasets: Define features (X) by excluding the ‘Date’ and ‘Close’ columns. Define the target variable (y) as the ‘Close’ price. Split the data into training and test sets (X_train, X_test, y_train, y_test) using an 80–20 ratio. Feature selection: Apply SelectKBest to choose the top 5 features based on their f_regression scores from the training data. Output: X_train_selected (transformed training features) |
3.2. Method
3.2.1. Training and Fine Tuning
3.2.2. Model Training with BO
3.2.3. Bayesian Optimization
- Objective Function: Let f(x) represent the objective function that encapsulates the performance metric of the Long Short-Term Memory (LSTM) model given the hyperparameter configuration x. This function serves as the basis for guiding the optimization process, where the aim is to find the hyperparameter values that maximize model performance.
- Surrogate Model (Gaussian Process): The surrogate model is a fundamental component of Bayesian optimization, allowing for the approximation of the objective function. In this research, a Gaussian Process (GP) was used as the surrogate model, which can be mathematically represented asThe Gaussian Process provides a distribution over functions rather than a single deterministic function. This probabilistic approach was employed to express uncertainty about the objective function, enabling effective exploration and exploitation during the optimization process.
- Acquisition Function: To decide where to sample next, we employ an acquisition function, specifically the Expected Improvement (EI), which is defined as
- Next Hyperparameters: The next set of hyperparameters to explore is determined by maximizing the acquisition function.
- Update Surrogate Model: After evaluating the objective function at the newly selected hyperparameter configuration , the surrogate model is updated to reflect this new information. The updated surrogate model can be expressed as
3.2.4. LSTM Model
- LSTM Output: Let LSTM(x) be the output of the LSTM model with the hyperparameter denoted by x.
- Training loss: Let the training loss, L(x), be the objective to minimize during training.
- Hyperparameters: Let x represent a set of hyperparameters for the LSTM model.
3.2.5. Fusion of Bayesian Optimization and LSTM
- Optimization Process:
- i.
- Integrate Bayesian optimization with the LSTM training process.
- ii.
- At each iteration of Bayesian optimization, select using the Expected Improvement acquisition function.
- iii.
- Train the LSTM model with hyperparameters .
- Overall Objective:Minimize the final performance metric of the LSTM model by tuning hyperparameters through Bayesian optimization
- Update Process:Update the surrogate model and LSTM model iteratively based on observed performance, adjusting the mean and variance of the GP predictive distribution.
Algorithm 2: Model Training with Bayesian Optimization |
1. Load the dataset from the CSV file. |
2. Preprocess the dataset: |
a. Drop missing values; b. Convert the‘Date’ column to date–time format; c. Set ‘Date’ as the index; d. Extract ‘Close’ prices. |
3. Normalize the dataset using MinMaxScaler. |
4. Create sequences and labels: |
a. Define the sequence length; b. For each time step in the dataset: |
i. Append sequence of ‘sequence_length’ to sequences list; ii. Append the next value to the label list. |
5. Convert sequences and labels to numpy arrays. |
6. Split the data into training and testing sets: |
a. Define the split ratio (e.g., 80–20); b. Split sequences and labels based on the split index. |
7. Define the hyperparameter search space: |
a. Set ranges for LSTM units, optimizers, batch sizes, epochs, learning rates, and dropout rates. |
8. Define the objective function: |
a. Initialize an empty Sequential model; b. Add LSTM layers with specified units and dropout; c. Compile the model with selected optimizer and loss function; d. Fit the model on training data with validation data and early stopping; e. Predict on test data and calculate the Mean Squared Error (MSE). |
9. Perform Bayesian optimization: |
a. Initialize Trials object; b. Run the optimization to find the best hyperparameters using the TPE algorithm. |
10. Retrieve the best hyperparameters. |
11. Build the final model with the best hyperparameters: |
a. Initialize a new Sequential model; b. Add LSTM layers with the best units and dropout; c. Compile the model with best optimizer and loss function. |
12. Train the final model on the training data with validation data and early stopping. |
13. Evaluate the final model on the test data: |
a. Calculate residuals; b. Calculate performance metrics (MSE, MAE, RMSE, R2, and MAPE). |
14. Plot results: |
a. Plot residuals; b. Plot training and validation accuracy; c. Plot training and validation loss. |
3.2.6. Performance Evaluation
4. Results
4.1. Performance of the Predictive Model
4.2. Impact and Performance of the Fusion Algorithm
4.3. Comparative Analysis
5. Discussion
6. Computational Setup
7. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Zhao, Y.; Yu, Y.; Shakeel, P.M.; Montenegro-Marin, C.E. Research on operational research-based financial model based on e-commerce platform. Inf. Syst. E-Bus. Manag. 2021, 21, 17. [Google Scholar] [CrossRef]
- Thakkar, A.; Chaudhari, K. Fusion in stock market prediction: A decade survey on the necessity, recent developments, and potential future directions. Inf. Fusion 2021, 65, 95–107. [Google Scholar] [CrossRef] [PubMed]
- Patel, K.R. Enhancing Global Supply Chain Resilience: Effective Strategies for Mitigating Disruptions in an Interconnected World. Bullet J. Multidisiplin Ilmu 2023, 2, 257–264. [Google Scholar]
- Sarker, I.H. Data science and analytics: An overview from data-driven smart computing, decision-making and applications perspective. SN Comput. Sci. 2021, 2, 377. [Google Scholar] [CrossRef]
- Kumar, G.; Jain, S.; Singh, U.P. Stock market forecasting using computational intelligence: A survey. Arch. Comput. Methods Eng. 2021, 28, 1069–1101. [Google Scholar] [CrossRef]
- Doering, J.; Kizys, R.; Juan, A.A.; Fito, A.; Polat, O. Metaheuristics for rich portfolio optimisation and risk management: Current state and future trends. Oper. Res. Perspect. 2019, 6, 100121. [Google Scholar] [CrossRef]
- Bartram, S.M.; Branke, J.; Motahari, M. Artificial Intelligence in Asset Management; CFA Institute Research Foundation: Charlottesville, VA, USA, 2020. [Google Scholar]
- Ghosh, I.; Jana, R.K. A granular machine learning framework for forecasting high-frequency financial market variables during the recent black swan event. Technol. Forecast. Soc. Chang. 2023, 194, 122719. [Google Scholar] [CrossRef]
- Bhandari, H.N.; Rimal, B.; Pokhrel, N.R.; Rimal, R.; Dahal, K.R.; Khatri, R.K. Predicting stock market index using LSTM. Mach. Learn. Appl. 2022, 9, 100320. [Google Scholar] [CrossRef]
- Gandhmal, D.P.; Kumar, K. Systematic analysis and review of stock market prediction techniques. Comput. Sci. Rev. 2019, 34, 100190. [Google Scholar] [CrossRef]
- Cavalcante, R.C.; Brasileiro, R.C.; Souza, V.L.; Nobrega, J.P.; Oliveira, A.L. Computational intelligence and financial markets: A survey and future directions. Expert Syst. Appl. 2016, 55, 194–211. [Google Scholar] [CrossRef]
- Gourieroux, C. Financial Econometrics: Problems, Models and Methods; Princeton University Press: Princeton, NJ, USA, 2018. [Google Scholar]
- Strielkowski, W.; Vlasov, A.; Selivanov, K.; Muraviev, K.; Shakhnov, V. Prospects and Challenges of the Machine Learning and Data-Driven Methods for the Predictive Analysis of Power Systems: A Review. Energies 2023, 16, 4025. [Google Scholar] [CrossRef]
- Zulfiqar, M.; Gamage, K.A.; Kamran, M.; Rasheed, M.B. Hyperparameter optimization of bayesian neural network using bayesian optimization and intelligent feature engineering for load forecasting. Sensors 2022, 22, 4446. [Google Scholar] [CrossRef] [PubMed]
- Bharadiya, J.P. Exploring the Use of Recurrent Neural Networks for Time Series Forecasting. Int. J. Innov. Sci. Res. Technol. 2023, 8, 2023–2027. [Google Scholar]
- Tang, Y.; Song, Z.; Zhu, Y.; Yuan, H.; Hou, M.; Ji, J.; Tang, C.; Li, J. A survey on machine learning models for financial time series forecasting. Neurocomputing 2022, 512, 363–380. [Google Scholar] [CrossRef]
- Chen, S.; Goo, Y.J.J.; Shen, Z.D. A hybrid approach of stepwise regression, logistic regression, support vector machine, and decision tree for forecasting fraudulent financial statements. Sci. World J. 2014, 2014, 968712. [Google Scholar] [CrossRef]
- Van Houdt, G.; Mosquera, C.; Nápoles, G. A review on the long short-term memory model. Artif. Intell. Rev. 2020, 53, 5929–5955. [Google Scholar] [CrossRef]
- Malsa, N.; Vyas, V.; Gautam, J. RMSE calculation of LSTM models for predicting prices of different cryptocurrencies. Int. J. Syst. Assur. Eng. Manag. 2021, 1–9. [Google Scholar] [CrossRef]
- Wu, D.; Wang, X.; Su, J.; Tang, B.; Wu, S. A labeling method for financial time series prediction based on trends. Entropy 2020, 22, 1162. [Google Scholar] [CrossRef]
- Siami-Namini, S.; Namin, A.S. Forecasting economics and financial time series: ARIMA vs. LSTM. arXiv 2018, arXiv:1803.06386. [Google Scholar]
- Zhang, X.; Liang, X.; Zhiyuli, A.; Zhang, S.; Xu, R.; Wu, B. At-lstm: An Attention-Based Lstm Model for Financial Time Series Prediction. In Proceedings of the 5th Annual International Workshop on Materials Science and Engineering, Changsha, China, 17–18 May 2019; IOP Publishing: Bristol, UK; Volume 569. [Google Scholar]
- Dixon, M.; London, J. Financial forecasting with α-RNNs: A time series modeling approach. Front. Appl. Math. Stat. 2021, 6, 551138. [Google Scholar] [CrossRef]
- Zhao, R.; Lei, Z.; Zhao, Z. Research on the application of deep learning techniques in stock market prediction and investment decision-making in financial management. Front. Energy Res. 2024, 12, 1376677. [Google Scholar] [CrossRef]
- Staffini, A. Stock price forecasting by a deep convolutional generative adversarial network. Front. Artif. Intell. 2022, 5, 837596. [Google Scholar] [CrossRef] [PubMed]
- Zheng, H.; Wu, J.; Song, R.; Guo, L.; Xu, Z. Predicting financial enterprise stocks and economic data trends using machine learning time series analysis. Appl. Comput. Eng. 2024, 87, 26–32. [Google Scholar] [CrossRef]
- Ghosh, B.P.; Bhuiyan, M.S.; Das, D.; Nguyen, T.N.; Jewel, R.M.; Mia, M.T.; Shahid, R. Deep Learning in Stock Market Forecasting: Comparative Analysis of Neural Network Architectures Across NSE and NYSE. J. Comput. Sci. Technol. Stud. 2024, 6, 68–75. [Google Scholar] [CrossRef]
- Sharma, R.; Sharma, R.; Hariharan, S. Stock Price Prediction Using ARIMA: A Study on AAPL, MSFT, NFLX, and GOOGL Stocks. In Proceedings of the 2024 International Conference on Computational Intelligence and Computing Applications (ICCICA), Panipat, India, 23–24 May 2024; Volume 1, pp. 47–50. [Google Scholar]
- Ahammad, I.; Sarkar, W.A.; Meem, F.A.; Ferdus, J.; Ahmed, M.K.; Rahman, M.R.; Sultana, R.; Islam, M.S. Advancing Stock Market Predictions with Time Series Analysis Including LSTM and ARIMA. Cloud Comput. Data Sci. 2024, 5, 226–241. [Google Scholar] [CrossRef]
- Md, A.Q.; Kapoor, S.; Juni, A.V.C.; Sivaraman, A.K.; Tee, K.F.; Sabireen, H.; Janakiraman, N. Novel optimization approach for stock price forecasting using multi-layered sequential LSTM. Appl. Soft Comput. 2023, 134, 109830. [Google Scholar] [CrossRef]
- Zhang, J.; Ye, L.; Lai, Y. Stock price prediction using CNN-BiLSTM-Attention model. Mathematics 2023, 11, 1985. [Google Scholar] [CrossRef]
- Chauhan, A.; Shivaprakash, S.J.; Sabireen, H.; Md, A.Q.; Venkataraman, N. Stock price forecasting using pso hypertuned neural nets and ensembling. Appl. Soft Comput. 2023, 147, 110835. [Google Scholar] [CrossRef]
Authors | Study Focus | Key Contributions | Relevance to Current Study |
---|---|---|---|
Tang et al. [16] | Demonstrated the efficacy of machine learning algorithms in stock price prediction | Showed that machine learning models can effectively predict stock prices, validating their application in financial forecasting. | Supports the use of machine learning models for enhancing the stock price prediction accuracy in the current study. |
Chen et al. [17] | Fraudulent financial statement detection using decision trees, logistic regression, and SVMs | Demonstrated that decision tree models (DT C5.0) achieve a higher classification accuracy compared to logistic regression and SVMs. | Highlights the early success of ensemble methods, providing a foundation for the application of advanced models in financial forecasting. |
Van et al. [18] | LSTM networks in financial time series analysis | Emphasized LSTM networks’ capacity to learn from extended sequences, which is critical for analyzing financial data. | Reinforced the importance of long-term memory in financial predictions, which is essential for modeling market trends and cycles. |
Malsa et al. [19] | LSTM networks’ performance in cryptocurrency price forecasting | LSTM networks outperformed traditional forecasting methods. | Demonstrated the practical superiority of LSTM networks, leading to theirbroader application in financial forecasting. |
Wu et al. [20] | LSTM networks in cryptocurrency market forecasting | Demonstrated LSTM networks’ adaptability to volatile markets like cryptocurrency. | Showcased the flexibility of LSTM networks in handling high volatility, contributing to their widespread adoption in financial forecasting. |
Siami-Namini and Namin [21] | LSTM vs. traditional ARIMA models for financial forecasting | Showed that LSTM models significantly reduce error rates compared to ARIMA, highlighting the advantages of deep learning models. | Validates the shift from traditional statistical models to deep learning approaches, setting the stage for more complex architectures. |
Zhang et al. [22] | Introduction of Attention-based LSTM (AT-LSTM) | Enhanced long-term dependency handling, improving prediction accuracy over standard LSTM models. | Introduced a more sophisticated LSTM variant that directly influences the architecture of modern financial forecasting models. |
Dixon and London [23] | Exponential smoothed RNNs (α-RNNs) in big data financial applications | Found α-RNNs effective with fewer parameters compared to GRUs and LSTMs. | Highlighted parameter efficiency, informing the design of more streamlined and scalable deep learning models. |
Zhao et al. [24] | Integrating CNN, BiLSTM, and attention mechanisms for financial time series forecasting | Achieved lower error metrics (MAE and RMSE) compared to traditional models. | Illustrates the progression to more integrated and complex model architectures, reflecting current trends in financial forecasting. |
Staffini [25] | DCGANs for stock price predictions | Showed significant improvements over ARIMA and LSTM networks, indicating the potential of generative models in financial forecasting. | Highlights the emergence of generative models, expanding the toolkit for financial predictions beyond conventional methods. |
Zheng et al. [26] | Explored advanced machine learning models (LSTM and CNN-BiLSTM) for stock price predictions | Demonstrated the effectiveness of the LSTM and CNN-BiLSTM models in capturing nonlinear market patterns, achieving a low Mean Squared Error (MSE) of 0.01286 and MAPE of 0.01984. | Highlights the potential of combining deep learning architectures to improve the accuracy of market trend predictions. |
Ghosh et al. [27] | Assessed the predictive performance of MLP, RNN, LSTM, and CNN models on stock price data from NSE and NYSE | Found that CNN models outperformed the others, achieving lower MAPE values across different companies and indicating superior predictive power for stock prices. | Provides insights into selecting CNN-based models for financial forecasting, supporting the use of effective architectures in experiments in the current study. |
Sharma et al. [28] | ARMA model’s effectiveness in technology stock price forecasting | MAPE values were between 1.10% and 1.50%, indicating reliable performance. | Provides a benchmark for evaluating newer models like Bayesian–LSTM against traditional approaches in financial forecasting. |
Date | Ticker | Open | High | Low | Close | Volume | Outstanding_Share |
---|---|---|---|---|---|---|---|
04/01/2005 | sh600000 | 0.77 | 0.77 | 0.75 | 0.76 | 3,808,939 | 900,000,000 |
05/01/2005 | sh600000 | 0.76 | 0.76 | 0.74 | 0.75 | 5,225,244 | 900,000,000 |
06/01/2005 | sh600000 | 0.75 | 0.75 | 0.73 | 0.74 | 4,298,099 | 900,000,000 |
07/01/2005 | sh600000 | 0.74 | 0.75 | 0.73 | 0.74 | 4,362,864 | 900,000,000 |
10/01/2005 | sh600000 | 0.75 | 0.77 | 0.74 | 0.77 | 7,115,260 | 900,000,000 |
11/01/2005 | sh600000 | 0.77 | 0.78 | 0.76 | 0.77 | 6,895,175 | 900,000,000 |
12/01/2005 | sh600000 | 0.78 | 0.78 | 0.76 | 0.77 | 2,797,449 | 900,000,000 |
13/01/2005 | sh600000 | 0.77 | 0.78 | 0.77 | 0.77 | 1,284,827 | 900,000,000 |
14/01/2005 | sh600000 | 0.77 | 0.78 | 0.76 | 0.76 | 1,731,048 | 900,000,000 |
17/01/2005 | sh600000 | 0.76 | 0.76 | 0.75 | 0.75 | 2,620,773 | 900,000,000 |
Variable | Statistics | p-Value | Critical Values |
---|---|---|---|
1% | |||
‘Close’ | −1.2411 | 0.6557 | −3.4329 |
‘Volume’ | −4.2250 | 0.0006 | −3.4329 |
Date | Open | High | Low | Close | Volume |
---|---|---|---|---|---|
04/01/2005 | 0.77 | 0.77 | 0.75 | 0.76 | 3,808,939 |
05/01/2005 | 0.76 | 0.76 | 0.74 | 0.75 | 5,225,244 |
06/01/2005 | 0.75 | 0.75 | 0.73 | 0.74 | 4,298,099 |
07/01/2005 | 0.74 | 0.75 | 0.73 | 0.74 | 4,362,864 |
10/01/2005 | 0.75 | 0.77 | 0.74 | 0.77 | 7,115,260 |
11/01/2005 | 0.77 | 0.78 | 0.76 | 0.77 | 6,895,175 |
12/01/2005 | 0.78 | 0.78 | 0.76 | 0.77 | 2,797,449 |
13/01/2005 | 0.77 | 0.78 | 0.77 | 0.77 | 1,284,827 |
14/01/2005 | 0.77 | 0.78 | 0.76 | 0.76 | 1,731,048 |
17/01/2005 | 0.76 | 0.76 | 0.75 | 0.75 | 2,620,773 |
BOLSTM_MSE | BOLSTM_MAE | BOLSTM_RMSE | BOLSTM_R2 | BOLSTM_MAPE | LSTM_MSE | LSTM_MAE | LSTM_RMSE | LSTM_R2 | LSTM_MAPE | |
---|---|---|---|---|---|---|---|---|---|---|
0 | 0.000005 | 0.001 | 0.002482 | 0.967518 | 0.030386 | 0.000154 | 0.014848 | 0.024132 | 0.744425 | 0.010763 |
1 | 0.000005 | 0.001 | 0.002287 | 1.089914 | 0.029556 | 0.000149 | 0.014846 | 0.02223 | 0.838599 | 0.010469 |
2 | 0.000006 | 0.001036 | 0.002321 | 0.996856 | 0.029135 | 0.000155 | 0.015384 | 0.02256 | 0.766998 | 0.01032 |
3 | 0.000006 | 0.000926 | 0.002148 | 0.944774 | 0.028863 | 0.000161 | 0.013746 | 0.020882 | 0.726925 | 0.010224 |
4 | 0.000005 | 0.000935 | 0.00225 | 1.038555 | 0.027138 | 0.000148 | 0.013889 | 0.021872 | 0.799082 | 0.009613 |
5 | 0.000005 | 0.000995 | 0.002326 | 0.936638 | 0.02825 | 0.000148 | 0.014773 | 0.022609 | 0.720665 | 0.010006 |
6 | 0.000006 | 0.000972 | 0.00218 | 1.007946 | 0.028629 | 0.000162 | 0.01443 | 0.02119 | 0.775531 | 0.010141 |
7 | 0.000006 | 0.00104 | 0.002356 | 0.899788 | 0.030853 | 0.000156 | 0.015439 | 0.022906 | 0.692312 | 0.010929 |
8 | 0.000005 | 0.000977 | 0.002243 | 0.931284 | 0.029808 | 0.000146 | 0.01451 | 0.021809 | 0.716546 | 0.010558 |
9 | 0.000005 | 0.000951 | 0.002279 | 1.007348 | 0.026721 | 0.000154 | 0.014127 | 0.022156 | 0.775071 | 0.009465 |
Authors | Methods | Dataset | Results |
---|---|---|---|
Proposed Model | Bayesian–LSTM | China Stock Price | MSE:0.00000535 |
MAE: 0.00102352 | |||
RMSE: 0.00231278 | |||
R2: 0.99752901 | |||
MAPE: 0.02930441 | |||
CAMBRIA Dataset | MAE: 7.60 | ||
MAPE: 2.06% | |||
RMSE: 1.065 | |||
MSE: 2.28 | |||
Financial Times Stock Exchange Milano Indice di Borsa | MAE: 0.0515 | ||
RMSE: 0.0437 | |||
MAPE: 0.0881 | |||
(National Stock Exchange of India) and NYSE (New York Stock Exchange) | MAPE 1.484% from MARUTI | ||
GOOGL Stocks | MAE: 0.66% | ||
RMSE: 0.44% | |||
MAPE: 0.90% | |||
Dhaka Stock Exchange | RMSE: 0.78048 | ||
MAE: 0.6226668 | |||
[24] | CNN, BiLSTM, Attention Mechanism | CAMBRIA Dataset, KRIRAN dataset, SHARMA dataset, JAMES dataset | MAE: 15.20 |
MAPE: 4.12% | |||
RMSE: 2.13 | |||
MSE: 4.56 | |||
[25] | DCGAN | Financial Times Stock Exchange Milano Indice di Borsa | RMSE: 0.112 |
MAE: 0.132 | |||
MAPE: 0.226 | |||
[27] | CNN | (National Stock Exchange of India) and NYSE (New York Stock Exchange) | MAPE values across companies like MARUTI, HCL, and AXIS BANK were5.30%, 6.33% and 7.32, respectively. |
[28] | ARIMA | AAPL, MSFT, NFLX, and GOOGL Stocks | The MAE, RMSE, and MAPE values range from 1.10% to 1.50%. |
[29] | ARIMA | Dhaka Stock Exchange | RMSE:4.336 |
MAE: 3.45926 | |||
[30] | Multi-Layered Sequential LSTM with the Adam Optimizer | GOOGL Stocks | MAPE: 1.79% |
RMSE:0.019 | |||
[31] | CNN–BiLSTM–Attention model | China Stock Price | MAPE: 1.56% |
RMSE:0.0076 | |||
[32] | PSO and ensemble of LSTM + GRU models | China Stock Price | MAE: 0.00102352 |
RMSE: 0.00231278 | |||
R2: 0.83752901 |
Test Case | Execution Time (Seconds) |
---|---|
Test Case 1 | 12.5 |
Test Case 2 | 15.3 |
Test Case 3 | 10.8 |
Test Case 4 | 18.6 |
Test Case 5 | 14.2 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Olaniyan, J.; Olaniyan, D.; Obagbuwa, I.C.; Esiefarienrhe, B.M.; Adebiyi, A.A.; Bernard, O.P. Intelligent Financial Forecasting with Granger Causality and Correlation Analysis Using Bayesian Optimization and Long Short-Term Memory. Electronics 2024, 13, 4408. https://doi.org/10.3390/electronics13224408
Olaniyan J, Olaniyan D, Obagbuwa IC, Esiefarienrhe BM, Adebiyi AA, Bernard OP. Intelligent Financial Forecasting with Granger Causality and Correlation Analysis Using Bayesian Optimization and Long Short-Term Memory. Electronics. 2024; 13(22):4408. https://doi.org/10.3390/electronics13224408
Chicago/Turabian StyleOlaniyan, Julius, Deborah Olaniyan, Ibidun Christiana Obagbuwa, Bukohwo Michael Esiefarienrhe, Ayodele A. Adebiyi, and Olorunfemi Paul Bernard. 2024. "Intelligent Financial Forecasting with Granger Causality and Correlation Analysis Using Bayesian Optimization and Long Short-Term Memory" Electronics 13, no. 22: 4408. https://doi.org/10.3390/electronics13224408