Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

A dual-path convolutional neural network combined with an attention-based bidirectional long short-term memory network for stock price prediction

  • Guiyan Zhao,

    Roles Formal analysis, Funding acquisition, Methodology, Software, Visualization, Writing – original draft

    Affiliation Qiaoxing School of Economics and Management, Fujian Polytechnic Normal University, Fuzhou, China

  • Yunfei Cheng ,

    Roles Data curation, Methodology, Software, Writing – review & editing

    zmcvbcyf@163.com

    Affiliation School of Mathematics and Statistics, Changchun University of Technology, Changchun, China

  • Jianhui Yang,

    Roles Resources, Supervision, Writing – review & editing

    Affiliation Qiaoxing School of Economics and Management, Fujian Polytechnic Normal University, Fuzhou, China

  • Jiayuan Ouyang

    Roles Conceptualization, Investigation, Validation

    Affiliation Qiaoxing School of Economics and Management, Fujian Polytechnic Normal University, Fuzhou, China

Abstract

The complexities of stock price data, characterized by its nonlinearity, non-stationarity, and intricate spatiotemporal patterns, make accurate prediction a substantial challenge. To address this, we propose the DCA-BiLSTM model, which combines dual-path convolutional neural networks with an attention mechanism (DCA) and bidirectional long short-term memory networks (BiLSTM). This model captures deep information and complex dependencies within time-series data. First, wavelet packet decomposition extracts high- and low-frequency features, followed by DCA for robust deep feature extraction, and finally, BiLSTM models bidirectional dependencies. Validated on datasets from Yahoo Finance, including Apple, Google, Tesla stocks, and the Nasdaq index, the model consistently outperforms traditional approaches. The DCA-BiLSTM achieves an of 0.9507 for Apple, 0.9595 for Google, 0.9077 for Tesla, and 0.9594 for the Nasdaq index, with significant reductions in error metrics across all datasets. These results demonstrate the model’s robustness and improved predictive accuracy, offering reliable insights for stock price forecasting.

1 Introduction

The financial market, as a complex macro system, is influenced by various macroeconomic, political, and social factors. Stock price fluctuations are widely regarded as effective indicators of overall economic health, and changes in the stock market significantly impact economic activities and corporate development. Consequently, researchers have long sought to develop effective models for predicting stock prices [1,2]. However, stock price prediction is inherently complex due to the multitude of factors affecting prices, which renders traditional models inadequate for capturing dynamic market patterns [3]. With advancements in big data processing technologies and AI algorithms, data science methodologies have achieved notable success in the financial domain. Accordingly, research leveraging diverse AI algorithms is expected to become a vital tool for financial researchers analyzing market dynamics, particularly in overcoming the limitations of traditional approaches.

Stock price prediction often involves intricate nonlinear relationships that traditional linear models struggle to capture effectively. Multiple factors, including company fundamentals, technical indicators, and market sentiment, influence price movements. AI algorithms, with their ability to process vast amounts of data and uncover latent patterns, provide a flexible and powerful solution. Unlike traditional economic models bound by rigid assumptions, these algorithms learn from historical data without constraints, adapting to complex data patterns. This adaptability is particularly valuable for real-time decision-making in financial trading.

In recent years, deep learning-based models for stock prediction have emerged as a focal area of research, with specific emphasis on convolutional neural networks [46], recurrent neural networks [7,8], and Transformers [9,10]. Hybrid approaches have garnered considerable attention, aiming to combine the strengths of multiple models to achieve higher predictive accuracy compared to traditional single-model approaches. Jaiswal and Singh introduced a novel hybrid convolutional-recurrent model that integrates a one-dimensional convolutional neural network (CNN) with a gated recurrent unit (GRU) model. In this model, CNN is used for feature extraction, while GRU manages temporal regression, showing superior performance over traditional models [11]. Rostamian and Hara proposed a model that incorporates long short-term memory (LSTM) with CNN, effectively combining LSTM’s robustness in temporal prediction with CNN’s strengths in feature extraction, leading to significant improvements over conventional models [12]. Nevertheless, despite the advancements offered by these hybrid models [13], there is still room for improvement. Some research advocates for the use of bidirectional recurrent neural network (RNN) models [14,15] and the integration of attention mechanisms [16] to further enhance feature extraction. Against this background, Lu et al. and Luo et al. presented attention-based models that adopt a CNN-BiLSTM fusion structure, where CNN is utilized for feature extraction and BiLSTM for regression prediction [17,18]. By incorporating attention mechanisms, these models effectively capture the influence of different temporal feature states on stock closing prices, thereby improving prediction accuracy. Wang proposed an alternative approach involving a temporal convolutional network that combines BiLSTM with enhanced Transformers, leveraging Transformers to capture global information and BiLSTM to learn bidirectional short-sequence information. Experimental results reveal its significant advantages over standalone Transformer models [19]. In summary, combining multiple models for prediction research offers distinct advantages over single-model approaches.

In stock prediction tasks, studies suggest that dual-channel or even multi-channel structures can significantly enhance model performance by performing deeper analyses of data features [2025]. To further improve the effectiveness of predictive models, this paper proposes an enhanced approach based on the traditional hybrid model. This model incorporates a neural network that integrates dual-path convolutional attention mechanisms with BiLSTM. Furthermore, to enhance the dual-path CNN’s ability to extract multi-scale features, we employ a wavelet packet coefficient decomposition method for dataset preprocessing. This allows the dual-channel CNN, operating at multiple scales, to extract more comprehensive features by separating high and low-frequency data through filtering. However, due to the relatively complex multi-scale feature structures extracted by the dual-channel CNN, directly using BiLSTM for regression poses challenges for datasets with longer periods. Therefore, we introduce an improved multi-scale feature fusion approach leveraging attention mechanisms to enhance performance on stock price datasets.

The structure of this paper is as follows: Section 2 provides a detailed overview of the proposed DCA-BiLSTM stock prediction model. Section 3 discusses the experimental validation and results. Section 4 concludes the paper, summarizing key findings.

2 Stock price prediction model

Given the nonlinearity, non-stationarity, and complex spatiotemporal relationships inherent in stock price data, this study introduces a stock price prediction model based on DCA-BiLSTM. The goal is to improve prediction accuracy by extracting and effectively leveraging highly correlated and diverse deep features during the regression process.

This study examines stock price prediction under a scenario involving n observation samples. For each target date i, the closing price on day i is treated as the dependent variable in regression analysis. The independent variable data includes metrics such as opening price, highest price, lowest price, price change, among others. By combining the independent variable data with the dependent variable data , we form the observation dataset .

The stock price prediction model proposed in this paper consists of three main stages:

  1. (1) Decomposing the raw input time series data into high-frequency domain signals and low-frequency domain signals using wavelet packet decomposition, and transforming them into shallow high-low frequency domain features through a feature extraction layer;
  2. (2) Using a dual-path convolutional neural network with an attention mechanism to integrate the shallow high-low frequency features into deep features with high reliability;
  3. (3) Modeling the extracted deep features from both forward and backward directions using a bidirectional long short-term memory network, and performing regression prediction through a fully connected layer.

The specific algorithmic framework is depicted in Fig 1. Below, we elaborate on the three stages involved in constructing this model.

thumbnail
Fig 1. The framework of the stock price prediction method based on the DCA-BiLSTM model.

https://doi.org/10.1371/journal.pone.0319775.g001

2.1 Wavelet packet-based feature decomposition

Stock price data often display marked nonlinearity and non-stationarity, resulting in varied statistical characteristics over time. This challenge complicates the ability of traditional linear models to accurately capture and predict stock price fluctuations. To address this, we employ wavelet packet decomposition, which effectively handles non-stationary time series signals and provides multi-scale information to capture the data’s nonlinear properties [26].

Wavelet packet analysis, a signal processing technique based on wavelet transforms, extends the capabilities of traditional methods in terms of time-frequency analysis. While conventional wavelet transforms focus on decomposing signals into approximation and detail components, with a primary emphasis on the low-frequency approximation, wavelet packet analysis provides a more comprehensive decomposition that includes both low- and high-frequency components. This approach extracts features from the entire scale domain, enhancing the scope of feature extraction. In this study, wavelet packet decomposition is applied to initially extract features, decomposing the signals into high- and low-frequency domain information. These features are then further processed by a dual-path convolutional neural network. The architecture of the wavelet packet decomposition layer within the network is illustrated in Fig 2.

thumbnail
Fig 2. The architecture of the wavelet packet decomposition layer.

https://doi.org/10.1371/journal.pone.0319775.g002

As depicted in Fig 2, the wavelet packet decomposition-based shallow feature extraction method uses time series data as input. During processing, multiple high-pass and low-pass linear filtering operations are performed, resulting in the extraction of wavelet packet coefficients from the time series data. These coefficients capture essential signal information and represent the decomposition of signal x at scale j in the wavelet packet function.

The calculation formula is as follows:where represents the original signal, while and denote the decomposed wavelet signals. Here, m represents the frequency band, where higher values correspond to higher-frequency bands; u represents the time-domain position, k indicates the time-domain shift of the filtering function, and Z denotes the set of integers. The symbols g ( ⋅ )  and h ( ⋅ )  represent the high-pass and low-pass filtering functions, respectively, which weight the wavelet packet function values across different frequency bands. Their formulas are as follows:

For the high-pass filtering function, denoted as g(a):

For the low-pass filtering function, denoted as h(a):

After passing through the feature extraction layer, the high-frequency time-domain signal and low-frequency time-domain signal are transformed into shallow high- and low-frequency features, respectively. While wavelet packet coefficients can effectively reflect the high- and low-frequency features in stock price samples, directly using these coefficients for stock price prediction is challenging [27,28]. The following sections present a model to address the handling of these extracted high- and low-frequency features for more effective application in stock price prediction.

2.2 Dual-path convolutional neural network with attention mechanism

Following wavelet packet decomposition, the input data is transformed into high- and low-frequency time-domain features, necessitating further extraction of deeper features. Given the multifaceted impact of factors such as market sentiment, macroeconomic indicators, and company fundamentals on stock prices, a model capable of effectively capturing these relationships is required. To address these challenges, this paper employs a DCA network architecture. This architecture integrates an attention mechanism into a dual-path convolutional neural network, allowing the extraction of highly reliable deep features from high- and low-frequency features via encoder structures with attention mechanisms [29].

The main contributions of this model are as follows:

  1. (1) Due to the differing feature scales of high- and low-frequency signals, optimizing the hyperparameters of different convolutional layers is essential. Therefore, this paper adopts a multi-path convolutional neural network structure;
  2. (2) To address the increased parameter size during convolution, which can prolong training times and reduce model efficiency, global average pooling is used after each convolutional layer to reduce feature dimensionality;
  3. (3) To enhance model accuracy and convergence rate, group normalization is applied to standardize each channel of the feature maps;
  4. (4) In the attention mechanism, scaled dot-product attention is employed to mitigate the complexity and reduce processing speed declines caused by multilinearity.

This network architecture, integrating attention mechanisms and dual-channel convolutional networks, aims to process time series inputs more effectively, enhance feature extraction, and support the BiLSTM layer in capturing critical information at various points in sequences. This architecture exhibits impressive deep feature extraction capabilities, as shown in Fig 3.

The multi-head self-attention mechanism includes several self-attention modules, each designed to compute attention weights for every element across positions in the input sequence, resulting in a weighted summation as the output. Illustrated in Fig 4, the architecture unfolds as follows:

  1. (1) First, attention scores are calculated by projecting the input sequence into different spaces using Query , Key , and Value matrices. The relevance of each element within the sequence is assessed by computing the attention scores based on the dot product between the query and key Ki pairs.
  2. (2) Next, attention weights are obtained by applying the softmax function to the attention scores, creating a distribution that highlights the significance of each part of the sequence.
  3. (3) A weighted summation is then produced by combining the input sequence with the calculated attention weights.
  4. (4) This weighted sum serves as the output of a single self-attention module.

In a multi-head self-attention mechanism, these steps are repeated independently across multiple attention heads, each with its own parameters. The outputs from each head are then combined, typically by concatenation, to produce the final result, enhancing the model’s ability to capture complex relationships within the sequence.

thumbnail
Fig 4. The architecture of the multi-head attention layer.

https://doi.org/10.1371/journal.pone.0319775.g004

2.3 Bidirectional long short-term memory

Stock prices embody information at different time scales, including short-term fluctuations and long-term trends. Simultaneously considering these scales in prediction is complex. Wavelet packet decomposition aids in modeling data across time scales by providing multi-scale information. BiLSTM, an improvement over LSTM networks, captures both long-term and short-term dependencies, thus addressing time scale issues more comprehensively.

BiLSTM combines forward and backward models, allowing it to capture contextual information bidirectionally [30]. Since a single LSTM only captures context in one direction, using a single LSTM may lead to the loss of features related to the opposite direction, particularly in sequential data like stock prices where context is often bidirectional. Hence, using BiLSTM helps in capturing dependencies more effectively. For stock price time-series data, which are sensitive to time and deeply dependent on past and future periods, BiLSTM improves predictive performance by capturing information across different time scales, including both short-term and long-term trends [31,32].

The architecture of a single LSTM unit is illustrated in Fig 5.

The relevant formula definitions are as follows:

where represents the input data at the current time step t. represents the input gate, controlling information flow into the memory cell . is the forget gate, deciding whether information from the previous time step t–1 should be retained in the memory cell . determines whether the memory cell information impacts the current hidden state . denotes the memory cell, embodying the neural state’s memory, and giving LSTM units the capacity to store, read, reset, and update historical information. The are weights associated with respective gates. Sigmoid and tanh are the commonly used activation functions [33].

3 Experiments

To assess the performance of the proposed stock price prediction model based on DCA-BiLSTM, datasets from four sources were selected: Apple stock, Google stock, Tesla stock, and the Nasdaq index data. The proposed model was compared with four alternative models: CNN-BiLSTM, DCA, CNN, and BiLSTM [3436].

3.1 Experimental environment and parameters

The experiments were conducted in a Windows 10 22H2 environment. Python 3 and the PyTorch neural network framework were used to implement the DCA-BiLSTM model. CUDA 11.8 was employed to accelerate model training.

3.2 Experimental datasets

The datasets for this study, sourced from Yahoo Finance (https://finance.yahoo.com/), contain historical daily price data for Apple, Google, Tesla stocks, and the Nasdaq index, with all data collection and usage complying with Yahoo Finance’s Terms of Service. Apple, as a global leader in technology, serves as a key market indicator. Google, one of the largest players in the internet industry, impacts the advertising sector and the broader IT field. Tesla, a pioneer in electric vehicles and renewable energy, offers insights into emerging markets, while the Nasdaq index, representing a range of technology companies, provides a broad view of technology sector trends.

thumbnail
Fig 7. Comparison of predicted values for different models (Apple stock).

https://doi.org/10.1371/journal.pone.0319775.g007

The study spans the following timeframes: Apple stock (2013/12/9-2023/12/7), Google stock (2014/1/13-2024/1/12), Tesla stock (2014/1/13-2024/1/12), and the Nasdaq index (2013/12/9-2023/12/7). The selected features include indicators such as opening price, highest price, lowest price, price change, percentage change, trading volume, turnover, market capitalization, and total market value, among others. These datasets comprise a diverse range of fundamental data, supporting the construction of multifactorial time-series samples and enabling comprehensive predictive analysis. The experimental models primarily utilize closing prices to forecast future market trends.

3.3 Evaluation metrics

In the experiment, we utilize four metrics to assess the predictive performance of the model, namely the coefficient of determination (), mean absolute error (MAE), mean absolute percentage error (MAPE), and root mean squared error (RMSE). The formulas for calculating these metrics are as follows:where represents the sample size of the test set, and , , and denote the true values, the mean of all true values in the test set, and the predicted values of the test set, respectively. These four metrics individually quantify the disparities between the model’s predicted values and the actual observations. indicates the model’s explanatory power, while MAE, MAPE, and RMSE quantify the magnitude of prediction errors. Higher values (closer to 1) and lower MAE, MAPE, and RMSE values generally indicate superior predictive performance.

thumbnail
Fig 8. Comparison of predicted values for different models (Google stock).

https://doi.org/10.1371/journal.pone.0319775.g008

thumbnail
Fig 9. Comparison of predicted values for different models (Tesla stock).

https://doi.org/10.1371/journal.pone.0319775.g009

thumbnail
Fig 10. Comparison of predicted values for different models (Nasdaq index).

https://doi.org/10.1371/journal.pone.0319775.g010

thumbnail
Table 1. Comparison of evaluation metrics for different models.

https://doi.org/10.1371/journal.pone.0319775.t001

3.4 Experimental results

To validate the proposed DCA-BiLSTM model, we compared it against four alternative models: CNN-BiLSTM, DCA, CNN, and BiLSTM. To enhance the predictive precision of the model, we allocated 90% of the dataset for training to refine parameter tuning, while the remaining 10% was reserved for testing to assess the model’s performance.

Fig 6 shows the loss function curves of these models, with the horizontal axis indicating the iteration count and the vertical axis representing the loss value. Figs 7 to 10 illustrate the closing price predictions of these models, with the horizontal axis representing the number of iterations and the vertical axis representing the closing price.

In Fig 6, the proposed DCA-BiLSTM model demonstrates relatively lower initial loss values across all datasets compared to the other models. With increasing iterations, all models exhibit a gradual reduction in loss values; however, the DCA-BiLSTM model maintains a more stable loss curve, highlighting its robustness and ability to achieve consistently lower loss levels across datasets.

Observing Figs 7 to 10 reveals that in the early iterations, the DCA-BiLSTM model trends closer to actual closing prices than the other models. As iterations increase, this proximity becomes more pronounced, while the predictions of the other models display larger deviations.

Table 1 presents the predictive results for each dataset.

The analysis of predictive performance across the Apple, Google, Tesla, and Nasdaq datasets reveals that the DCA-BiLSTM model consistently outperforms other models. For Apple stock, it achieves an of 0.9507, with superior accuracy across MAE, MAPE, and RMSE metrics. On the Google stock dataset, the model stands out with an of 0.9595 and the lowest MAE, RMSE, and MAPE, demonstrating precise price predictions. Similarly, for Tesla stock, the DCA-BiLSTM model achieves the highest at 0.9077, indicating strong predictive performance across all metrics. Finally, on the Nasdaq index, the model attains an of 0.9594 and outperforms in error metrics, underscoring its robustness and accuracy across varied stock datasets.

4 Conclusion and future work

This paper proposes a stock price prediction model based on the DCA-BiLSTM approach. The method begins with decomposing the time-series data using wavelet packet decomposition to extract high- and low-frequency features. Subsequently, a neural network architecture integrating convolutional layers and attention mechanisms performs deep feature extraction. A bidirectional LSTM model is then applied to capture forward and backward context within the data. Experimental results demonstrate the DCA-BiLSTM model’s accuracy and stability in stock price prediction, underscoring its capacity to capture complex temporal relationships and improve prediction precision.

Nonetheless, several challenges remain for future research. First, stock markets are influenced by numerous external factors such as political events, natural disasters, and economic policies, which existing models may inadequately address. Future studies should explore integrating these factors to enhance model robustness and applicability. Additionally, although our model performs well in short-term predictions, its accuracy and stability diminish in long-term forecasting due to factors such as economic cycles and industry trends that are difficult to capture. Therefore, future research should focus on addressing long-term prediction challenges through specialized modeling techniques. Lastly, current models face overfitting risks, especially when handling large datasets. Future studies should explore strategies to prevent overfitting and improve model generalization and robustness.

In conclusion, while this study has achieved promising results, stock price prediction remains a complex and evolving field that demands continuous exploration and refinement of predictive techniques. Future research will continue to investigate innovative methodologies to better capture stock market patterns, providing investors with more accurate and reliable forecasting insights.

References

  1. 1. Gandhmal DP, Kumar K. Systematic analysis and review of stock market prediction techniques. Comput Sci Rev. 2019;34:100190.
  2. 2. Jiang W. Applications of deep learning in stock market prediction: recent progress. Expert Syst Appl. 2021;184:115537.
  3. 3. Obthong M, Tantisantiwong N, Jeamwatthanachai W, Wills G. A survey on machine learning for stock price prediction: algorithms and techniques. In: 2nd International Conference on Finance, Economics, Management and IT Business, 5 May 2020–6 May 2020.
  4. 4. Mehtab S, Sen J. Stock price prediction using convolutional neural networks on a multivariate timeseries. arXiv preprint. arXiv:2001.09769. 2020.
  5. 5. Wu JMT, Li Z, Srivastava G, Frnda J, Diaz VG, Lin JCW. A CNN-based stock price trend prediction with futures and historical price. In: 2020 International Conference on Pervasive Artificial Intelligence (ICPAI). IEEE; 2020: pp. 134–9.
  6. 6. Cheung L, Wang Y, Lau ASM, Chan RMC. Using a novel clustered 3D-CNN model for improving crop future price prediction. Knowledge-Based Systems. 2023; 260:110133.
  7. 7. Yang F, Chen J, Liu Y. Improved and optimized recurrent neural network based on PSO and its application in stock price prediction. Soft Comput 2023;27(6):3461–76.
  8. 8. Fayeem A, Kumar A, Sagar R, Aggarwal A, Jain D. Stock price prediction: recurrent neural network in financial market. Int J Mod Trends Sci Technol. 2022;8:259–64.
  9. 9. Muhammad T, Aftab AB, Ibrahim M, Ahsan MM, Muhu MM, Khan SI et al. Transformer-based deep learning model for stock price prediction: a case study on Bangladesh stock market. Int J Comput Intell Appl. 2023:2350013.
  10. 10. Costa LD, Machado AMC. Prediction of stock price time series using transformers. In: Anais do II Brazilian Workshop on Artificial Intelligence in Finance. SBC; 2023: pp. 85–95.
  11. 11. Jaiswal R, Singh B. A hybrid convolutional recurrent (CNN-GRU) model for stock price prediction. In: 2022 IEEE 11th International Conference on Communication Systems and Network Technologies (CSNT). IEEE; 2022: pp. 299–304.
  12. 12. Rostamian A, O’Hara JG. Event prediction within directional change framework using a CNN-LSTM model. Neural Comput Applic 2022;34(20):17193–205.
  13. 13. Putra R A, Nurmawati E. Prediction-based stock portfolio optimization using bidirectional long short-term memory (BiLSTM) and LSTM. Sci J Inform 2024;11(3):609–20.
  14. 14. Woźniak M, Wieczorek M, Siłka J. BiLSTM deep neural network model for imbalanced medical data of IoT systems. Future Gener Comput Syst. 2023;141:489–99.
  15. 15. Masalimov K, Muslimov T, Munasypov R. Real-time monitoring of parameters and diagnostics of the technical condition of small unmanned aerial vehicle’s (UAV) units based on deep BiGRU-CNN models. Drones 2022;6(11):368.
  16. 16. Guo MH, Xu TX, Liu JJ, Liu ZN, Jiang PT, Mu TJ. Attention mechanisms in computer vision: a survey. Comput Vis Media 2022;8(3):331–68.
  17. 17. Lu W, Li J, Wang J, Qin L. A CNN-BiLSTM-AM method for stock price prediction. Neural Comput Applic. 2021;33:4741–53. https://link.springer.com/article/10.1007/s00521-020-05532-z
  18. 18. Luo A, Zhong L, Wang J, Wang Y, Li S, Tai W. Short-term stock correlation forecasting based on CNN-BiLSTM enhanced by attention mechanism. IEEE Access. 2024;12:29617–32.
  19. 19. Wang S. A stock price prediction method based on BiLSTM and improved transformer. IEEE Access. 2023;11:104211–23.
  20. 20. Jiang X, Zhu Y, Liu Y, Wang N, Yi L. MC-DC: an MLP-CNN based dual-path complementary network for medical image segmentation. Comput Methods Programs Biomed. 2023;242:107846.
  21. 21. Tong J, Wang C. A dual tri-path CNN system for brain tumor segmentation. Biomed Signal Process Control. 2023;81:104411.
  22. 22. Zhang Y, Yang S, Sun Y, Liu S, Li X. Attention-guided multi-path cross-CNN for underwater image super-resolution. SIViP 2022;16:155–63
  23. 23. Dang F, Chen H, Zhang P. DPT-FSNet: dual-path transformer based full-band and sub-band fusion network for speech enhancement. In: ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE; 2022: pp. 6857–61.
  24. 24. Sun Y, Zheng H, Zhang G, Ren J, Xu H, Xu C. DP-ViT: a dual-path vision transformer for real-time sonar target detection. Remote Sensing 2022;14(22):5807.
  25. 25. Lee Y, Kim J, Willette J, Hwang SJ. Mpvit: Multi-path vision transformer for dense prediction. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022: pp. 7287–96.
  26. 26. Ferkous K, Guermoui M, Bellaour A, Boulmaiz T, Bailek N. Enhancing photovoltaic energy forecasting: a progressive approach using wavelet packet decomposition. Clean Energy 2024;8(3):7287–96.
  27. 27. Franco de Godói L, de Oliveira Nóbrega EG. Unsupervised machinery prognostics approach based on wavelet packet decomposition and variational autoencoder. J Braz Soc Mech Sci Eng 2024;46(2):97.
  28. 28. Taranenko YK, Oliinyk OY. Optimizing the algorithm of wavelet packet signal filtering. Cybern Syst Anal. 2024;60(1):135–45.
  29. 29. Halder R, Chatterjee R. CNN-BiLSTM model for violence detection in smart surveillance. SN Comput Sci 2020;1(4):201. https://link.springer.com/article/10.1007/s42979-020-00207-x
  30. 30. El Zaar A, Benaya N, Bakir T, Mansouri M, El Allati A. Prediction of US 30-years-treasury-bonds movement and trading entry point using the robust 1DCNN-BiLSTM-XGBoost algorithm. Expert Syst 2024;41(1):e13459.
  31. 31. Pandey V, Lilhore UK, Walia R, Alroobaea R, Alsafyani M, Baqasah AM, et al. Enhancing heart disease classification with M2MASC and CNN-BiLSTM integration for improved accuracy. Sci Rep 2024;14(1):24221. pmid:39414872
  32. 32. Dalal S, Lilhore UK, Simaiya S, Radulescu M, Belascu L. Improving efficiency and sustainability via supply chain optimization through CNNs and BiLSTM. Technol Forecast Soc Change. 2024;209:123841.
  33. 33. Tian J, Shen C, Wang B, Ren C, Xia X, Dong R, et al. EVADE: targeted adversarial false data injection attacks for state estimation in smart grid. IEEE Trans Sustain Comput. 2024;(1):1–13. 10.1109/TSUSC.2024.3492290
  34. 34. Anantanasuwong K, Chaivisuttangkun S. Do investors benefit from DCA? Evidence from the stock exchange of Thailand. Chula Med J. Chula Bus Rev. 2019;41(2):84–101.
  35. 35. Hoseinzade E, Haratizade S. CNNpred: CNN-based stock market prediction using a diverse set of variables. Expert Syst Appl. 2019;129:273–85.
  36. 36. Pramesti MI, Indikawati FI, Prahara A. Multivariate time series stock price data prediction in the banking sector in Indonesia using Bidirectional Long Short-Term Memory (biLSTM). Signal Image Process Lett. 2022;4(2):28–37.