Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

A novel agricultural commodity price prediction model integrating deep learning and enhanced swarm intelligence algorithm

  • Kaixuan Sun,

    Roles Conceptualization, Data curation, Methodology, Project administration, Resources, Software, Supervision, Writing – original draft, Writing – review & editing

    Affiliation School of Economics and Management, Huainan Normal University, Huainan, China

  • Qi Yao,

    Roles Funding acquisition, Supervision, Validation

    Affiliations School of Information Management, Central China Normal University, Wuhan, China, Management School, Wuhan College, Wuhan, China

  • Yanhui Li

    Roles Funding acquisition, Supervision, Validation

    yhlee@ccnu.edu.cn

    Affiliations School of Information Management, Central China Normal University, Wuhan, China, Management School, Wuhan College, Wuhan, China

Abstract

The volatility of agricultural commodity prices significantly affects market stability and financial market dynamics, especially during periods of economic uncertainty and global shocks. Accurate price prediction, however, remains challenging due to the complex, nonlinear characteristics of agricultural markets and the diverse range of influencing factors. To overcome these challenges, this study develops a novel price forecasting framework that combines advanced time series decomposition, swarm intelligence optimization, and deep learning techniques. The proposed framework employs successive variational mode decomposition (SVMD) to deconstruct the raw price data into multiple components, effectively capturing the underlying nonlinear patterns and dynamic features. These components are then fed into a CNN-augmented BiLSTM model, enhanced with an attention mechanism to extract both temporal dependencies and intricate data relationships. To fine-tune the model’s hyperparameters, this study introduces a multiple strategies dung beetle optimisation algorithm (MSDBO), which integrates four strategic modifications to improve the balance between global search, local exploration, and convergence efficiency. Using historical data from corn and wheat markets as case studies, the experimental findings demonstrate that the proposed SVMD-MSDBO-CNN-BiLSTM-A model significantly outperforms nine baseline approaches. Specifically, it reduces the Mean Absolute Percentage Error (MAPE) by 25.78% and 37.57%, respectively, and enhances directional accuracy (Dstat) by 1.15% and 14.53% compared to the top single models.

1 Introduction

The volatility of agricultural commodity prices is a critical aspect of global financial market dynamics and economic stability. Agricultural commodities, as key components of the broader commodity markets, play an essential role in ensuring food security, supporting economic growth, and facilitating global trade. However, the price movements of these commodities are inherently nonlinear, driven by factors such as macroeconomic shocks, geopolitical tensions, supply-demand imbalances, and climate variability [13]. These complexities have been exacerbated by escalating global trade tensions and supply chain vulnerabilities, which have altered traditional price transmission mechanisms and introduced new variables into price formation processes [4,5]. Moreover, given their close integration with financial systems, disruptions in agricultural commodity prices can generate spillover effects, amplifying risks for investors, policymakers, and market regulators [6].

Recent global events, including health emergencies, inflationary pressures, and geopolitical risks, have further complicated price dynamics [7]. These events have heightened market uncertainty and volatility, underscoring the urgent need for robust, data-driven forecasting frameworks capable of capturing the nonlinear characteristics of agricultural commodity prices. Such models are critical for improving risk management, financial market regulation, and understanding price dynamics under volatile conditions [8,9].

Nevertheless, predicting agricultural futures prices remains a highly complex task. These include economic globalization, climate variability, and financial speculation, each contributing to distorted prices and increased market volatility [1013]. Governmental policies, such as export restrictions, subsidies, and import tariffs, further complicate market predictions, adding layers of unpredictability [14]. Recent challenges, like the COVID-19 pandemic and ongoing geopolitical tensions, have disrupted supply chains, exacerbating the volatility of agricultural markets [15,16].

The nonlinear and interconnected nature of agricultural commodity price fluctuations poses a significant challenge for accurate forecasting, especially amid the growing volatility in global financial and commodity markets. Addressing these challenges requires advanced forecasting models that can capture non-linear patterns and temporal dependencies, which are essential for improving risk management and supporting market stability.

The main contributions and innovations of this study can be summarized in a more concise and academically improved manner as follows:

(a) This study pioneers a hybrid forecasting system for agricultural commodity prices, combining an intelligent optimization algorithm with a deep learning model stack. The proposed SVMD-MSDBO-CNN-BiLSTM-A integrates data decomposition, feature selection, parameter optimization, and forecasting, outperforming nine benchmark models across two datasets.

(b) MSDBO is utilized to optimize six significant hyperparameters of the BiLSTM model, including learning rate, time step, neuron count in hidden layers, training batch size, maximum iterations, and nodes in the fully connected layer. Experimental results indicate a significant improvement in BiLSTM’s forecasting effectiveness.

(c) This study presents MSDBO, which incorporates modified circular chaotic mapping, rat swarm optimization (RSO), golden sine optimization, and an adaptive bidirectional Gaussian perturbation strategy. These modifications enhance the algorithm’s global search capabilities and local exploitation efficiency, while also preventing it from getting trapped in local optima.

The remainder of this article is organized as follows. Sect 2 reviews the relevant literature. Sect 3 introduces the proposed forecasting framework. Sect 4 describes data pre-processing and analysis. Sect 5 presents the experimental results and discussion. Finally, Sect 6 concludes the study and suggests future research directions.

2 Literature review

2.1 Traditional statistical learning, machine learning, and deep learning methods

Current research in agricultural commodity price forecasting has explored a wide range of methods to develop effective prediction models. These include statistical learning methods, machine learning techniques, deep learning approaches, and hybrid decomposition-ensemble models. Traditional statistical models, such as autoregressive conditional heteroskedasticity (ARCH) [17] and autoregressive integrated moving average (ARIMA) along with its variants [18], have been widely used. However, these models often struggle with nonlinear data, lack generalization capability, and are limited in extracting effective information from time series data. On the other hand, machine learning methods like back propagation neural network (BPNN) [19], extreme learning machine (ELM) [20], and support vector machine (SVM) [21] are prevalent in price forecasting. Despite their widespread application, these methods face challenges such as susceptibility to overfitting.

Recent advancements in deep learning, propelled by big data, have significantly advanced forecasting capabilities. Deep learning models, such as recurrent neural network (RNN) [22], long short-term memory (LSTM) [23], gated recurrent unit (GRU) [24], and Bi-directional long short-term memory (BiLSTM) [25], have proven effective for long-term time series forecasting. They are advantageous over traditional machine learning models due to their interconnected neuron structure in hidden layers [26]. However, deep learning models are not without limitations, often challenged by hyperparameter sensitivity and risks of overfitting or settling in local optima.

2.2 Hybrid forecasting models: Weight-based and stack-based

Financial time series, including those for agricultural commodities, are typically complex, chaotic, and nonlinear. Directly forecasting raw data often leads to unsatisfactory results, and no single model can comprehensively predict agricultural commodity prices.

To address these challenges, this study introduces a decomposition-reconstruction framework for agricultural price forecasting. It leverages decomposition denoising techniques to simplify the original time series into less complex subsequences, thereby reducing external noise interference. Prior research has explored various decomposition methods, including empirical wavelet transform (EWT) [27], empirical mode decomposition (EMD) [28], ensemble empirical mode decomposition (EEMD) [29], and successive variational mode decomposition (SVMD) [30].

In agricultural price forecasting, hybrid models are generally divided into two categories: weight-based and stack-based models. For instance, [31] employed three denoising methods—singular spectral analysis (SSA), EMD, and variational mode decomposition (VMD)—to reduce noise in original price data. They then used a combination of statistical learning (ARIMA), machine learning (support vector regression - SVR), and deep learning (RNN, GRU, LSTM) models for forecasting agricultural commodity prices. The weights for each model were optimized using artificial bee colony algorithm (ABC). Similarly, [32] applied EWT, SSA, and VMD for data decomposition, followed by individual forecasts using ARIMA, exponential smoothing (ETS), BPNN, and ELM. Particle swarm optimization (PSO) was then used in conjunction with cuckoo search (CS) for assigning weights to these models.

Stack-based models in price forecasting typically use outputs from basic models as features for higher-level models [33]. For instance, [34] utilized convolutional neural networks (CNNs) to extract features from electricity price data, subsequently inputting these features into a BiLSTM for forecasting. However, research on deep learning stack-based models specifically for agricultural commodity prices is limited.

Deep learning models are increasingly recognized for their suitability in forecasting volatile agricultural commodity prices, yet their effectiveness is heavily influenced by hyperparameters. Optimizing these parameters through intelligent algorithms is crucial for enhanced performance [35]. Recent studies have focused on using optimization algorithms for this purpose, such as genetic algorithm (GA) [36], differential evolutionary algorithm (DE) [37], and sparrow search algorithm (SSA) [38]. However, no single algorithm is universally applicable across all fields, as indicated by the ‘no free lunch’ theorem [39]. For example, [40] employed EWT for input feature decomposition and crisscross optimization algorithm (CSO) for LSTM parameter optimization, enhancing the model’s generalization capability. [41] used DE algorithm to optimize LSTM hyperparameters, achieving superior forecasting accuracy. Similarly, [42] developed a hybrid model combining modified ensemble empirical mode decomposition (MEEMD) with LSTM, optimized by improved whale optimization algorithm (IWOA), outperforming 11 benchmark models.

Although these hybrid forecasting frameworks demonstrate promising performance, several critical challenges remain. First, while decomposition techniques such as EMD, EEMD, and VMD have been widely adopted in agricultural price forecasting, they each have specific drawbacks. EMD suffers from mode mixing, and EEMD—while addressing this issue—introduces stochastic noise. VMD improves frequency separation but may face challenges in handling highly nonstationary signals [43]. In contrast, SVMD exhibits superior decomposition capability, particularly in preserving mode integrity, reducing endpoint effects, and effectively isolating meaningful intrinsic components from noisy time series [44]. These advantages make SVMD a more robust and suitable choice for agricultural price data, which are often nonlinear and volatile. Second, deep learning-based hybrid models often involve numerous hyperparameters; however, many existing studies only tune a limited subset, neglecting important factors such as time steps, hidden neuron count, or activation functions, which significantly influence model accuracy and stability [45]. Third, while optimization algorithms like GA, PSO, and DE have been employed to enhance model performance, their standard versions are susceptible to local optima and often inefficient when navigating complex, high-dimensional hyperparameter spaces [46]. These limitations hinder the robustness, efficiency, and generalization of many existing hybrid forecasting systems.

Based on the aforementioned literature reviews, this study introduces an innovative stack-based deep learning system for forecasting agricultural commodity prices, focusing on parameter optimization and feature extraction. The system, named SVMD-MSDBO-CNN-BiLSTM-A, integrates SVMD, MSDBO, CNN, attention mechanism, and BiLSTM networks. SVMD is employed to break down the original commodity price data into simpler subsequences. CNN and attention mechanism work in tandem to extract relevant information and features from the time series data. MSDBO is merged with BiLSTM to create MSDBO-BiLSTM model, designed for precise forecasting of agricultural commodity prices. MSDBO fine-tunes several parameters of BiLSTM model, including learning rate, number of neurons in the hidden layer, nodes in the fully connected layer, time steps, training batch size, and maximum training iterations. Table 1 lists the abbreviations and their corresponding full names of all models used in this study, and Table 2 subsequently presents a comparative analysis of their advantages and disadvantages.

thumbnail
Table 1. List of abbreviations and their full names used in the study.

https://doi.org/10.1371/journal.pone.0337103.t001

thumbnail
Table 2. Advantages and disadvantages of different forecasting models.

https://doi.org/10.1371/journal.pone.0337103.t002

3 Forecasting model

3.1 Framework of developed combined forecasting model

The proposed price forecasting model comprises four main steps, as illustrated in Fig 1.

Step 1: Data collection.

Collect daily future price data for corn and wheat to use as research samples.

Step 2: Data decomposition.

Preprocess the dataset obtained in Step 1. Employ SVMD to decompose the original dataset into several components and divide it into training, validation and test sets.

Step 3: Train MSDBO-CNN-BiLSTM-A models.

This involves initializing the population using improved circular chaos mapping, adopting RSO algorithms for behavioral simulation, and integrating GSA for enhanced optimization. The adaptive bidirectional Gaussian perturbation strategy is applied to avoid local optima. Utilize the training and validation sets from Step 2 to determine the optimal model parameters.

Step 4: Forecasting.

Input the test set data into the trained MSDBO-CNN-BiLSTM-A model to generate the final price forecasting values.

3.2 SVMD

SVMD [30] extends the Variational Mode Decomposition (VMD) method by adaptively extracting intrinsic mode functions without predefining their number. Instead of decomposing a signal into a fixed number of modes as in VMD, SVMD successively extracts modes by introducing additional constraints, eliminating the need to specify the number of modes a priori.

Assume the input signal f(t) can be decomposed into L–th order modes uL(t) and the residual signal fr(t).

(1)

where uL(t) is the L–th order mode; fr(t) is the residual signal, containing the obtained mode and the unprocessed part fu(t), which can be expressed as follows:

(2)

where ui(t) is the i-th order mode and the 1st order mode is found by making fr(t) in be 0. When decomposing the signal to steer the modes predominantly toward the center frequency, the relevant criterion is used as the guiding constraints [47], as shown in the following equation:

(3)

where is the center frequency of the L-th mode and denotes convolution. To minimize spectral overlap between the residual and the current mode, a correlation criterion is introduced:

(4)

where is the impulse response of the filter, its frequency response can be expressed as follows:

(5)

By minimizing the j1 and j2 constraints, the L th order modes may not be effectively distinguished from the first L-1 order modes. Therefore, based on the idea of establishing constraint j2, the frequency response of the used filter shown as follows:

(6)

Thus, the constraints established are expressed as follows:

(7)

When performing the decomposition, the following constraints are established to ensure the signal’s complete reconstruct:

(8)

Therefore, the task of extracting the modal components can be expressed as a constrained minimization problem:

(9)

where α is the parameter for balancing j1, j2 and j3.

3.3 CNN

CNNs, proposed by LeCun et al. [48], are advanced variants of traditional back propagation networks. They are classified by the dimension of convolution: 1D-CNNs for sequences, 2D-CNNs for images, and 3D-CNNs for videos and volumetric data. Key principles of CNNs include local receptive fields, weight sharing, and pooling, which collectively reduce the number of parameters and enhance computational efficiency. A typical CNN consists of three components: the convolutional layer, the pooling layer, and the fully connected layer. The convolutional layer extracts features through convolution operations, while the pooling layer downsamples these features to reduce dimensionality and computational cost. Fig 2 illustrates the structure of a CNN.

(10)

Where pt is the the output value after convolution; tanh is the activation function; xt is the input vector; wt is the weight of the convolution kernel; and bt is the bias of the convolution kernel.

3.4 Attention mechanism

The attention mechanism, inspired by the human brain’s focus system, operates on a principle of probability weighting [49]. Its core concept is to prioritize important information by assigning it higher weights, while less significant information receives lower weights. In the context of forecasting price series, the attention mechanism is incorporated to improve the feature extraction capabilities of CNNs when analyzing decomposed subsequences and residuals. The formula for the attention mechanism is outlined as follows:

(11)(12)(13)

where: hi is the hidden state at time step i from the input sequence. W and b are the learnable weight matrix and bias vector of the scoring network, respectively. ei is the unnormalized attention score for hi. (or wi) is the normalized attention weight assigned to hi. t is the total length of the input sequence. y is the context vector, a weighted representation of the most relevant parts of the input sequence.

3.5 BiLSTM

LSTM, a variant of RNN, was developed by Hochreiter & Schmidhuber to overcome the issue of long-term dependencies that RNN models face [50]. A BiLSTM consists of two LSTM layers, one processing data forwards and the other backwards, hence the term ‘bidirectional’. The architecture of LSTM includes four main components: a forget gate, an input gate, an output gate, and a cell state. The structure of a single-layer LSTM is illustrated in Fig 3.

The sigmoid function is defined as:

(14)

The forgetting gate is expressed as follows:

(15)

The inputting gate is defined as follows:

(16)

The cell state update is a two-step process. First, a candidate cell state gt is created using a layer:

(17)

The state value of the memory cell is written as follows:

(18)

The output of the output gate is expressed as follows:

(19)

The output of the hidden layer is defined as follows:

(20)

The backward LSTM calculation in the BiLSTM structure is similar to the forward LSTM, and the BiLSTM is determined as follows.

(21)(22)

where hf is the output of the forward LSTM network and hb is the output of the reverse LSTM network. The final output of the hidden layer is shown as follows:

(23)

where, σ is the sigmoid activation function; is the hyperbolic tangent activation function; the matrices Wf, Wi, Wc, Wo, wf1, wf2, wb1, wb2 are the recurrent weight matrices; why is the hidden output weight matrix; bf, bi, bc, bo, by are the bias vectors; and yt is the final hidden layer output.

Traditional neural networks often fail to fully utilize time series data for agricultural commodity prices, resulting in suboptimal forecasting performance. In contrast, BiLSTM, with its bi-directional architecture, improves forecasting accuracy by using two LSTM layers working in opposite directions to leverage both past and future information. This helps address data insufficiency issues commonly seen in standard LSTM models [51]. The structural design of BiLSTM model is depicted in Fig 4.

3.6 Improved dung beetle optimizer

The Dung Beetle Optimizer (DBO) is a swarm intelligence algorithm inspired by dung beetles’ behaviors such as rolling, dancing, stealing, foraging, and reproduction [52]. While the original DBO exhibits promising global search ability, it suffers from slow convergence and imbalanced exploration–exploitation. To address these limitations, we propose a an improved MSDBO with the following enhancements:

(a) Improved circle chaotic mapping.

Chaotic mapping is known for its thoroughness and randomness in exploring solutions. Utilizing chaotic mapping functions to generate initial positions for population individuals in an algorithm can improve the uniformity and coverage of the initial solution space. This technique has been successfully applied in areas such as artificial neural networks, modeling of natural phenomena, and nonlinear circuits, demonstrating significant effectiveness [53]. Common types of chaos maps include logistic, tent, and circle chaos maps. Among these, circular chaotic mapping is particularly notable for its stability and comprehensive coverage of chaotic values [54]. The formula for the improved circular chaotic mapping is detailed in Eq (24).

(24)

Where, α = 0.4204, β=0.0305.

(b) Fusion RSO and DBO.

DBO has been augmented by incorporating rat swarm optimization (RSO) algorithm, specifically to enhance the ball-rolling behavior of dung beetles. RSO algorithm, inspired by the chasing and fighting behaviors observed in rat swarms, is a recent development in group intelligence algorithms [55]. Its simplicity and straightforward structure, devoid of complex mathematical operations, make it effective in both global searching and local exploitation. By adopting RSO’s chasing and fighting strategies in place of the dung beetle’s traditional ball-rolling and dancing behaviors, the optimization process is further improved. The prey-chasing behavior integral to RSO is concisely represented in Eqs (25)(27).

(25)(26)(27)

Where: xb(t) is the current optimal position of the rat population; xi(t) is the current optimal position of the rat population; xi a is the random number in [1, 5]; rand is the random number in [0, 1]; t is the current iteration number; and Tmax is the maximum number of iterations.

In the modeling of RSO, the fight process between the rats and prey is defined in Eq (28).

(28)

Where: xi(t + 1) is the location of the –th rat population.

(c) Fusion of GSA and DBO.

The optimization capabilities of DBO have been enhanced through the integration of golden sine algorithm (GSA), a novel metaheuristic proposed by Tanyildizi and Demir in 2017 [56]. This algorithm uses a sine function to model optimization problems. By incorporating the golden section coefficient, GSA effectively searches around potential solutions, enhancing the local exploitation capability and accelerating the convergence process of the overall algorithm. This method replaces the stealing behavior typically observed in dung beetles. The adaptation of this modified local search strategy within the algorithm is outlined in Eq (29).

(29)

Where, R1 and R2 are random numbers between [0, 2π ] and [0, π], respectively. R1 denotes the distance moved, and R2 denotes the direction of movement; denotes the position of the i-th dung beetle optimal individual at the t-th iteration; x1, and x2 are the coefficients obtained by introducing the number of golden divisors, which can reduce the search space and make the individual gradually close to the optimal value, these parameters can be expressed as x1, x2 , ; and τ is the golden section number, with the utilization of irrational number 3.

(d) Adaptive bidirectional gaussian perturbation strategy.

To prevent the algorithm from getting trapped in local optima, an adaptive bidirectional gaussian perturbation strategy is employed [57]. This strategy updates the positions of the current individuals in the algorithm, allowing it to effectively escape from local optimal solutions. The mechanism of this perturbation is detailed in Eq (30). Additionally, Gaussian perturbation factor U is defined in Eq (31).

(30)(31)

3.7 Pseudocode of the proposed MSDBO

To improve accessibility, the overall procedure of the proposed MSDBO is summarised in Algorithm 1. The four core behaviours—(i) Ball rolling (RSO-enhanced), (ii) Breeding, (iii) Foraging, and (iv) Stealing (GSA-enhanced)—are executed iteratively. Optional modules such as improved circle chaotic initialization and adaptive bi-directional Gaussian perturbation are inserted at the corresponding stages.

Algorithm 1. Improved Multi-Strategy Dung Beetle Optimizer (MSDBO).

3.8 Convergence evaluation of MSDBO

Four benchmark functions, commonly used for evaluating algorithm performance, were employed in this work. These functions were selected based on the study by GhaemiDizaji et al.[58]. The convergence behavior of the proposed multi-strategy enhanced dung beetle optimizer (MSDBO) on these benchmark functions is illustrated in Fig 5. Among them, F1 and F5 are unimodal functions, shown in Fig 5(a) and Fig 5(b), respectively, while F9 and F22 are multimodal functions, presented in Fig 5(c) and Fig 5(d).

thumbnail
Fig 5. Convergence curves for two unimodal benchmark functions and two multimodal functions.

https://doi.org/10.1371/journal.pone.0337103.g005

For the unimodal functions F1 and F5, MSDBO demonstrates a significantly faster convergence rate during the initial iterations, rapidly approaching the optimal solution. This performance clearly surpasses that of the original DBO and the GA, indicating the superior global exploration and local exploitation capabilities of MSDBO. In the case of the multimodal functions F9 and F22, MSDBO continues to exhibit robust optimization behavior, effectively escaping local optima and maintaining stable convergence. These results further confirm the algorithm’s strong local search ability and robustness in avoiding premature convergence.

4 Experimental

4.1 Data description

The Chicago Board of Trade (CBOT) provides globally recognized benchmark prices for bulk agricultural commodities, reflecting market supply-demand dynamics, macroeconomic shocks, and investor sentiment. These price data are essential for analyzing nonlinear price movements and capturing high-frequency volatility, offering a robust foundation for understanding price dynamics, improving risk assessment, and supporting financial market stability [31]. In this context, this study utilizes daily closing price data of corn and wheat traded on the CBOT to conduct an in-depth analysis . The dataset spans from January 1970 to May 2021, comprising 12,850 price observations for corn and 12,856 for wheat. The data is segmented into three parts: the first 80% is designated as the training set, the subsequent 10% as the validation set, and the final 10% as the test set. Fig 6(a) and 6(b) show the corn and wheat price datasets, respectively. Table 3 summarises the main statistical information of the data.

thumbnail
Table 3. Descriptive statistics for agricultural commodity price prediction.

https://doi.org/10.1371/journal.pone.0337103.t003

4.2 Evaluation criteria

The model’s forecasting performance is assessed using four key metrics: root mean square error (RMSE), mean absolute error (MAE), mean absolute percentage error (MAPE), and a directional accuracy Dstat.

(32)(33)(34)(35)

Where N is the number of output samples; denotes the forecasting value; yi denotes the actual value; and =1 only when and = 0 otherwise.

4.3 Decomposition and denoising

To mitigate the nonlinearity and high volatility in the price data, thereby enhancing prediction accuracy, the original corn data was decomposed using SVMD, resulting in 9 Intrinsic Mode Functions (IMFs) and a residual series. Similarly, the wheat data decomposition yielded 10 IMFs and a residual series. Figs 7 and 8 display the original corn and wheat data along with their respective SVMD decompositions.

5 Forecasting analysis

5.1 Comparable models and parameter setting

To assess the accuracy and stability of the proposed model, its performance is compared against nine benchmark models, including variants based on SVMD, EMD, and EEMD combined with BPNN, LSTM, BiLSTM, CNN-BiLSTM, and optimization strategies such as DBO and MSDBO.

BPNN, LSTM, and BiLSTM were also included as standalone benchmark models. Their parameters were carefully tuned, with BiLSTM time steps set between 1 and 48, hidden layers set to two layers, up to 200 training epochs, 1–128 hidden neurons, batch sizes from 1 to 120, and learning rates between 0.001 and 0.5. Full parameter settings are listed in Table 4.

5.2 Comparison of the proposed model with other established comparable models

The proposed SVMD-BiLSTM model is evaluated using RMSE, MAPE, MAE, and Dstat, and compared with nine alternative models. Forecasting results for corn and wheat are presented in Figs 9 and 10, while corresponding error metrics are shown in Figs 11 and 12. Detailed performance comparisons are summarized in Tables 5 and 6, followed by an in-depth discussion.

thumbnail
Fig 9. Performance evaluation of forecast combination models for corn.

https://doi.org/10.1371/journal.pone.0337103.g009

thumbnail
Fig 10. Performance evaluation of forecast combination models for wheat.

https://doi.org/10.1371/journal.pone.0337103.g010

thumbnail
Fig 11. RMSE, MAE, , and MAPE of combination models for corn price forecasting.

https://doi.org/10.1371/journal.pone.0337103.g011

thumbnail
Fig 12. RMSE, MAE, , and MAPE of combination models for wheat price forecasting.

https://doi.org/10.1371/journal.pone.0337103.g012

thumbnail
Table 5. Comparative of forecasting performance by different models for corn prices.

https://doi.org/10.1371/journal.pone.0337103.t005

thumbnail
Table 6. Comparative of forecasting performance by different models for wheat prices.

https://doi.org/10.1371/journal.pone.0337103.t006

(a) SVMD-BiLSTM model shows superior results on both datasets compared to SVMD-BPNN and SVMD-LSTM, achieving the lowest RMSE, MAE, MAPE, and maximum Dstat. Specifically, for the corn dataset, the MAPE values for SVMD-BPNN, SVMD-LSTM, and SVMD-BiLSTM are 1.9051%, 1.1969%, and 0.8312%, respectively. These results establish BiLSTM as the most effective individual forecasting model in this comparison.

(b) A comparison between SVMD-CNN-BiLSTM and SVMD-BiLSTM on both datasets shows that the former consistently outperforms the latter across all evaluation metrics. For instance, on the corn dataset, SVMD-CNN-BiLSTM achieves an RMSE of 4.8075, MAE of 2.952, MAPE of 0.7267%, and Dstat of 74.2991%. These results highlight the effectiveness of CNN in extracting features and enhancing forecasting accuracy.

(c) Further improvement is observed with the incorporation of an attention mechanism in SVMD-CNN-BiLSTM-A, which leads to better performance on both datasets. On the wheat dataset, it achieves an RMSE of 3.9701, MAE of 3.1016, MAPE of 0.6377%, and Dstat of 79.7665%, confirming the contribution of the attention mechanism in improving model precision when integrated with CNN.

(d) When comparing SVMD-CNN-BiLSTM-A with SVMD-MSDBO-CNN-BiLSTM-A, the latter yields the best performance on both datasets. On the corn dataset, RMSE, MAE, and MAPE are reduced by 7.90%, 9.76%, and 10.15%, respectively, while Dstat increases by 0.96%. These findings demonstrate that integrating optimization algorithms with deep hybrid models significantly enhances forecasting effectiveness.

5.3 Comparison of different decomposition methods

This section evaluates the performance of various signal decomposition techniques—including EMD, EEMD, EWT, VMD, and SVMD—within a consistent forecasting framework. Fig 13 illustrates the scatter plot of the forecasting and true values for different decomposition methods based on the same forecasting framework. Tables 7 and 8 demonstrate that SVMD consistently outperforms other decomposition methods in forecasting accuracy for both corn and wheat prices, while Table 9 lists the corresponding p-values from Wilcoxon rank-sum tests, statistically validating these improvements.

thumbnail
Fig 13. Scatter plots of actual and forecasted values of different decomposition methods for corn and wheat datasets.

https://doi.org/10.1371/journal.pone.0337103.g013

thumbnail
Table 7. Comparative results of by different decomposition methods for corn prices.

https://doi.org/10.1371/journal.pone.0337103.t007

thumbnail
Table 8. Comparative results of by different decomposition methods for wheat prices.

https://doi.org/10.1371/journal.pone.0337103.t008

(b) Similarly, for the wheat dataset, SVMD again delivers the best performance across all metrics, with RMSE of 3.9701, MAE of 3.1016, MAPE of 0.6377%, and Dstat of 79.7665%. Compared with the EEMD-based model, SVMD improves RMSE by 19.98%, MAE by 20.73%, MAPE by 21.40%, and Dstat by 14.53%. When compared to the newly included EWT and VMD models, SVMD still offers noticeable advantages, further validating its effectiveness.

(c) These findings highlight that the SVMD-based decomposition approach provides the most significant improvements in forecasting performance. Its superior ability to extract relevant features contributes to lower prediction errors and better directional accuracy, making it highly effective for agricultural futures forecasting tasks.

(d) The statistical significance of performance differences was rigorously examined using the Wilcoxon rank-sum test, with all p-values documented in Table 9. A p-value exceeding the 0.05 threshold indicates no statistically significant difference between decomposition methods, whereas values below 0.05 confirm significant performance disparities. Our analysis demonstrates SVMD’s decisive superiority in agricultural markets: extreme significance over EMD (corn: p = 3.17 × 10−5<0.05; wheat: ) and consistent advantages against VMD (corn: p = 0.0147<0.05; wheat: ). These systematically robust outcomes validate SVMD’s dual strengths in agricultural price forecasting - numerical dominance and statistical rigor.

thumbnail
Table 9. Comparison of P-values between SVMD-MSDBO-CNN-BiLSTM-A and other methods.

https://doi.org/10.1371/journal.pone.0337103.t009

5.4 Comparison of different optimization algorithm

This section analyzes the impact of optimization method selection on forecasting accuracy, with particular focus on demonstrating the enhanced performance of the improved DBO algorithm. Fig 14 illustrates the scatter plot of the forecasting and true values for different optimization methods based on the same forecasting framework. Tables 10 and 11 show the forecasting results of the forecasting model framework based on three different optimization methods for two different datasets. By analyzing the results in Tables 10 and 11, the analysis can be summarized as follows.

thumbnail
Fig 14. Scatter plots of actual and forecasted values of different optimization algorithms for BiLSTM models in corn and wheat datasets.

https://doi.org/10.1371/journal.pone.0337103.g014

thumbnail
Table 10. Comparison of forecasting performance by different models for corn prices.

https://doi.org/10.1371/journal.pone.0337103.t010

thumbnail
Table 11. Comparison of forecasting performance by different models for wheat prices.

https://doi.org/10.1371/journal.pone.0337103.t011

(a) Tables 10 and 11 demonstrate that the MSDBO based forecasting framework achieves minimum values for RMSE, MAE and MAPE and maximum values for Dstat on both data. For example, on the corn dataset, for MAPE, the MSDBO based forecasting framework obtains an reduction of 8.68% and 6.05% compared to the GA and DBO based forecasting frameworks, respectively. These results suggest that MSDBO can better optimize the model parameters and improve the forecasting accuracy of the model.

(b) Further examination of Tables 10 and 11 demonstrates that MSDBO-based framework outperforms DBO-based framework. This means that four strategies can be used to improve DBO. For example, within the wheat dataset, the application of MSDBO results in a 10.34% reduction in RMSE, an 8.81% reduction in MAE, a 9.03% reduction in MAPE, and a 3.54% improvement in Dstat. Besides that, the efficacy of optimization algorithms combined with stack-based models in price forecasting is further demonstrated.

5.5 Comparison of the forecasting speed

In this section, we analyze the forecasting speed of different models under a unified CNN-BiLSTM architecture. The experimental results are shown in Table 12. The main findings can be summarized as follows:

thumbnail
Table 12. Comparison of prediction efficiency by different models.

https://doi.org/10.1371/journal.pone.0337103.t012

(a) The SVMD-MSDBO-CNN-BiLSTM-A model is selected as the performance benchmark, achieving execution times of 405.33 seconds and 426.85 seconds on the Corn and Wheat datasets, respectively. Despite incorporating multilevel signal decomposition and metaheuristic optimization, its overall runtime remains within an acceptable range, demonstrating good practical feasibility.

(b) Compared to traditional signal decomposition methods, the SVMD-MSDBO-CNN-BiLSTM-A model exhibits superior computational efficiency. On the Corn dataset, it achieves runtime reductions of 8.34%, 6.96%, and 4.33% over EMD, EEMD, and VMD-based models, respectively. On the Wheat dataset, the corresponding reductions are 8.80%, 6.39%, and 4.78%.

(c) Although the EWT-MSDBO-CNN-BiLSTM-A model demonstrates slightly faster execution times—improving by 0.61% and 3.65% on the Corn and Wheat datasets respectively—it shows inferior forecasting performance. This indicates that although EWT offers modest computational advantages, it is less capable of capturing deep temporal features, thus reducing overall prediction accuracy.

(d) When compared with models employing alternative optimization algorithms (such as GA and DBO), SVMD-MSDBO-CNN-BiLSTM-A achieves substantial runtime reductions. On the Wheat dataset, it reduces time by 26.71% and 17.59% compared to SVMD-GA-CNN-BiLSTM-A and SVMD-DBO-CNN-BiLSTM-A, respectively. On the Corn dataset, the improvements are 26.72% and 17.27%, respectively. These results demonstrate the effectiveness of MSDBO in balancing forecasting accuracy and execution efficiency.

(e) Structurally simplified versions—such as SVMD-MSDBO-BiLSTM and SVMD-MSDBO-CNN-BiLSTM—achieve faster runtime, reducing execution time by 24.72% and 15.07% on the Corn dataset, respectively. However, their forecasting performance is significantly degraded, indicating that simplification comes at the cost of predictive power and limits their suitability for high-stakes applications.

5.6 Model interpretability analysis

To enhance the interpretability of our hybrid model in agricultural forecasting, we analyze the attention weights within the BiLSTM-Attention module. This mechanism dynamically assigns weights to input features—the original price series, decomposed Intrinsic Mode Functions (IMFs), and residual—revealing the model’s prioritization. High-frequency IMFs capture short-term noise (e.g., weather disruptions or speculation), while low-frequency IMFs and residuals reflect long-term trends, including seasonal cycles and economic influences, bridging model outputs to market drivers.

We first assess the attention weight distribution of the top-performing SVMD-MSDBO-CNN-BiLSTM-A model on corn and wheat datasets. Weights, computed as softmax-normalized averages over test samples and timesteps (summing to 1), highlight the relative importance of the original price, IMFs (from high to low frequencies), and residual. We then compare decomposition variants—EMD-MSDBO-CNN-BiLSTM-A and SVMD-MSDBO-CNN-BiLSTM-A—across datasets to evaluate how decomposition affects feature prioritization. EMD (empirical) and SVMD (variational) were chosen for their distinct approaches—EMD’s data-driven nature versus SVMD’s optimized separation—to test weight consistency robustly.

Fig 15 shows attention weight distributions across datasets and methods, with subplots: (a) Corn with EMD, (b) Corn with SVMD, (c) Wheat with EMD, and (d) Wheat with SVMD.

thumbnail
Fig 15. Comparative attention weights: corn and wheat with EMD and SVMD.

https://doi.org/10.1371/journal.pone.0337103.g015

For SVMD-MSDBO-CNN-BiLSTM-A:

(a) In corn (EMD), weights are uniform, with the original price at 0.181, followed by high-frequency IMFs (e.g., IMF1 0.130, IMF3 0.141), tapering to IMF8 (0.046), indicating focus on short-term noise like weather shocks, aligning with corn’s volatility.

(b) In corn (SVMD), weights decline steeply, with original price at 0.080, peaking at IMF1 (0.167) and IMF3 (0.165), decaying to IMF10 (0.026), emphasizing noise filtering and trend shift.

(c) In wheat (EMD), decay is gradual (original 0.165, IMF1 0.245), with less high-frequency focus (e.g., IMF2 0.070), reflecting stability and trend drivers like seasonal cycles.

(d) In wheat (SVMD), weights concentrate on original price (0.119) and mid-to-low IMFs (e.g., IMF2 0.123, IMF8 0.104), with residual (IMF11 0.097), focusing on structural trends.

Comparing methods, EMD-MSDBO-CNN-BiLSTM-A shows uniform distributions: corn [(a)] from 0.181 (original) to 0.046 (IMF8); wheat [(c)] from 0.165 (original) to 0.048 (IMF8), suggesting mode mixing overemphasizes noise. SVMD-MSDBO-CNN-BiLSTM-A [(b) and (d)] exhibits steeper profiles, highlighting SVMD’s advantage in reducing aliasing and enhancing trend focus.

Across datasets, low-frequency IMFs and residuals exceed 0.4 aggregate, reinforcing trend contributions (e.g., seasonal and economic factors). Differences include corn’s high-frequency bias (volatility) versus wheat’s low-frequency dominance (stability), underscoring SVMD’s edge in volatile scenarios.

5.7 Practical implications

The high predictive accuracy and interpretability of the proposed SVMD-MSDBO-CNN-BiLSTM-A framework offer significant practical benefits for key stakeholders in agricultural commodity markets. The attention-based interpretability mechanism offers critical insights into the model’s decision-making process. As detailed in Sect 5.6 and illustrated in Fig 15, the framework demonstrates a distinct ability to discern and leverage market signals across different frequency domains. Specifically, it differentiates between high-frequency components such as IMF1 through IMF3, which represent short-term market noise, and low-frequency along with residual components that reflect underlying long-term trends. This capability enables tailored decision-support across sectors.

For traders and financial institutions:

(a) Short-term trading signals: the model’s facility in capturing high-frequency fluctuations (e.g., IMF1–IMF3) allows traders to detect signals from transient market disturbances, speculative activity, or sudden supply-demand shifts. Predictions coupled with elevated attention weights on these components can inform short-term trading strategies and optimize timing in futures contracts.

(b) Portfolio risk management: accurate volatility forecasts can enhance Value-at-Risk (VaR) models. Institutions can dynamically adjust hedging strategies when the model indicates heightened attention to high-frequency components, serving as an early warning for potential volatility episodes.

For Farmers and agricultural producers:

(a) Optimal sales timing: the model’s emphasis on mid- to long-term trends (e.g., low-frequency IMFs and residues) helps identify seasonal and macroeconomic patterns. As Fig 15 illustrates, the pronounced attention weight on trend components in wheat data lends reliability to medium-term forecasts, supporting decisions on selling windows and storage strategies to maximize profit.

(b) Production and inventory planning: forecasts of sustained upward trends may justify expanding cultivation acreage or investing in additional storage capacity, providing a data-driven basis for operational planning.

For policymakers and government agencies:

(a) Food security early-warning systems: the model can be integrated into early-warning systems that monitor long-term trend components. Predicting critical price inflation enables proactive measures such as strategic reserve releases or trade policy adjustments to stabilize markets.

(b) Evidence-based policy formulation: model interpretability offers quantitative evidence on drivers of price volatility. Persistent attention to high-frequency components (e.g., IMF1) suggests short-term speculative influences, calling for market oversight policies. Conversely, dominance of low-frequency drivers indicates structural or supply-side issues, warranting interventions in production support, climate adaptation, or infrastructure.

6 Conclusion

This study introduces a novel hybrid forecasting model for agricultural commodity prices, integrating the SVMD denoising technique, the BiLSTM deep learning method with an attention mechanism, and the improved MSDBO optimization algorithm. The BiLSTM-Attention-CNN model leverages BiLSTM units, an attention mechanism for feature prioritization, and a CNN component to capture nonlinear dynamics and temporal dependencies in daily price data for corn and wheat from the Chicago Board of Trade (CBOT), evaluated against ten benchmark configurations. The results demonstrate enhanced forecasting accuracy, surpassing existing methods, and provide valuable insights into price dynamics through the attention mechanism, supporting risk management, policy-making, and market stability in volatile global commodity markets. The key findings of this research are illustrated as follows:

(a) SVMD-MSDBO-CNN-BiLSTM-A model consistently shows lower RMSE, MAE, and MAPE values, and high Dstat in both corn and wheat datasets, the four values achieved on corn data are 3.5701, 2.4611, 0.6169% and 73.9097% respectively, indicating its superior forecasting capability compared to other models.

(b) On both datasets, SVMD-CNN-BiLSTM-A achieves better performance compared to SVMD-BiLSTM. On the corn data, RMSE, MAE and MAPE achieved a reduction of 21.58%, 18.58%, 17.40% and Dstat achieved an improvement of 2.00%. This result confirms that incorporating CNN for feature extraction and the attention mechanism for weight assignment significantly improves the model’s forecasting accuracy.

(c) Utilizing EMD, EEMD, EWT, VMD and SVMD for data decomposition, SVMD-based hybrid forecasting model (MSDBO-CNN-BiLSTM-A) achieves the best results, for example, on the wheat data, the smallest RMSE, MAE and MAPE, and the largest Dstat were obtained, which were 3.9701, 3.1016, 0.6377% and 79.7665%, respectively. The results show that the importance of data decomposition in enhancing forecasting accuracy.

(d) On both datasets, the MSDBO-based prediction model obtained better prediction performance. For instance, on wheat data, MAPE obtained 13.78% and 10.34% reduction, respectively. This result indicates that MSDBO outperforms both GA and DBO in optimizing BiLSTM model’s hyperparameters, demonstrating the efficacy of the improved DBO method. And furthermore, the effectiveness of optimization algorithms combined with stack-based models in price forecasting is demonstrated.

(e) Across both datasets, the MSDBO-based forecasting model consistently delivered superior predictive performance. For instance, on the wheat dataset, the MAPE was reduced by 13.78% and 10.34% compared to models optimized using GA and DBO, respectively. These results underscore the enhanced capability of MSDBO in fine-tuning the hyperparameters of BiLSTM architectures, validating the effectiveness of the improved DBO strategy. Furthermore, they highlight the strong potential of integrating advanced metaheuristic optimization techniques with stacked deep learning models for accurate and robust price forecasting.

(f) Across datasets, low-frequency IMFs and residuals contribute over 0.4 of aggregate attention weight, highlighting their role in capturing long-term trends (e.g., seasonal and macroeconomic factors). Corn shows a higher high-frequency bias (volatility adaptation), while wheat exhibits low-frequency dominance (stability), confirming SVMD’s robust interpretability under varying market conditions. This capability offers clear practical value: traders can exploit high-frequency signals for short-term strategies, farmers can rely on trend components for sales timing and production planning, and policymakers can use frequency dominance patterns to guide market stabilization and food security interventions.

While the proposed SVMD-MSDBO-CNN-BiLSTM-A model demonstrates robust performance in forecasting agricultural commodity prices, several limitations warrant consideration. The current attention mechanism provides valuable insights into feature prioritization but lacks a comprehensive quantification of interactions among predictive variables—such as market trends, seasonality, and macroeconomic indicators—or their individual contributions to price movements. Additionally, the model does not incorporate exogenous factors, such as weather conditions, geopolitical events, or global supply chain disruptions, which are critical drivers of commodity price volatility and remain unaccounted for in the present framework. Furthermore, the generalizability of the model across diverse economic systems, such as energy or metals markets, remains untested, potentially limiting its adaptability to varying market dynamics.

Future research should address these limitations to enhance model efficacy and interpretability. A detailed exploration of the relative importance of predictive features, including market trends, seasonal cycles, and macroeconomic indicators, could be pursued using advanced interpretability techniques, such as SHAP (SHapley Additive exPlanations) or LIME (Local Interpretable Model-agnostic Explanations), to disentangle their interactive effects on price fluctuations. Integrating exogenous variables—e.g., weather data, geopolitical risk indices, and supply chain metrics—into the model framework could further improve forecasting accuracy by capturing external influences. Moreover, extending the methodology to other commodities or financial instruments, such as energy or metals markets, would validate its robustness and adaptability across diverse economic contexts, broadening its practical applicability and providing a foundation for cross-market predictive analytics.

References

  1. 1. Khalfaoui R, Goodell JW, Mefteh-Wali S, Chishti MZ, Gozgor G. Impact of climate risk shocks on global food and agricultural markets: a multiscale and tail connectedness analysis. International Review of Financial Analysis. 2024;93:103206.
  2. 2. Wang J, Tang W. Construction and analysis of chinese macro-financial stability index. Comput Econ. 2024;66(2):1625–46.
  3. 3. Urak F, Bilgic A, Florkowski WJ, Bozma G. Confluence of COVID-19 and the Russia-Ukraine conflict: effects on agricultural commodity prices and food security. Borsa Istanbul Review. 2024;24(3):506–19.
  4. 4. Anderson K. Trade-related food policies in a more volatile climate and trade environment. Food Policy. 2022;109:102253.
  5. 5. Kumar P, Kumar Singh R. Strategic framework for developing resilience in Agri-Food Supply Chains during COVID 19 pandemic. International Journal of Logistics Research and Applications. 2021;25(11):1401–24.
  6. 6. Hu X, Zhu B, Zhang B, Zhou S. Do internal and external risk spillovers of the food system matter for national food security?. Economic Modelling. 2024;136:106747.
  7. 7. Woode JK, Owusu Junior P, Idun AA-A, Kawor S, Bambir J, Adam AM. Time-varying and frequency-based spillover connectedness between cryptocurrencies and non-ferrous industrial metals in light of market plummets. Comput Econ. 2024;66(3):2225–64.
  8. 8. Yang T, Dong Q, Du M, Du Q. Geopolitical risks, oil price shocks and inflation: evidence from a TVP–SV–VAR approach. Energy Economics. 2023;127:107099.
  9. 9. Zhao J, Wang B, Dong K, Shahbaz M, Ni G. How do energy price shocks affect global economic stability? Reflection on geopolitical conflicts. Energy Economics. 2023;126:107014.
  10. 10. Huchet-Bourdon M. Agricultural commodity price volatility: an overview. OECD Food, Agriculture and Fisheries Papers. 2011;
  11. 11. Sharma R, Kamble SS, Gunasekaran A, Kumar V, Kumar A. A systematic literature review on machine learning applications for sustainable agriculture supply chain performance. Computers & Operations Research. 2020;119:104926.
  12. 12. Makkonen A, Vallström D, Uddin GS, Rahman ML, Haddad MFC. The effect of temperature anomaly and macroeconomic fundamentals on agricultural commodity futures returns. Energy Economics. 2021;100:105377.
  13. 13. Zhang Y, Ding S, Scheffel EM. A key determinant of commodity price co-movement: the role of daily market liquidity. Economic Modelling. 2019;81:170–80.
  14. 14. Roy RP, Sinha Roy S. Commodity futures prices pass-through and monetary policy in India: does asymmetry matter?. The Journal of Economic Asymmetries. 2022;25:e00229.
  15. 15. Umar Z, Gubareva M, Teplova T. The impact of Covid-19 on commodity markets volatility: analyzing time-frequency relations between commodity prices and coronavirus panic levels. Resour Policy. 2021;73:102164. pmid:36567729
  16. 16. Tsang YP, Fan Y, Feng ZP, Li Y. Examining supply chain vulnerability via an analysis of ESG-prioritized firms amid the Russian-Ukrainian conflict. Journal of Cleaner Production. 2024;434:139754.
  17. 17. Zhang Y, chang H, Saliba C, Hasnaoui A. Metallic natural resources commodity prices volatility in the pandemic: evidence for silver, platinum, and palladium. Resources Policy. 2022;78:102924.
  18. 18. Kriechbaumer T, Angus A, Parsons D, Rivas Casado M. An improved wavelet–ARIMA approach for forecasting metal prices. Resources Policy. 2014;39:32–41.
  19. 19. Ji Z, Niu D, Li M, Li W, Sun L, Zhu Y. A three-stage framework for vertical carbon price interval forecast based on decomposition–integration method. Applied Soft Computing. 2022;116:108204.
  20. 20. Du P, Guo J, Sun S, Wang S, Wu J. Multi-step metal prices forecasting based on a data preprocessing method and an optimized extreme learning machine by marine predators algorithm. Resources Policy. 2021;74:102335.
  21. 21. Xiao C, Xia W, Jiang J. Stock price forecast based on combined model of ARI-MA-LS-SVM. Neural Comput & Applic. 2020;32(10):5379–88.
  22. 22. Wang J, Lv M, Li Z, Zeng B. Multivariate selection-combination short-term wind speed forecasting system based on convolution-recurrent network and multi-objective chameleon swarm algorithm. Expert Systems with Applications. 2023;214:119129.
  23. 23. Frasca M, Torre DL, Pravettoni G, Cutica I. Combining convolution neural networks with long-short term memory layers to predict Parkinson’s disease progression. International Transactions in Operational Research. 2023.
  24. 24. Zhang S, Luo J, Wang S, Liu F. Oil price forecasting: a hybrid GRU neural network based on decomposition–reconstruction methods. Expert Systems with Applications. 2023;218:119617.
  25. 25. Li Y, Sun K, Yao Q, Wang L. A dual-optimization wind speed forecasting model based on deep learning and improved dung beetle optimization algorithm. Energy. 2024;286:129604.
  26. 26. Chang Z, Zhang Y, Chen W. Electricity price prediction based on hybrid model of adam optimized LSTM neural network and wavelet transform. Energy. 2019;187:115804.
  27. 27. Gilles J. Empirical wavelet transform. IEEE Trans Signal Process. 2013;61(16):3999–4010.
  28. 28. Huang NE, Shen Z, Long SR, Wu MC, Shih HH, Zheng Q, et al. The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis. Proc R Soc Lond A. 1998;454(1971):903–95.
  29. 29. Wu Z, Huang NE. Ensemble empirical mode decomposition: A noise-assisted data analysis method. Adv Adapt Data Anal. 2009;01(01):1–41.
  30. 30. Nazari M, Sakhaei SM. Successive variational mode decomposition. Signal Processing. 2020;174:107610.
  31. 31. Sarris A, Conforti P, Prakash A. The use of organized commodity markets to manage food import price instability and risk. Agricultural Economics. 2010;42(1):47–64.
  32. 32. Zeng L, Ling L, Zhang D, Jiang W. Optimal forecast combination based on PSO-CS approach for daily agricultural future prices forecasting. Applied Soft Computing. 2023;132:109833.
  33. 33. Wu B, Wang L, Zeng Y-R. Interpretable wind speed prediction with multivariate time series and temporal fusion transformers. Energy. 2022;252:123990.
  34. 34. Kim J, Oh S, Kim H, Choi W. Tutorial on time series prediction using 1D-CNN and BiLSTM: A case example of peak electricity demand and system marginal price prediction. Engineering Applications of Artificial Intelligence. 2023;126:106817.
  35. 35. Gülmez B. Stock price prediction with optimized deep LSTM network with artificial rabbits optimization algorithm. Expert Systems with Applications. 2023;227:120346.
  36. 36. Chen Y, Dong Z, Wang Y, Su J, Han Z, Zhou D, et al. Short-term wind speed predicting framework based on EEMD-GA-LSTM method under large scaled wind history. Energy Conversion and Management. 2021;227:113559.
  37. 37. Zhu H, Wang J, Liu C, Shi W, Cai Q. An MBD-driven order remaining completion time prediction method based on SSA-BiLSTM in the IoT-enabled manufacturing workshop. International Journal of Production Research. 2023;62(10):3559–84.
  38. 38. Mayer DG, Kinghorn BP, Archer AA. Differential evolution – an easy and efficient evolutionary algorithm for model optimisation. Agricultural Systems. 2005;83(3):315–28.
  39. 39. Wolpert DH, Macready WG. No free lunch theorems for optimization. IEEE Trans Evol Computat. 1997;1(1):67–82.
  40. 40. Meng A, Wang P, Zhai G, Zeng C, Chen S, Yang X, et al. Electricity price forecasting with high penetration of renewable energy using attention-based LSTM network trained by crisscross optimization. Energy. 2022;254:124212.
  41. 41. Peng L, Liu S, Liu R, Wang L. Effective long short-term memory with differential evolution algorithm for electricity price prediction. Energy. 2018;162:1301–14.
  42. 42. Yang S, Chen D, Li S, Wang W. Carbon price forecasting based on modified ensemble empirical mode decomposition and long short-term memory optimized by improved whale optimization algorithm. Sci Total Environ. 2020;716:137117. pmid:32074939
  43. 43. Sun C, Pei M, Cao B, Chang S, Si H. A study on agricultural commodity price prediction model based on secondary decomposition and long short-term memory network. Agriculture. 2023;14(1):60.
  44. 44. Agarwal S, Sharma S, Faisal KN, Sharma RR. Time-series forecasting using SVMD-LSTM: a hybrid approach for stock market prediction. Journal of Probability and Statistics. 2025;2025(1):9464938.
  45. 45. Song H, Mu S, Tuxunaili Y, Wang S. Peanut price prediction based on GWO-BiLSTM. In: Proceedings of the 2024 7th International Conference on Signal Processing and Machine Learning. 2024. p. 42–8. https://doi.org/10.1145/3686490.3686496
  46. 46. Mubarak H, Abdellatif A, Ahmad S, Zohurul Islam M, Muyeen SM, Abdul Mannan M, et al. Day-Ahead electricity price forecasting using a CNN-BiLSTM model in conjunction with autoregressive modeling and hyperparameter optimization. International Journal of Electrical Power & Energy Systems. 2024;161:110206.
  47. 47. Ma H, Xu Y, Wang J, Song M, Zhang S. SVMD coupled with dual-threshold criteria of correlation coefficient: a self-adaptive denoising method for ship-radiated noise signal. Ocean Engineering. 2023;281:114931.
  48. 48. Lecun Y, Bottou L, Bengio Y, Haffner P. Gradient-based learning applied to document recognition. Proc IEEE. 1998;86(11):2278–324.
  49. 49. Lin Y, Chen K, Zhang X, Tan B, Lu Q. Forecasting crude oil futures prices using BiLSTM-Attention-CNN model with Wavelet transform. Applied Soft Computing. 2022;130:109723.
  50. 50. Hochreiter S, Schmidhuber J. Long short-term memory. Neural Comput. 1997;9(8):1735–80. pmid:9377276
  51. 51. Xiong J, Peng T, Tao Z, Zhang C, Song S, Nazir MS. A dual-scale deep learning model based on ELM-BiLSTM and improved reptile search algorithm for wind power prediction. Energy. 2023;266:126419.
  52. 52. Xue J, Shen B. Dung beetle optimizer: a new meta-heuristic algorithm for global optimization. J Supercomput. 2022;79(7):7305–36.
  53. 53. Varol Altay E, Alatas B. Bird swarm algorithms with chaotic mapping. Artif Intell Rev. 2019;53(2):1373–414.
  54. 54. Wu R, Huang H, Wei J, Ma C, Zhu Y, Chen Y, et al. An improved sparrow search algorithm based on quantum computations and multi-strategy enhancement. Expert Systems with Applications. 2023;215:119421.
  55. 55. Dhiman G, Garg M, Nagar A, Kumar V, Dehghani M. A novel algorithm for global optimization: rat swarm optimizer. J Ambient Intell Human Comput. 2020;12(8):8457–82.
  56. 56. Tanyildizi E, Demir G. Golden sine algorithm: a novel math-inspired algorithm. Adv Electr Comp Eng. 2017;17(2):71–8.
  57. 57. Xiao S, Wang H, Wang W, Huang Z, Zhou X, Xu M. Artificial bee colony algorithm based on adaptive neighborhood search and Gaussian perturbation. Applied Soft Computing. 2021;100:106955.
  58. 58. GhaemiDizaji M, Dadkhah C, Leung H. OHDA: an opposition based high dimensional optimization algorithm. Applied Soft Computing. 2020;91:106185.