Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

RETRACTED: Natural gas price prediction based on artificial intelligence models

Retraction

After this article [1] was published, the following concerns were noted:

  • The articles cited as References 3, 4, 5, 7, 8, 9, 10, 16, 17, 18, 20, 29, 32, 34, 49, 55, 57, 60, 61, 73, and 74 do not appear to support the corresponding cited statements.
  • The article cited as Reference 3 was retracted before [1] was published.
  • Potential non-compliance with the PLOS policy on Artificial Intelligence Tools and Technologies.

The first author acknowledged that errors occurred in verifying references during manuscript preparation and that the authors failed to declare the use of artificial intelligence tools for language polishing and text sorting.

The PLOS One Editors retract this article due to the cumulative issues which raise concerns about its reliability and provenance.

XL and MS agreed with the retraction. MT, YF, TF, and ZW either did not respond directly or could not be reached.

23 Feb 2026: The PLOS One Editors (2026) Retraction: Natural gas price prediction based on artificial intelligence models. PLOS ONE 21(2): e0343466. https://doi.org/10.1371/journal.pone.0343466 View retraction

Abstract

The natural gas supply crisis triggered by the Russia–Ukraine conflict has laid bare the energy market’s extreme vulnerability in the face of geopolitical risk, highlighting the need for accurate multi-step gas price forecasting. However, most AI-based energy price studies have a gap: they focus on single-step prediction or homogeneous model comparisons, lacking analysis of performance degradation in multi-step dynamic frameworks. This study takes daily natural gas price data from the Henry Hub in the United States from 1997 to 2024 as the research object, constructs a multi-step prediction framework with a step size ranging from 1 to 4 days, and systematically compares the prediction performances of four artificial intelligence models: feedforward neural network, support vector machine, random forest, and long short-term memory network. The quantitative results show that, across all prediction cycles, the long short-term memory model has the lowest error rate. For example, in one-step forecasting, its Mean Absolute Percentage Error is 8.53%. Practically, the findings matter. Taking European governments facing natural gas shortages in the Russia-Ukraine conflict as an example, LSTM models can be used for multi-step prediction to forecast price fluctuations 2–4 days in advance, optimizing import reserve strategies to avoid supply disruptions; energy traders can use this to design robust futures arbitrage portfolios. In summary, the research provides a scientific basis and reference for government energy security policy-making and institutional investor trading.

Introduction

Natural gas has long played a central role in the global energy structure [13]. However, the Russia-Ukraine conflict has exposed the extreme sensitivity and vulnerability of the natural gas market to geopolitical risks, triggering substantial disruptions in global energy supply chains and significant price volatility [46]. The conflict has not only tightened natural gas supply and caused sharp price fluctuations but has also profoundly affected national energy security strategies, international trade patterns, and energy dependency structures [710]. As a critical energy resource, natural gas prices not only reflect market supply and demand conditions but also directly influence corporate investment decisions, industrial production costs, and household energy consumption, while providing essential guidance for government energy policy formulation and adjustment [1116]. Consequently, accurately forecasting natural gas prices has become a core concern for global energy markets, policymakers, and investors. Enhancing the precision of price forecasts can help investors manage risks and optimize resource allocation, provide policymakers with scientific support to strengthen energy security, and ultimately stabilize international energy markets, mitigating potential economic disruptions caused by sudden geopolitical events [1720].

The earliest method for predicting natural gas was the traditional econometric approach. Salehnia et al. (2013) utilized natural gas spot price data from the U.S. Henry Hub to construct and compare three nonlinear models: Local Linear Regression (LLR), Dynamic Local Linear Regression (DLLR), and artificial neural network (ANN). The empirical results demonstrated that the DLLR model achieved a lower Mean Absolute Percentage Error (MAPE) compared to the ANN model [21]. Xin et al. (2022) employed spatial convergence models (including the Spatial Durbin Model, SDM) to analyze China’s natural gas consumption patterns and found that while absolute convergence existed, significant regional economic disparities substantially hindered the convergence process [22]. Alam et al. (2023) employed multiple time series prediction tools such as the Autoregressive Integrated Moving Average (ARIMA) model, Simple Exponential Smoothing (SES) model, and K-Nearest Neighbors (K-NN) method to conduct in-depth research on Indian oil, coal, and natural gas price fluctuations from the pre- and post-COVID-19 periods. They found that the K-NN method generally performs well in predicting these energy prices, and that the SES model can also provide better prediction results under specific conditions [23]. Ribeiro et al. (2023) used a mixed-methods framework for European gas markets, identifying pricing mechanisms (β = 0.68) and pipeline infrastructure (elasticity = 0.43) as convergence drivers—but their spatial econometric model focused on structural analysis rather than predictive accuracy [24].

These studies collectively highlight the core limitations of econometric models: Constrained by linearity assumptions and stationarity requirements, they cannot effectively capture nonlinear features (e.g., stochastic price jumps from geopolitical conflicts or supply-demand shocks) [2528]. Additionally, differencing/normalization procedures cause information loss, and subjective variable/lag order selection amplifies uncertainty. This aligns with broader limitations of theory-driven paradigms—for example, Sedaghat et al. (2016)’s Marsh Funnel model (assuming Newtonian fluid behavior) showed 3.45% prediction error for non-Newtonian fluids, mirroring econometric models’ failure in complex natural gas markets [29].

To address econometric models’ shortcomings, single artificial intelligence models emerged, leveraging data-driven nonlinear mapping to capture complex price dynamics without pre-specified assumptions. For instance, Hailemarim and Smyth (2019) used a Structural Heterogeneous Autoregressive Vector (SHAVAR) model to analyze monthly U.S. natural gas price volatility (1978–2018), quantifying shock contributions via impulse response analysis and Forecast Error Variance Decomposition (FEVD) [30]. Su et al. (2019a) applied a least squares regression boosting algorithm, which significantly improved the coefficient of determination and reduced Mean Absolute Error (MAE), narrowing the gap between predicted and actual prices [31].

To address the limitations of single-model approaches, scholars have incorporated fuzzy logic systems to manage the ambiguous information inherent in market environments [32]. For instance, Azaden et al. (2012) applied fuzzy linear regression to forecast Iran’s industrial natural gas prices, comprehensively integrating market factors beyond traditional regression methods and thereby providing policymakers with a more robust foundation for decision-making [33]. Manavalan et al. (2024) proposed Neutral Interval-Valued Sets (NIVSs), a fuzzy logic advancement that quantifies “neither-this-nor-that” neutral information via logarithmic transformation (L operator). For instance, it converts “policy easing” into an interval membership degree [0.7, 0.8], and aggregates expert judgments via operators like NIVWA to support natural gas investment decisions [34].

Natural gas price forecasting requires balancing ambiguous information handling and time series patterns. Hybrid models combining multiple approaches effectively mitigate single-method limitations. Consequently, model combination has increasingly gained favor and is widely applied in addressing the challenges of natural gas price prediction [3539]. For example, using a data-driven weighted hybrid model combining Support Vector Regression (SVR), Long Short-Term Memory network (LSTM), and an Integrated Prediction and Smoothing System (IPSS), Wang et al. (2020) predicted the daily spot prices of natural gas in the U.S. for the period from June 2018 to May 2019. The model performed excellently in predicting natural gas prices [40]. Abdollahi H et al. (2020) constructed a decomposition prediction hybrid model that successfully captured nonlinear and volatile time series features. They not only made effective predictions but also tested the data to ensure model stability and reliability [41]. Wang et al. (2021) combined fully adaptive noise ensemble empirical mode decomposition with a gated recurrent unit, optimized via the particle swarm optimization algorithm; their model predicted weekly prices with high accuracy [42]. Li et al. (2021) introduced an integrated prognostic model that combines variational mode decomposition, employs particle swarm optimization and incorporates deep learning models. They empirically predicted the natural gas prices at the Henry Center in the United States and demonstrated good predictive performance [39]. Even non-energy studies (e.g., Ayben Koy and Colak (2023), who combined KSS, ESTAR, and MLP-ANN to predict stock indices and CDS spreads with sub-1% deviation) highlight hybrid models’ potential—but their applicability to multi-step natural gas price forecasting is untested [43]. Nevertheless, additional investigations are needed in order to ascertain whether this hybrid model is capable of achieving an equilibrium between prediction precision and stability [4446].

Constrained by assumptions of linear relationships and normality distributions, econometric methods exhibit limited predictive power in complex real-world economic environments. While combined forecasting models enhance accuracy through integrating single models, their model selection process and weight allocation remain complex and constrained by inter-model correlations. A systematic comparison of these three method categories (econometric, AI, and hybrid) further highlights their trade-offs in assumptions, performance, and interpretability, as summarized in Table 1.

thumbnail
Table 1. Comparison of Natural Gas Price Forecasting Methods: Econometric, AI, and Hybrid Models.

https://doi.org/10.1371/journal.pone.0336582.t001

By contrast, Artificial Intelligence (AI) models such as LSTM demonstrate superior predictive precision and generalization capabilities. As shown in Table 1, these models leverage robust data processing, nonlinear mapping, and adaptive learning to capture intricate economic dynamics (e.g., long-term price dependencies and sudden volatility) without relying on pre-specified theoretical assumptions [4754], making them well-suited for natural gas price forecasting amid geopolitical and supply-demand uncertainties.

Exploring model interpretability is critical but presents substantial challenges: deep learning models feature intricate architectures, making their internal decision-making processes difficult to interpret when handling complex data. Moreover, different model structures and algorithms exhibit significant variations in their impact on interpretability, and there remains a lack of unified evaluation standards [5559].

To address these deficiencies, this paper focuses on the following research questions:

  1. RQ1: How do different AI models (FNN, SVM, RF, LSTM) perform in natural gas price forecasting compared to traditional methods?
  2. RQ2: Can deep learning models, particularly LSTM, achieve a balance between accuracy and robustness under conditions of price volatility and geopolitical uncertainty?

Building on these research questions, several important gaps remain. First, few studies systematically evaluate econometric and AI models under geopolitical shocks, raising concerns about prediction robustness. Second, although LSTM and related methods demonstrate strong accuracy, their lack of interpretability hampers broader adoption. Finally, hybrid approaches show promise but remain constrained by complexity and the absence of standardized design principles. This study addresses these gaps through empirical evaluation of AI models and robustness testing of LSTM.

This study focuses on Henry Hub daily natural gas prices (1997–2024), applying Feedforward Neural Network (FNN), Support Vector Machine (SVM), Random Forest (RF), and Long Short-term Memory network (LSTM). Each model is trained on standardized data and evaluated using multiple error metrics across different forecasting horizons. The main contributions are threefold: (1) providing a systematic comparison of AI and traditional models, (2) assessing LSTM’s performance during major historical shocks, and (3) offering guidance for future hybrid and explainable forecasting frameworks.

The remainder of this paper is organized as follows. Section 2 introduces the methodological framework of FNN, SVM, RF, and LSTM. Section 3 describes the dataset, preprocessing steps, and evaluation metrics. Section 4 presents empirical results and discussion, while Section 5 reports robustness checks. Section 6 concludes with key findings, limitations, and future research directions.

Research methodology

The selection of feedforward neural network (FNN), support vector machine (SVM), random forest (RF), and long short-term memory network (LSTM) models for natural gas price forecasting in this study was based on multi-faceted considerations. FNN can rapidly train and predict relatively stable natural gas price data segments with non-complex nonlinear features, providing a benchmark reference for subsequent model performance evaluations. The natural gas market is influenced by multiple comprehensive factors, characterized by high-dimensional data and complex nonlinear relationships between variables [60,61]. SVM effectively addresses the “curse of dimensionality” through kernel functions, which map low-dimensional nonlinear problems into high-dimensional space for linear processing [62,63]. As an ensemble learning approach, RF constructs multiple decision trees to enhance prediction accuracy and model generalization capability. In natural gas price forecasting, where data volatility is significant due to numerous uncertain factors, individual decision trees may fail to capture complex variations comprehensively. Conversely, the multiple decision trees in RF enable data learning and analysis from diverse perspectives, reducing dependency on individual data points and enhancing stability and anti-interference capabilities [6466]. LSTM’s long- and short-term memory characteristics make it effective in handling long-range dependencies in time series data. Natural gas prices exhibit distinct long-term trends and short-term fluctuations, with interrelated price movements across different time scales. Through the coordinated action of forget, input, and output gates, LSTM selectively retains or discards historical information to better capture relationships between long-term trends and short-term fluctuations [67,68]. Although a gated recurrent unit (GRU) is a simpler variant of gated recurrent neural networks, its memory capacity and ability to capture complex relationships may be insufficient for intricate natural gas price time series data. While Transformer models demonstrate superior performance in processing long sequences and capturing complex dependencies, their high computational complexity results in inefficient training and prediction under limited computational resources. Collectively, the selected models of FNN, SVM, RF, and LSTM ensure research quality while enabling more effective and accurate forecasting of natural gas prices.

FNN model

FNN is considered to be fundamental to the field of artificial neural networks. Its typical composition involves an input layer for data, multiple hidden layers, and an output layer for learning. The fundamental functioning of the model is as follows: At the onset of the forward propagation process, the external input signal enters the input layer first. Subsequently, it is conveyed from the input layer to the hidden layer, where the information is subject to processing operations. A quantity of the hidden layers is flexibly configured according to the specific task requirements. The FNN training process aims to make the network output as close as possible to the actual output. To start with, in general practice, a suitable loss function needs to be defined in order to assess the deviation between the output generated by the network and the real-world output. The working principle is shown below (Fig 1):

In the diagram, input vectors , the superscript represents the -th sample point, the hidden layer outputs a vector , the value outputted by the output layer is , the forecasted output is , the weight matrix that runs from the input layer to the hidden layer is signified by V, and the weight matrix from the hidden layer to the output layer is denoted by W. The activation functions of each layer are and , respectively.

Positive propagation process

The input signal migrates from the input layer, passing through the hidden layer and arriving at the output layer. If the output signal generated in the output layer satisfies the given requirements, the algorithm concludes; otherwise, it initiates the backpropagation process. Regarding the output layer, the integration function and output of the m-th neuron are as follows:

(1)(2)

For hidden layers, there are

(3)(4)

The discrepancy between the output produced by the mesh fitting and the anticipated output is defined as follows:

(5)

When inputting the n-th sample, the objective function of the network is as follows:

(6)

When there are N training samples in the network, the total error is .

The error of the m-th output neuron is , while the chain principle of differentiation includes the following:

(7)

The following can thus be calculated: , , so

(8)

The k-th hidden layer neuron error is , the chain principle of differentiation includes the following:

(9)

It can be calculated that ,

(10)

Backpropagation process

The criterion for weight adjustment is progressively decreasing the error. As a result, this needs to be adjusted in line with the negative gradient of the weights. The values for adjusting the weight W that bridges the hidden layer to the output layer, as well as the weight V that connects the input layer to the hidden layer, are outlined below:

(11)(12)

Among them, is the learning rate given in advance, and the weights of each layer are iteratively updated for the (t + 1)-th time as follows:

(13)(14)

SVM model

SVM represents a statistical search methodology grounded in the tenets of structural risk minimization and high-dimensional space theory. In principle, this entails finding an optimal hyperplane by training a finite dataset to maximize the classification margin between two classes of sample data, while guaranteeing that the nearest sample point to the hyperplane is as close as possible. When confronting nonlinear problems, SVM models successfully grapple with the “curse of dimensionality” in high-dimensional regions by means of kernel functions, circumventing complex inner product computations in high-dimensional feature spaces and substantially reducing computational complexity. The utilization of kernel functions is not only the bedrock of the SVM model but also the crucial technology behind its extensive application. Currently, the prevalently used kernel functions primarily comprise four types:

Linear kernel function

(15)

A linear kernel function is a relatively simple kernel function without additional parameter variables.

Polynomial kernel function

(16)

In the equation, d is the polynomial order.

When the sample dimension is relatively high and the polynomial kernel function order is relatively large, computational operations become especially complex, and the calculation results are liable to errors.

Sigmoid kernel function

(17)

The sigmoid kernel function has two parameters— and , , <0—and is predominantly applied in the BP algorithm. In the equation, signifies the center of the dataset relative to the kernel function, and σ indicates the function’s width parameter.

Gaussian Radial Basis Function (RBF)

(18)

In the equation, represents the center of the kernel function, while g denotes the function’s width parameter. The RBF only has this single parameter. The code discussed in this article employs the Gaussian RBF as its kernel function.

Linear separable SVM

Assume that the training sample set for a linear regression model is given by . In this context, denotes the n-th feature vector, each of which is characterized by m features, i.e., , i = 1, 2,...,n, and denotes the n-th dependent variable. For binary classification, and i = 1, 2,...,n, denotes the negative and positive classes, respectively.

As shown in Fig 2, the plane is two-dimensional, and the three dashed lines in the figure could represent linear functions:

(19)

In the above equation, w can be understood as a weight vector, and the following function can be used to represent this:

(20)

If there exists a hyperplane in the feature space that partitions the learning objectives into positive and negative classes, the distance from any sample point to the hyperplane is at least 1, i.e.,

(21)

Assume that if the hyperplane is known, the geometric distance from any data point in the dataset to the hyperplane can be calculated:

(22)

The geometric distance from a support vector to the hyperplane is given by the following:

(23)

Transform the binary classification problem into an optimization problem, formulated as follows:

(24)

Divide both sides of Equation 24 by and set ; then, it becomes

(25)

Maximizing the geometric distance from the support vectors to the hyperplane is equivalent to maximizing , which translates to solving for the minimum of , formulated as follows:

(26)

To control classification errors within a certain range, slack variables are introduced, transforming the optimization problem into a quadratic programming problem, formulated as follows:

(27)

Here, c represents the penalty factor, which has an impact on the model’s generalization capacity, that is, the model’s complexity and adaptability. The solution is obtained by means of the Lagrangian function:

(28)

Let stand for the Lagrange multiplier, and the dual form can be written as follows:

(29)

Solve Equation 29, compute and obtain , to derive the feature space hyperplane as , and ultimately ascertain the binary classification decision function:

(30)

Nonlinear SVM

The basic principle of the nonlinear SVM model is similar to that of its linearly separable equivalent. However, its key difference lies in the introduction of kernel function technology. Through the use of kernel functions, nonlinear sample datasets can be made linearly separable in the kernel space. This expands the application scope of the SVM model and enables it to effectively handle numerous classification problems involving high-dimensional and nonlinear datasets.

The formula for the nonlinear SVM, with the Lagrangian function introduced, is as follows:

(31)

The objective function of the high-dimensional nonlinear SVM is as follows:

(32)

The symmetric function is .

After substituting the RBF,

(33)

Thus, the classification model is formulated as follows:

(34)

RF model

RF is an ensemble learning method, and one of its fundamental building blocks is a decision tree known as a Classification Regression Tree (CART). A CART employs a binary recursive partitioning method, which partitions the sample set into two subsets according to specific rules via a binary tree structure until non-divisible leaf nodes are generated, thereby enhancing the overall prediction accuracy and model generalization ability.

In the RF training process (Fig 3), each time a decision tree is established, a subset of data is obtained through bootstrap sampling to participate in decision tree training. Because of the resampling process, around one-third of the data are excluded in creating the current decision tree. These excluded data are referred to as Out of Bag (OOB) data, and they can be utilized to assess the decision tree effectiveness and determine a pattern’s prediction error rate, known as the OOB data error. When training an individual decision tree, the following techniques are typically employed in choosing effective, discrete nature partitioning:

Information gain

If we consider a discrete attribute ‘a’ that has V potential values, we can assess the “purity enhancement” achieved by utilizing information gain for its partition node; the formula is as follows:

(35)

Among them, is the massage entropy of sample set D and the optimal partition attribute .

(36)

Gini index

The level of purity within dataset D is gauged through utilizing the Gini value, as shown in the following equation:

(37)

The lower the Gini (D) value turns out to be, the greater the purity of the data value. The Gini index for attribute ‘a’ is specified by the following equation:

(38)

Thus, in candidate nature set A, the nature that minimizes the Gini index after partitioning is selected as the optimal partitioning nature, as follows:

(39)

In the process of determining continuous characteristics, provided with a sample set D, along with continuous attributes, and supposing there are n distinct values emerging, we need to arrange these values in ascending order and label them as . Based on the partition point t, D can be divided into subsets and . Of these, the former contains samples with values no greater than t on attribute a, while the latter contains those with values greater than t. Next, the same information gain criterion as with discrete attributes can be used to select the optimal partition point t, as shown in the following equation:

(40)

LSTM model

The LSTM model is distinctive type of recurrent neural mesh, which is a typical representative of deep learning network models with important significance for solving RNN model gradient problems. The LSTM model, a type of gated RNN, is constructed to maintain stable gradient information flow and adjustable connection weights. It effectively accumulates and selectively forgets information through self-cycling and dynamically changing the time scale, providing an excellent solution for the long-term problems of hidden variable models in terms of long-term information preservation and short-term input.

The core part of the LSTM consists of a series of memory cells. Every single cell contains three integral internal components known as gates: namely, the forget gate, the input gate, and, equally importantly, the output gate. These are responsible for governing information flow inside the cell and overseeing state updates. This allows the LSTM to preserve valuable long-term information while efficiently sifting out superfluous data (Fig 4).

Forget gate

The forget gate determines that, in natural gas price prediction, the model needs to forget certain unrelated past price fluctuations. This gate accepts the hidden state from the previous time step (previous period) and the current input , and outputs a scalar matrix in the range of 0–1 through a sigmoid activation function, indicating how much information the current cell needs to forget.

(41)

is the weight matrix, is the bias, and is sigmoid activation function, which aims to compress the output value between 0 and 1.

Input gate

The input gate regulates the process through which new inputs are blended into the current unit state. At the same time, the model refreshes its forecasts for prospective prices, relying on contemporary market fluctuations and time-distinctive features. It consists of two steps:

The first part generates a gate value from 0 to 1 through the sigmoid activation function, indicating which values need to be updated.

The second section produces new potential values using the Tanh function, the incorporation of the aforementioned element into the cell state is subject to the regulation of the input gate. The calculation formulas for these two steps are as follows:

(42)(43)

is input the activation value of the door.

are candidate values.

In the presence of the control exerted by the forget and input gates, the LSTM functions in a particular way. It commences by multiplying the cell state of the previous time step with the output of the forget gate. Then, it combines the product of the input gate and the candidate value to form the updated cell state.

The formula is expressed as follows:

(44)

Among them, is the updated unit state, representing the memory cell content.

Output gate

This is responsible for determining output at the present time step, thus enabling the model to predict future prices based on the current cell state and the output gate’s value. Initially, this value is calculated using a sigmoid function, followed by the transformation of the new cell state using a Tanh function. The hidden state for the current time step, denoted by the symbol ht, is then obtained by multiplying the output gate’s value with the transformed cell state. This formula is expressed as follows:

(45)(46)

is the activation value of the output gate that determines which part of the cell state will have an effect on the hidden state at the next moment, while is that at the current moment.

Study design

Data and preprocessing.

Presently, the international spot price of natural gas principally comprises the following several significant types:

First, the natural gas price at the Henry Hub in the United States; second, the natural gas price at the National Balancing Point (NBP) in the United Kingdom; third, the natural gas price at the Title Transfer Facility (TTF) in the Netherlands; and fourth, the landed price of liquefied natural gas (LNG) in Japan. Among them, the latter three all have certain limitations, and the first is the most representative.

The natural gas price at the Henry Hub in the United States is a key benchmark price for the US market. Located in Louisiana, this hub is very active in natural gas trading activities. The United States has an extensive and complex network of natural gas pipelines. With its superior geographical location and complete infrastructure, the Henry Hub has become important in the US natural gas market. Many natural gas producers, traders and end-users conduct transactions here, and the prices reflect the supply and demand situation. The natural gas futures market at the Henry Hub is a mature financial instrument, and its price has a significant impact on global natural gas market trends, thus providing a crucial reference index for global natural gas trading.

The natural gas price at the Henry Hub in the United States as research object.

The American Henry Center natural gas price was selected as the object of study. This center includes natural gas from local rich reserves, a pipeline network, a high methane content and low levels of impurities. With widely active trading, and many participants in both future and spot-coordinated development, not only is the north American pricing a benchmark, but it also affects the global price. The developed pipeline transportation network connects local and neighboring markets and, although exports are currently limited, with the progress of LNG technology, it is anticipated that the global supply and demand will be balanced. In the energy industry, stable prices are of great importance. The daily Henry Center natural gas price data utilized in this study are freely available on the Energy Information Administration (EIA) website (http://www.eia.gov/dnav/ng/hist/rngwhhdD.htm). Operating under the auspices of the U.S. Department of Energy (DOE), the EIA is a professional energy information statistics and analysis agency responsible for the collection, organization and release of energy data, both within the United States and worldwide. In this research, we analyzed daily data spanning from January 8, 1997 to August 7, 2024, comprising a total of 6,930 observations. To facilitate the operation of the neural network model, data from January 8, 1997 to January 8, 2018 were designated as the training set, consisting of 5,285 observations. Those from January 9, 2018 to August 7, 2024 were designated as the out-of-sample test set, comprising 1,645 observations. This approach aligns with the prevailing practice of dividing samples into four-fifths for training and one-fifth for testing. The model was constructed using data prior to January 2018, and predictions were subsequently made for the period from January 9, 2018 to August 7, 2024.

To study natural gas future prices with univariate time series characteristics, we conducted an empirical analysis of the natural gas futures dataset. During the analysis, we identified missing values in some time series observations. To address these missing values, we selected the mean imputation method. To prevent overfitting and ensure stability across experiments, we implemented multiple regularization techniques. Random seeds were fixed to ensure reproducibility. Dropout layers were added to the LSTM networks to reduce overfitting. Additionally, an early stopping mechanism with a patience of 10 epochs was used to halt unstable training before convergence began to deteriorate.

Data normalization.

To reduce the impact of noise in the natural gas market and improve prediction accuracy, we normalized Henry Hub natural gas prices [69]. Multiple normalization methods can enhance neural network training, including the approach adopted in this work, which scales data to the [0,1] range using the following equation:

(47)

Inverse normalized value.

The standardized processing of natural gas price results in negative values for some data, which are actually the relative value of the standardized mean, not the true meaning of the original data. Therefore, inverse normalization operations must be performed when the practical meaning and magnitude of the data are required. The formula for this is as follows:

(48)

Four common prediction models (FNN, SVM, RF, LSTM) were used here, and through four different prediction steps, four criteria for measuring errors were subsequently chosen and utilized to evaluate their effectiveness: the MAPE, MAE, Root Mean Square Percentage Error (RMSPE), and Root Mean Square Error (RMSE). These four unique metrics measure the predictive accuracy of each pattern. Below are definitions for each of these metrics:

(49)(50)(51)(52)

In these formulas, and are the predicted and true values of time t, respectively. N denotes the total number of data points. Attention should be paid to the fact that the MAPE and RMSPE embody the relative error existing between the actual and forecasted values. On the flip side, MAE and RMSE epitomize the absolute error that lies between the true and predicted data.

Results and discussion

Evaluations of various models

As shown in Table 2, the overall performance of LSTM is optimal, followed by FNN, RF, and SVM in descending order. For one-step prediction, the LSTM achieved the lowest MAPE of 8.53%, approximately 3.47%, 6.58%, and 6.58% lower than FNN, SVM, and RF, respectively. These results highlight the superior generalization and adaptability of the LSTM model in forecasting volatile natural gas prices.

thumbnail
Table 2. A comparison of the prediction results obtained from different models.

https://doi.org/10.1371/journal.pone.0336582.t002

The superiority of LSTM can be attributed to its internal gating mechanism and ability to capture both short- and long-term temporal dependencies in the data. Natural gas prices exhibit strong nonlinear, seasonal, and shock-sensitive behaviors. LSTM’s cell state and forget gates effectively filter out transient noise while preserving key temporal information, enabling the model to learn intrinsic fluctuation patterns and respond robustly to sudden market changes. In contrast, FNN lacks temporal memory, processing each observation independently; while it performs well on relatively stable segments, it cannot leverage inter-temporal dynamics. RF, although nonlinear and ensemble-based, may overfit when faced with noisy or limited samples, leading to instability across forecast horizons. SVM, primarily designed for classification, struggles to adapt to continuous, nonstationary sequences even when kernel functions are applied, resulting in its weakest performance.

Therefore, LSTM’s combination of nonlinear mapping capacity, sequential memory, and noise resistance explains its superior predictive accuracy and robustness in modeling natural gas price dynamics.

Discussion

Fig 5 demonstrates the partial predicted values of FNN, SVM, RF, and LSTM at one-, two-, three-, and four-step levels. It can be intuitively seen that the prediction curve of the LSTM model fits the true value curve most closely, echoing the conclusion—drawn from earlier error index calculations—that this model has the best prediction performance. Across different prediction horizons, the LSTM model effectively captures the changing trends of natural gas prices. For instance, in one-step prediction, the LSTM model’s predicted values closely follow the true value fluctuations, reflecting its sensitivity and accurate prediction ability for short-term price changes.

thumbnail
Fig 5. Partial predicted values of FNN, SVM, RF, and LSTM at one-, two-, three-, and four-step levels.

https://doi.org/10.1371/journal.pone.0336582.g005

In contrast, the SVM model’s prediction curve deviates significantly from the true values. This may be because it primarily applies to dimensionality reduction in high-dimensional spaces. For two-dimensional time series data like natural gas prices, its advantages cannot be fully leveraged. When facing drastic price fluctuations, the SVM model struggles to accurately fit price changes, leading to notable discrepancies between predicted and true values.

The RF model’s prediction performance lies between LSTM and SVM. While this model has certain nonlinear prediction capabilities, the extensive use of decision trees in its training process complicates it. When handling single-event sequences (such as natural gas price time series), it is prone to overfitting, affecting prediction accuracy. As shown in the figure, during periods of significant fluctuations (e.g., mid-2008), the RF model’s prediction curve fails to track true value changes effectively.

The FNN model performs relatively well, reflecting price trends in some periods. However, due to the high nonlinearity and volatility of natural gas prices, the FNN model still has limitations in handling complex price fluctuations. During the rapid price increase in the first half of 2021, although the FNN model’s predictions also rose, there were still gaps in both the magnitude and trend compared to the true values.

Our findings reaffirm the superior performance of LSTM compared with conventional machine learning approaches such as SVM and RF. This result aligns with Ghosh et al. (2022) [70], who demonstrated that LSTM networks outperform traditional methods, including random forests, in forecasting the intraday directional movements of S&P 500 constituent stocks [69]. Interestingly, our experiments also revealed that FNN outperformed SVM. While this outcome was consistent with our initial expectations, it contrasts with the conclusions of Yu et al. (2017). We attribute this discrepancy to the incorporation of regularization strategies in our FNN implementation—such as dropout layers, weight decay, and early stopping—which enhanced its generalization ability when handling complex and volatile natural gas price data. In contrast, the SVM model used for comparison lacked similar optimization, which may explain its relatively weaker predictive performance [71].

Through a detailed analysis of Fig 5, we can more intuitively understand the performance differences of various models in natural gas price prediction, further validate our conclusions based on error metrics, and provide a stronger foundation for model selection and improvement.

  1. During the period from September 21 to October 28, 2005, the United States was hit by two extreme weather events with great destructive power—Katrina and Rita. This sudden disaster dealt a heavy blow to the Gulf of Mexico. It is imperative to note that this area is regarded as one of the most significant natural gas-producing regions in the US, and many production facilities were severely damaged by the hurricane. At the same time, the market demand for natural gas increased as winter approached; residents needed large amounts of natural gas to maintain warm living conditions, and some industrial companies that rely on natural gas for energy increased their purchases to meet winter production needs. However, with demand rising sharply, the supply of natural gas was greatly limited by these damaged production facilities. This severe imbalance between supply and demand prompted a sharp rise in natural gas prices, bringing a huge shock to the market.
  2. Between June 19 and July 2, 2008, a series of changes took place in the international energy market. At that time, the price of other energy sources, such as oil, showed a rising trend, which led some users who used other energy sources to turn to natural gas, which was more advantageous relative to price. Consequently, this resulted in increased demand. In addition, in some areas, gas production equipment has entered maintenance periods, which has led to a temporary reduction in gas supply in these areas. Whilst maintenance equipment is intended to guarantee the stability and sustainability of future natural gas production, it is an irrefutable fact that natural gas supply is constrained during maintenance. Under the dual factors of increasing demand and decreasing supply, natural gas prices have been supported to a certain extent, showing a rising trend.
  3. On February 17, 2021, natural gas prices rose again. The phenomenon under discussion is primarily attributable to a combination of two factors. On one side, it was winter, and the cold weather increased residents’ demand for natural gas heating. Under low temperatures and pressure, the frequency and length of the use of home heating equipment increase significantly, generating marked increases in demand for natural gas. On the other hand, the global economy is gradually recovering after experiencing the impact of the epidemic. As economic recovery advanced, industrial production activities gradually recovered and accelerated, and the energy demands of industrial enterprises also increased. As an important energy source, natural gas is widely used in these fields, so an increase in industrial demand also promotes increased natural gas prices.
  4. In the early months of 2022, after the Russia–Ukraine, the global natural gas market encountered a precipitous decline, marking a period of significant economic downturn, with Europe taking the hardest hit and suffering substantial losses. During this time frame, natural gas prices showed extreme volatility. On the one hand, aiming to lessen their dependence on Russian natural gas, European countries swiftly redirected their procurement endeavors towards alternative energy sources, notably liquefied natural gas (LNG) from the US. This sudden alteration immediately disrupted and remolded market competition, becoming a crucial factor in gas price fluctuations. On the other hand, the combination of frequent extreme weather events, ongoing geopolitical crises, and persistent imbalances between supply and demand has functioned as a potent catalyst, ceaselessly disrupting natural gas price stability. Concurrently, due to the globalization-driven energy ripple effect, the cost of gas in the Asia–Pacific region has also been impacted. Affected by the upward trend of global oil prices, prices here have also seen an upward climb, casting a shadow over regional energy consumption and economic development.

Robustness test

The various forecasting time frames.

We have demonstrated our step-by-step approach to predicting gas prices. All four models effectively forecast prices and promptly capture changes in the natural gas market. Among these models, LSTM demonstrated the highest accuracy in predictions, with FNN coming in second. To evaluate prediction performance stability across the four models, we adjusted the prediction lengths to the fifth, sixth, seventh, and eighth steps. The outcomes are detailed below in Table 3.

thumbnail
Table 3. Evaluations of forecasts from various prediction models across different time frames.

https://doi.org/10.1371/journal.pone.0336582.t003

Model performance comparison: In the multi-model comparison system, including FNN, SVM, RF and LSTM, the LSTM model stands out, showing excellent prediction accuracy. This model performs five-, six-, seven-, or eight-step prediction well on the four key evaluation indicators: MAPE, MAE, RMSPE and RMSE. For example, the MAPE is as low as 10.63, which is significantly lower than the 11.71 for FNN, 11.94 for SVM and 12.63 for the RF; it is evident that 9.03 is lower than the corresponding values observed in other models. This shows that the LSTM pattern can more accurately capture the underlying rules in data, minimize the deviation from the actual value of the future trend, and has strong adaptability and reliability in complex prediction tasks, providing a solid cornerstone of data support for accurate decision-making.

Prediction range and error correlation: On the whole, with the gradual expansion of the prediction step length, the prediction error of each model generally shows an upward trend; this phenomenon is highly consistent across different models. Taking the LSTM model as an example, from five- to eight-step prediction, MAPE gradually increased from 10.63 to 9.03, MAE from 0.39 to 0.34, and RMSE from 0.89 to 0.83; other models such as FNN, SVM and RF follow a similar trend. In the field of prediction, a longer time span will inevitably introduce more uncertainty factors and cumulative errors, presenting severe challenges for prediction accuracy and prediction model optimization. The practical application of step size selection provides key directional guidance, emphasizing the pursuit of lengthy, more carefully balanced prediction precision and range, in order to meet the diversified reality of requirement and decision targets.

Alternative indicators of Gas prices.

In this section, an alternative proxy variable for gas price, specifically the flat-ui gas price, is examined. We substituted the Henry Hub Natural Gas Spot Price (measured in USD per million BTU) with this proxy variable. The dataset encompasses the period from January 8, 1997 to August 7, 2024, with a total of 6,933 observations. The data from January 8, 1997 to January 8, 2018 were designated as the training set (within-sample) with 5,285 observations, while those from January 9, 2018 to August 7, 2024 served as the test set (out-of-sample) with 1,648 observations. This approach aligns with the prevailing convention of allocating four-fifths of the data for training and one-fifth for testing. As follows, the analysis undertaken using the flat-ui configuration confirmed that the LSTM model exhibited the lowest prediction error, as shown in Table 4.

thumbnail
Table 4. Evaluation forecasts generated by various prediction models for flat-ui.

https://doi.org/10.1371/journal.pone.0336582.t004

A comparative analysis of the prognostication horizon dimensions.

Rossi and Inoue (2012) posited that the dogmatical selection distinguishing prognostication horizon dimensions can result in significantly varied out-of-sample outcomes, thus underscoring the selection of the optimal window size as a pivotal component of effective assessment in this regard [72]. The present study aims to build on the aforementioned concept by examining the Henry Hub Natural Gas Spot Price in USD per million BTU from January 4, 2000 to August 27, 2024, encompassing a total of 6,195 observations. The training dataset spans from January 4, 2000 to December 29, 2017, consisting of 4,531 observations, while the test dataset covers January 2, 2018 to August 27, 2024, with 1,663 data points, adhering to the standard practice of dividing the sample into four-fifths for cultivating and one-fifth for testing.

As demonstrated in Table 5, in the context of one-step prediction, the LSTM pattern emerges as the most precise and stable pattern base for predicting recent trends in gas prices, in terms of the MAPE, MAE, RMSPE and RMSE indicators. It far exceeds the FNN, SVM and RF models, providing a more reliable choice for short-term natural gas price forecast, as shown in Table 5.

thumbnail
Table 5. Comparisons of forecasts generated by various models using different time frames for predictions.

https://doi.org/10.1371/journal.pone.0336582.t005

A comparative analysis of forecasting window sizes.

For a comprehensive assessment of the predictive ability of the FNN, LSTM, SVM, and RF models, the ARIMA pattern was selected as the benchmark model (Table 6). The model has gained significant renown within the domain of time series prediction, providing a reliable reference for measuring the performance of other models with its ability to effectively capture data trends and seasonality. Table 6 lists the prediction results of the FNN, LSTM, SVM and RF models based on ARIMA; key assessment metrics are covered, including MSE and MAE. It is evident that the text presents a balanced and objective analysis of the merits and drawbacks of each model when applied to particular datasets, providing a powerful basis for further optimizing model selection and parameter adjustment and helping researchers to explore the applicability and effectiveness of different models in practical application scenarios. The best ARIMA parameters were calculated: (0,1,2)

thumbnail
Table 6. Evaluation of various forecasting models against an alternative benchmark.

https://doi.org/10.1371/journal.pone.0336582.t006

As illustrated in Table 6, across the four prediction intervals of the five models, the LSTM model has the smallest prediction error, fully showing that even in the presence of multiple competitive models, its prediction effect is still excellent and it is reliably able to forecast trends in natural gas prices.

Conclusion

To enhance the forecasting precision, this study employs four artificial intelligence models—FNN, SVM, RF, and LSTM—to predict international natural gas prices, thereby enriching the methodological toolkit for energy price prediction. Given the strong nonlinearity and drastic volatility of natural gas prices, machine learning models adept at handling nonlinear patterns are prioritized over traditional econometric models, which excel in linear prediction. The empirical results demonstrate that the LSTM model outperforms the others, with its specific contributions outlined as follows:

1. Methodological Contributions and Model Superiority

First, for the first time, this study systematically compares the performance of FNN, SVM, RF, and LSTM in natural gas price forecasting, identifying the unique advantages of the latter in capturing strong nonlinearity and short-term price fluctuations. This provides new empirical evidence for selecting time series prediction models in the energy sector. Unlike traditional econometric models, which struggle with high-volatility markets, deep learning models like LSTM are validated to be more suitable for forecasting natural gas prices, filling a gap in the literature on short-term prediction model comparisons for this specific context. The LSTM model’s superiority stems from two key factors: (1) its ability to effectively fit the strongly nonlinear and volatile time series characteristics of natural gas prices, which traditional linear models cannot adequately represent; (2) its strong short-term prediction capability, which is aligned with this study’s focus on the relationship between long- and short-term memory functions and short-term price dynamics. The hypothesized strong short-term correlation in prices is confirmed, underpinning the model’s effective forecasting performance.

2. Practical implications for Decision-Makers and Investors

The LSTM model’s enhanced predictive power offers actionable, stakeholder-specific insights, drawing on empirical findings from energy economics, which links short-term price forecasts to trade strategy optimization, which emphasizes policy alignment with user needs) [73, 74]: Building on energy economics research on market linkage and fiscal sustainability, this study demonstrates that LSTM forecasts provide practical benefits for both market participants and policymakers. For energy traders, short-term predictive accuracy supports more agile decisions in spot markets and enhances hedging strategies by capturing short-term price memory, thereby reducing risks from transient volatility. For government ministries, particularly energy and finance, the forecasts offer a foundation for proactive measures such as adjusting strategic reserves, refining import budgets, and recalibrating subsidy policies. Together, these applications show that integrating LSTM into decision-making not only strengthens market efficiency and fiscal resilience but also promotes greater stability in energy systems facing global uncertainty.

3. Predictive Accuracy and Market Dynamics

This study also identified variability in prediction accuracy across different time frames. Notably, the forecasting performance in 2005, 2008, 2021, and 2024 was suboptimal, with significant errors attributed to drastic geopolitical shifts and international events during these periods (e.g., financial crises, geopolitical conflicts). Such disruptions introduced extreme volatility that challenged the model’s ability to fit historical patterns, underscoring AI forecasts’ sensitivity to exogenous shocks.

4. Short-Term Correlation and Market Behavior

The LSTM model’s results confirm the short-term correlation of natural gas prices, implying that prices exhibit a “memory” of recent trends. This finding suggests that major events may have prolonged impacts on price dynamics. As a risk mitigation strategy, institutional investors often replicate previous trading strategies to avoid losses, reflecting the practical relevance of short-term predictability in market behavior.

Limitations and future directions

Beyond these contributions, it is important to critically reflect on the limitations of AI-based forecasting. The “black-box” nature of deep learning models constrains transparency and interpretability, raising concerns about accountability when predictions fail under extreme shocks such as the Russia–Ukraine conflict. Furthermore, the ethical dimension of AI applications in energy markets should not be overlooked: the opacity of models could exacerbate information asymmetry or facilitate market manipulation, thereby challenging market integrity. These concerns highlight the necessity of developing regulatory mechanisms that promote transparency, such as disclosure of input features and prediction logic.

Linking this study to the broader energy economics literature, several empirical works have emphasized the role of exogenous shocks, demand–supply dynamics, and fiscal conditions in shaping price volatility. Integrating these insights with AI approaches can strengthen both theoretical grounding and policy relevance. Future research may thus explore hybrid frameworks that combine the predictive accuracy of deep learning with the interpretability of econometric or structural models, bridging gaps between AI innovation and established energy economics research.

Supporting information

Acknowledgments

We appreciate the feedback from the anonymous reviewers.

References

  1. 1. Cabrales S, Valencia C, Ramírez C, Ramírez A, Herrera J, Cadena A. Stochastic cost-benefit analysis to assess new infrastructure to improve the reliability of the natural gas supply. Energy. 2022;246:123421.
  2. 2. Alkathery MA, Chaudhuri K, Nasir MA. Implications of clean energy, oil and emissions pricing for the GCC energy sector stock. Energy Economics. 2022;112:106119.
  3. 3. Khalil RG, Damrah S, Bajaher M, Shawtari FA. Unveiling the relationship of ESG, fintech, green finance, innovation and sustainability: case of Gulf countries. Environ Sci Pollut Res Int. 2023;30(54):116299–312. pmid:37910364
  4. 4. Ottesen A, Damrah S. Evaluating Kuwait’s Electric Vehicle Lag: In-Depth Analysis of Owner Experiences, Dealer Perspectives, and Policy Recommendations with Local Data Insights. Proc. 2024;101(1):20.
  5. 5. Atyeh M, Damrah S. Examining macroeconomic factors and environmental sustainability: An OLS regression analysis of the impact of exports, imports, governmental budget, and oil products on CO2 emissions in Jordan. Edelweiss Appl Sci Technol. 2024;8(4):1420–7.
  6. 6. Tang Y, Chen XH, Sarker PK, Baroudi S. Asymmetric effects of geopolitical risks and uncertainties on green bond markets. Technol Forecast Soc Change. 2023;189:122348.
  7. 7. Kou X, Luo X, Chu W, Zhang Y, Liu Y. Multi-gas pollutant detection based on sparrow search algorithm optimized ALSTM-FCN. PLoS One. 2024;19(9):e0310101. pmid:39269976
  8. 8. AlShaikh-Mubarak GA, Kotb E, Alabdalall AH, Aldayel MF. A survey of elastase-producing bacteria and characteristics of the most potent producer, Priestia megaterium gasm32. PLoS One. 2023;18(3):e0282963. pmid:36913358
  9. 9. Satrovic E, Cetindas A, Akben I, Damrah S. Do natural resource dependence, economic growth and transport energy consumption accelerate ecological footprint in the most innovative countries? The moderating role of technological innovation. Gondwana Res. 2024;127:116–30.
  10. 10. Bajja S, Radoine H, Celik A, Dakyaga F, Damrah S. Human capital and manufacturing activities under environmentally-driven urbanization in the MENA region. Front Environ Sci. 2024;11.
  11. 11. Atyeh M, Damrah S. Investigating the Performance of the Private Equity Sector in the GCC. In: Proceedings of the World Conference on Management, Business, and Finance. 2023;1(1):2.
  12. 12. Umar M, Riaz Y, Yousaf I. Impact of Russian-Ukraine war on clean energy, conventional energy, and metal markets: Evidence from event study approach. Resour Policy. 2022;79:102966.
  13. 13. Wiggins S, Etienne XL. Turbulent times: Uncovering the origins of US natural gas price fluctuations since deregulation. Energy Econ. 2017;64:196–205.
  14. 14. Izzeldin M, Muradoğlu YG, Pappas V, Petropoulou A, Sivaprasad S. The impact of the Russian-Ukrainian war on global financial markets. Int Rev Financ Anal. 2023;87:102598.
  15. 15. Karkowska R, Urjasz S. How does the Russian-Ukrainian war change connectedness and hedging opportunities? Comparison between dirty and clean energy markets versus global stock indices. J Int Financ Markets Inst Money. 2023;85:101768.
  16. 16. Ottesen A, Banna S, Alzougool B, Damrah S. A greener Kuwait: how electric vehicles can lower CO2 emissions. LSE Middle East Centre Kuwait Programme Paper Series. 2023, 18 p. Available from: http://eprints.lse.ac.uk/id/eprint/120091
  17. 17. Bani-Mustafa A, Al Qudah A, Damrah S, Alameen M. Does culture influence whether a society justifies tax cheating? J Financ Crime. 2020;30(4):955–65.
  18. 18. Damrah S, Elian MI, Atyeh M, Shawtari FA, Bani-Mustafa A. A Linear Mixed Model Approach for Determining the Effect of Financial Inclusion on Bank Stability: Comparative Empirical Evidence for Islamic and Conventional Banks in Kuwait. Mathematics. 2023;11(7):1698.
  19. 19. Zhang S, Wang L. The Russia-Ukraine war, energy poverty, and social conflict: An analysis based on global liquified natural gas maritime shipping. Appl Geogr. 2024;166:103263.
  20. 20. Zhang J, Xu Y, Lu K, Gong Z, Weng Z, Shu P, et al. Differences in gas exchange, chlorophyll fluorescence, and modulated reflection of light at 820 nm between two rhododendron cultivars under aluminum stress conditions. PLoS One. 2024;19(6):e0305133. pmid:38935623
  21. 21. Salehnia N, Falahi MA, Seifi A, Mahdavi Adeli MH. Forecasting natural gas spot prices with nonlinear modeling using Gamma test analysis. J Nat Gas Sci Eng. 2013;14:238–49.
  22. 22. Guan X, Lu X, Wen Y. Is China’s Natural Gas Consumption Converging? Empirical Research Based on Spatial Econometrics. Energies. 2022;15(24):9448.
  23. 23. Alam MS, Murshed M, Manigandan P, Pachiyappan D, Abduvaxitovna SZ. Forecasting oil, coal, and natural gas prices in the pre-and post-COVID scenarios: Contextual evidence from India using time series forecasting tools. Resour Policy. 2023;81:103342. pmid:36815943
  24. 24. Ribeiro VM, Soutinho G, Soares I. Natural Gas Prices in the Framework of European Union’s Energy Transition: Assessing Evolution and Drivers. Energies. 2023;16(4):2029.
  25. 25. van Goor H, Scholtens B. Modeling natural gas price volatility: The case of the UK gas market. Energy. 2014;72:126–34.
  26. 26. Mishra P. Natural Gas Price Forecasting: A Novel Approach. In: IAENG Transactions on Engineering Technologies: Special Volume of the World Congress on Engineering 2012. Dordrecht: Springer Netherlands; 2013, p. 627–39.
  27. 27. Swamy PAVB, Tavlas GS, Hall SGF, Hondroyiannis G. Estimation of Parameters in the Presence of Model Misspecification and Measurement Error. Stud Nonlinear Dyn Econom. 2010;14(3).
  28. 28. Hasanov FJ, Javid M, Mikayilov JI, Shabaneh R, Darandary A, Alyamani R. Macroeconomic and sectoral effects of natural gas price: Policy insights from a macroeconometric model. Energy Econ. 2025;143:108233.
  29. 29. Sedaghat A, Omar MAA, Damrah S, Gaith M. Mathematical modelling of the Marsh Funnel for measuring rheological properties of drilling nanofluids for energy efficient environment. In: 2016 Eleventh International Conference on Ecological Vehicles and Renewable Energies (EVER). IEEE, 2016, p. 1–4.
  30. 30. Hailemariam A, Smyth R. What drives volatility in natural gas prices? Energy Econ. 2019;80:731–42.
  31. 31. Su M, Zhang Z, Zhu Y, Zha D. Data-Driven Natural Gas Spot Price Forecasting with Least Squares Regression Boosting Algorithm. Energies. 2019;12(6):1094.
  32. 32. Firouzkouhi N, Amini A, Bani-Mustafa A, Mehdizadeh A, Damrah S, Gholami A, et al. Generalized fuzzy hypergraph for link prediction and identification of influencers in dynamic social media networks. Exp Syst Appl. 2024;238:121736.
  33. 33. Azadeh A, Sheikhalishahi M, Shahmiri S. A hybrid neuro-fuzzy simulation approach for improvement of natural gas price forecasting in industrial sectors with vague indicators. Int J Adv Manuf Technol. 2012;62(1–4):15–33.
  34. 34. M. M, Damrah S, Falahah IA, Al-Husban AA, Palanikumar M. Selection process based on new type neutrosophic interval-valued set applied to logarithm operator. Int J Neutrosophic Sci. 2024;24(4):335–51.
  35. 35. Jin J, Kim J. Forecasting Natural Gas Prices Using Wavelets, Time Series, and Artificial Neural Networks. PLoS One. 2015;10(11):e0142064. pmid:26539722
  36. 36. Čeperić E, Žiković S, Čeperić V. Short-term forecasting of natural gas prices using machine learning and feature selection algorithms. Energy. 2017;140:893–900.
  37. 37. Duan Y, Zhang J, Wang X. Henry Hub monthly natural gas price forecasting using CEEMDAN–Bagging–HHO–SVR. Front Energy Res. 2023;11.
  38. 38. Zheng Y, Luo J, Chen J, Chen Z, Shang P. Natural gas spot price prediction research under the background of Russia-Ukraine conflict - based on FS-GA-SVR hybrid model. J Environ Manage. 2023;344:118446. pmid:37352627
  39. 39. Li J, Wu Q, Tian Y, Fan L. Monthly Henry Hub natural gas spot prices forecasting using variational mode decomposition and deep belief network. Energy. 2021;227:120478.
  40. 40. Wang J, Lei C, Guo M. Daily natural gas price forecasting by a weighted hybrid data-driven model. J Petrol Sci Eng. 2020;192:107240.
  41. 41. Abdollahi H. A novel hybrid model for forecasting crude oil price based on time series decomposition. Appl Energy. 2020;267:115035.
  42. 42. Wang J, Cao J, Yuan S, Cheng M. Short-term forecasting of natural gas prices by using a novel hybrid method based on a combination of the CEEMDAN-SE-and the PSO-ALS-optimized GRU network. Energy. 2021;233:121082.
  43. 43. Koy A, Çolak AB. Predicting Stock Market Index and Credit Default Swap Spreads Using Artificial Intelligence and Determining Nonlinear Relations. Arch Adv Eng Sci. 2025;3(3):151–62.
  44. 44. Li M-J, Song C-X, Tao W-Q. A hybrid model for explaining the short-term dynamics of energy efficiency of China’s thermal power plants. Appl Energy. 2016;169:738–47.
  45. 45. Yang X, Liu Y, Gao Y, Zhang Y, Gao Z. An active learning kriging model for hybrid reliability analysis with both random and interval variables. Struct Multidisc Optim. 2014;51(5):1003–16.
  46. 46. Hong L, Li H, Gao N, Fu J, Peng K. Random and multi-super-ellipsoidal variables hybrid reliability analysis based on a novel active learning Kriging model. Comput Methods Appl Mech Eng. 2021;373:113555.
  47. 47. Xin F, Wang P, Hu H, Wang Q, Li L. An efficient system reliability analysis method for flap mechanism under random-interval hybrid uncertainties. Struct Multidisc Optim. 2024;67(8).
  48. 48. Liu J, Shi Y, Ding C, Beer M. Hybrid uncertainty propagation based on multi-fidelity surrogate model. Comput Struct. 2024;293:107267.
  49. 49. Kumar A, Maurya S, Tiwari G. Reliability analysis of deep tunnels in spatially varying brittle rocks using interval and random field modelling. Int J Rock Mech Min Sci. 2024;181:105836.
  50. 50. Wang W, Gao H, Zhou C, Zhang Z. Reliability analysis of motion mechanism under three types of hybrid uncertainties. Mech Mach Theor. 2018;121:769–84.
  51. 51. Kang J, Lee HJ, Jeong SH, Lee HS, Oh KJ. Developing a Forecasting Model for Real Estate Auction Prices Using Artificial Intelligence. Sustainability. 2020;12(7):2899.
  52. 52. Gradojevic N, Kukolj D. Unlocking the black box: Non-parametric option pricing before and during COVID-19. Ann Oper Res. 2022;:1–24. pmid:35233127
  53. 53. Kalampokas T, Tziridis K, Kalampokas N, Nikolaou A, Vrochidou E, Papakostas GA. A Holistic Approach on Airfare Price Prediction Using Machine Learning Techniques. IEEE Access. 2023;11:46627–43.
  54. 54. Giudici P, Piergallini A, Recchioni MC, Raffinetti E. Explainable Artificial Intelligence methods for financial time series. Phys A Stat Mech Appl. 2024;655:130176.
  55. 55. Biltayib BM, Damrah S, Al-Fakeh B, Al-Kanderi A. Validation of directional drilling well design trajectory with a case study in Kuwait. In: 2015 International Conference on Sustainable Mobility Applications, Renewables and Technology (SMART). IEEE; 2015, p. 1–7.
  56. 56. Rumbe G, Hamasha M, Al Mashaqbeh S. A comparison of Holts-Winter and Artificial Neural Network approach in forecasting: A case study for tent manufacturing industry. Results Eng. 2024;21:101899.
  57. 57. Damrah S, DarAssi MH, AbuHour Y. Mathematical Modeling for Exploring the Spread of Cyberattacks Through IoT Devices. In: International Conference on Science, Engineering Management and Information Technology; Cham: Springer Nature Switzerland; 2023, p. 17–27.
  58. 58. M. M, Damrah S, Ali MMA, Al-Husban AA, Palanikumar M. Type-I extension Diophantine neutrosophic interval valued soft set in real life applications for a decision making. Int J Neutrosophic Sci. 2024;24(4):151–64.
  59. 59. DarAssi MH, Damrah S, AbuHour Y. A mathematical study of the omicron variant in a discrete-time Covid-19 model. Eur Phys J Plus. 2023;138(7).
  60. 60. Jiang C, Zhu H, Wang X. Decoupling Control of Outer Rotor Coreless Bearingless Permanent Magnet Synchronous Generator Based on Fuzzy Neural Network Inverse System. IEEE Trans Transp Electrific. 2023;9(3):3908–17.
  61. 61. Geng X, Liu W, Liu X, Wen G, Xue M, Wang J. Optimal Torque Control of the Launching Process with AMT Clutch for Heavy-Duty Vehicles. Machines. 2024;12(6):363.
  62. 62. Que Z, Lin C-J. One-Class SVM Probabilistic Outputs. IEEE Trans Neural Netw Learn Syst. 2025;36(4):6244–56. pmid:38717884
  63. 63. Zhu H, Hao H-K, Lu C. Enhanced support vector machine-based moving regression strategy for response prediction and reliability estimation of complex structure. Aerosp Sci Technol. 2024;155:109634.
  64. 64. Cavalheiro LP, Bernard S, Barddal JP, Heutte L. Random forest kernel for high-dimension low sample size classification. Stat Comput. 2023;34(1).
  65. 65. Lowe CN, Charest N, Ramsland C, Chang DT, Martin TM, Williams AJ. Transparency in Modeling through Careful Application of OECD’s QSAR/QSPR Principles via a Curated Water Solubility Data Set. Chem Res Toxicol. 2023;36(3):465–78. pmid:36877669
  66. 66. Qiao L, Inoue J, Zhu J. Machine learning guided constitutive model and processing map for Fe2Ni2CrAl1.2 multi-principle element alloys. J Mater Res Technol. 2024;29:353–63.
  67. 67. Hochreiter S, Schmidhuber J. Long short-term memory. Neural Comput. 1997;9(8):1735–80. pmid:9377276
  68. 68. Sherstinsky A. Fundamentals of Recurrent Neural Network (RNN) and Long Short-Term Memory (LSTM) network. Physica D. 2020;404:132306.
  69. 69. Fang T, Zheng C, Wang D. Forecasting the crude oil prices with an EMD-ISBM-FNN model. Energy. 2023;263:125407.
  70. 70. Ghosh P, Neufeld A, Sahoo JK. Forecasting directional movements of stock prices for intraday trading using LSTM and random forests. Financ Res Lett. 2022;46:102280.
  71. 71. Yu L, Zhang X, Wang S. Assessing Potentiality of Support Vector Machine Method in Crude Oil Price Forecasting. EURASIA J Math Sci Technol Educ. 2017;13(12).
  72. 72. Rossi B, Inoue A. Out-of-Sample Forecast Tests Robust to the Choice of Window Size. Journal of Business & Economic Statistics. 2012;30(3):432–53.
  73. 73. Damrah S, Satrovic E, Atyeh M, Shawtari FA. Employing the Panel Quantile Regression Approach to Examine the Role of Natural Resources in Achieving Environmental Sustainability: Does Globalization Create Some Difference? Mathematics. 2022;10(24):4795.
  74. 74. Damrah S, Satrovic E, Shawtari FA. How does financial inclusion affect environmental degradation in the six oil exporting countries? The moderating role of information and communication technology. Front Environ Sci. 2022;10:1013326.