Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Shear wave velocity prediction using Long Short-Term Memory Network with generative adversarial mechanism

  • Xingan Fu,

    Roles Conceptualization, Methodology, Writing – original draft

    Affiliations College of Mathematics and Science, Chengdu University of Technology, Chengdu, China, Geomathematics Key Laboratory of Sichuan Province, Chengdu University of Technology, Chengdu, China

  • Youhua Wei ,

    Roles Funding acquisition, Project administration, Supervision

    weiyouhua@cdut.edu.cn

    Affiliation Geomathematics Key Laboratory of Sichuan Province, Chengdu University of Technology, Chengdu, China

  • Yun Su,

    Roles Data curation, Resources, Supervision

    Affiliation Chengdu North Petroleum Exploration and Development Technology Company Limited, Chengdu, China

  • Haixia Hu,

    Roles Conceptualization, Validation, Visualization

    Affiliation Geomathematics Key Laboratory of Sichuan Province, Chengdu University of Technology, Chengdu, China

  • Ji Zhang,

    Roles Writing – review & editing

    Affiliation Geomathematics Key Laboratory of Sichuan Province, Chengdu University of Technology, Chengdu, China

  • Quan Wang

    Roles Writing – review & editing

    Affiliation Geomathematics Key Laboratory of Sichuan Province, Chengdu University of Technology, Chengdu, China

Abstract

Shear wave velocity (Vs) serves as a crucial petrophysical parameter for subsurface characterization, yet its acquisition remains challenging. While long short-term memory (LSTM) networks have emerged as the predominant solution for Vs prediction by synthesizing contextual relationships among conventional logging curves, existing implementations often overlook characteristic discrepancies between training and prediction datasets, leading to suboptimal performance. This study proposes an enhanced LSTM architecture integrated with a generative adversarial mechanism (LSTM-GAM) to address this limitation. The framework employs a dual-component structure: 1) A primary LSTM backbone that captures contextual dependencies across multi-logging sequences, and 2) An adversarial module where the generator minimizes reconstruction errors while the discriminator identifies essential feature representations common to both training and predictive data. This synergistic architecture not only preserves sequential correlations but also enhances cross-domain adaptability through adversarial feature alignment. We validate the model’s efficacy using logging data from two vertical wells in the South China Sea. Comparative experiments demonstrate the proposed LSTM-GAM achieves superior prediction accuracy with a mean absolute error (MAE) of 59.4 m/s and determination coefficient (R²) of 0.9064, outperforming conventional LSTM network. Further ablation studies reveal consistent performance improvements across varied input configurations, confirming the method’s enhanced generalization capability for Vs prediction tasks. The technical advancement provides an effective data-driven solution for shear wave velocity estimation in complex geological environments.

1. Introduction and literature review

1.1 Introduction

Shear wave velocity is a crucial geophysical parameter for studying reservoir AVO (Amplitude Versus Offset) characteristics, playing a significant role in the evaluation of underground reservoirs [1,2]. When the velocity can be accurately determined within a certain depth range in certain wells, it greatly aids in the detection of oil deposits and enhances the oil drilling with lower risk and optimal strategy [3]. However, there is a severe conflict between the availability of shear wave velocity because of various reasons (expensive core sampling testing and unavailable dipole acoustic logging tools et al) [4,5]. Moreover, the petrophysical approach, which involves making an equivalent medium replacement based on the physical properties of underground rocks and calculating the corresponding theoretical P- and S- wave velocities [6], faces limitations. The accuracy of this method heavily relies on the availability of precise petrophysical parameters, including porosity, pore shape, and mineral composition [7]. Unfortunately, these parameters are often difficult to obtain or are subjectively set, which compromises the accuracy of shear wave velocity predictions using petrophysical methods. Consequently, petrophysical analysis tends to depend excessively on specific geological conditions and the prior expertise of professionals, limiting its broader applicability. Given these challenges, there is a pressing need to develop high-precision methods for predicting shear wave velocity to enhance the accuracy of underground reservoir exploration. A promising approach involves synthesizing shear wave velocity data from remaining conventional logging curves (Fig 1), which could provide a more reliable and accessible means of obtaining this vital parameter.

1.2 Literature review

Shear wave velocity is intrinsically related to the matrix mineral type, rock porosity, fluid inclusions, and environment of underground rock formations [810]. Since both P-wave velocity and S-wave velocity are influenced by these factors and exhibit a strong correlation, many researchers use P-wave velocity to regress the empirical formula of S-wave velocity based on data statistics [6,9,1013]. For instantly, Pickett and Castagna established empirical relations for estimating shear wave velocities in limestone and clastic rocks, respectively [9,14]. While the empirical formula is a fast and convenient VS estimation method, the generalization and accuracy of this method are limited due to varations in rock types and geographical locations [15]. Furthermore, to improve prediction accuracy, it is essential to incorporate additional logging curves related to shear wave velocity, which necessitates innovation in modeling approaches. For example, Akhundi achieved a lower prediction error by including porosity and density logging, and many researchers have attempted to integrate more logging parameters to enhance the accuracy of S-wave velocity predictions (Anemangely, 2017; Dvorkin, 2008) [16,17]. The essence of the problem lies in identifying the influencing factors of S-wave velocity and accurately inferring its true value [3]. With the development of technology, machine learning has emerged as a powerful tool for geological parameter estimation [1821]. For instance, Bagheripour et al. (2015) and Anemangely et al. (2019) utilized support vector regression models (SVR) to predict Vs and achive a higher accuracy compared to traditional empirical formulas [22,1]. The SVR algorithm works by mapping features into high-dimensional space and establishing hyperplanes, enabling the separation and recombination of primary features for predicting [23]. To simplify implementation and get high accuracy, Nourafkan and Kadkhodaie-Ilkhchi (2015) employed a fuzzy inference system to predict the velocity, resulting a strong robustness model [24]. Similarly, Wang and Peng (2019) applied the Extreme Learning Machine (ELM) to predict Vs, achiving a model with high computational efficiency [25]. Although traditional machine learning can achieve good numerical fitting and representational accuracy, it may struggle to capture complex or implicit relationships within the data, limiting its sustainability in more intricate scenarios.

Moreover, compared to the low-level and figurative features learned by traditional machine learning methods, deep neural networks are more capable of establishing profound mappings between inputs and outputs. Deep learning excels at extracting key information from the deep level of variables relations and has made great achievements in speech recognition, natural language processing, and reinforcement learning [26]. Kheirollahi et al. (2023) used standard logging curves as inputs of ANN to predict the shear wave velocity in carbonate formations, achieving a lower RMSE error compared to linear regression and ensemble learning [27]. Additionally, Zhang et al. (2018) employed LSTM to approximate missing logging data, resulting in more realistic and usable data [28]. Similarly, Jun Wang et al. (2020) and Jiachun You et al. (2021) incorporated the background information of logging data with LSTM network, ultimately achieving superior results of VS prediction compared to ANN [7]. The LSTM-based models exhibit strong learning capabilities for time-series data, effectively mitigating the risk of overfitting. Moreover, Chen et al. (2022) integrated the feature extraction of CNN into LSTM network, achieving a higher accuracy than the original model [29]. The three thresholds of the LSTM or GRU network greatly increase the ability of the network to assimilate contextual information using adjacent multiple samples, thus effectively improving the precision [30]. Notably, LSTM network demonstrates remarkable computational efficiency through parallel cells. Because its cells have long-term memory ability, LSTM network can retain key effective information and deal with nonlinearity caused by time variability [31]. In summary, the LSTM network exhibits excellent performance in sequence processing [32].

A comprehensive review of the relevant literature reveals that the neural network has the highest accuracy among current ways for Vs prediction (Table 1). While various advanced techniques such as board learning systems and integrated LSTM approaches make a lot of contributions to enhancing model performance, they are still not involved in the feature discrepancy between the trained and predicted data. Summarily, two main challenges remain unresolved:(1) Precision challenge: how to design an overall network capable of representing the S-wave velocity mapping model across diverse geological formations; (2) Generalization challenge: how to mitigate feature differences between training and prediction to eanhance the adaptability of models. To address these dual challenges, this paper proposes a novel framework combining long short-term memory networks with a generative adversarial mechanism. The proposed architecture employs LSTM as its core component for contextual modeling and feature extraction. The features extracted by the LSTM network are subsequently processed by a generator to produce enhanced representations for both training and prediction phases. Finally, a discriminator is employed to classify these features’ origins and progressively minimize the distribution gap between training and prediction. Through this innovative approach, the accuracy of shear wave velocity is substantially improved, thereby facilitating more precise oil and gas exploration.

thumbnail
Table 1. the main structural differences among different algorithms.

https://doi.org/10.1371/journal.pone.0325271.t001

The organizational structure of this paper is as follows. First, we present the theoretical methodologies employed in this study, including Artificial Neural Networks (ANN), Long Short-Term Memory (LSTM), and the novel GAN-LSTM framework proposed herein. Subsequently, we detail the logging dataset utilized, outline the workflow of the proposed algorithm, and define the evaluation metrics. Next, we train the GAN-LSTM network using measured log data and compare its predictive accuracy with several methods discussed in the literature review. Finally, we assess the prediction performance of our approach under varying input length parameters. Experimental results demonstrate the proposed model achieves superior outcomes, highlighting its effectiveness in shear wave velocity estimation.

2. Method and theory

2.1 Artificial Neural Network (ANN)

The artificial neural network (ANN) comprises three essential components: an input layer, hidden layers, and an output layer [33]. In this framework, well log parameters correlated with shear wave velocity are fed into the input layer, followed by the establishment of complex mapping relationships between the log data and shear wave velocity through the hidden layers, ultimately yielding predictions at the output layer (Fig 2). Adjacent hidden layers relate to weights, complemented by activation functions to introduce non-linearity. The final hidden layer extracts characteristic variables through a linear transformation process. To approximate predicted value with actual measurements, a loss function is implemented to quantify the discrepancy between predicted and observed values. Through iterative optimization, the trained model becomes a predictive tool for shear wave velocity estimation.

The formula in two adjacent layers of the network can be expressed as:

(1)

Where, on behalf of the j neuron of the l layer, refer to the i neuron of the l − 1 layer, is the weight matrix of the l layer, represents the corresponding bias, and the sigmoid activation function is represented as .

2.2 Long Short-term Memory Network (LSTM)

The LSTM (Long Short-Term Memory) network, introduced by Hochreiter and Schmidhuber [34], enhances the traditional ANN by transforming hidden layer information into a format that can propagate along the sequence direction. The LSTM network consists of multiple LSTM units (Fig 3), which are used to build the mapping relationship of each sequence information. Each unit incorporates three gate structures, which are used to select the hidden layer information of the mapping relationship from the previous sequence giving into the next hidden layer of the following sequence. The input gate identifies the valuable information to participate into the calculation; The forget gate control discards irrelevant information from the previous time step; The output gate determines which hidden layer information should be passed to the subsequent unit. By leveraging these mechanisms, the LSTM network provides a robust theoretical foundation for extracting sequential features from logging curves, making it highly effective for shear wave prediction. Unlike traditional ANN models, which focus solely on current input variables, the LSTM-based prediction model captures contextual information across sequences, enabling a more comprehensive understanding of the data. This capability makes the LSTM network particularly advantageous for extracting sequential patterns, thereby improving the accuracy and reliability of predictions.

The unit of LSTM includes input gate (it), forgetting gate (ft), status update layer (ct), output gate (ot), hidden layer (ht−1) at t-1 time, hidden layer ht at t time, and input xt at t time. The calculation of LSTM cell can be expressed as the following formulas:

(2)(3)(4)(5)(6)

Where Wi, Wf, Wc, Wo and bi, bf, bc, bo are the weight matrix and bias term of the input gate, the forgot gate, the update gate, and the output gate respectively. “σ” and “tanh” are sigmoid and tanh activation functions, respectively; “ ” expression dot product operation; [a, b] indicates that the tail of vectors a is connected to the head of vectors b.

Taking an input length of 5 as an example, Fig 4 illustrates the overall architecture of the LSTM network, which comprises five LSTM units structured as depicted in Fig 2. The input data consists of logging curves from five adjacent measurement points. Information from each LSTM unit is sequentially propagated to the next unit, culminating in a fully connected layer at the final stage, where all data is integrated for output. Unlike traditional ANN models, which rely on isolated samples for predictions, the LSTM network facilitates information transfer across sequential samples. This unique capability enhances the network’s ability to capture contextual relationships and improves its overall effectiveness in predicting shear wave velocities across diverse geological formations.

2.3 Long Short-term Memory Network Combined with Generative Adversarial Net (GAN-LSTM)

The generative adversarial network (GAN), introduced by Goodfellow in 2014 [2], was initially designed for image generation and enhancement. At its core, the GAN framework operates through a dual-component system: the generator, which guides the creation of data, and the discriminator, which evaluates and corrects discrepancies between the generated data and the target data. This interplay enables the network to produce data that aligns with specific desired characteristics, effectively searching for optimal network parameters under given constraints.

To address the challenge of accuracy degradation caused by feature distribution mismatches between training and prediction datasets, the GAN-LSTM network is proposed. This hybrid architecture comprises three key components: a feature extractor, a generator, and a discriminator. The feature extractor, implemented as an LSTM network, extracts relevant features from both the training and prediction datasets, which are then fed into the generator and discriminator (Fig 5). The discriminator’s role is to differentiate whether the feature vectors originate from the training or prediction data. When the trained discriminator struggles to distinguish the source of the feature data, it indicates that the LSTM-extracted features are shared across both datasets, thereby mitigating feature bias. Finally, the generator leverages these shared features to predict shear wave velocity, producing results that are better adapted to the prediction dataset.

thumbnail
Fig 5. diagram of the GAN-LSTM network.

Note: In the Fig 4, , , are the set of network parameters of feature extractor, the generator and the discriminator respectively; and are loss functions for the generator and the discriminator.

https://doi.org/10.1371/journal.pone.0325271.g005

The mathematical description of the principle of the GAN-LSTM network is given below. Assuming the training set is , and the prediction set is . Where, is the independent variables of training data or prediction data, are the corresponding labels. Assuming the training data and the prediction data have the same output space . However, due to the differences between the two data sources, the total feature space or the single feature distribution of the two sets are different (,), so the conditional probability distribution is also different . The purpose of this study is to use the training data along the prediction data to learn the migration model f, to achieve the target prediction of .

To achieve accurate prediction results, we integrated the LSTM network with a generative adversarial network (GAN-LSTM). The architecture of the GAN-LSTM network is illustrated in Fig 6. The LSTM component serves as the feature extractor, capturing the eigenvectors of the input data. Specifically, the LSTM establishes temporal dependencies and reflections, while the eigenvectors encapsulate the sequence variance and nonlinear relationships derived from the training set. However, these features may not inherently generalize well to the prediction set, making it crucial to minimize the discrepancies between the two datasets.

The discriminator evaluates whether the extracted features originate from the training set or the prediction set. The training set is labeled as 1, and the prediction set as 0, with the discriminator outputting a probability between these two labels. The generator, on the other hand, is guided by the measured shear wave velocity from the training set as its target. The network employs the mean absolute error (MAE) as the generator’s loss function and the binary cross-entropy (BCE) as the discriminator’s loss function.

During training, the generator network aims to minimize the prediction error on the training set, while the discriminator works to reduce the feature distribution differences between the training and prediction sets. This dual optimization ensures that the features used by the generator for shear wave prediction become increasingly similar across both datasets. Therefore, the GAN-LSTM framework achieves a balance between feature alignment and training accuracy, enhancing its generalization capability.

The generator loss function is expressed as:

(7)

The discriminator loss function can be shown as follows:

(8)

Finally, we use the multi-task neural network optimization strategy [35] to integrate the above loss functions:

(9)

Where, is the number of the training set, is the number of the total samples, is the measured value of shear wave, is the predicted value, is the label 0 or 1, is the correct probability; , respectively represent the standard deviation of , .

3. Data collection and preparation

3.1 Data description and processing

The logging data used in this study are from two different drilled wells, Well A and Well B, in Yinggehai Basin, South China Sea. Data from the two wells included seven standard logs: shear velocity (Vs), P-wave velocity (Vp), gamma ray (GR), volume of clay (V clay), density (RHOB), water saturation (SW), and resistivity (RS). These drilling data (logging curves) are used for geological information detection and oil and gas development. The buried depth of the main gas reservoir of Well A is about 4600m ~ 5000m, mainly low permeability sandstone, and the reservoir type is the delta sandstone. The main gas reservoir of Well B is about 4050m ~ 4200m with an apparent gas structure and the reservoir type is submarine fan sandstone. To verify the effectiveness of the proposed GAN-LSTM network, Well A and Well B are divided into the training well and prediction well, respectively. Data from well A are used for training, and data from well B is used for prediction. The statistical indicators of the logging curves are shown in Table 2. The depth of Well A ranges from 3302m to 5042m, the sampling interval is 0.125m, and a total of 13924 samples were measured. The depth of Well B ranges from 3335m to 4076m, and the total sampling points are 5924.

thumbnail
Table 2. Statistical characteristics of variables of two wells.

https://doi.org/10.1371/journal.pone.0325271.t002

Table 2 is the statistical index description of the conventional logging data of two Wells in the working area. The log curves of two wells are shown in Figs 7 and 8, respectively.

To suit the training of the network, we apply the following reversible transformation to standardize the data:

(10)

Where, and represent the mean value and standard deviation of the logging curve, respectively. is the normalization of . The mean and standard deviation can be used to restore normalized data to the original data. The original distribution features of the processed data are saved, and the range of the data is narrowed.

Various logging attributes exhibit correlations with shear wave velocity. For instance, an increase in water saturation (SW) may result in a decrease in shear wave velocity, while volume of clay (VCL) and gamma ray (GR) can reflect the clay content within the formation. Additionally, the ratio of compressional wave velocity (VP) to shear wave velocity (VS) tends to show a slight increase in shale-rich strata. Furthermore, both compressional and shear wave velocities are commonly influenced by factors such as rock matrix composition, porosity, and the inclusions. Resistivity logs, on the other hand, provide insights into the characteristics of pore fluids within the subsurface. Collectively, these logging curves demonstrate intrinsic relationships not only with each other but also with shear wave velocity. To statistically analyze these relationships, we present cross-plots of each logging curve from Well A against shear wave velocity, as illustrated in Fig 9.

thumbnail
Fig 9. Correlation of logging curves with shear wave velocity.

https://doi.org/10.1371/journal.pone.0325271.g009

Fig 10 illustrates the workflow for transforming logging curves into input matrices. Six logging curves are sampled using a fixed window length to capture actual data points. These data points are then sequentially organized into sample matrices, moving from top to bottom with a step size of 1. Each sample matrix is paired with the corresponding shear wave velocity value at the final time point within the matrix, serving as the label for training. Once the model is trained completely, it is applied to the prediction set to generate the complete shear wave velocity prediction curve.

3.2 Parameter settings and algorithm flowchart

The network structure parameters, including the number of hidden layers and hidden units, are determined through an iterative trial-and-error process. Initially, the number of hidden units is fixed, and the number of hidden layers is incrementally increased until a structure is identified where further layers do not significantly reduce the loss value. Subsequently, the number of hidden neurons is adjusted within the determined number of layers to achieve a stable loss value. The specific parameters of the proposed GAN-LSTM network are detailed in Table 3.

In this architecture, the LSTM network serves as the feature extractor and comprises a 2-layer LSTM network with 10 hidden units. It takes as input six different logging attributes from seven adjacent samples and outputs 10 neurons representing distinct features. The generator consists of a single layer with 12 hidden units. Its role is to transform the feature vectors extracted by the LSTM into predicted values. Parallel to the generator, the discriminator is structured as a 4-layer artificial neural network (ANN) with 8 hidden units. It evaluates the discrepancy between the feature vectors extracted from the training set and the prediction set. As the discriminator’s loss decreases, the feature variables extracted by the LSTM network increasingly capture the shared characteristics of both datasets. This ensures that the generator’s predictions are better generalized to the prediction set, ultimately enhancing the model’s performance.

After configuring the network parameters, the workflow of the GAN-LSTM network (Fig 11) is designed to estimate shear wave velocity. The process begins with the LSTM network extracting and generating critical feature vectors for shear wave prediction. These feature vectors are then utilized to simultaneously minimize prediction errors on the training set and reduce feature discrepancies between the training and prediction sets. The detailed procedure is as follows: First, the LSTM network extracts features from the input data. These features are subsequently fed into both the discriminator and the generator, where the respective loss functions, and , are calculated, respectively. aims to minimize the differences between the feature distributions of the two datasets, while ensures that the predicted values closely approximate the actual values on the training set. The overall loss L balances these two objectives, enabling the network to predict shear wave velocity effectively while maintaining feature consistency across datasets.

3.3 Evaluation metrics

R2 and MAE are selected as evaluation metrics to measure the quality of the method involved in this paper. R2 indicates the correlation between the predicted value and the measured value and judges whether there is an excessive deviation between them. MAE is the mean absolute error between the predicted value and the measured actual value, which can evaluate how close the predicted value is to the measured value. The formula for R2 and MAE can be expressed as follows:

(11)(12)

Where, is the estimated value by the prediction method, is the measured shear wave velocity, is the mean of measured values, and n is the number of predicted samples.

4. Results and discussion

4.1 Experiment 1

To find out the difference between the LSTM network and the GAN-LSTM network on the prediction set, we selected two point with the same measured value in two wells as a comparison reference for the applied method. The first point is located at the 2092nd sample point of Well A with a depth of 3527m, and the value is 2092.36m/s. The second one is the 1315th sample point of Well B with a depth of 3500m, and the value is 2092.58m/s. While the LSTM network and GAN-LSTM network are used to predict these two points, five feature vectors generated for prediction are shown in Table 4. The Euclidean squared distance of sample point 1 and sample point 2 are calculated. The results generated by the two methods are 0.52 and 0.04 respectively. That shows the GAN-LSTM network can effectively reduce the distance between the eigenvectors of the training set and the prediction set.

The experimental results demonstrate that the proposed GAN-LSTM network can generate highly similar feature vectors between the prediction set and training set while maintaining training accuracy, thereby significantly enhancing prediction generalization capability. To comprehensively evaluate the performance of our proposed network, we conducted comparative studies using both ANN and LSTM networks as benchmarks. The ANN architecture consists of four layers with ten hidden units per layer, while the LSTM network employs a two-layer structure with twelve neurons in each layer and an input sample length of seven. The detailed parameter configuration of the GAN-LSTM network is presented in Table 2. All networks were optimized using the Adam optimizer for parameter fine-tuning.

Fig 12 illustrates the prediction performance of the three networks on the prediction set after being trained with Well A data. While all three methods demonstrate reasonable capability in capturing the general trend of the actual values, the GAN-LSTM network exhibits superior performance with prediction results that are substantially closer to the measured curve, particularly in the depth intervals near 3500m and 3900m, where its advantages are most pronounced.

thumbnail
Fig 12. Predictive results of three neural networks in well B.

https://doi.org/10.1371/journal.pone.0325271.g012

The training specifications for both LSTM and GAN-LSTM networks are detailed in Table 5. It should be noted that the testing time, which represents the application of trained model parameters to generate predictions from input variables, is relatively brief and thus negligible. The GAN-LSTM architecture incorporates an additional discriminative network component based on ANN structure, resulting in a greater number of trainable parameters compared to the standard LSTM network, consequently requiring longer training times. All training procedures are conducted using PyTorch 1.10.2, with hardware configuration comprising an NVIDIA GeForce RTX 3050 Ti GPU and an Intel Core i5 CPU.

thumbnail
Table 5. Network deployment environment and training time.

https://doi.org/10.1371/journal.pone.0325271.t005

(Mark the upper and lower parts of the curve in red for comparison)

As illustrated in the red frame of Fig 12, the prediction results demonstrate a progressive improvement in consistency with the actual curve from sub Fig (a) to (c). Notably, sub Fig (c) exhibits superior performance in handling areas with significant fluctuations compared to sub Figs (a) and (b). Furthermore, at the 3700m mark, both sub Figs (a) and (b) show slightly inferior performance relative to sub Fig (c), potentially indicating a substantial divergence in characteristics between the training and prediction datasets. Overall, the predictive performance of sub Fig (c) is markedly superior to that of the preceding two subFigs.

Neural networks have demonstrated their effectiveness in predicting shear wave velocity with minimal deviation. To further illustrate this, two specific intervals are highlighted in red for detailed comparison, as shown in Figs 13 and 14. The results clearly indicate that the predictions generated by the GAN-LSTM model align more closely with the true values compared to those from ANN or LSTM. This improvement is attributed to the GAN-LSTM’s ability to reduce the disparity between prediction and training feature vectors, thereby significantly enhancing prediction accuracy. Moreover, Fig 14 reveals that the sequence predicted by the proposed GAN-LSTM model follows the actual data more closely, even in regions where the curve exhibits significant variations. This underscores the importance of extracting common essential features in addressing the accuracy decline caused by feature discrepancies. By incorporating the discriminator component, the GAN-LSTM network effectively mitigates the influence of non-generalizable features on the prediction set, achieving remarkable predictive performance.

thumbnail
Fig 13. Local amplification comparison in Well B (the upper part marked in

Fig 12).

https://doi.org/10.1371/journal.pone.0325271.g013

thumbnail
Fig 14. Local amplification comparison in Well B (the lower part marked in

Fig 12).

https://doi.org/10.1371/journal.pone.0325271.g014

A comparative analysis of the prediction results from (a) to (c) reveals a progressive improvement in the fitting accuracy to the actual curve. Specifically, the prediction curve in (b) demonstrates superior smoothness and adaptability compared to (a), which can be attributed to the LSTM network’s capability to capture contextual information within the sample length. More significantly, the GAN-LSTM network in (c) effectively mitigates the bottleneck effect caused by the characteristic discrepancy between training and prediction datasets during model application. This enhanced performance stems from the network’s ability to simultaneously account for contextual relationships and address uneven feature distribution, making it particularly suitable for practical forecasting scenarios.

Fig 15(a) presents the frequency histogram of the actual VS curve, while Figs 15(b)15(d) illustrate the frequency distributions of the prediction results obtained from ANN, LSTM, and GAN-LSTM, respectively. A comparative analysis reveals that the GAN-LSTM prediction results demonstrate remarkable similarity to the actual distribution curve in Fig 15(a), particularly in the velocity ranges around 1800 and 2500. This observation provides substantial evidence that GAN-LSTM effectively adapts to the characteristic distribution of the actual data, outperforming the other two methods in achieving a more accurate alignment between predicted and actual distributions.

thumbnail
Fig 15. Comparison of the probability distribution of the three prediction methods.

https://doi.org/10.1371/journal.pone.0325271.g015

Fig 16 presents a comparative analysis of relative errors generated by the three methods, with the scatter plots (from top to bottom) representing the ratio of predicted deviations to actual values. The ANN results in sub Fig (a) exhibit a substantial number of data points with relative errors exceeding 0.2, which can be attributed to its inability to capture contextual relationships among multiple sample points. In contrast, sub Figs (b) and (c) demonstrate that GAN-LSTM produces significantly less fluctuation, thereby enhancing its practical application value.

thumbnail
Fig 16. The relative error of the three forecasting methods.

https://doi.org/10.1371/journal.pone.0325271.g016

For quantitative performance evaluation, we employed MAE (Mean Absolute Error) and R² (Coefficient of Determination) metrics to assess the accuracy of the methods discussed in the literature review. Table 6 provides a comprehensive quantitative comparison of prediction results for both the training well (Well A) and prediction well (Well B) across different methods. Each method was trained on Well A and subsequently applied to Well B for prediction. The analysis reveals that while the LSTM network achieves the lowest MAE (45.87) in the training well, marginally better than GAN-LSTM’s 46.84, the GAN-LSTM network attains the highest R² value of 0.968. More importantly, in the prediction well, GAN-LSTM demonstrates superior performance with MAE and R² values of 59.41 and 0.906, respectively. These results collectively indicate that the proposed network in this study significantly outperforms other methods in both prediction precision and generalization capability. The comparative performance metrics of various methods in the prediction well are visually presented in Fig 17.

thumbnail
Table 6. Comparison of accuracy under different methods in two wells.

https://doi.org/10.1371/journal.pone.0325271.t006

As evidenced by Fig 17, the GAN-LSTM model demonstrates superior performance in Well B, exhibiting significantly lower MAE and higher R² values compared to other methods. While the performance difference between LSTM and GAN-LSTM in Well A appears marginal, the GAN-LSTM network achieves substantially higher prediction accuracy in the prediction well. These results conclusively demonstrate that the GAN-LSTM architecture significantly enhances the prediction accuracy of actual shear wave velocity measurements. This improvement can be attributed to the network’s unique capability to utilize the prediction set’s independent variables to minimize the feature distance from the training set, thereby effectively reducing prediction inaccuracies caused by feature distribution dissimilarities.

The quantitative analysis in Fig 17 reveals that GAN-LSTM achieves optimal performance metrics, with the highest R² and lowest MAE values among all methods. Notably, the prediction results of GAN-LSTM show the closest alignment with actual values, as evidenced by the minimal discrepancy between the R² and MAE values of the prediction set and those of the training set. This close correspondence between prediction and training set metrics further confirms the superior generalization capability of the GAN-LSTM approach compared to conventional methods.

Fig 18 demonstrates that the prediction results of the GAN-LSTM method exhibit a significantly tighter clustering around the y = x line, corresponding to its optimal performance metrics of the lowest MAE and highest R² values (as presented in Table 5). This distribution pattern strongly indicates an exceptional consistency between the network’s predicted values and the actual measurements. The GAN-LSTM predictions maintain remarkable agreement with actual values across all depth intervals, with no observable dispersion or disorder in the predicted data points. In contrast, other methods display more scattered data points, suggesting potential prediction biases at various depth levels. This dispersion pattern may reflect underlying characteristic discrepancies between the prediction data and training data, which the GAN-LSTM network appears to effectively mitigate through its advanced architecture.

Fig 19 presents a comprehensive comparison of the proposed GAN-LSTM model with 1DCNN and GRU neural network architectures. The 1DCNN architecture employs a multi-stage feature extraction process, initially projecting the six logging attributes into a 135-dimensional space, followed by the formation of five feature vectors through three one-dimensional convolutional layers (with a kernel length of 3 and stride of 3) for final shear wave velocity determination. The GRU model, while maintaining identical structural parameters to GAN-LSTM (comprising two layers with 10 hidden units each, an input size of 6, and an input length of 7), simplifies the gating mechanism by combining the input and forget gates.

The experimental results reveal distinct performance characteristics among the models. As shown in Fig 19(a), the 1DCNN demonstrates limited capability in extracting contextual information from adjacent samples, resulting in the poorest shear wave velocity prediction performance. Fig 19(b) illustrates that while the GRU model achieves reduced training time through parameter optimization, this comes at the cost of slightly diminished prediction accuracy due to its simplified gating architecture. In contrast, Fig 19(c) demonstrates the superior prediction performance of the proposed GAN-LSTM model on Well B, which achieves the highest prediction accuracy among the three models. This performance advantage underscores GAN-LSTM’s enhanced capabilities in context extraction and feature difference compensation compared to both CNN and GRU architectures.

4.2 Experiment 2

To conduct an in-depth analysis of the GAN-LSTM network’s performance, we systematically compared its prediction capabilities with the LSTM network across varying input lengths. This comparative study, when integrated with the findings from Experiment 1, offers a more comprehensive evaluation framework. We examined four distinct input lengths (5, 7, 9, and 11) while maintaining identical network parameters as those used in Experiment 1 for both architectures. The comparative prediction results, illustrated in Fig 20, demonstrate that the GAN-LSTM network achieves significantly enhanced prediction accuracy in Well B across all tested input lengths compared to its LSTM counterpart. This consistent performance improvement not only validates the network’s effectiveness but also highlights its robustness to variations in input sequence length.

thumbnail
Fig 20. Prediction of LSTM and proposed GAN-LSTM network under different input in Well B.

(Input length from left to right is set as 5, 7, 9, 11). (Mark the upper and lower parts of the curve in red for comparison).

https://doi.org/10.1371/journal.pone.0325271.g020

For detailed performance analysis, Figs 21 and 22 present localized comparisons of the marked regions (highlighted in red) from Fig 20. Visual inspection clearly reveals that the GAN-LSTM predictions maintain closer alignment with the true value curve than those of the LSTM network. Notably, as shown in Fig 22, the GAN-LSTM network exhibits superior stability in regions with substantial variations, where the LSTM network’s performance tends to degrade. This comparative analysis substantiates that the proposed method offers both enhanced prediction accuracy and greater stability across diverse geological conditions.

thumbnail
Fig 21. Local magnification of LSTM and GANLSTM’s prediction under different input length.

(upper part in Fig 20). (The input length of (a), (b), (c), (d) is set as 5,7,9,11, respectively).

https://doi.org/10.1371/journal.pone.0325271.g021

thumbnail
Fig 22. Local magnification of LSTM and GANLSTM’s prediction under different input length.

(lower part in Fig 20). (The input length of (a), (b), (c), (d) is set as 5,7,9,11, respectively).

https://doi.org/10.1371/journal.pone.0325271.g022

Fig 20 presents a comparative analysis of prediction results from LSTM and GAN-LSTM networks across varying input parameters (with input lengths of 5, 7, 9, and 11 from left to right, respectively). The results within the red boxes (subFigs a-d) reveal that the LSTM predictions (represented by orange curves) exhibit significantly greater deviation from the actual measurements (green curves) compared to other results. Notably, the magnitude of deviation appears positively correlated with the fluctuation intensity of the curves. This phenomenon suggests the existence of feature discrepancies between the training and prediction datasets at certain intervals. The GAN-LSTM network, through its domain transformation mechanism, demonstrates superior capability in addressing these feature inconsistencies, thereby substantially improving prediction accuracy under such challenging conditions.

Figs 2122 demonstrate that the GAN-LSTM network’s prediction results exhibit superior alignment with the actual curve across various input parameters compared to the LSTM network. While both architectures incorporate contextual information within the sample length, resulting in smooth-adjusted and trend-adapted curves, the GAN-LSTM network significantly outperforms its counterpart by effectively mitigating prediction inaccuracies caused by uneven feature distribution. This enhanced capability to adapt to practical conditions makes GAN-LSTM particularly effective in regions with heterogeneous features.

As illustrated in Fig 20, the comparative analysis of MAE and R² metrics between the two networks under different parameters in the prediction well reveals distinct performance characteristics. Both LSTM and GAN-LSTM demonstrate the fundamental capability to capture the general trend of the actual curve, validating their applicability for shear wave velocity prediction. However, the GAN-LSTM predictions (represented by the orange curve) consistently show closer alignment with actual measurements, particularly in regions with significant variations (Fig 21). This superior performance stems from GAN-LSTM’s ability to reduce the feature discrepancy between prediction and training datasets, resulting in enhanced accuracy and robustness.

The quantitative analysis presented in Table 7 provides further evidence of GAN-LSTM’s advantages. While both methods show comparable accuracy in Well A (training well), GAN-LSTM achieves significantly better R² and MAE metrics in Well B (prediction well) across all tested input lengths. These results collectively demonstrate that GAN-LSTM maintains the prediction accuracy of the training set while substantially improving the performance on the prediction set. The network’s consistent superiority under varying input lengths confirms its enhanced generalization capability and practical applicability in shear wave velocity prediction tasks.

thumbnail
Table 7. Evaluation of prediction in Well B under different input window length.

https://doi.org/10.1371/journal.pone.0325271.t007

Figs 23 and 24 present column graph visualizations derived from Table 6, facilitating a comprehensive comparison of LSTM and GAN-LSTM performance across different input parameters during prediction. Fig 23 reveals that the proposed GAN-LSTM method consistently achieves higher R² values than LSTM across all tested input lengths, indicating superior prediction consistency. Complementary to this finding, Fig 24 demonstrates that GAN-LSTM maintains lower MAE values compared to LSTM for various input configurations. These quantitative comparisons lead to two significant conclusions: first, the proposed network achieves substantially reduced MAE in prediction; second, it demonstrates markedly higher R² values than the conventional LSTM network.

thumbnail
Fig 23. Comparison of R2 of results between LSTM and GANLSTM.

https://doi.org/10.1371/journal.pone.0325271.g023

thumbnail
Fig 24. Comparison of MAE of results between LSTM and GANLSTM.

https://doi.org/10.1371/journal.pone.0325271.g024

The experimental results collectively indicate that the GAN-LSTM architecture significantly enhances prediction accuracy across diverse input lengths, thereby ensuring both effectiveness and stability in practical applications. The graphical comparisons clearly demonstrate that regardless of sample input variations, the proposed network maintains superior performance metrics, with consistently lower MAE errors and higher R² values than its LSTM counterpart. This performance advantage stems from the network’s dual capability: not only does it effectively capture contextual information within the curves, but it also addresses the critical issue of feature distribution discrepancies between training and prediction sets. This comprehensive approach significantly enhances the model’s generalization capability, making it particularly suitable for practical situations.

While this study has made significant progress in addressing the feature discrepancy between training and prediction sets, several limitations and areas for future improvement warrant attention:

First, the parameters within the deep learning network for shear wave velocity prediction lack explicit physical interpretation, and their actual significance in the final model remains unclear. This interpretability gap represents a fundamental challenge in applying deep learning to geophysical problems. Second, the integration of theoretical mechanisms into the shear wave velocity prediction framework merits further investigation, as such incorporation could potentially enhance both model performance and physical plausibility. Third, emerging deep learning paradigms, such as instance learning [37,38] and other advanced architectures, may offer novel approaches to address current limitations.

Although this research represents an advancement over traditional LSTM through substantial modifications, the core architecture remains fundamentally similar. To better meet the demands of production practice, future work should focus on developing innovative shear wave prediction models that not only improve accuracy but also enhance interpretability and physical consistency. These advancements could potentially lead to more reliable and practical solutions for shear wave velocity prediction in various geological contexts.

5. Conclusion

Shear wave velocity logging holds significant position in petroleum exploration. To address the scarcity of actual VS curves, we have developed a neural network-based method for high-precision prediction of shear wave curves. Through comprehensive literature review and experimental analysis, we have identified that LSTM networks possess superior capability in extracting environmental context information, making them more suitable for time-varying sequences compared to ANN, thereby achieving higher prediction accuracy. However, logging data characteristics exhibit inherent variability across both temporal (depth) and spatial (well location) dimensions. To address the feature discrepancy between training and prediction sets, we propose a novel architecture combining long short-term memory networks with generative adversarial networks (GAN-LSTM). This approach demonstrates superior performance, with similar frequency distributions to actual curves and exhibiting lower relative errors and higher consistency compared to other neural network methods. Experimental results substantiate these advantages: in Experiment 1, GAN-LSTM achieves optimal metrics in the prediction well (MAE = 59.4101, R² = 0.9064), surpassing all comparative methods. Experiment 2 further demonstrates the network’s robustness, showing significantly better prediction results across various input lengths compared to conventional LSTM networks. Notably, GAN-LSTM exhibits exceptional performance in regions with substantial geological variations. These findings collectively demonstrate that the proposed network architecture successfully extracts essential features and significantly enhances practical application accuracy in shear wave velocity prediction.

Supporting information

S1. Data_wellA&wellB.xlsx. The part of Information Data.

https://doi.org/10.1371/journal.pone.0325271.s001

(XLSX)

References

  1. 1. Anemangely M, Ramezanzadeh A, Amiri H. Machine learning technique for the prediction of shear wave velocity using petrophysical logs. Journal of Petroleum Science and Engineering. 2019;174:306–27.
  2. 2. Gholami A, Amirpour M, Ansari HR. Porosity prediction from pre-stack seismic data via committee machine with optimized parameters. Journal of Petroleum Science and Engineering. 2022;210:110067.
  3. 3. You J, Cao J, Wang X. Shear wave velocity prediction based on LSTM and its application for morphology identification and saturation inversion of gas hydrate. Journal of Petroleum Science and Engineering. 2021;205:109027.
  4. 4. Sohail GM, Hawkes CD. An evaluation of empirical and rock physics models to estimate shear wave velocity in a potential shale gas reservoir using wireline logs. Journal of Petroleum Science and Engineering. 2020;185:106666.
  5. 5. Khatibi S, Aghajanpour A. Machine learning: A useful tool in geomechanical studies, a case study from an offshore gas field. Energies. 2020;13(14):3528.
  6. 6. Ali M, Ma H, Pan H. Building a rock physics model for the formation evaluation of the Lower Goru sand reservoir of the Southern Indus Basin in Pakistan. Journal of Petroleum Science and Engineering. 2020;194:107461.
  7. 7. Wang J, Cao J, Yuan S. Shear wave velocity prediction based on adaptive particle swarm optimization optimized recurrent neural network. Journal of Petroleum Science and Engineering. 2020;194:107466.
  8. 8. Lee MW. A simple method of predicting S-wave velocity. Geophysics. 2006;71:F161–4.
  9. 9. Pickett GR. Acoustic character logs and their applications in formation evaluation. Journal of Petroleum Technology. 1963;15:659–67.
  10. 10. Castagna JP, Batzle ML, Eastwood RL. Relationships between compressional‐wave and shear‐wave velocities in clastic silicate rocks. GEOPHYSICS. 1985;50(4):571–81.
  11. 11. Carroll RD. The determination of the acoustic parameters of volcanic rocks from compressional velocity measurements. International Journal of Rock Mechanics and Mining Sciences & Geomechanics Abstracts. 1969;6(6):557–79.
  12. 12. Castagna JP, Batzle ML, Eastwood RL. Relationships between compressional-wave and shear-wave velocities in clastic silicate rocks. Geophysics. 1985;50(4):571–81.
  13. 13. Brocher TM. Empirical relations between elastic wavespeeds and density in the Earth’s crust. Bulletin of the Seismological Society of America. 2005;95(6):2081–92.
  14. 14. Li J, Gao G. Digital construction of geophysical well logging curves using the LSTM deep-learning network. Frontiers in Earth Science. 2023;10:1041807.
  15. 15. Anemangely M, Ramezanzadeh A, Tokhmechi B. Shear wave travel time estimation from petrophysical logs using ANFIS-PSO algorithm: A case study from Ab-Teymour Oilfield. Journal of Natural Gas Science and Engineering. 2017;38:373–87.
  16. 16. Dvorkin JP. Yet another Vs equation. GEOPHYSICS. 2008;73(2):E35–9.
  17. 17. Wang H, Xu Y, Tang S, Wu L, Cao W, Huang X. Well log prediction while drilling using seismic impedance with an improved LSTM artificial neural networks. Front Earth Sci. 2023;11.
  18. 18. Zerrouki AA, Aïfa T, Baddari K. Prediction of natural fracture porosity from well log data by means of fuzzy ranking and an artificial neural network in Hassi Messaoud oil field, Algeria. Journal of Petroleum Science and Engineering. 2014;115:78–89.
  19. 19. Iturrarán-Viveros U, Parra JO. Artificial Neural Networks applied to estimate permeability, porosity and intrinsic attenuation using seismic attributes and well-log data. Journal of Applied Geophysics. 2014;107:45–54.
  20. 20. Bagheripour P, Gholami A, Asoodeh M. Support vector regression based determination of shear wave velocity. Journal of Petroleum Science and Engineering. 2015;125:95–9.
  21. 21. Zhang Q, Jin Y, Wang G. Improving stability and safety in concrete structures against high-energy projectiles: a machine learning perspective. Frontiers in Materials. 2024;11:1416918.
  22. 22. Nourafkan A, Kadkhodaie-Ilkhchi A. Shear wave velocity estimation from conventional well log data by using a hybrid ant colony–fuzzy inference system: A case study from Cheshmeh–Khosh oilfield. Journal of Petroleum Science and Engineering. 2015;127:459–68.
  23. 23. Wang P, Peng S. On a new method of estimating shear wave velocity from conventional well logs. Journal of Petroleum Science and Engineering. 2019;180:105–23.
  24. 24. Feng F, Ghorbani H, Radwan AE. Predicting groundwater level using traditional and deep machine learning algorithms. Front Environ Sci. 2024;12.
  25. 25. Kheirollahi H, Manaman NS, Leisi A. Robust estimation of shear wave velocity in a carbonate oil reservoir from conventional well logging data using machine learning algorithms. Journal of Applied Geophysics. 2023;211:104971.
  26. 26. Zhang DX, Chen YT, Meng J. Synthetic well logs generation via Recurrent Neural Networks. Petroleum Exploration and Development. 2018;45(4):629–39.
  27. 27. Chen T, Gao G, Wang P. Prediction of shear wave velocity based on a hybrid network of two-dimensional convolutional neural network and gated recurrent unit. Geofluids. 2022;2022.
  28. 28. Liu Y, Gao C, Zhao B. Shear-wave velocity prediction based on the CNN-BiGRU integrated network with spatiotemporal attention mechanism. Processes. 2024;12(7):1367.
  29. 29. Li H, Li S, Ghorbani H. Data-driven novel deep learning applications for the prediction of rainfall using meteorological data. Frontiers in Environmental Science. 2024;12:1445967.
  30. 30. Lin B, Wei C, Feng F. A vehicle velocity prediction method with kinematic segment recognition. Applied Sciences. 2024;14(12):5030.
  31. 31. Onalo D, Adedigba S, Khan F. Data driven model for sonic well log prediction. Journal of Petroleum Science and Engineering. 2018;170:1022–37.
  32. 32. Hochreiter S, Schmidhuber J. Long short-term memory. Neural Comput. 1997;9(8):1735–80. pmid:9377276
  33. 33. Goodfellow I, Pouget-Abadie J, Mirza M. Generative adversarial nets. Advances in Neural Information Processing Systems. 2014;27.
  34. 34. Kendall A, Gal Y, Cipolla R. Multi-task learning using uncertainty to weigh losses for scene geometry and semanticsC. Proceedings of the IEEE conference on computer vision and pattern recognition; 2018. p. 7482–7491.
  35. 35. Chen T, Gao G, Wang P. Prediction of shear wave velocity based on a hybrid network of two‐dimensional convolutional neural network and gated recurrent unit. Geofluids. 2022;2022(1):9974157.
  36. 36. Zhou R, Hu X, Yuan B, et al. Lithology classification system for well logging based on bidirectional gated recurrent unitC. //2021 4th International conference on artificial intelligence and big data (ICAIBD). IEEE, 2021. p. 599–603.
  37. 37. Waqas M, et al. Exploring multiple instance learning (MIL): A brief survey. Expert Systems with Applications. 2024. p. 123893.
  38. 38. Waqas M, Tahir MA, Al-Maadeed S, Bouridane A, Wu J. Simultaneous instance pooling and bag representation selection approach for multiple-instance learning (MIL) using vision transformer. Neural Computing and Applications. 2024;36(12):6659–80.