Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Accurate time-series forecasting of floating platform motion via a reinforced fusion CNN–BiLSTM–attention model

  • Huiyuan Zheng,

    Roles Conceptualization, Formal analysis, Investigation, Methodology, Software, Visualization, Writing – original draft

    Affiliation National Ocean Technology Center, Tianjin, China

  • Shicheng Wang,

    Roles Formal analysis, Investigation, Methodology, Visualization

    Affiliation National Ocean Technology Center, Tianjin, China

  • Shihua Li,

    Roles Conceptualization, Formal analysis, Methodology, Writing – original draft

    Affiliation National Ocean Technology Center, Tianjin, China

  • Kuan Lu ,

    Roles Conceptualization, Funding acquisition, Investigation, Methodology, Project administration, Supervision, Writing – review & editing

    lukuan0322@163.com

    Affiliations National Ocean Technology Center, Tianjin, China, The Key Laboratory of Ocean Observation Technology of the Ministry of Natural Resources, Tianjin, China

  • Xin Wang,

    Roles Methodology, Project administration, Supervision

    Affiliations National Ocean Technology Center, Tianjin, China, The Key Laboratory of Ocean Observation Technology of the Ministry of Natural Resources, Tianjin, China

  • Yuzheng Liu,

    Roles Investigation, Software

    Affiliation National Ocean Technology Center, Tianjin, China

  • Xiao Wu

    Roles Methodology, Software

    Affiliation National Ocean Technology Center, Tianjin, China

Abstract

Accurate motion prediction of floating platforms is critical for ensuring operational safety in offshore engineering applications or marine equipment testing. However, the strong nonlinearity and non-stationary characteristics induced by complex marine environments pose significant challenges to conventional prediction models. This study proposes a reinforced hybrid neural network (CNN-BiLSTM-Attention) integrated with advanced signal processing techniques to address these challenges. The methodology combines complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN) for multi-scale signal analysis, coupled with temporal feature engineering through sliding window optimization. And the architecture innovatively integrates convolutional neural networks for spatial pattern extraction, bidirectional long short-term memory networks for temporal dependency modeling, and attention mechanisms for dynamic feature weighting. By analyzing datasets generated via hydrodynamic simulations, this study elucidates the model’s physical interpretability and establishes a closed-loop validation framework between data-driven methods and physics-based models. Finally, the predictive performance of the model is evaluated using motion datasets of the proportional platform in the water pool test under different working conditions, demonstrating its broad applicability and transferability by assessed using a dual-stage EWMA control line. Overall, the proposed CNN-BiLSTM-Attention model and its data-physics integrated validation method provide a reliable, interpretable and transferable solution for floating platform motion prediction, which can break through the limitations of single analysis methods, and provide a new research idea for integrating data-driven and physics-based methods in the field of ocean engineering.

Introduction

The offshore floating platform exhibits strongly nonlinear and non-stationary motion characteristics under complex marine environmental conditions, e.g., coupled wind, wave, and current interactions with structures [1,2]. As the primary testing platform of the National Marine Test Site (Weihai), the “GuoHaiShi 1” requires precise motion prediction to ensure the safety and reliability of maritime experiments [3]. In this context, accurate motion prediction plays a pivotal role in:

Operational safety assurance: by providing early warning of extreme platform motions and enabling proactive attitude adjustments or emergency shutdowns to minimize risks to personnel and critical equipment;

Data quality guarantee: through the maintenance of stable platform conditions essential for acquiring high‑precision sensor measurements and for rigorous validation of hydrodynamic models;

Design optimization support: since long‑term statistics of prediction errors and analyses under extreme sea states furnish objective guidance for improving structural configurations and for the development of more resilient floating platforms capable of withstanding severe wind and wave loading.

However, traditional prediction methods based on hydrodynamic equations or empirical models may face challenges such as modeling inaccuracies and limited generalization capabilities [4,5]. With the development of computer science, deep learning based on neural networks has become increasingly mature and has powerful nonlinear solving capabilities, making it possible to use deep learning to address problems related to marine structures. In recent years, deep learning-based time-series prediction techniques have demonstrated significant advantages in nonlinear dynamic modeling [6].

While components such as CEEMDAN, convolutional neural networks (CNN), bidirectional LSTM (BiLSTM) and attention mechanisms are each well established in time-series forecasting, recent maritime and geoscience studies show that carefully tailored hybridizations of these methods substantially improve short-term motion and load prediction under irregular, nonlinear sea states. For example, a wavelet-PCA preprocessing step combined with an optimized DC-LSTM yielded marked gains in short-term ship motion forecasting by reducing input dimensionality and improving LSTM convergence and stability [7]. Similarly, multi-stage spatial–temporal hybrids that couple CNN feature extraction with multi-recurrent architectures (and advanced hyperparameter optimization) have demonstrated superior 6-DOF prediction accuracy compared with single-model baselines for ship-motion tasks [8]. Complementary evidence shows that combining different signal-decomposition schemes with LSTM (e.g., EMD/EEMD/CEEMDAN variants) improves semi-submersible short-term motion forecasts by isolating multi-scale components and reducing model variance, while CEEMDAN-based preprocessing has also been successfully applied to mooring-tension prediction in semi-submersible platforms [911]. Outside marine motion prediction, attention augmented recurrent frameworks have proven effective for complex oceanographic fields (e.g., SST and pan-Arctic sea-ice forecasting), demonstrating the mechanism’s ability to weight temporally varying features and enhance generalization [12,13]. Finally, for robust online assessment and alarm/monitoring tasks, recent adaptive EWMA control strategies and engineering fault-detection frameworks (which combine attention mechanisms, multimodule feature extraction, and adaptive strategies) point to practical procedures for model performance tracking and anomaly detection in operational marine systems [1416].

Usually, these are very cross-application prospect through these contents: Convolutional Neural Networks (CNNs) effectively extract spatial-localized features from multisensor data [17,18], Bidirectional Long Short-Term Memory networks (BiLSTMs) capture bidirectional temporal dependencies in motion response time series data of offshore floating structure platform [19,20], and Attention Mechanisms adaptively prioritize weights of features that affect platform motion at critical time steps [21,22]. Taken together, this study therefore proposes a CEEMDAN-filtered and CNN-BiLSTM-Attention hybrid network, employing a technical framework of signal decomposition feature screening multimodal fusion to overcome the limitations of single-model approaches, thereby providing a novel solution for high-precision prediction of six-degree-of-freedom motions in offshore floating structures. Simultaneously, it is incremental relative to prior work and specifically targeted to the six-DOF prediction, closed-loop validation, and EWMA-based robustness assessment required for the “GuoHaiShi 1” platform as a case study. And the reinforced fusion strategy, coupled with a physics-informed validation framework, offers incremental advancements in model interpretability and demonstrate better transferability, specifically adapted to the non-stationary, multi-scale dynamics characteristics of floating platforms.

Methodology

CEEMDAN

To address the non-stationary characteristics of raw motion signals, an improved Complete Ensemble Empirical Mode Decomposition with Adaptive Noise (CEEMDAN) is employed. Traditional Empirical Mode Decomposition (EMD) may induce mode mixing when processing motion signals of floating platforms, causing fluctuations of distinct temporal scales to couple within the same Intrinsic Mode Function (IMF) component [23]. CEEMDAN mitigates mode mixing and enhances the physical interpretability of decomposition by introducing an adaptive noise injection strategy and complete ensemble averaging [24]. Its core advancements are manifested in three aspects:

Adaptive Noise Injection: Gaussian white noise with adaptively adjusted intensity, rather than fixed-magnitude noise, is added at each decomposition stage.

Progressive Residual Decomposition: A stepwise decomposition mode for residual signals ensures that noise energy aligns with the characteristic scales of the signal.

Ensemble Averaging Optimization: Residual noise interference is eliminated through multiple ensemble averaging iterations.

Let the original motion signal be denoted as x(t). The CEEMDAN decomposition procedure is as follows:

  1. 1). Noise Injection and Initial Decomposition

Generate N sets of auxiliary signals containing adaptive noise:

(1)

where is the white noise satisfying the N (0,1) distribution, and ε0 = 0.2σxx is the standard deviation of the raw signal). EMD decomposition of each to obtain the ensemble of the first order IMF components, and calculate the ensemble average:

(2)
  1. 2). Residual progressive decomposition

Calculate the first-order residual signal:

(3)

EMD decomposition of (Ej(·) for order j IMF extraction) to obtain the ensemble average of the second order IMF:

(4)

And so on, the k-order residue is:

(5)

: The (k-1)th order residual signal, i.e., the residual from the previous decomposition step. : The ensemble average of the intrinsic mode function (IMF). It is the k-th order IMF component obtained by averaging after multiple noise injections and EMD decompositions, representing the signal’s oscillation mode at this scale.

Repeat this process until the residual rk (t) is a monotonic function and cannot continue the decomposition.

  1. 3). Adaptive adjustment of the noise coefficient

Set the noise energy attenuation coefficient:

(6)

: The kth-order noise energy attenuation coefficient, used to control the intensity of injected noise. As the decomposition order k increases, the exponent attenuates to prevent low-frequency modes from being contaminated by high-frequency noise. : Initial noise coefficient, representing the baseline intensity for noise injection. : Attenuation factor, a constant set to 0.5 here. This parameter controls the decay rate of the noise coefficient as the order increases; values less than 1 ensure exponential attenuation characteristics.

As the order of decomposition increases, the noise intensity decayed exponentially, avoiding low-frequency modes contaminated by high-frequency noise.

  1. 4). Key parameter settings

Number of noise additions N: According to the Heisenberg uncertainty principle, take N = 200 times to ensure statistical significance. The ensemble size, common CEEMDAN practice (N range 100–500), determines the statistical stability of the reconstructed IMFs. A larger N reduces stochastic variability of the noise-assisted decomposition but increases computational cost.

Initial noise coefficient ε0: dynamically adjustment by signal-to-noise ratio (SNR), ε0∈[0.1σx, 0.3σx]. The interval reflects an empirical SNR-driven selection range: smaller values risk residual mode mixing, while larger values may inject excessive perturbation.

Stop criterion: terminate the decomposition when the number of residual signal extreme points is 2, which effectively halts further mode splitting and prevents over-decomposition of low-frequency content.

Pearson correlation coefficient

The linear correlation of each IMF component with the original motion sequence was quantified using Pearson’s correlation coefficient r:

(7)

The threshold r was set to remove the low correlation noise component and retain the key IMF component that reflects the motion frequency characteristics of the platform. The larger the correlation coefficient between the two, the stronger the correlation between the two variables; the weaker the correlation, the IMF correlation with the original data.

Time sliding window

In time-series prediction tasks, capturing temporal dependencies within data is critical [25]. The sliding window model, a widely adopted data preprocessing approach, constructs fixed-length temporal segments to encapsulate both short-term and long-term trends in historical data, thereby providing structured input for deep learning models.

The sliding window method partitions time-series data into consecutive, overlapping subsequences using a fixed-size window. Let the time-series dataset be denoted as , where N represents the total number of timesteps. Given a window size W, each input sample Xt can be expressed as:

(8)

The corresponding target output Yt is the predicted value at timestep t + W:

(9)

By incrementally shifting the window, each timestep generates a new training sample, enabling serialized processing of sequential data. The selection of window size W significantly impacts prediction performance: smaller W values may fail to capture long-term dependencies, while excessively large W risks introducing redundant information, escalating computational complexity, and inducing overfitting.

Furthermore, as physical parameters may exhibit substantial differences in numerical ranges, directly inputting raw data into deep learning models could destabilize gradient updates. To mitigate scale-related biases and accelerate model convergence, standard deviation normalization is applied prior to sliding window processing:

(10)

where μ and σ denote the mean and standard deviation of the entire dataset, respectively.

Convolutional neural network

The Convolutional Neural Network (CNN) is a deep learning architecture specifically designed for processing grid-structured data. By leveraging local connectivity and weight-sharing mechanisms, CNNs significantly reduce computational overhead while enhancing spatial hierarchical feature extraction capabilities [26]. These networks automatically learn multi-level features from raw data, compress information via pooling and convolutional operations, suppress irrelevant noise, and improve model robustness [27].

Floating structure platform motion signals exhibit pronounced localized fluctuations (e.g., transient responses induced by wave impacts), which traditional fully connected neural networks may struggle to capture effectively. The one-dimensional CNN (1D-CNN) addresses this limitation by exploiting local receptive fields and weight-sharing principles to autonomously extract spatially correlated features from multi-sensor time-series data. Its advantages include:

  1. 1). Localized Feature Extraction: Convolutional kernels slide along the temporal axis to identify abrupt signal variations within short time intervals (e.g., 5–10 seconds).
  2. 2). Shift invariance: Max-pooling operations enhance robustness against phase shifts in signal features.
  3. 3). Dimensionality Reduction: Hierarchical abstraction minimizes the temporal modeling complexity for subsequent BiLSTM modules.

Let the input data represent the normalized multi-channel time-series signals reconstructed via CEEMDAN and segmented using sliding windows, where T denotes timesteps and C corresponds to sensor channels. The CNN architecture comprises multiple hierarchical layers, each dedicated to distinct feature extraction and learning tasks, structured as follows:

Input Layer: Receives preprocessed signal data.

Convolutional Layer: Extracts localized patterns through filter kernels (W) scanning the input. The convolution operation is defined as:

(11)

where X is the input feature map, W the kernel weights, b the bias term, and * denotes the convolution operator.

Activation Function: Employs Rectified Linear Units (ReLUs) to introduce nonlinearity, mitigate gradient vanishing, and enhance representational capacity [28].

(12)

Pooling Layer: Reduces feature map dimensions via max-pooling or average-pooling operations, improving computational efficiency and alleviating overfitting [29].

Fully Connected Layer (FC): Maps abstracted features to task-specific outputs (e.g., regression or classification).

Output Layer: Utilizes activation functions (e.g., Sigmoid or Softmax) to generate probabilistic predictions or numerical forecasts.

Bidirectional long short-term memory

The Bidirectional Long Short-Term Memory network (BiLSTM) is an enhanced recurrent neural network (RNN) architecture designed to address the gradient vanishing problem inherent in traditional RNNs during long-sequence modeling [30]. Its core feature lies in a bidirectional information propagation mechanism that simultaneously captures forward and backward temporal dynamics, significantly enhancing the model’s ability to model temporal dependencies [31]. BiLSTM has demonstrated exceptional performance in natural language processing and speech recognition, and recent applications in time-series prediction, anomaly detection, and fault diagnosis highlight its unique advantages. Future research will explore its deployment in platform anchor-system monitoring:

The foundational unit of BiLSTM is the LSTM cell, which regulates information flow via gating mechanisms (input gate, forget gate, output gate):

Forget Gate: Determines which information from the previous cell state should be retained or discarded, generating weights between 0 and 1 via a Sigmoid function.

Input Gate: Filters relevant information from current inputs to update the cell state, combining a Sigmoid layer (for weight selection) and a Tanh layer (for generating new candidate values).

Output Gate: Controls the transfer of cell state information to hidden layer outputs, ensuring task-relevant feature propagation.

This architecture enables LSTMs to effectively capture long-term dependencies, circumventing the memory degradation caused by gradient attenuation in traditional RNNs.

Floating platform motions exhibit strong historical dependencies and future underlying state correlations (e.g., motion hysteresis due to periodic wave-induced effects). While unidirectional LSTMs only model past-to-present temporal relationships, BiLSTM integrates forward (past→future) and backward (future→past) propagation paths to achieve global temporal modeling of motion characteristics under complex marine environments. Key advantages include:

  1. 1). Bidirectional Causality: The forward LSTM captures cumulative wave excitation effects, while the backward LSTM deciphers inertial response decay patterns.
  2. 2). Long-Term Dependency Preservation: Cell state mechanisms mitigate gradient vanishing issues inherent in conventional RNNs.
  3. 3). Environmental Coupling Modeling: Simultaneous learning of nonlinear temporal mappings from coupled wind-wave-current interactions.

In BiLSTM, bidirectional processing is achieved by stacking two independent LSTM layers:

Forward LSTM Layer: Processes input sequences chronologically (t1 → tn) to extract historical influences on the current state.

Backward LSTM Layer: Processes sequences in reverse (tn → t1)to infer latent correlations between future states and the present.

Let denote the high-order feature sequence output by the CNN module. The internal computations of the BiLSTM unit proceed as follows:

  1. 1). Forward LSTM Unit

At each timestep t, the hidden state and cell state are computed:

(13)

where σ denotes the sigmoid function, ⊙ represents element-wise multiplication, and constitutes the forward weight matrix.

  1. 2). Backward LSTM Unit

The feature sequence is processed in reverse temporal order through symmetric computations:

(14)

Yielding the backward hidden state sequence .

  1. 3). Bidirectional State Fusion

Global temporal characteristics are synthesized through concatenation-based fusion of bidirectional information:

(15)

where , respectively represent forward and backward hidden states, with || denoting concatenation along the feature dimension.

  1. 4). Optimization Strategies
  2. (1). Temporal-aware Dropout

Variational dropout regularization (rate = 0.2) is implemented between LSTM layers, maintaining consistent dropout masks across timesteps within each batch to balance model generalization and feature preservation.

  1. (2). Gradient Clipping

A gradient norm threshold θ = 1 is established to prevent explosion from oceanic data noise:

(16)

In time-series applications, the bidirectional nature of BiLSTM enables effective multi-scale feature extraction, analyzing both localized fluctuations (e.g., short-term trends) and global patterns (e.g., periodic regularities) within sequential data. Furthermore, it enhances contextual awareness by leveraging bidirectional dependencies—for instance, the backward layer can uncover dependencies between the current state and future events in sequences with delayed effects. Additionally, BiLSTM improves robustness in anomaly detection through bidirectional consistency checks (e.g., reconstruction error analysis), identifying anomalies overlooked by unidirectional models. These capabilities support critical applications such as fault diagnosis, operational state monitoring, and early warning systems. Future research will focus on exploring these applications in greater depth.

Attention mechanism

In floating platform motion prediction, the contributions of features at different time nodes to forecasting targets vary significantly (e.g., transient wave impact peaks, resonant phase shifts). Traditional time-series models treat all historical features equally, risking the dilution of critical information by noise. The Attention Mechanism addresses this by dynamically assigning feature weights, enabling the model to focus on temporally relevant segments [32]. Its core advantages include:

  1. (1). Key Event Prioritization: Amplifies representation weights for pivotal events such as transient wave impacts and motion resonance points.
  2. (2). Long-Range Dependency Modeling: Overcomes distance limitations in BiLSTM hidden state propagation by directly encoding cross-cycle temporal relationships.
  3. (3). Multi-Scale Feature Fusion: Leverages multi-head mechanisms to concurrently capture physical patterns across varying temporal resolutions (e.g., second-level wave cycles, minute-level slow-drift motions). This study employs 4 attention heads.

Let the BiLSTM output be denoted as HBiLSTM=[h1,h2,...,hT′]∈RT′×2H, the multi-head attention computation proceeds as follows:

  1. 1). Linear Projection for Query/Key/Value:

Map the input sequence to M subspaces (M = 4, number of attention heads):

(17)

where

  1. 2). Scaled Dot-Product Attention:

The output of the m-th attention head:

(18)

The scaling factor mitigates gradient vanishing caused by large dot-product magnitudes in high-dimensional spaces.

  1. 3). Multi-Head Concatenation and Linear Transformation:

Concatenate and project outputs from all heads:

(19)

where, ensures dimensional consistency with the input.

  1. 4). Residual Connections and Layer Normalization:

Preserve original temporal information while enhancing feature representation:

Algorithm structure

Structure design.

The raw time-series data of floating platform motions undergo data reconstruction via CEEMDAN filtering, standard normalization, and sliding window processing, followed by the construction of a CNN-BiLSTM-Attention prediction model. This model adopts a cascaded decomposition reconstruction prediction architecture, incorporating the following innovations. The overall structure of the model is shown in Fig 1.

  1. 1). Data Preprocessing Level:

CEEMDAN-Based Signal Stabilization: Non-stationary signals are decomposed into quasi-stationary Intrinsic Mode Function (IMF) components via CEEMDAN, reducing model learning complexity.

Physics-Informed Noise Filtering: IMFs are selectively retained based on Pearson correlation coefficients to achieve physical mechanism-driven noise filtering and mitigate overfitting risks inherent in purely data-driven approaches.

  1. 2). Spatiotemporal Feature Coordination Framework:

CNN Layer: Extracts spatial correlations from multi-channel sensor data.

BiLSTM Layer: Captures temporal dependencies across short- and long-term sequences.

Attention Mechanism: Dynamically assigns weights to features for adaptive focusing on critical events.

Inter-Module Fusion: Fully connected (FC) layers are inserted between neural network to enhance feature integration and data exchange, thereby reinforcing the model’s the network fusion and representational capacity.

Parameter setting.

The target prediction parameters of the model are set as the roll and heave degrees of freedom of the floating platform’s response motion. Additionally, multiple feature variables influencing platform motion are incorporated, including “Current Force”, “Wind Force”, “Diffraction Force”, “Linear Damping Force”, “Drift Force”, “Mooring Force”, “RadiationForce” and so on.

During the data preprocessing stage, the noise amplitude for CEEMDAN is set to 0.1σx, with added white noise following a normal distribution N(0,1) for empirical mode decomposition. When selecting intrinsic mode function (IMF) components, the screening criterion is set as the Pearson correlation coefficient R2 > 0.2 between the IMF component and the original data. After reconstructing the IMF data, standardization and sliding window segmentation are applied, with a window size of 300.

For the CNN-BiLSTM-Attention model configuration, CNN employs three types of one-dimensional convolutional kernels, each with 64 filters. The first convolutional layer uses a kernel size of kernal_size1 = 3, the second layer kernal_size1 = 5 and the third layer kernal_size layer kernal_size1 = 7. MaxPooling1D layers are applied between convolutional layers, with pooling window sizes set to kernel_size = 1 and kernel_size = 2, respectively. After the third convolutional layer, batch normalization (BatchNorm) with a size of 3 × 64 is implemented, followed by a dropout regularization layer (dropout = 0.2) to prevent neuron overfitting.

In the BiLSTM module, both the forward and backward LSTM layers are configured with 128 hidden units, consisting of two LSTM layers with a dropout of 0.2. The attention mechanism employs four attention heads with dropout set to 0.2. A fully connected layer is added between each module and the model is trained for 25 epochs with a batch size of 64.

To enhance the performance of the original CNN_BiLSTM_Attention model, several optimizations were implemented, as summarized in the following and shown in Table 1.

thumbnail
Table 1. Parameter setting and robustness optimization strategy of traning.

https://doi.org/10.1371/journal.pone.0342081.t001

Evaluation index.

The model evaluates the prediction performance using MAE (mean absolute error), RMSE (root mean square error), MSE (mean squared error), R² Score (goodness of fit) and the comparison between the predicted value and the true value. It also outputs the time history curve of the predicted value and the true value, the regression prediction curve, and the Huber Loss curve during each epochs of training.

(20)(21)(22)

Additionally, a feature importance analysis based on gradient propagation is introduced to evaluate the impact of each feature on the prediction results by calculating the sensitivity of neural network outputs to input features. In neural networks, the gradient (∂output/∂input) reflects the extent to which small variations in input features affect the output. A larger absolute gradient value indicates a more significant influence of that feature on the output, signifying higher feature importance. Using the backpropagation algorithm, gradients are propagated layer by layer from the output layer back to the input layer.

In this temporal model (CNN-BiLSTM-Attention mechanism model), the gradient is backpropagated through multiple components, including the attention weights, the temporal dependencies in BiLSTM, and the local convolutions in CNN. This process naturally integrates the feature processing logic across different layers of the model. For each input sample, the absolute values (or squared values) of the gradients for all features at every time step are computed. These values are then aggregated—such as by taking the mean or maximum—to obtain a feature importance score. Finally, the scores are averaged across all samples to derive a global importance ranking, which quantifies the relative contribution of each feature to the prediction target parameters.

Case study

The “GuoHaiShi 1” developed by the National Ocean Technology Center, is a multifunctional comprehensive marine testing platform designed specifically for validating marine scientific research and technological innovations. Equipped with state-of-the-art marine testing instruments and comprehensive living support facilities, the platform is capable of supporting diverse maritime testing applications, including marine environmental monitoring, ocean energy development, and marine equipment evaluation. Its primary objective is to provide researchers and engineers with stable, secure, and efficient experimental conditions alongside a habitable working environment.

The ‘GuoHaiShi 1’ offshore test platform is a steel non-powered floating mooring structure with a gross tonnage of 432 tonnes, adopting a double-floating hull type, with a hull length of 30m, a width of 21m, a depth of 4.5m, a design draught of 2.2m, and the actual deployment draught of 2.07m, arranged in the sea area with a maximum water depth of 70m. The actual structure is shown in Fig 2.

The National Marine Technology Center of China conducted year-round wave observations in the waters where the platform is deployed, utilizing wave sensors mounted on multi-parameter integrated buoys. Additionally, the flow velocities of both major and minor tidal currents are measured using Acoustic Doppler Current Profilers [33]. Based on the field observation data and historical records, the wind, wave, and current parameters for different return periods in the area are estimated. The estimation results are presented in Table 2 [34].

thumbnail
Table 2. The calculation results of different recurrence periods of wind and waves.

https://doi.org/10.1371/journal.pone.0342081.t002

Numerical simulation analysis

Numerical simulation setting.

As shown in the Fig 3, this is a 1:1 model of the “GuoHaiShi 1” offshore test platform used for finite element calculation, featured a twin-hull structureor. In the model, four 165m anchor chains are used to connect the anchor points and the platform connection points to simulate the actual anchor chains. The main size parameters and hydrostatic parameters of the platform are shown in Table 3

thumbnail
Table 3. Main dimensional parameters and hydrostatic parameters of the platform.

https://doi.org/10.1371/journal.pone.0342081.t003

thumbnail
Fig 3. Finite element model of the “Guohaishi 1” offshore test platform.

https://doi.org/10.1371/journal.pone.0342081.g003

The one-in-a-century sea state in Table 2 is set as the model input condition, in which the Jonswap short-peaked irregular wave spectrum is used as the incident wave load. At the same time, the simulation step is set to be 0.01s in the model, and the data output is 0.05s, and the simulation duration is 1000s. The specific parameter settings in the case study can be found in the article [35].

Data preprocessing.

The raw roll data of the floating platform generated by the hydrodynamic analysis was filtered using the CEEMDAN module, decomposed to obtain 11 IMFs components (Fig 4 and Fig 5) and the Pearson correlation coefficients of each IMF component with the raw data were calculated as shown in the Table 4 and Table 5 below, and the same operation was performed for the rest of the features.

thumbnail
Table 4. IMFs components of ceemdan decomposition of roll.

https://doi.org/10.1371/journal.pone.0342081.t004

thumbnail
Table 5. IMFs components of ceemdan decomposition of heave.

https://doi.org/10.1371/journal.pone.0342081.t005

Based on the results of CEEMDAN, the high-frequency IMFs (IMF1-IMF6) correspond to external excitation responses such as ocean waves, while the low-frequency IMFs (IMF7-IMF11) reflect the platform’s inherent oscillatory modes. The specific physical interpretations of these components require further investigation and are not analyzed in depth within this study. Due to hardware limitations preventing add each IMF component as each feature to the CNN-BiLSTM-Attention model for calculation, those with Pearson correlation coefficients below 0.2 were excluded from analysis. The remaining components were then reconstructed to form refined input datasets for the deep learning model.

Result analysis.

For the roll degree of freedom response of the floating platform, four different algorithms were used to train the reconstructed data. The time-history curves and regression curves of the predicted and actual values in the test set, as well as the training loss of the training and validation sets, are shown in the corresponding Figs 68 for roll motion and Figs 911 for heave motion. It can be observed that after 25 training epochs, all four models stabilize and demonstrate excellent predictive performance.

thumbnail
Fig 6. Time history curves of roll motion in different models.

https://doi.org/10.1371/journal.pone.0342081.g006

thumbnail
Fig 7. Regression curves of roll motions in different models.

https://doi.org/10.1371/journal.pone.0342081.g007

thumbnail
Fig 8. The per-epoch training loss curves of the roll motion for different models.

https://doi.org/10.1371/journal.pone.0342081.g008

thumbnail
Fig 9. Time history curves of heave motion in different models.

https://doi.org/10.1371/journal.pone.0342081.g009

thumbnail
Fig 10. Regression curves of actual and predicted heave motions in different models.

https://doi.org/10.1371/journal.pone.0342081.g010

thumbnail
Fig 11. The per-epoch training loss curves of the roll motion.

https://doi.org/10.1371/journal.pone.0342081.g011

The Table 6, 7 below presents the evaluation metrics for the four models. Among them, the CEEMDAN-CNN-BiLSTM-Attention model achieves the highest R2 value and the lowest RMSE, MSE, and MAE. However, the performance improvement of the CNN-BiLSTM-Attention model after applying CEEMDAN filtering is not as significant as expected, with RMSE decreasing by only 9.69%. This may be due to the input data being simulated by hydrodynamic analysis, where the floating platform’s response under this sea state does not exhibit extreme values and is of high quality with few anomalies.

thumbnail
Table 6. Comparison of evaluation indicators of different models in roll motion.

https://doi.org/10.1371/journal.pone.0342081.t006

thumbnail
Table 7. Comparison of evaluation indicators of different models in heave motion.

https://doi.org/10.1371/journal.pone.0342081.t007

However, compared with the other two models—those lacking the CNN and Attention mechanisms — the CEEMDAN-CNN-BiLSTM-Attention model achieves a 27.70% and 10.32% reduction in RMSE, a 62.88% and 21.61% reduction in MSE, and a 32.72% and 14.58% reduction in MAE, respectively, which indicates a notable performance enhancement. The results suggest that the integration of CEEMDAN can improve model performance to some extent, as the signal decomposition technique effectively extracts the nonlinear and non-stationary features in roll motion. Furthermore, the combination of CNN layers and the attention mechanism further enhances the model’s ability to capture spatiotemporal features.

For the heave degree of freedom response simulation, all four models also exhibit outstanding performance. However, omitting the filtering step results in a better fit. Additionally, compared with the other two models, the three evaluation metrics show a certain degree of error reduction. This suggests that CEEMDAN may introduce noise in heave motion prediction, as heave motion inherently exhibits stronger linear characteristics, making direct modeling more effective. Moreover, the performance gain from the Attention mechanism is relatively significant in this case, but necessitating further feature importance analysis to clarify the weight distribution within the attention mechanism.

The feature importance scores obtained through gradient-based propagation analysis are shown in the Fig 12 below. It can be observed that CEEMDAN significantly alters the differentiation of several key features (e.g., Diffraction Force, Linear Damping Force, and Hydrostatic Pressure Force), indicating that this preprocessing method effectively enhances the model’s sensitivity to complex features, thereby improving its ability to capture nonlinear dynamic characteristics.

thumbnail
Fig 12. Feature importance score and the upper part is roll motion and the lower part is heave motion.

https://doi.org/10.1371/journal.pone.0342081.g012

The CNN model automatically extracts localized, meaningful features from time-series data, particularly those assigned high importance in gradient propagation (e.g., Diffraction Force, Radiation Force, and Linear Damping Force). Through multi-scale convolutional kernels, CNN captures both the short-term impact and long-term cumulative effects of wave diffraction. Additionally, weight sharing in convolutional kernels reduces excessive dependence on a single feature, preventing the model from falling into local optima. For instance, the absence of CNN leads to an exaggerated feature effect of Linear Damping Force and Hydrostatic, resulting in the CEEMDAN-BiLSTM-Attention model underestimating predictions relative to actual values, as seen in the regression plot (Fig 10).

The Attention mechanism improves overall model performance by dynamically weighting and reinforcing key features. For example, Diffraction Force and Linear Damping Force dominate the prediction process, and the Attention mechanism dynamically adjusts weight allocation for complex physical forces such as radiation and damping, leading to an approximate 10% reduction in RMSE. However, the effect of attention weight distribution differs between the two degrees of freedom.

Linear Damping Force and Hydrostatic Pressure Force consistently receive high importance scores across all models, indicating that the response motion of the floating platform is primarily governed by damping effects and hydrostatic restoring forces. This aligns with fluid dynamics principles, where damping force and buoyancy play crucial roles in oscillation attenuation, confirming that the model correctly and effectively captures the physical mechanisms underlying the relationship between features and prediction targets. The significance of Diffraction Force and Radiation Force is also evident, highlighting the critical role of wave-structure interactions—a classical problem in potential flow theory—in motion prediction.

And thus, to further clarify the dominant physical drivers revealed by the feature-importance analysis, a quantitative comparison was performed for the three major hydrodynamic contributions, Diffraction Force, Linear Damping Force, and Hydrostatic Restoring Force, which consistently exhibit the highest importance scores across all models. For the heave DOF, these three components account for 60–82% of the total normalized importance, indicating that vertical motion is primarily governed by the balance between wave excitation, fluid-induced energy dissipation, and buoyancy-driven restoring effects. In particular, Linear Damping contributes approximately 17–36%, highlighting its central role in suppressing excessive heave oscillations, while Hydrostatic Restoring typically contributes 16–18%, consistent with the platform’s stiff vertical restoring characteristics. Diffraction Force contributes 26–30%, reflecting the strong influence of incident wave scattering on the vertical radiation–diffraction response. For the roll DOF, although additional nonlinear external loads such as wind and current drag, Diffraction Force and Linear Damping Force remain dominant with combined average contributions of 40% across models, while Hydrostatic Restoring (metacentric restoring moment) accounts for 21%. This distribution aligns with classical seakeeping theory: roll motion is weakly hydrostatically restored but highly sensitive to damping and wave-induced moments [36]. The models therefore correctly attribute higher importance to damping-related features, which physically govern roll decay and stability in irregular seas.

The physical coupling mechanisms are also consistent with hydrodynamic theory. In heave, the restoring force acts directly in the vertical direction, while diffraction governs the energy input from incident waves; the strong importance of these features indicates that the deep-learning models have effectively learned the radiation-diffraction coupling that characterizes vertical motion. In roll, the interaction between wave-induced moments (diffraction) and cross-coupled radiation damping generates asymmetric energy transfer during oscillation. The elevated importance scores for damping-related features demonstrate that the models accurately capture this coupling, especially under CEEMDAN preprocessing, which enhances sensitivity to low-frequency and nonlinear components. It can clearly be seen that these quantitative findings confirm that the learned feature relationships are physically interpretable and consistent with established hydrodynamic principles governing floating platforms.

In summary, the effectiveness of CEEMDAN exhibits a degree-of-freedom dependency, showing superior performance in complex motion patterns such as roll. CNN not only extracts localized features and reduces noise but also enhances high-importance features and establishes a dynamic balance mechanism among them, ultimately improving overall model performance. The Attention mechanism enhances the model’s sensitivity to dynamic features, where adaptive weight allocation further optimizes model performance. These findings underscore that feature importance analysis elucidates the physical interpretability of these prediction models, forming a closed-loop validation between data-driven and mechanism or physics-based models paradigms.

Analysis of measured proportional model

Introduction of pool model.

In order to verify the actual usability of the algorithm, a scale model pool test was carried out to obtain measured data for training. Based on the prototype dimensions of the floating offshore test platform and the deployment sea conditions, in conjunction with the experimental capabilities of the wave basin at the Marine Dynamic Environment Laboratory of the National Ocean Technology Center, the physical model test was finalized with a scale ratio of 1:30. The scaling coefficient is 30, resulting in a platform model with a length of 1.00 m and a width of 0.70 m. The specific scaling scheme is detailed in the accompanying Table 8, and the fabricated test platform model is shown in the corresponding Fig 13.

thumbnail
Fig 13. Scaled model of the floating offshore test platform.

https://doi.org/10.1371/journal.pone.0342081.g013

The scaled model test system consists of the following components:

Experimental Environment Simulation Devices: Including a wave generator, wind generator, and an equivalent current force simulation device.

Platform Model: Comprising the scaled platform structure and its mooring system.

Measurement and Data Acquisition System: This includes a wave height sensor (Fig 14), anemometers (Fig 15), six-degree-of-freedom (6-DOF) acquisition system (Fig 16) implemented using the FZMotion Motion Capture System, and tension sensors.

thumbnail
Fig 16. Non-contact six-degree-of-freedom acquisition system.

https://doi.org/10.1371/journal.pone.0342081.g016

These measurement sensors are integrated into a data acquisition system to ensure real-time and synchronized data collection.

Pool test settings.

The test platform model was deployed at a location 85 m from the wave generator and 25 m from the wind generator. A wave height sensor was positioned 5 m in front of the model. To ensure accurate wind speed measurements, a calibrated thermal anemometer was installed at the test platform location and removed after completing wind speed calibration. An underwater tension sensor was installed at the mooring chain connection point, while the non-contact 6-DOF motion measurement system was divided into two parts: one mounted on the moving platform approximately 5m diagonally above the wave-facing side, and the other mounted onshore using a tripod. To facilitate attitude measurements, marker points were installed on the platform model before testing. Additionally, an equivalent current force simulation device, consisting of pulleys, fishing lines, and weights, was installed. The tension application point was maintained at the same horizontal level as the platform model’s center of gravity. The experimental layout is illustrated in Fig 17.

The wind-wave-current environmental parameters for the water pool test were determined based on similarity criteria, covering conditions from working sea states to a once-in-a-century event sea conditions. In all test scenarios, wind, waves, and currents acted in the same direction on the test platform. The key experimental conditions were established as follows Table 9.

During the experiment, the wind field was simulated first. After inputting the corresponding wind parameters into the wind generation system, the wind was initiated. Once the wind speed and direction stabilized, weights were applied to the equivalent simulation device to simulate current forces. After the test model stabilized under the combined action of wind and current, the wave generator was activated to simulate irregular waves. The irregular waves followed the JONSWAP spectrum, which was validated by comparison with measured spectra. Each test condition was repeated three times, and the resulting datasets were used to segment the data into testing subsets. The actual test environment and data acquisition system are shown in Figs 18 and 19.

thumbnail
Fig 18. Field test arrangement of the floating platform scaling model.

https://doi.org/10.1371/journal.pone.0342081.g018

thumbnail
Fig 19. On-site test diagram of data acquisition computing system.

https://doi.org/10.1371/journal.pone.0342081.g019

Test result analysis.

To validate the model’s applicability and further investigate the coupling effects among different degrees of freedom of the floating platform, the scaled model pool experiment incorporated the platform’s remaining five degrees of freedom (excluding roll) and the forces on two mooring lines as input features, while roll remained the prediction target. The response dataset of the floating platform under the 100-year return period condition was used as the training set for model training. Additionally, this dataset was proportionally divided into 27 test subsets, while datasets from the other 13 test conditions were also segmented into different test subsets.

To enhance the model’s generalizability, adjustments were made to activation functions and regularization dropoot coefficients. A robustness assessment framework tailored to diverse design sea conditions was proposed, centered on a dual-stage Exponentially Weighted Moving Average (EWMA) control line design. Defined as a statistical monitoring technique that assigns exponentially decreasing weights to past observations, EWMA is chosen for this study because it is sensitive to small shifts and trends in residuals while giving greater weight to recent data points — this property is desirable for monitoring model transferability when environmental conditions drift slowly over time. EWMA also reduces volatility relative to raw residual tracking, making thresholding decisions less noisy.

First, residual features were constructed, using RMSE as the input source, which could be either standardized or non-standardized. The dual-stage EWMA control limit was established to characterize the maximum tolerable RMSE.

The control limit consists of two progressive stages. The first stage tracks residual trends by computing time-varying statistics of the residual sequence using an exponentially weighted moving average:

(23)

where is the decay factor that controls the weighting of recent data.

The second stage constructs a threshold control limit based on statistical characteristics within a sliding window:

  1. 1). Windowed Statistical Calculation:

A sliding window of length W is maintained to compute the mean and standard deviation within the window. In this study, the window length was set to 24, corresponding to the 24 test subsets mentioned earlier.

(24)
  1. 2). Dynamic Threshold Calibration:

Combining the central limit theorem and control theory, the threshold Ut is customised as:

(25)

where X is a control coefficient (typically set between 3 and 5), and the correction factor compensates for the weighted bias introduced by the EWMA method. If the sample size is large enough, the threshold can also be calculated using the IQF rule or the 3σ rule

  1. 3). Applicability Decision Rule:

The model is considered applicable if the residual values satisfy the condition:

(26)

where and represent the RMSE of the sample and the maximum tolerable RMSE, respectively.

As shown in the Fig 20, the predicted and actual values of the roll response of the floating platform in the pool test exhibit strong agreement, with R2 = 0.9896 and original testdata RMSE = 0.2546. From the trends of the Train Loss and Validation Loss curves, a more pronounced separation between the two indicates a higher likelihood of overfitting. The newly implemented adjustment strategy effectively reduces the model’s overfitting tendency while maintaining excellent fitting performance on new datasets.

thumbnail
Fig 20. Regression curve and the per-epoch training loss curves.

https://doi.org/10.1371/journal.pone.0342081.g020

Fig 21 presents the evaluation results based on the dual-stage EWMA control limit, where the test set comprises floating platform motion response datasets from 13 different test conditions. The calculated maximum tolerable RMSE is 3.20, and test sets with RMSE values below this threshold are considered suitable for this model. Similarly using the IQF rule and the 3σ rule to calculate the thresholds obtained were 2.569 and 2.665, respectively. Statistical analysis indicates that approximately 117 test sets are deemed applicable, while 24 test sets exhibit poor fitting performance, yielding an applicability rate of 82.98%. This demonstrates the outstanding capability of this model, considering that it was trained using only a single test condition. If a larger and more representative dataset were available, the estimation accuracy of the simulation could be further improved.

thumbnail
Fig 21. Two-stage EWMA control line and RMSE distribution of the test set.

https://doi.org/10.1371/journal.pone.0342081.g021

The Fig 22 illustrates the feature importance scores for various samples. The first 28 samples originate from the motion response dataset of a floating platform under the same sea state, which have been used for the training of the model, where it can be observed that the distribution of feature scores remains largely consistent under identical conditions. In contrast, the cross-test-condition dataset exhibits a distinct feature score distribution. Regarding individual features, sway and heave generally receive higher scores and are respectively defined as Position X and Position Z, while roll corresponds to Rotation Y. The direction of wind, waves, and currents is from west to east, i.e., along the x-axis. It is evident that heave directly influences the vertical motion of the platform, potentially causing relative displacement between the buoyancy center and the center of gravity, thereby inducing roll. The strong coupling between heave and roll observed in the Fig aligns with the heave-roll coupling effect in hydrodynamics. Additionally, under the action of wind, waves, and currents, the platform experiences lateral displacement in the X direction (sway), which, due to the constraints imposed by the mooring chains, can indirectly induce rolling motion around the Y-axis. This phenomenon is consistent with the mechanism in which wave loads transmit sway motion to roll. However, considering the RMSE, certain samples exhibit relatively poor fitting performance, likely due to the model overemphasizing the weight of heave during the fitting process, necessitating further adjustments in subsequent analyses.

thumbnail
Fig 22. Importance Score of each feature and each sample.

https://doi.org/10.1371/journal.pone.0342081.g022

Figs 2325 illustrate the fitting performance of the test set with R2 = 0.5348 in the model. Fig 23 presents the time-series comparison between the actual and predicted values. Fig 24 (Left) utilizes cumulative summation to analyze the overall data trend, while Fig 24 (Right) displays the frequency distribution histogram. Fig 25 shows the anomaly distribution of the data.

thumbnail
Fig 23. Time series curve of true value and predicted value.

https://doi.org/10.1371/journal.pone.0342081.g023

thumbnail
Fig 24. Cumulative sum graph (Left) and Frequency distribution histogram (Right).

https://doi.org/10.1371/journal.pone.0342081.g024

From the data trend and frequency distribution, it can be observed that the overall trend and numerical distribution of the predicted values align well with the actual values, only the numerical values are slightly smaller. Regarding the anomaly distribution, the deviation patterns of the predicted and actual values are largely consistent, with a mean difference of 0.98, a maximum error of 8.64, and a standard deviation of 2.61. These results indicate that prediction errors are relatively uniform across all data points, meaning the model exhibits consistent error distribution without large fluctuations. This further supports the model’s robust applicability and strong generalization ability. Moreover, this modeling approach can be transferred to other application domains, such as fault diagnosis and operational health monitoring of equipment, demonstrating its potential for broader interdisciplinary applications.

Discussion

Analysis of degree-of-freedom dependent performance

The study further reveals that the model’s performance is sensitive to the specific degree of freedom (DOF) being predicted, particularly when comparing Heave and Roll motions. The proposed hybrid model demonstrates superior performance in predicting Roll motion, which is characterized by strong nonlinearity, viscous damping effects, and resonance sensitivity. The Attention mechanism effectively captures these dynamic shifts by weighting critical time steps. However, for Heave motion, which is primarily governed by linear hydrostatic restoring forces and exhibits stronger periodicity, the inclusion of complex signal decomposition and attention mechanisms resulted in diminishing returns. As evidenced by the results in the analysis, the CEEMDAN process occasionally introduced boundary effects or mode mixing that slightly degraded the prediction of linear Heave responses. This suggests that while the proposed architecture excels in handling highly coupled, nonlinear interactions, simpler modeling approaches may suffice for DOFs dominated by linear physical mechanisms.

Limitations of the proposed methodology

Despite the demonstrated robustness of the CEEMDAN-CNN-BiLSTM-Attention framework, certain limitations must be acknowledged for practical engineering applications:

  1. Computational Complexity and Real-Time Constraints: The iterative nature of the CEEMDAN algorithm significantly increases the computational cost during the data preprocessing stage. Unlike the end-to-end inference of the neural network, the decomposition process requires substantial calculation time, which may introduce latency in scenarios requiring seconds-level real-time control or immediate early warning systems.
  2. Absence of Extensive Benchmarking: The current study has not yet conducted a comprehensive comparison with other standard benchmark methods, such as Transformer-based architectures (e.g., Temporal Fusion Transformers, Informer, or Attention-LSTM), hybrid Physics-Informed Neural Networks (PINNs), or alternative recurrent baselines like CNN-GRU and BiGRU. Future work will address this by incorporating these advanced models into the comparative analysis. Furthermore, small-scale ablation studies and sensitivity tests regarding the hyperparameters of these baselines will be conducted to ensure a rigorous and robust validation.

Conclusion

In response to the nonlinear and non-stationary motion characteristics of offshore floating platform responses, this study proposes a time-series motion prediction model based on CNN-BiLSTM-Attention, with CEEMDAN filtering and sliding window preprocessing as a preprocessing steps. The model employs CNN to extract spatial-local features from input data, utilizes BiLSTM to capture bidirectional temporal dependencies (resolving motion hysteresis effects induced by wave excitation), and leverages an Attention mechanism to adaptively prioritize feature weights at critical timesteps. Additionally, a two-stage EWMA control line is introduced as a robustness evaluation framework to assess the model’s applicability under different operational conditions.

While the individual components of this architecture (CEEMDAN, CNN, BiLSTM) are well-established in the literature, the distinct contribution of this study lies in the strategic, application-oriented fusion of these methods and the rigorous validation framework developed specifically for the complex hydrodynamics of offshore structures. This workflow effectively bridges the gap between “black-box” data-driven inference and classical hydrodynamic theory, offering the following three specific innovations:

  1. Application-oriented reinforced fusion workflow. The study integrates CEEMDAN-driven modal decomposition with a cascaded CNN–BiLSTM–Attention architecture using a lightweight fusion layer. The contribution lies in demonstrating how this coordinated workflow enhances prediction stability and physical interpretability for floating-platform motion.
  2. Physics-aligned interpretability via gradient-based feature attribution. Gradient-propagation importance analysis is applied and quantitatively summarized to link learned representations with hydrodynamic processes such as damping, hydrostatic restoring, and diffraction/radiation effects. This establishes a structured connection between data-driven inference and classical hydrodynamics.
  3. Dual-stage EWMA applicability and robustness framework. A two-stage EWMA control-line mechanism is introduced for evaluating model residual trends under varying environmental conditions. This framework provides a practical, transparent criterion for determining cross-condition applicability.

Finally, analyses of both the hydrodynamic simulation results and physical pool test data indicate that all four models demonstrate excellent predictive performance for the target variable, with the CEEMDAN-CNN-BiLSTM-Attention model exhibiting superior performance.The constructed model accurately and effectively reflects the physical mechanisms underlying the relationship between features and the prediction target rather than relying solely on data-driven approaches.Ultimately, validation with platform motion datasets of various conditions confirms the model’s applicability and generalization capability. This model therefore holds superior applicability, feasibility and practical value for precise prediction of the floating platform’s motion attitude, which is critical for ensuring the safety and reliability of offshore experiments.

Acknowledgments

The National Ocean Technology Center(NOTC), The Key Laboratory of Ocean Observation Technology of the Ministry of Natural Resources are credited for their support in the pool test. And Institute of Oceanographic Instrumentation, Qilu University of Technology (Shandong Academy of Sciences) and School of Mechanical Engineering, Shandong University are credited for their software support and guidance.

References

  1. 1. Chen M, Yuan G, Li B, Li CB, Ouyang M, Li L, et al. Dynamic analysis of lift-off operation of offshore wind turbine jacket foundation from the transportation barge. Ocean Engineering. 2024;301:117443.
  2. 2. Ramezani M, Choe D-E, Heydarpour K, Koo B. Uncertainty models for the structural design of floating offshore wind turbines: A review. Renewable and Sustainable Energy Reviews. 2023;185:113610.
  3. 3. Chen Z, Liu X, Ji X, Gui H. Real-Time Prediction of Multi-Degree-of-Freedom Ship Motion and Resting Periods Using LSTM Networks. JMSE. 2024;12(9):1591.
  4. 4. Yuan HJK, Zhang W. Research progress of resolvent analysis in fluid mechanics. Chinese Journal of Theoretical and Applied Mechanics. 2024;56(10):2799–814.
  5. 5. Cui K, Yang J, Chang S, Tian Z, Xiao Y, Li G. Rapid prediction method for high-pressure capturing wing surface flow field based on proper orthogonal decomposition and surrogate model. Chinese Journal of Theoretical and Applied Mechanics. 2025;57(4):883–94.
  6. 6. Wen S-C, Yang C-H. Time series analysis and prediction of nonlinear systems with ensemble learning framework applied to deep learning neural networks. Information Sciences. 2021;572:167–81.
  7. 7. Gong J, Xu J, Xu L, Hong Z. Enhancing motion forecasting of ship sailing in irregular waves based on optimized LSTM model and principal component of wave-height. Front Mar Sci. 2025;12.
  8. 8. Zhang L, Feng X, Wang L, Gong B, Ai J. A hybrid ship-motion prediction model based on CNN–MRNN and IADPSO. Ocean Engineering. 2024;299:117428.
  9. 9. Fan G, Yu P, Wang Q, Dong Y. Short-term motion prediction of a semi-submersible by combining LSTM neural network and different signal decomposition methods. Ocean Engineering. 2023;267:113266.
  10. 10. Liu C, Xin Y, Zhang Y, Yang Y, Sun L, Chen C. Real-Time Mooring Tension Prediction for Semi-Submersible Aquaculture Platforms by an EEMD-TCN-SA Neural Network Approach. JMSE. 2025;13(2):327.
  11. 11. Chen Y, Yuan L, Zan Y, Li Z. Forecasting mooring tension of offshore platforms based on complete ensemble empirical mode decomposition with adaptive noise and deep learning network. Measurement. 2025;239:115515.
  12. 12. Gong J, Xu J, Xu L, Hong Z. Enhancing motion forecasting of ship sailing in irregular waves based on optimized LSTM model and principal component of wave-height. Front Mar Sci. 2025;12.
  13. 13. Wei J, Hang R, Luo J-J. Prediction of Pan-Arctic Sea Ice Using Attention-Based LSTM Neural Networks. Front Mar Sci. 2022;9.
  14. 14. Yang Q, Liu G, Bao Y, Chen Q. Fault Detection of Wind Turbine Generator Bearing Using Attention-Based Neural Networks and Voting-Based Strategy. IEEE/ASME Trans Mechatron. 2022;27(5):3008–18.
  15. 15. Noor-Ul-Amin M, Kazmi MW, Alkhalaf S, Abdel-Khalek S, Nabi M. Machine learning based parameter-free adaptive EWMA control chart to monitor process dispersion. Sci Rep. 2024;14(1):31271. pmid:39732838
  16. 16. Liu G, Si J, Meng W, Yang Q, Li C. Wind Turbine Fault Detection With Multimodule Feature Extraction Network and Adaptive Strategy. IEEE Trans Instrum Meas. 2023;72:1–13.
  17. 17. Li Z, Liu F, Yang W, Peng S, Zhou J. A Survey of Convolutional Neural Networks: Analysis, Applications, and Prospects. IEEE Trans Neural Netw Learn Syst. 2022;33(12):6999–7019. pmid:34111009
  18. 18. Jin T, Yan C, Chen C, Yang Z, Tian H, Wang S. Light neural network with fewer parameters based on CNN for fault diagnosis of rotating machinery. Measurement. 2021;181:109639.
  19. 19. Alizadegan H, Rashidi Malki B, Radmehr A, Karimi H, Ilani MA. Comparative study of long short-term memory (LSTM), bidirectional LSTM, and traditional machine learning approaches for energy consumption prediction. Energy Exploration & Exploitation. 2024;43(1):281–301.
  20. 20. Lindemann B, Müller T, Vietz H, Jazdi N, Weyrich M. A survey on long short-term memory networks for time series prediction. Procedia CIRP. 2021;99:650–5.
  21. 21. Cui J, Shi Y, Huang J, Yang X, Gao J, Li Q. Dual temporal attention mechanism-based convolutional LSTM model for industrial dynamic soft sensor. Meas Sci Technol. 2024;35(11):115103.
  22. 22. Liao N, Guan J. Multi-scale Convolutional Feature Fusion Network Based on Attention Mechanism for IoT Traffic Classification. Int J Comput Intell Syst. 2024;17(1).
  23. 23. Yang D, Liu J, Song W, Yang C, Tuo J, Wang F. A load forecasting method for industrial customers based on the ICEEMDAN algorithm. Power System Protection and Control. 2022;50(4):36–43.
  24. 24. Qiu H, Hu R, Chen J, Yuan Z. Short-Term Electricity Load Forecasting Based on Complete Ensemble Empirical Mode Decomposition with Adaptive Noise and Improved Sparrow Search Algorithm–Convolutional Neural Network–Bidirectional Long Short-Term Memory Model. Mathematics. 2025;13(5):813.
  25. 25. Dang X-H, Shah SY, Zerfos P. seq2graph: Discovering Dynamic Non-linear Dependencies from Multivariate Time Series. In: 2019 IEEE International Conference on Big Data (Big Data), 2019:1774–83.
  26. 26. Khan SH, Iqbal R. A comprehensive survey on architectural advances in deep CNNs: challenges, applications, and emerging research directions. arXiv preprint. 2025.
  27. 27. Zhang Y, Shi X, Zhang H, Cao Y, Terzija V. Review on deep learning applications in frequency analysis and control of modern power system. International Journal of Electrical Power & Energy Systems. 2022;136:107744.
  28. 28. Hu H, Liu A, Guan Q, Qian H, Li X, Chen S, et al. Adaptively Customizing Activation Functions for Various Layers. IEEE Trans Neural Netw Learn Syst. 2023;34(9):6096–107. pmid:35007200
  29. 29. Zafar A, Aamir M, Mohd Nawi N, Arshad A, Riaz S, Alruban A, et al. A Comparison of Pooling Methods for Convolutional Neural Networks. Applied Sciences. 2022;12(17):8643.
  30. 30. Mienye ID, Swart TG, Obaido G. Recurrent Neural Networks: A Comprehensive Review of Architectures, Variants, and Applications. Information. 2024;15(9):517.
  31. 31. Wei Z, Shaohua J, Gang B, Yang C, Chengyang P, Haixing X. A Method for Sound Speed Profile Prediction Based on CNN-BiLSTM-Attention Network. JMSE. 2024;12(3):414.
  32. 32. Chaudhari S, Mithal V, Polatkan G, Ramanath R. An Attentive Survey of Attention Models. ACM Trans Intell Syst Technol. 2021;12(5):1–32.
  33. 33. Wang H, Li M, Lu K, Liu Y, Li S. Hydrodynamic characteristics experimental of anchored offshore test platform under wind wave current coupling. Ship Engineering. 2024;46(08):138–47.
  34. 34. Li S, Lu K, Wang H, Zheng H. Research on survivability test method for new types of floating offshore platforms under extreme sea conditions. China Meas Test. 2024;50(08):21–60.
  35. 35. Zheng H, Wang S, Li S, Wang X, Lu K, Li D, et al. Research on the stability of the ballast water tank on the offshore floating platform based on numerical simulation. PLoS One. 2025;20(10):e0334494. pmid:41100480
  36. 36. Petacco N, Pollicardo A, Gualeni P. An insight about roll damping within the second generation intact stability criteria. J Ocean Eng Mar Energy. 2024;10(4):1011–21.