Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

A bearing fault diagnosis method based on hybrid artificial intelligence models

Abstract

The working state of rolling bearing severely affects the performance of industrial equipment. Addressing the issue of that the difficulty of incipient weak signals feature extraction influences the rolling bearing diagnosis accuracy, an efficient bearing fault diagnostic technique, a proposition is forwarded for hybrid artificial intelligence models, which integrates Improved Harris Hawks Optimization (IHHO) into the optimization of Deep Belief Networks and Extreme Learning Machines (DBN-ELM). The process employs Maximum Second-order Cyclostationary Blind Deconvolution (CYCBD) to filter out noise from the vibration signals emitted by bearings; secondly, considering the issue with the conventional Harris Hawks Optimization (HHO) algorithm which tends to prematurely converge to local optima, the differential evolution mutation operator is introduced and the escape energy factor is improved from linear to nonlinear in IHHO; then, a double-layer network model based on DBN-ELM is proposed, to avoid the number of hidden layer nodes of DBN from human experience interference, and IHHO is used to optimize DBN structure, which is denoted as IHHO-DBN-ELM method; with the optimal structure is obtained by using a combined IHHO optimized DBN and ELM; in conclusion, the proposed IHHO-DBN-ELM approach is applied to the bearing fault detection using the Western Reserve University’s bearing fault dataset. The outcome of the experiments demonstrates that IHHO-DBN-ELM technique successfully extracts fault characteristics from the raw time-domain signals, thereby offering enhanced diagnostic accuracy and superior generalization capabilities.

Introduction

With the continuous advancement of modern industrial production systems, ensuring the safety and reliability of mechanical equipment has become increasingly critical. When a machine or component fails, it disrupts production processes, leading to significant financial losses. Moreover, severe malfunctions may endanger human lives and result in irreversible consequences. As key components of industrial rotating machinery, rolling bearings often operate under extreme conditions, such as high temperatures and heavy loads, making them particularly prone to failure. Therefore, timely and accurate detection of bearing faults is essential for maintaining efficient and safe industrial operations [1].

The early symptoms of rolling bearing faults typically appear as vibrations exhibiting nonlinear and non-stationary characteristics. These fault-induced signals are often embedded in background noise, making it challenging to extract meaningful diagnostic features. To address this issue, many researchers employ blind deconvolution algorithms to enhance signal quality by suppressing noise and recovering underlying fault signatures. A fault diagnostic approach was introduced by Zhang et al. [2], which integrated Continuous Vibration Separation (CVS) methodology with the application of Minimum Entropy Deconvolution (MED). The MCKD (Maximum Correlated Kurtosis Deconvolution) algorithm was introduced by MCDONALD et al. [3], with its primary focus on utilizing correlation kurtosis for optimization, thus emphasizing concealed fault characteristics amidst noise. In a subsequent study, Cui et al. [4] merged Variational Mode Decomposition (VMD) with MCKD to enhance fault detection. This approach strengthened the faint pulse signals of rolling element faults through the application of MCKD. However, MCKD has strict requirements on several input parameters such as filter length, maximum number of iterations and shift order, and is less effective in extracting shock components. These achievements have achieved certain results in signal denoising, but there are also fatal drawbacks. The CYCBD algorithm, which stands for Maximum Second-order Cyclostationarity Blind Deconvolution, was introduced by BUZZONI and collaborators in [5]. This method relies on the principle of maximizing the second-order cyclostationary statistic. Compared with other deconvolution algorithms, the proposed method demonstrates superior performance in extracting impulses under noisy conditions, effectively enhancing fault-related features in the signal. Given these advantages, this study employs the CYCBD to refine raw bearing vibration signals and suppress noise interference.

After ideal noise reduction effect is obtained for the bearing vibration signal, many scholars further extract the features and input them into the intelligent diagnosis model, which includes capsule network [6], ANN [7], BPNN [8]. However, these shallow models have poor fitting ability during dealing with complex nonlinear data, which resulting in low diagnostic accuracy [912].

In recent years, advanced deep learning architectures, including Convolutional Neural Networks (CNNs), Deep Belief Networks (DBNs), Autoencoders (AEs), and Recurrent Neural Networks (RNNs), have demonstrated remarkable success in bearing fault detection applications [1317]. Among these, DBNs exhibit particular strengths in learning complex nonlinear relationships from input data through their unique unsupervised layer-wise training mechanism. This architecture endows DBNs with exceptional nonlinear mapping capabilities and robust adaptability. As a result, DBNs have proven particularly effective at extracting subtle fault features from mechanical vibration signals, enabling highly accurate identification of incipient bearing defects. A hybrid Multi model fusion method for fault diagnosis, incorporating deep learning, which utilized different DBN combinations to achieve decision-level multimodal fusion, and was suggested by Che et al. [18]. An inventive Optimized Adaptive Deep Belief Network (OABN) was put forward by Gao et al. [19]. OABN undergoes initial training through minimal batch stochastic gradient descent. Subsequently, it is refined with the combination of backpropagation (BP) and conjugate gradient descent, thereby exhibiting superior classification capabilities. In reference [20], Yan and collaborators introduced a technique for unbalanced fault diagnosis utilizing Deep Belief Network (DBN). The results showed that a cracking classification accuracy was obtained in unbalanced data. Simultaneously, Extreme Learning Machine (ELM) possesses the distinct virtue of neural networks approximating nonlinear functions compared to SVM. Moreover, ELM has a fast learning rate and powerful generalization ability. It is currently widely utilized in fault classification fields and achieved ideal results [2123]. Therefore, combining the advantages of both DBN and ELM algorithms, so lots of scholars have established the DBN-ELM model, Gao et al. [24] utilized the DBN-ELM for power quality disturbance (PQD) classification. DBN extracted features from the denoised PQD signal, which improved the PQD classification efficiency through the ELM classifier. In the work by Cui et al. [25], a technique was introduced for the fault detection in gas turbines utilizing DBN-ELM architecture. The method was verified by experiments that DBN-ELM was more accurate than using DBN or ELM respectively. The DBN-ELM exhibits exceptional capabilities in data feature extraction and model classification efficiency, which led to the development of a fault diagnostic model that leverages these strengths.

The aforementioned approaches offer valuable insights for bearing feature extraction using Deep Belief Networks (DBNs). However, the performance of DBNs is critically dependent on the number of neurons in their hidden layers. Achieving optimal trade-offs between model stability and predictive accuracy, without resorting to empirical guesswork, remains a significant challenge in practical applications. Currently, swarm intelligence optimization algorithms have been extensively studied by many scholars [2631], which is adopted in the selection of the optimal parameters of the models. Referring to Zhang et al.‘s work [32], they introduced an innovative approach for diagnosing gearbox bearing faults. This approach fuses a Single-Dimensional Convolutional Neural Network (1DCNN) alongside a Support Vector Machine (SVM) for an integrated solution, where the SVM component is meticulously fine-tuned using Particle Swarm Optimization (PSO). Then, the experiments showed that the proposed 1DCNN-PSO-SVM improved the diagnostic accuracy. In the work of Ke et al. [33], they introduced a strategy that harnesses Genetic Algorithm (GA) to fine-tune the hyperparameters of Deep Convolutional Neural Networks (DCNNs), demonstrating exceptional efficiency in the context of modular multilevel converters (MMCs). Separately, in reference [34], Jin and colleagues employed Deep Belief Network (DBN) refined by an enhanced Grey Wolf Optimization (GWO) algorithm for the purpose of identifying faint faults in bearings. Their study confirmed the effective diagnosis of subtle faults in axle box bearings.

In light of the preceding research, this study introduces a technique forbearing fault diagnosis, which harnesses the improved Harris Hawks Optimizer (IHHO) to refine Deep Belief Networks (DBN). Firstly, CYCBD method is applied to denoise the raw bearing vibration signals, followed by time-domain feature dimension reduction using Principal Component Analysis (PCA). Secondly, Harris Hawks Optimizer is improved by incorporating both a differential evolution mutation strategy and a nonlinear escape energy mechanism, namely IHHO. This dual modification significantly improves the algorithm’s global exploration capability while effectively preventing premature convergence to local optima. The optimized IHHO algorithm then determines the optimal number of hidden units in the DBN-ELM framework. Finally, the selected time-domain features are fed into the optimally configured DBN-ELM model for fault classification. Experimental results demonstrate that our hybrid AI model achieves superior diagnostic accuracy with excellent generalization capability, though requiring longer training times due to its numerous parameters.

Related works

Harris hawks optimizer

Harris Hawk Optimizer (HHO), introduced by Heidari et al. [35] in 2019, is an innovative optimization strategy that mimics the hunting dynamics of hawks pursuing their prey. HHO follows three indispensable stages: exploration, the shift from exploration to exploitation, and finally, exploitation.

Exploration: In the exploration stage, Harris hawks foraging includes two mechanisms, randomly perching in tall trees or hunting according to other members and the location of the prey. A mathematical formulation is constructed as depicted in equation (1).

(1)

Here, and denote the location of an individual at the present and subsequent iterations, correspondingly. signifies the average strategic location of every participant, symbolizes the prey’s position, represents a selectively chosen individual’s position, denotes the total number of iterations, and are random values that lie within the range of 0 and 1.

The shift from exploration to exploitation: The shift of Harris’ hawks from discovery to utilization primarily revolves around the prey’s escape energy, as depicted in equation (2).

(2)

Here, represents the initial escape energy of the prey, is the present iteration count, and is upper limit for iterations.

Exploitation: Harris hawks exhibit a range of four strategic hunting techniques: soft besiege, hard besiege, soft besiege with progressive rapid dives and hard besiege with progressive rapid dives.

  1. (1) Soft besiege: In the scenario where , the prey, despite possessing adequate energy to evade, succumbs to the pursuit. Harris hawk skillfully encircles its target before executing an unexpected assault. The position is updated as follows:
(3)(4)

where, , and is the jumping distance during the escape of the rabbit.

  1. (2) Hard besiege: In the scenario where , the prey is depleted of energy and cannot escape. So, in HHO, the prey is exhausted and can attack it with little need to surround the prey. The update of position is shown in (5).
(5)
  1. (3) Soft besiege with progressive rapid dives: In the scenario where , it is possible for prey to escape because of sufficient energy. Harris hawks dives quickly around the prey several times, and always adjust its position to attack it. The formula is as follows:
(6)(7)(8)

Here, represents the dimension, is the random vector by inside , while denotes the Levy flight function.

  1. (4) Hard besiege with progressive rapid dives: In the scenario where , the prey has insufficient energy. A hard encirclement to raid the prey is formed in HHO, and the position is updated as (9).
(9)

Deep belief network (DBN)

DBN is constructed with multiple strata of Restricted Boltzmann Machines (RBMs). Within a RBM architecture, you’ll find a visible unit denoted as ‘v’ and a concealed unit referred to as ‘h’. The role of the visible unit v is to process incoming data, whereas the hidden unit h undertakes the task of extracting high-level features. The architectural layout of an RBM is depicted in Fig 1.

The connection weights and biases between RBM layers are initialized in an unsupervised greedy way. Then, each layer of RBM is trained from bottom to top. The energy defined between layers is as (10).

(10)

Here, denotes the parameter of the RBM, with and denoting the biases for the visible and hidden layers individually, while represents the weight matrix. The joint probability distribution function can be calculated by (10) as shown in (11).

The connection weights and biases between RBM layers are initialized in an unsupervised greedy way. Then, each layer of RBM is trained from bottom to top. The energy defined between layers is as (11).

(11)

The probabilities governing the activation of each visible layer element and hidden layer unit are depicted in equations (12) and (13) respectively.

(12)(13)

Here, is the Logistic function.

The parameters of RBM are refined through optimizing the maximum likelihood estimation equation derived from the training dataset, thus proceeding as equations (14)-(16).

(14)(15)(16)

In this context, denotes the Learning rate, signifies the anticipated values for the training dataset, while stands for the mathematical expectation of the reconstructed model.

Extreme learning machine

Extreme Learning Machine (ELM) represents an efficient machine learning algorithm specifically designed for Single Hidden Layer Feedforward Neural Networks (SLFNs). Unlike conventional approaches, ELM employs a unique batch training methodology characterized by randomly initialized hidden layer weights and biases.ELM’s training phase concludes through the determination of the hidden layer’s output weights. As mathematically formulated in Equation (17), it expresses the network output.

(17)

Here, is the activation function of the hidden layers, and are the weights for the input and output, respectively. is the bias of the hidden layer , and represents the target vector, is the sequence number of the target vector. The ELM training process can be simplified as shown in (18).

(18)

Where, denotes the output from the nodes in hidden layers, and represents the desired output. can be obtained from (19).

(19)

Here, represents the generalized inverse of .

The proposed method

IHHO

During iterative optimization, the conventional Harris Hawks Optimizer (HHO) exhibits progressive loss of population diversity, significantly impairing its capacity to avoid local optima convergence. To mitigate this limitation, a mutation mechanism is integrated during the exploitation phase, adopting the well-known DE/best/2 mutation operator from [36] to supplant the initial position update approach outlined in the equation. The enhanced position update formulation is depicted in (20).

(20)(21)(22)

Here, , , and are distinct integers chosen from . In order to avoid the phenomenon of premature maturity, the factor with adaptive variation is increased, and the initial value of is set to 0.5. Therefore, the position update in the exploration phase is changed from (1) to (23):

(23)

The escape energy factor of the prey plays a crucial role in balancing the global search and local development phases.

Upon the condition that , the algorithm initiates the exploration phase at the outset of the iterative process, hawks increase the global search range. When , hawks narrow the search for local development during the development phase.

During the exploration phase of , the algorithm expands the comprehensive search territory, whereas in the development phase of , the hawks refine their focus for localized progress. With the progressive rise in iterations, the value in (3) decreases linearly with the change of in the conventional HHO algorithm. It is noteworthy that the algorithm typically exhibits robust global search capabilities during the initial phases of its iterative process, while the later iterations focus more on the local search capability. Therefore, the transition parameter is improved as shown in (24).

(24)(25)

As the number of iterations progresses, the convergence factor diminishes non-linearly from 1 to 0. This evolution of and is graphically depicted in Fig 2. Diverging from the linear trajectory of , the non-linear transformation of facilitates a broader exploratory scope during the initial iteration stages, thereby augmenting the algorithm’s global search capabilities to prevent premature localization into a suboptimal solution. As the iterations progress, the parameter value declines at an accelerated pace, facilitating precise local searches and enhancing the algorithm’s convergence velocity. Consequently, the enhanced non-linear escape energy facilitates a more stable transition between the early exploration phase and the subsequent exploitation phase, thereby being more beneficial for enhancing both the convergence velocity and solution precision of HHO.

Performance analysis of IHHO algorithm

To rigorously evaluate the performance enhancement of our proposed Improved Harris Hawks Optimizer (IHHO), comparative experiments are conducted using Harris Hawk Optimizer (HHO), Whale Optimization Algorithm (WOA) [37,38], Grey Wolf Optimizer (GWO) and Genetic Algorithm (GA). These independent replication trials are executed on a set of six benchmark functions detailed in Table 1, sourced from the CEC (International Conference on Evolutionary Computing) Test Suite. These functions are relatively straightforward and facilitate convenient implementation. In addition, ~ are single-maxima benchmark functions, and ~ are versus complex multi-maxima counterparts. A single-maximum function serves to gauge the localized search capability of a swarm algorithm, whereas a multi-maxima function challenges its global exploration capacity.

The experimental configuration entails setting particular parameters such as a population size of 30 for each algorithm and a maximum iteration limit of 300. The logarithmic spiral parameter b in WOA is fixed at 1, while the search factor is configured to diminish linearly from a value of 2 down to zero.

The collaborative dynamics of GWO are influenced by a cooperative factor vector, along with its convergence attributes. The genetic algorithm’s mutation rate is configured at 0.1, while the probabilities for crossover and selection also play their roles. Each experimental function undergoes 30 isolated evaluations. The graphical representation of the fitness progression for these test functions can be observed in Figs 3–8.

From Figs 3–8, the following analysis and conclusions can be drawn:

  1. (1) The convergence curves of the functions show that GA has the weakest overall optimization capability and the optimization accuracy under most of the test functions is extremely low.
  2. (2) HHO and IHHO both have better convergence speed and optimization accuracy than WOA, GWO, and GA algorithms, proving the advantages of HHO and IHHO.
  3. (3) In terms of convergence precision, IHHO outperforms all other algorithms and demonstrates the swiftest rate of convergence for some of the test functions. For test function and, it can be seen that the convergence accuracy of IHHO is significantly improved. Besides, only the optimal value of IHHO is 0 for multi-peaked function ~, especially the function converges to 0 quickly, and reaches the optimal value in about 100 generations, which shows that the convergence speed becomes faster. can intuitively see that there are multiple inflection points in the whole iteration, which proves that the improved algorithm can prevent falling into local optimum.

After executing 30 times for each of the five algorithms, a comparison is made regarding the best outcome, mean value, and standard deviation derived from each test function. The standard deviation serves as an indicator of the algorithm’s stability, whereas the mean assesses the functional convergence precision. These experimental findings are counted as Table 2.

Compared to other algorithms, as revealed in Table 2, IHHO demonstrates the most superior optimization capability. The optimal value of function ,,,, is close to 0, and the optimization effect is the best. Similarly, the average value in the middle is the lowest, indicating that IHHO has high convergence accuracy, especially, IHHO standard deviation of and is very low, indicating that it has high robustness. Obviously, the optimization effect and convergence speed of IHHO are higher than those of the other four algorithms.

DBN-ELM optimized by IHHO algorithm

The construction of DBN-ELM model involves ELM and DBN, as depicted in Fig 9. The bottom DBN comprises three RBMs, which undergo unsupervised, greedy training for feature extraction. The visual layer, denoted as v, takes in the training set data with the neuron count equivalent to the dataset’s dimension. The top layer is the ELM classification stage, where the input layer is fed by the concealed layer’s output of the third RBM configuration. This innovative DBN-ELM hybrid architecture synergistically combines the complementary advantages of both components: the Deep Belief Network (DBN) demonstrates exceptional capability in hierarchical feature extraction, while the ELM contributes remarkable learning efficiency and robust generalization performance.The integrated framework significantly enhances both classification accuracy and computational efficiency during model training.

The diagnostic performance of the DBN-ELM model is critically dependent on the optimal configuration of hidden layer neurons. Conventional approaches typically determine this hyperparameter through empirical estimation, often resulting in suboptimal model performance and inconsistent diagnostic outcomes. To address this limitation and achieve enhanced model accuracy with accelerated convergence, we implement the Improved Harris Hawks Optimizer (IHHO) for systematic optimization of the hidden layer architecture. The optimization process proceeds through the following key steps:

  1. (1) The optimization process begins with parameter initialization for IHHO, which involves configuring three critical operational parameters: the population size, the maximum iteration limit, and the boundaries. Initialize the positions of Harris hawks population randomly, designating these positions as the neuron count in each concealed layer of the model.
  2. (2) The feature vector obtained after data preprocessing is input into DBN for training, and the cross-entropy loss is computed by back-propagation. DBN is fine-tuned and saved. Then, the parameters of DBN are utilized to initialize the feature space of the ELM, and the output weight vector β for ELM is computed.
  3. (3) Utilize the mean precision deviation of the 10-fold cross-validation on the genuine labels and the categorical designations of DBN-ELM as the IHHO fitness function.
  4. (4) Update prey energy. Utilize the energy of the prey and a random value r to select and implement the appropriate updating tactic. Subsequently, revise the positioning of the hawk population.
  5. (5) Evaluate if the present iteration counts have attained the terminating criterion; if met, conclude the loop and yield the outcome; otherwise, proceed back to stage 2) for further execution.
  6. (6) Return the optimal individual position prey position and its fitness value. Set the optimal individual position prey position as the structure parameter of DBN-ELM to construct the model.

Overview of the proposed method

The proposed IHHO-DBN-ELM fault detection strategy’s comprehensive architectural progression is depicted in Fig 10. This approach encompasses two distinctive phases: pre-operational training and real-time diagnosis. The pre-operational training segment focuses on optimizing the DBN architecture for heightened effectiveness and efficiency. The online diagnosis part is tested with the model obtained by offline training, and the bearing diagnosis result is obtained. The specific operation process is completed by the following four steps:

  1. (1) Obtain the original signal.
  2. (2) Adopt CYCBD algorithm to denoise the data, and the bearing signal is selected and normalized in time domain. Proceed with segmenting the dataset into separate training and testing sets.
  3. (3) Construct IHHO-DBN-ELM framework, applying IHHO technique to optimize the architectural configuration of DBN-ELM network. Employ the training subset to train the model, thereby preserving the optimal parameters.
  4. (4) Validate the performance of the trained DBN-ELM with the test set.

Data preprocessing and feature extraction

Experimental device

The proposed IHHO-DBN-ELM algorithm is verified by the bearing data set published of Case Western Reserve University. Experimental device consists of five parts: 1.5 kw electric motor, fan and bearing, drive end bearing torque transducer & encoder, dynamometer, which is shown in Fig 11. In the experiment, single-point faults such as inner ring, outer ring, and rolling body with a diameter of 0.007 mm were caused by electric sparks, and the acceleration and vibration data at the sampling frequency of 12khz were collected. The load is 0 ~ 2HP. Table 3 shows the detailed information of the experimental data on load 0HP.

Data preprocessing

Owing to the intricate operational conditions, the pulse components caused by rotating machinery faults are often interfered by random pulses, periodic harmonics, etc., which affect the fault diagnosis and classification. Therefore, CYCBD algorithm is selected to perform noise reduction preprocessing on the experimental data.

CYCBD approach introduces a novel blind deconvolution technique that involves iterative resolution of the feature vector h, to optimize the filter and thereby diminish noise disturbances while emphasizing fault characteristics. The dataset of 2000 samples from four distinct fault scenarios (normal, inner fault, rolling fault, and outer fault) under 0 load was selected. These were subjected to noise reduction treatment using the CYCBD method. As depicted in Fig 12, the left side displays the observed noisy signal after noisy reduction, while the right side exhibits the estimated source. It is demonstrates enhanced periodic patterns and pulse clarity, significantly mitigating noise interference on the right side of Fig 12, which is more conducive to extracting fault features.

Feature selection

Time domain features are often applied in fault diagnosis because of their simplicity and easy availability. Therefore, temporal features are selected to train the proposed IHHO-DBN-ELM. The time domain statistical method mainly consists of waveform observation and time domain numerical calculation. Table 4 exhibits the details of common time-domain characteristic formulas.

If 14 time-domain features in Table 4 are formed into a feature matrix, the amount of computation will increase and the classification efficiency will be affected. Hence, the dataset’s dimensionality is reduced through the utilization of Principal Component Analysis (PCA), and four kinds of time-domain characteristic parameters whose cumulative contribution rate is greater than 95% are selected. The four time-domain features are mean, peak, standard deviation, and amplitude sum of squares. That is, the four parameters preserved through PCA dimensionality reduction retain 95% of the important information in all parameters, but can greatly improve computational efficiency.

Experiments

Bearing fault diagnosis based on IHHO-DBN-ELM

To comprehensively evaluate the performance of the proposed IHHO-DBN-ELM framework, we established an experimental protocol using 240,000 preprocessed bearing vibration records collected under no-load conditions (0 HP). The dataset was strategically partitioned with an 80:20 ratio for training and testing purposes, respectively. The optimization parameters were configured as follows: population size is 30, maximum iterations is 300, and searching space boundaries belong to [100, 500] for hidden neuron counts. The visibility layer count is fixed at 400. Through IHHO optimization, we identified an optimal DBN-ELM architecture of 400-129-375-375-165-4. The classification performance is quantitatively demonstrated through the confusion matrix presented in Fig 13, which reveals the model’s diagnostic accuracy across different fault categories.

As depicted in Fig 13, the proposed IHHO-DBN showcases remarkable performance with commendable precision in its identification capabilities. Notably, a solitary misclassification occurred where a test specimen was misidentified, assigning it to a rolling fault instead of the intended inner fault. The recognition accuracy of the entire test set reached 99.17%, which shows that the fault types can be effectively identified.

To demonstrate the effectiveness performance of the proposed method, identical datasets are utilized for both training and testing phases for IHHO-DBN-ELM and HHO-DBN-ELM. The examination of the test set outcomes can be observed in Figs 14 and 15. The diagnostic accuracy of HHO-DBN-ELM is 93.33%, while the diagnostic accuracy of IHHO-DBN-ELM reached 99.17%. Observations reveal that IHHO-DBN-ELM has effectively pinpointed optimal hyperparameters through tuning the quantity of hidden layer neurons and enhancing the training approach, thereby maximizing the exceptional capabilities of DBN-ELM.

Comparison experiments of optimized DBN-ELM models

To substantiate the model’s credibility further, the mean diagnostic precision was ascertained through five replicate tests employing IHHO-DBN-ELM, HHO-DBN-ELM, GWO-DBN-ELM, and DBN-ELM frameworks. These comparisons are depicted in Table 5. Observations reveal that the initial DBN-ELM’s average accuracy falls below 90%. Conversely, the swarm algorithm-optimized models exhibit notably superior diagnostic precision compared to the base DBN-ELM. Specifically, the IHHO-DBN-ELM achieves an average accuracy of 97.53%, consistently outperforming HHO-DBN-ELM across all experiments, demonstrating that the IHHO algorithm optimizes DBN-ELM’s performance more effectively than its baseline counterpart. Consequently, the proposed IHHO-DBN-ELM method demonstrates the highest diagnostic efficacy.

Comparison experiment of common AI diagnostic models

As shown in Table 6, compared with BP neural network, SVM model, RBF model and CNN model proposed by scholars in recent years, the experimental results demonstrate significant performance advantages of the proposed deep learning framework over conventional shallow models. Specifically, our IHHO-DBN-ELM architecture achieves a remarkable classification accuracy of 96.73%, representing a substantial improvement over traditional shallow networks which typically attain only 85–88% accuracy on similar diagnostic tasks. While maintaining competitive computational efficiency with an average inference time of 0.73 seconds per test sample, well within the sub-second requirement for real-time industrial applications, the proposed method exhibits superior diagnostic precision compared to both individual deep networks and shallow models. These findings suggest that the IHHO-DBN-ELM framework not only provides enhanced classification performance but also maintains practical applicability, indicating strong potential for implementation in industrial fault diagnosis systems.

thumbnail
Table 6. Experimental comparison results of common AI diagnostic models.

https://doi.org/10.1371/journal.pone.0327646.t006

Comparative experiments under different data sets

Accuracy is an important indicator for evaluating IHHO-DBN-ELM intelligent fault diagnosis. The higher the accuracy, the better the diagnosis effect. In order to enrich the coverage of sample selection, this paper adjusts the sampling frequency, truncates the original bearing data into a sample length according to different points, and obtains 3000, 6000, and 10000 pieces of data to form a test set for the next operation. The ratio of the number of healthy and unhealthy data is 1:3, and as depicted in Table 7, it is evident that Table 7 that the IHHO-DBN intelligent fault diagnosis system has a high classification accuracy under different test set quantities, meeting the expected goals.

thumbnail
Table 7. Accuracy test of artificial intelligence diagnosis model.

https://doi.org/10.1371/journal.pone.0327646.t007

Conclusions

This study proposes an advanced bearing fault detection method that integrates an Improved Harris Hawks Optimization (IHHO) algorithm with a Deep Belief Network-Extreme Learning Machine (DBN-ELM) architecture. The proposed framework addresses two critical challenges in rotating machinery fault diagnosis: (1) noise contamination that masks incipient fault features in vibration signals; (2) suboptimal network architecture in conventional deep learning approaches. The main work of this paper includes two parts: methodology and experimental validation, which are described as follows:

In aspect of the methodology:(1) Signal Preprocessing: The raw vibration signals are first enhanced using Maximum Correlated Kurtosis Deconvolution (CYCBD) to amplify fault-related periodic impulses while suppressing noise interference. (2) Feature Learning: The denoised time-domain signals serve as input to the DBN, which automatically extracts discriminative fault features through its hierarchical structure of Restricted Boltzmann Machines. (3) Model Optimization: IHHO algorithm determines the optimal DBN-ELM architecture by introducing a differential evolution mutation operator to preserve population diversity and implementing a nonlinear escape energy factor to dynamically balance exploration and exploitation. (4) Fault Classification: The optimized DBN-ELM model performs final fault pattern recognition.

In aspect of experimental validation: Evaluation on the Case Western Reserve University (CWRU) bearing dataset demonstrates superior performance: 99.17% classification accuracy, 97.53% average precision, which has significant improvement over conventional methods.

While achieving excellent diagnostic performance, the method presents two limitations: strong dependency on effective noise removal during preprocessing and increased computational requirements compared to traditional techniques. Future work will focus on optimizing computational efficiency for real-time applications and enhancing generalization capability under varying operational conditions.

References

  1. 1. Li J, Luo W, Bai M, Song M. Fault diagnosis of high-speed rolling bearing in the whole life cycle based on improved grey wolf optimizer-least squares support vector machines. Digital Signal Processing. 2024;145:104345.
  2. 2. Zhang L, Hu N. Fault diagnosis of sun gear based on continuous vibration separation and minimum entropy deconvolution. Measurement. 2019;141:332–44.
  3. 3. McDonald GL, Zhao Q, Zuo MJ. Maximum correlated Kurtosis deconvolution and application on gear tooth chip fault detection. Mechanical Systems and Signal Processing. 2012;33:237–55.
  4. 4. Cui H, Guan Y, Chen H. Rolling Element Fault Diagnosis Based on VMD and Sensitivity MCKD. IEEE Access. 2021;9:120297–308.
  5. 5. Buzzoni M, Antoni J, D’Elia G. Blind deconvolution based on cyclostationarity maximization and its application to fault identification. Journal of Sound and Vibration. 2018;432:569–601.
  6. 6. Yu T, Gan Q, Feng G, Han G. A new fuzzy cognitive maps classifier based on capsule network. Knowledge-Based Systems. 2022;250:108950.
  7. 7. Sun L, Chen S, Zhu J, Li J. Substation Equipment Temperature Prediction Method considering Local Spatiotemporal Relationship. Scientific Programming. 2022;2022:1–14.
  8. 8. Xiao M, Zhang W, Zhao Y, Xu X, Zhou S. Fault diagnosis of gearbox based on wavelet packet transform and CLSPSO-BP. Multimed Tools Appl. 2022;81(8):11519–35.
  9. 9. Song H, Yang W. GSCCTL: a general semi-supervised scene classification method for remote sensing images based on clustering and transfer learning. International Journal of Remote Sensing. 2022;43(15–16):5976–6000.
  10. 10. Wei H, Zhang Q, Shang M, Gu Y. Extreme learning Machine-based classifier for fault diagnosis of rotating Machinery using a residual network and continuous wavelet transform. Measurement. 2021;183:109864.
  11. 11. Lei Y, Yang B, Jiang X, Jia F, Li N, Nandi AK. Applications of machine learning to machine fault diagnosis: A review and roadmap. Mechanical Systems and Signal Processing. 2020;138:106587.
  12. 12. Wang D, Zhao X. Affective video recommender systems: A survey. Front Neurosci. 2022;16:984404. pmid:36090291
  13. 13. Gong W, Chen H, Zhang Z, Zhang M, Wang R, Guan C, et al. A Novel Deep Learning Method for Intelligent Fault Diagnosis of Rotating Machinery Based on Improved CNN-SVM and Multichannel Data Fusion. Sensors (Basel). 2019;19(7):1693. pmid:30970672
  14. 14. Zhang X, He C, Lu Y, Chen B, Zhu L, Zhang L. Fault diagnosis for small samples based on attention mechanism. Measurement. 2022;187:110242.
  15. 15. Niu G, Wang X, Golda M, Mastro S, Zhang B. An optimized adaptive PReLU-DBN for rolling element bearing fault diagnosis. Neurocomputing. 2021;445:26–34.
  16. 16. Zhao K, Jiang H, Liu C, Wang Y, Zhu K. A new data generation approach with modified Wasserstein auto-encoder for rotating machinery fault diagnosis with limited fault data. Knowledge-Based Systems. 2022;238:107892.
  17. 17. Han S, Zhong X, Shao H, Xu T, Zhao R, Cheng J. Novel multi-scale dilated CNN-LSTM for fault diagnosis of planetary gearbox with unbalanced samples under noisy environment. Meas Sci Technol. 2021;32(12):124002.
  18. 18. Che C, Wang H, Ni X, Lin R. Hybrid multimodal fusion with deep learning for rolling bearing fault diagnosis. Measurement. 2021;173:108655.
  19. 19. Gao S, Xu L, Zhang Y, Pei Z. Rolling bearing fault diagnosis based on SSA optimized self-adaptive DBN. ISA Trans. 2022;128(Pt B):485–502. pmid:35177261
  20. 20. Yan J, Hu Y, Guo C. Rotor unbalance fault diagnosis using DBN based on multi-source heterogeneous information fusion. Procedia Manufacturing. 2019;35:1184–9.
  21. 21. Pang S, Yang X, Zhang X, Sun Y. Fault diagnosis of rotating machinery components with deep ELM ensemble induced by real-valued output-based diversity metric. Mechanical Systems and Signal Processing. 2021;159:107821.
  22. 22. Li J, Lin M. Ensemble learning with diversified base models for fault diagnosis in nuclear power plants. Annals of Nuclear Energy. 2021;158:108265.
  23. 23. Lan Y, Hu J, Huang J, Niu L, Zeng X, Xiong X, Wu B. Fault diagnosis on slipper abrasion of axial piston pump based on extreme learning machine. Meas J Int Meas Confed. 2018;124:378–85.
  24. 24. Gao Y, Li Y, Zhu Y, Wu C, Gu D. Power quality disturbance classification under noisy conditions using adaptive wavelet threshold and DBN-ELM hybrid model. Electric Power Systems Research. 2022;204:107682.
  25. 25. Cui J, Liu Y, Yu M, Jiang X. gas turbine fault diagnosis based on depth learning and information fusion. Mach Des Manuf. 2019;12:28–31.
  26. 26. Xu X, Hao J, Zheng Y. Multi-objective artificial bee colony algorithm for multi-stage resource leveling problem in sharing logistics network. Computers & Industrial Engineering. 2020;142:106338.
  27. 27. Zhang Q, Zhou Y. Recent Advances in Non-Gaussian Stochastic Systems Control Theory and Its Applications. Int J Netw Dyn Intell. 2022;1(1):111–9.
  28. 28. Xue J, Shen B. A novel swarm intelligence optimization approach: sparrow search algorithm. Systems Science & Control Engineering. 2018;8(1):22–34.
  29. 29. Hou J, Su H, Yu C, Chen F, Li P, Xie H, et al. Consistent Subspace Identification of Errors-in-Variables Hammerstein Systems. IEEE Trans Syst Man Cybern, Syst. 2022;53(4):2292–303.
  30. 30. Moayedi H, Nguyen H, Kok FL. Nonlinear evolutionary swarm intelligence of grasshopper optimization algorithm and gray wolf optimization for weight adjustment of neural network. Eng Comput. 2019;37(2):1265–75.
  31. 31. Xu X, Lin Z, Li X, Shang C, Shen Q. Multi-objective robust optimisation model for MDVRPLS in refined oil distribution. Int J Prod Res. 2022;60(22):6772–92.
  32. 32. Zhang X, Han P, Xu L, Zhang F, Wang Y, Gao L. Research on bearing fault diagnosis of wind turbine earbox based on 1DCNN-PSO-SVM. IEEE Access. 2020;8:192248–58.
  33. 33. Ke L, Zhang Y, Yang B, Luo Z, Liu Z. Fault diagnosis with synchrosqueezing transform and optimized deep convolutional neural network: An application in modular multilevel converters. Neurocomputing. 2021;430:24–33.
  34. 34. Jin Z, He D, Wei Z. Intelligent fault diagnosis of train axle box bearing based on parameter optimization VMD and improved DBN. Engineering Applications of Artificial Intelligence. 2022;110:104713.
  35. 35. Ayinla SL, Amosa TI, Ibrahim O. Optimal control of DC motor using leader-based Harris Hawks optimization algorithm. Franklin Open. 2024;6:100058.
  36. 36. Opara KR, Arabas J. Differential Evolution: A survey of theoretical analyses. Swarm Evol Comput. 2019;44:546–58.
  37. 37. El-Fergany AA, Hasanien HM, Agwa AM. Semi-empirical PEM fuel cells model using whale optimization algorithm. Energy Conversion and Management. 2019;201:112197.
  38. 38. Xu X, Liu W, Yu L. Trajectory prediction for heterogeneous traffic-agents using knowledge correction data-driven model. Information Sciences. 2022;608:375–91.