Figures
Abstract
As a crucial component in rotating machinery, bearings are prone to varying degrees of damage in practical application scenarios. Therefore, studying the fault diagnosis of bearings is of great significance. This article proposes the Kepler algorithm to optimize the weights of neural networks and improve the diagnostic accuracy of the model. At the same time, combined with attention mechanisms, the model will focus on useful information, ignore useless information, and efficiently extract key features. Finally, using third-party bearing data and inputting it into the fault diagnosis model, it was verified that Kepler algorithm and attention mechanism can improve the diagnostic accuracy. Meanwhile, the algorithm proposed in this paper was compared with other algorithms to verify its feasibility and superiority.
Citation: Guang YJ, Gen XS, Meng Meng S, Hui YW, Yan F, Jie YH (2025) Bearing fault diagnosis based on Kepler algorithm and attention mechanism. PLoS One 20(9): e0331128. https://doi.org/10.1371/journal.pone.0331128
Editor: Yirui Wang, Ningbo University, CHINA
Received: April 25, 2025; Accepted: August 11, 2025; Published: September 4, 2025
Copyright: © 2025 Guang et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The data and code are stored in the following link: https://github.com/Rocky206282/KOA-and-attention.
Funding: This work is supported by funds from the Natural Science Foundation of Fujian Province (Grant No. 2024J01946) and the Ningde City’s Major Technical Requirement Project for “Unveiling the List and Leading the Way” (No. ND2024J004) and the collaborative innovation center project of Ningde Normal University (Grant No. 2023ZX01). Thank you for the project support of the "Mechanical Engineering" discipline construction platform.
Competing interests: There is no conflict of interest in this paper.
1. Introduction
As a crucial component in rotating machinery, bearings have the advantages of compact structure, light weight, strong load-bearing capacity, and low cost. In practical application scenarios, bearings are subjected to complex dynamic heavy load forces and are susceptible to varying degrees of damage, resulting in different types of failure [1]. Meanwhile, the majority of mechanical equipment failures are directly caused by bearing failures, accounting for 40% −50% [2,3]. Therefore, studying the fault diagnosis of bearings is of great significance. In recent years, this issue has received widespread attention from scholars both domestically and internationally.
When diagnosing bearing faults, traditional fault diagnosis methods and artificial intelligence fault diagnosis methods can be used. The traditional fault diagnosis method is to use sensors to collect data on bearing operation, manually extract features from the operation data, and finally use the extracted features to train a machine learning model that can distinguish different types of fault features [4–6]. At present, traditional fault diagnosis mainly includes K-nearest neighbor algorithm, support vector machine, and artificial neural network. Due to its simple principle, K-Nearest Neighbor (KNN) algorithm is often used in early bearing fault diagnosis.Song et al. [7] proposed a fault detection method based on standardized KNN, which characterizes the distance between data and its neighbors through a standardized distance,this method requires a large amount of computation and is inefficient because for each text to be classified, its distance to all known samples must be calculated. In order to reduce computational complexity and improve efficiency, the support vector machine (SVM) method was adopted in later fault diagnosis.Chen et al. [8,9] proposed an early fault diagnosis method based on orthogonal neighborhood preservation embedding and Adaboost SVM algorithm. The orthogonal neighborhood preservation embedding method is used to eliminate redundant information in the original multi domain feature set, and then SVM is improved into Adaboost SVM for early fault diagnosis. However, this method relies heavily on the operator’s professional knowledge. In order to reduce the dependence of the model on the operator’s professional knowledge,Rex et al. [10,11]proposed a hybrid method for extracting and classifying gear faults by integrating Hu invariant moments and artificial neural networks. However, this method is inefficient and labor-intensive. In short, traditional fault diagnosis methods have obvious drawbacks, as they require a large amount of manpower and specialized knowledge in the corresponding field, resulting in low accuracy. In addition, there may be errors in the diagnosis process due to subjective factors.
Meanwhile, artificial intelligence can also be applied in the field of bearing fault diagnosis, achieving excellent results in the field of fault diagnosis [12–15]. Among them, neural networks, as a branch of artificial intelligence, are widely used in fault detection of bearings [16–19]. Jin Zhihao et al. [20]used neural networks for bearing fault detection. This method takes the original time-domain vibration signal as input, converts the data form using Welch power spectrum while suppressing high-intensity noise, and then trains the convolutional neural network with the obtained power spectrum. Finally, the trained model is used for bearing fault diagnosis. However, this method requires preprocessing of irrelevant data, and the diagnosis process is relatively complicated. In order to reduce the tedious data preprocessing process, Gao Feng et al. [21]established a neural network-based fault diagnosis model to achieve adaptive extraction of fault features. Although this method can reduce the preprocessing process, the neural network structure used is relatively simple, which makes the diagnosis process more complicated and requires multiple training of the model. In order to reduce the number of training models, Chang Miao et al. [22] proposed a fault diagnosis algorithm based on an improved neural network.Improving the structure of the neural network model by adding a new convolutional layer before the fully connected layer improves the structure of the neural network. However, this method does not optimize the weights of the neural network, and when dealing with numerous data with small differences, the neural network cannot concentrate and efficiently capture feature data.
In order to solve the weight optimization problem in the neural network structure and improve the attention of the neural network, this paper uses Kepler algorithm to optimize the weights in the neural network, and applies attention mechanism to enhance the attention of the neural network. The optimized neural network model is applied to bearing fault diagnosis, which improves the efficiency and accuracy of bearing fault diagnosis.
2. Basic knowledge
2.1. Kepler algorithm
In dealing with complex problems and optimization fields, some traditional optimization methods, such as gradient descent and genetic algorithms, have achieved good results in many problems. However, when dealing with complex multimodal optimization problems and highly nonlinear problems, they still expose slow convergence speed and local optima. Therefore, new optimization algorithms have emerged to better solve complex problems in modern science and engineering. The inspiration for Kepler Optimization Algorithm (KOA) comes from Kepler’s laws of planetary motion, which apply the laws of planetary motion to the algorithm and design a new type of swarm intelligence optimization algorithm. Its emergence provides engineers with a novel and unique way to solve optimization problems, especially in solving complex numerical optimization and machine learning parameter tuning, demonstrating good performance. The advantages of KOA algorithm:
- (1). Strong global search capability: The KOA algorithm can comprehensively explore the search space by simulating the orbital motion of planets, avoiding getting stuck in local optimal solutions
- (2). Fast convergence speed: KOA algorithm can find the optimal solution in a short time, improving diagnostic efficiency
- (3). Good parameter optimization effect: By optimizing the hyperparameters of neural network, the KOA algorithm significantly improves the accuracy and efficiency of bearing fault diagnosis.
Traditional metaheuristic algorithms such as genetic algorithm (GA), particle swarm optimization algorithm (PSO), and ant colony algorithm (ACO) have shown excellent performance in solving complex optimization problems, but they have some shortcomings:
- (1). Easy to get stuck in local optima: These algorithms may converge too early during the search process and cannot find the global optimum
- (2). Slow convergence speed: Traditional metaheuristic algorithms have a slow convergence speed when dealing with large-scale problems, which affects diagnostic efficiency
It is precisely because the KOA algorithm has the above advantages that the KOA algorithm superior over traditional metaheuristics for this specific bearing diagnosis task. In this article, we apply the Kepler algorithm to optimize the weights of neural network. Optimizing the weights of neural network through Kepler algorithm can converge to the optimal value more quickly, thereby improving the training speed and prediction accuracy of the model.
The basic principle of KOA [23] is derived from Kepler’s laws of planetary motion. The three laws of KOA summarize the laws of planetary motion around the sun. KOA has created a mathematical model based on position and velocity by simulating the trajectory of planetary motion and the interaction of universal gravity [24]. The two main aspects of algorithms include:
- (1). Gravitational attraction: Planets move around their orbits due to the universal gravitational force of the Sun. KOA adjusts their search direction by calculating the gravitational interaction between particles (representative solutions), enhancing their global search capability.
- (2). Orbital characteristics: Different orbital features (such as ellipses, parabolas, etc.) are used to simulate different search states, making the optimization process diverse and flexible.
The basic principles of KOA algorithm [25,26] are described in detail as follows:
2.1.1. Definition of gravity.
As the largest celestial body in the solar system, the sun maintains the orbits of planets in elliptical orbits through universal gravity, as shown in Fig 1. The orbital velocity of a planet is inversely proportional to its distance from the sun: the closer the distance, the higher the velocity. These dynamics can be explained by the law of universal gravitation, which describes that the gravitational force between objects is proportional to their mass and inversely proportional to the square of their distance. The expression of universal gravitation is shown in formulas (1) and (2).
Among them, is the gravitational constant;
is the eccentricity of a planet’s orbit,it is a random value between 0 and 1;
representing the Euclidean distance between the sun and planets;
and
represent the standardized values of
and
;
is a minimum value; The value of
is between 0 and 1, which is a randomly generated value that provides more variation for the gravity value during the optimization process.
2.1.2. Calculate the speed of an object.
When a planet approaches the sun, its velocity increases due to the strong gravitational pull of the sun; When a planet moves away from the sun, its speed decreases due to the weakening of gravity, and this dynamic behavior can be modeled through equations [27]. This model consists of two parts: adjusting the distance between solutions to adjust the velocity of planets close to the sun, in order to enhance search diversity; Reduce these distances to decrease the speed of planets away from the sun, improve the problem of insufficient solution diversity, and enhance solution diversity. The velocity expression of a planet can be represented by formulas (3) – (14).
Among them, represents the velocity of the object
in time
,
represents the planet
,
and
is a randomly generated value in the interval [0,1].
and
are two vectors, random values between 0 and 1;
and
represent solutions randomly selected from the group;
and
represents the mass of
and
;
is the constant of universal gravitation;
is a small value used to prevent zero division error;
representing the distance between the optimal solution
and the object
at time
;
representing the semi major axis of the elliptical orbit of the object
at time
;
representing the orbital period of the object
;
is a value randomly generated based on a normal distribution.
2.1.3. Jumping out of local optima.
KOA draws inspiration from the natural behavior of planets in the solar system rotating around the sun and introduces a marker to change the search direction, which can effectively escape from local optimal areas and enhance the comprehensive exploration capability of the entire space [28].
2.1.4. Update target location.
As shown in Fig 2, KOA divides the simulation of the natural motion of celestial bodies around the sun in elliptical orbits into two stages: exploration and development. In the exploration phase, KOA explores areas far from the sun to find new solutions. During the development phase, KOA focuses on utilizing known solutions close to the sun. The update of the target position can be represented by formula (15):
Among them, is the new position of the object
at time
;
representing the speed required for an object
to reach a new location;
representing the best position of the sun discovered so far;
used as a flag to change search direction.
2.1.5. Update distance from the sun.
KOA optimizes exploration and development by simulating the natural changes in distance between the sun and planets, and adjusts its operating mode based on the values of adjustment parameters. The distance between the planet and the sun can be updated using formulas (16) and (17).
Among them, is an adaptive factor used to control the distance between the sun and the current planet at time
,
is a linearly decreasing factor from 1 to −2.
In summary, the entire calculation process of Kepler algorithm is shown in Fig 3:
2.2 Neural networks
Neural networks [29]mimic the data processing process of human neurons to process data. Neural networks generally include convolution operations, which have the characteristics of sparse connections, parameter sharing, and translation invariance. Neural networks are mainly composed of several fully connected layers,convolutional layers, and pooling layers, which will be introduced in this article.
2.2.1 Fully Connected Layer.
The fully connected layer [30] is composed of multiple M-P neurons, as shown on the right side of Fig 4, with a chain like structure between layers. Usually, the number of neurons in the next layer (layer width) is chosen to be less than or equal to the number of neurons in the previous layer, in order to compress the high-dimensional data of the convolutional neural network into lower dimensional data (feature extraction) for subsequent classifier classification.
M-P neuron is a mathematical model abstracted from biological neurons, and the output expression of a single neuron is shown in equation (18).
Among them, represents the input component,
is the bias,
is the weight on the corresponding arc on the right side of the neuron model. The input data passes through the arc and is multiplied by the weight, then input to the neuron for summation, and finally activated by the hidden unit (activation function) to obtain the output of the neuron. From the calculation process, it can be seen that weights will affect the final result, so optimizing their weights is very necessary.
2.2.2. Convolutional layer.
The convolutional layer uses convolution operations between the input matrix and the weight matrix instead of matrix multiplication [31]. Convolution operation is a mathematical operation on two time-varying functions, where and are the real variable functions of, known as kernel functions. Due to the fact that the data processed by computers is discrete, the discrete expressions for convolution operations are shown in formulas (19) and (20).
In the formula, represents the current convolution time point,
represents the time scale of the time series, and
is the length of the time series,
is the convolution kernel and also a type of weight.Different weights will greatly affect the final calculation result. From the calculation process, it can be seen that weights will affect the final result, so optimizing their weights is very necessary.This article focuses on one-dimensional time series, and the calculation process can refer to Fig 5.
2.2.3. Pooling layer.
The pooling layer replaces the input with local statistics [32], and commonly used statistics in the pooling process include maximum mean and norm. The formula for the pooling process can be represented by formula (21).
Among them, represents the
th output component of the pooling layer
.
represents the
th input component of the layer
, is the pooling width,
is the pooling stride,
is the pooling stride,
is the sum of the lengths of the input sequence, and
represents rounding up. The maximum statistical value for the pooling process shown in Fig 6 is selected, with a pooling width of 2 and a pooling stride of 2. From the calculation process of the pooling layer, it can be seen that the pooling layer needs to select the maximum value in the data. If the data size of a long time series is relatively close during the data processing, it will be difficult for the pooling layer to select suitable data during the operation process, and it will also take a lot of time to perform the operation. To address this issue, attention mechanisms can be introduced to ignore redundant irrelevant data and focus attention on useful feature data.
2.3. Long short term memory network(LSTM)
There are many types of neural networks, and long short-term memory networks are one of them. Long Short Term Memory [33](LSTM) is a special type of neural network, which is a time recurrent neural network mainly used for processing and modeling sequence data, such as text, speech, time series, etc. Due to the fact that the bearing fault data used in this article is time series data, an LSTM network is employed. Due to the presence of fully connected layers and convolutional layers in the LSTM network structure, weights are used for mathematical operations in each layer, which can affect the final calculation results. Therefore, optimizing the weights of LSTM is crucial.
2.4. Attention mechanism
Due to the inability of the pooling layer in LSTM networks to handle massive amounts of duplicate data, this paper introduces an attention mechanism to address this issue. The attention mechanism is essentially a resource allocation mechanism that changes the way resources are allocated based on the importance of the attention target, tilting resources more towards the attention object [34]. Adding attention mechanism to object detection tasks can enhance the representation ability of the model, reduce the interference of invalid targets, improve the detection effect of attention targets, and improve the accuracy of model detection. According to the characteristics of attention mechanism, it can solve the pain points of LSTM network processing duplicate data. It can make LSTM network more focused and efficient, and can match LSTM network very well.The model is shown in Fig 7 below:
3. Fault diagnosis model based on Kepler algorithm and attention mechanism
Based on the above description, it can be seen that neural networks have two major pain points. In view of this, this paper proposes a new diagnostic model based on LSTM network, optimizes its weights using Kepler algorithm, and combines attention mechanism to improve its ability to process duplicate data, solving its two major pain points. This model extracts features from a large amount of one-dimensional vibration data through convolution, and then reduces the dimensionality through pooling. The pooled data uses attention mechanism to fully extract its features, and then outputs the final diagnostic result. Meanwhile, throughout the entire diagnostic process, the weights of the neural network will be optimized using the Kepler algorithm. The fault diagnosis model used in this article has a structure as shown in Fig 8:
4. Experimental analysis
4.1. Experimental data
The most objective way to evaluate the superiority of the algorithm proposed in this article is to use a third-party standard database [35]and compare the prediction results of this algorithm with current mainstream algorithms. This article selects rolling bearing data from Xi’an Jiaotong University, and the fault data of the bearings is obtained through artificial damage setting and accelerated experiments [36]. The experimental collection platform is shown in Fig 9. The experimental equipment includes an AC motor, a motor speed controller, a motor speed controller, two support bearings (heavy-duty load roller bearings), a hydraulic loading system, etc. During the acceleration testing experiment, three different operating conditions were set, and five bearings were tested under each operating condition. The operating conditions set are as follows:
- (1). 2100 revolutions per minute (35 Hz) and 12 kilonewtons;
- (2). 2250 revolutions per minute (37.5 Hz) and 11 kilonewtons;
- (3). 2400 revolutions per minute (40 Hz) and 10 kilonewtons.
The final collected experimental results are as follows, with a total of 9 experimental data states. When diagnosing bearing faults, undamaged normal bearing data was added, with a total of 10 data states. The normal bearing data was named category 10, and the naming of the bearing data is detailed in Table 1 below. The damage to the bearings after the experiment is shown in Figs 10–12.
4.2. Model parameter
1. Data preprocessing.
All input features are normalized using [min max normalization] before training, and reduced to the [0,1] interval.
Samples with missing values have been removed to ensure data integrity.
All classification labels have been encoded with integers (such as category labels encoded as 1, 2, 3, etc.).
2. Training/testing set partitioning.
The original dataset is randomly divided into a training set and a testing set, with a segmentation ratio of 80%/20%.
3. Neural network training parameters.
Maximum number of iterations (Epochs): 500
Batch size: 32 samples per batch
This model adopts a double-layer one-dimensional convolution structure (convolution kernel size of 3x1, channel numbers of 32 and 64 respectively), integrates SE attention mechanism (channel compression ratio of 1/4), and is combined with single-layer LSTM (unit number of 6) and self attention module. The final classification output is achieved through fully connected layers and Softmax function. During training, the Adam optimizer is used with an initial learning rate of 0.01, which decays to 0.1 after every 100 rounds of training. The L2 regularization coefficient is 0.01, and the maximum training is 500 rounds. Before each round of training, the sample order will be randomly shuffled.
Input the fault data of the bearings into the model in this article, and the output structure size of each layer of the neural network is shown in Table 2 below:
4.3. Result analysis
To verify the effectiveness of the algorithm, a comparison was made between the model with Kepler algorithm and the model without Kepler algorithm to verify that Kepler algorithm can improve the accuracy of the model. At the same time, compare the models with and without attention mechanism to verify that the attention mechanism can extract important information features and improve the accuracy of the model.
This article uses four algorithms, namely KOA LSTM Attention (Algorithm 1), LSTM Attention (Algorithm 2), KOA LSTM (Algorithm 3), and LSTM (Algorithm 4), to predict fault classification results. The results are compared and summarized in Figs 13 and 14.
The accuracy of the four algorithms is 98.32%, 95.38%, 97.47%, and 94.12%, respectively. Algorithm 1 has a 2.94% higher accuracy than Algorithm 2, and Algorithm 3 has a 3.35% higher accuracy than Algorithm 4. The comparison of the two sets of results fully demonstrates that Kepler algorithm can optimize the weights of neural networks and improve the accuracy of the model. The accuracy of
Algorithm 1 is 0.85% higher than that of Algorithm 3, and the accuracy of Algorithm 2 is 1.26% higher than that of Algorithm 4. The comparison results of the two sets of data fully demonstrate that the attention mechanism can effectively process massive duplicate data, improve the model’s focus, and enhance the accuracy of the improved model.
To further analyze the performance of each model, we calculated the F1 score of four algorithms in 10 independent experiments, and the results are shown in Table 3. And a box plot (F1 macro Boxplot) and a mean ± standard deviation bar chart (F1 macro Mean ± Std) were plotted in Fig 15. From the box plot, it can be seen that the F1 macro distribution of the KOA-LSTM Attention model is generally higher than other models, with smaller variance and more stable performance. The mean bar chart also shows that the model has the highest mean, and the difference is statistically significant.
Meanwhile, based on the prediction results of the test sets of the four algorithms shown in Fig 16, it can be concluded that the KOA-LSTM Attention algorithm has the fastest prediction results approaching the true values, which verifies the superiority of the proposed algorithm in this paper.
Meanwhile,independent experiments and performance comparisons were conducted for all three scenarios mentioned above, quantifying the respective contributions of KOA and attention mechanisms. The experimental results are shown in Table 4:
Due to the fact that the innovation of this article mainly focuses on the joint design of KOA mechanism and attention mechanism, the analysis of ablation experiments highlights the role of these two core parts in improving model performance and the comprehensive gain brought by their mutual cooperation.
The confusion matrices of the four algorithms are shown in Fig 17
The precision, recall, and F1 score data for each category are shown in Table 5
Through the F1 scores of each category, it can be seen that some small sample categories or easily mixed categories (such as categories 1 and 8) have lower scores, mainly due to uneven distribution of categories or similar features. With the improvement of the model, the accuracy and F1 of all categories have improved, especially KOA-LSTM Attention, which performs the best in each category and has the strongest overall performance. Combining the confusion matrix, further identify the main sources of misjudgment for certain categories and guide subsequent model optimization.
To further validate the feature extraction capability of the model, we have added t-SNE visualizations for each model’s feature extraction in Fig 18, which intuitively demonstrate the performance of each model in category differentiation from the perspective of dimensionality reduction distribution.
5. Conclusion
This article addresses the two major pain points of LSTM networks by applying Kepler algorithm and combined attention mechanism to solve their weight and attention problems. The optimized model is applied to the fault diagnosis of bearings, achieving efficient fault diagnosis. Finally, the accuracy of the four algorithms is compared to verify the feasibility and superiority of the algorithm proposed in this article.
References
- 1.
Sun W e i f e n g. Design and implementation of rolling bearing vibration signal analysis and data processing system. Harbin Institute of Technology. 2023.
- 2. Thorsen OV, Dalva M. Failure identification and analysis for high-voltage induction motors in the petrochemical industry. IEEE Trans on Ind Applicat. 1999;35(4):810–8.
- 3. Report IC. Report of large motor reliability survey of industrial and commercial installations. Industry Applications IEEE Transactions on. 1987;ia–23(4):153–8.
- 4. Glowacz A, Glowacz W, Glowacz Z, Kozik J. Early fault diagnosis of bearing and stator faults of the single-phase induction motor using acoustic signals. Measurement. 2018;113:1–9.
- 5. Anand R, Rath A, Sahoo PK, Jain P, Panda G, Wang X, et al. Winograd Transform-Based Fast Detection of Heart Disease Using ECG Signals and Chest X-Ray Images. IEEE Access. 2025;13:57119–40.
- 6. Sahu B, Panigrahi A, Pati A, Das MN, Jain P, Sahoo G, et al. Novel Hybrid Feature Selection Using Binary Portia Spider Optimization Algorithm and Fast mRMR. Bioengineering (Basel). 2025;12(3):291. pmid:40150755
- 7.
Song B, Tan S, Shi H, Zhao B. Fault detection and diagnosis via standardized k nearest neighbor for multimode process. 2020.
- 8. Chen F, Cheng M, Tang B, Chen B, Xiao W. Pattern recognition of a sensitive feature set based on the orthogonal neighborhood preserving embedding and adaboost_SVM algorithm for rolling bearing early fault diagnosis. Meas Sci Technol. 2020;31(10):105007.
- 9. Panda P, Bisoy SK, Panigrahi A, Pati A, Sahu B, Guo Z, et al. BIMSSA: enhancing cancer prediction with salp swarm optimization and ensemble machine learning approaches. Front Genet. 2025;15:1491602. pmid:39834551
- 10. Rex FMT, Andrews A, Krishnakumari A, Hariharasakthisudhan P. A hybrid approach for fault diagnosis of spur gears using hu invariant moments and artificial neural networks. Metrology and Measurement Systems: Metrologia I Systemy Pomiarowe. 2020;3:27.
- 11. Houkan A, Sahoo AK, Gochhayat SP, Sahoo PK, Liu H, Khalid SG, et al. Enhancing Security in Industrial IoT Networks: Machine Learning Solutions for Feature Selection and Reduction. IEEE Access. 2024;12:160864–83.
- 12. Yang J, Li X, Jiang Y, Qiu G, Buckdahn S. Target recognition system of dynamic scene based on artificial intelligence vision. J Intelligent Fuzzy Systems. 2018;35(4):4373–83.
- 13. Mazher Iqbal JL., Senthil Kumar M, Mishra G, Asha G R, Saritha A N, Ramesh JVN, et al. Facial emotion recognition using geometrical features based deep learning techniques. Int J Comput Commun Control. 2023;18(4).
- 14. Kumar MS, Dipankar D, Sivaji B. Machine translation using recurrent neural network on statistical machine translation. J Intelligent Syst. 2018.
- 15. He D, Xu Y, Jin Z, Liu Q, Zhao M, Chen Y. A zero-shot model for diagnosing unknown composite faults in train bearings based on label feature vector generated fault features. Appl Acoustics. 2025;232:110563.
- 16. He D, Wu J, Jin Z, Huang C, Wei Z, Yi C. AGFCN:A bearing fault diagnosis method for high-speed train bogie under complex working conditions. Reliability Engineering System Safety. 2025;258:110907.
- 17. Wan A n p i n g, Yang J i e, Wang J i n g l i n, Dan T i a n m i n, Miao X u, Huang J i a y o n g. Multi sensor fusion convolutional neural network fault diagnosis method for aircraft engine bearings. Chinese J Electrical Engineering. 2022;42(13):9.
- 18. Xiong X, Wang J, Zhang Y, Guo Q, Zong S. A two-dimensional convolutional neural network optimization method for bearing fault diagnosis. Chinese J Electrical Eng. 2019;39(15):10.
- 19. Lu X, Zhang C, Gao J, Xu Y, Shao X. A bearing fault diagnosis algorithm based on convolutional neural network and catboost. Mechanical Electrical Eng. 2023;40(5):715–22.
- 20. Jin Z, Zhang X, Zhang Y, Zhang K. Rolling bearing fault diagnosis using Welch power spectrum combined with convolutional neural network. Mechanical Design Manufacturing. 2024;2:271–5.
- 21. Qu J, Yu L, Yuan T, Tian Y, Gao F. Adaptive fault diagnosis algorithm for rolling bearings based on one-dimensional convolutional neural network. J Instrumentation. 2018;39(7):10.
- 22. Chang M, Shen Y a n x i a. Wind turbine bearing fault diagnosis strategy based on improved convolutional neural network. Power System Protect Control. 2021;49(6):131–7.
- 23. Houssein EH, Abdalkarim N, Samee NA, Alabdulhafith M, Mohamed E. Improved Kepler Optimization Algorithm for enhanced feature selection in liver disease classification. Knowledge-Based Systems. 2024;297:111960.
- 24. El Ghouate N, Bencherqui A, Mansouri H, Maloufy AE, Tahiri MA, Karmouni H, et al. Improving the Kepler optimization algorithm with chaotic maps: comprehensive performance evaluation and engineering applications. Artif Intell Rev. 2024;57(11).
- 25. Abdel-Basset M, Mohamed R, Azeem SAA, Jameel M, Abouhawwash M. Kepler optimization algorithm: A new metaheuristic algorithm inspired by Kepler’s laws of planetary motion. Knowledge-Based Systems. 2023;268:110454.
- 26. Yao J, Yang J, Zhang C, Zhang J, Zhang T. Autonomous Underwater Vehicle Trajectory Prediction with the Nonlinear Kepler Optimization Algorithm–Bidirectional Long Short-Term Memory–Time-Variable Attention Model. JMSE. 2024;12(7):1115.
- 27. Cai L, Zhao S, Meng F, Zhang T. Adaptive K-NN metric classification based on improved Kepler optimization algorithm. J Supercomput. 2024;81(1).
- 28.
Suganuma M, Shirakawa S, Nagao T. A genetic programming approach to designing convolutional neural network architectures. 2017.
- 29.
Gao Z. Stock Price Prediction Based on Joint LSTM and Fully Co-nnected Layer. In: International Conference on Computational Finance and Business Analytics. Cham: Springer; 2024.
- 30. Chegeni MK, Rashno A, Fadaei S. Convolution-layer parameters optimization in Convolutional Neural Networks. Knowledge-Based Systems. 2023;261:110210.
- 31. Zhang Y, Huang Y. Research on malicious software classification and recognition based on machine learning. Technology News. 2018;16(30):3.
- 32. Wang Y, Yan J, Ye X, Qi Z, Wang J, Geng Y. GIS partial discharge pattern recognition via a novel capsule deep graph convolutional network. IET Generation Trans & Dist. 2022;16(14):2903–12.
- 33. Zhang R, Zhu Z, Yuan M, Guo Y, Song J, Shi X, et al. Regional Residential Short-Term Load-Interval Forecasting Based on SSA-LSTM and Load Consumption Consistency Analysis. Energies. 2023;16(24):8062.
- 34. Xiao H, Fu L, Shang C, Bao X, Xu X, Guo W. Ship energy scheduling with DQN-CE algorithm combining bi-directional LSTM and attention mechanism. Appl Energy. 2023;347:121378.
- 35. Kingdom L, Tianyu H, Biao W, Naipeng L, Tao Y, Jun Y. Interpretation of Xjtu sy rolling bearing accelerated life test dataset. J Mechanical Eng. 2019;16(6).
- 36. Wang B, Lei Y, Li N, Li N. A Hybrid Prognostics Approach for Estimating Remaining Useful Life of Rolling Element Bearings. IEEE Trans Rel. 2020;69(1):401–12.