Figures
Abstract
Understanding and controlling grain growth kinetics in steels is crucial for optimizing mechanical properties during thermomechanical processing. However, traditional empirical models often fail to account for the complex, nonlinear interactions between alloying elements and processing parameters. In this study, we introduce a novel machine learning (ML) based framework that predicts austenitic grain growth behaviour directly from chemical composition and process conditions, utilizing a comprehensive dataset of 1039 experimentally validated samples. Among various algorithms tested, the XGBoost model demonstrated exceptional predictive capability, achieving an R2 value of 0.9728 after hyperparameter optimization. Feature selection methods (Pearson correlation, CfsSubset, ReliefF) and SHAP-based explainable AI analyses were employed to identify the most influential parameters, revealing temperature, initial grain size, and holding time as dominant factors. Experimental validation was conducted on 316L stainless steel samples annealed at 1100 °C. The predicted grain sizes showed strong agreement with experimental measurements, and the observed hardness variations followed the expected Hall–Petch behaviour. This study demonstrates the first integrated ML and experimental approach for predicting grain growth kinetics in steels, offering a powerful tool for alloy design and process optimization. Future work will extend this framework to additional process variables and alloy systems.
Citation: Demirci S, Şahin DÖ, Demirci S, Tünçay MM, Attallah MM (2026) Predicting grain growth kinetic in steels using machine learning and XAI for mechanical properties. PLoS One 21(1): e0341053. https://doi.org/10.1371/journal.pone.0341053
Editor: Arun Devaraj, Pacific Northwest National Laboratory, UNITED STATES OF AMERICA
Received: June 12, 2025; Accepted: December 31, 2025; Published: January 16, 2026
Copyright: © 2026 Demirci et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The design strategy mentioned in the paper was implemented in Python. Source codes, datasets, and algorithms are available in https://github.com/DurmusSahin/GrainGrowthStell.
Funding: The author(s) received no specific funding for this work.
Competing interests: The authors have declared that no competing interests exist.
1. Introduction
Steel has long been a fundamental material in engineering and industrial applications due to its excellent mechanical properties, affordability, and versatility [1–3]. As an alloy primarily composed of iron and carbon, steel often includes additional elements such as manganese, chromium, nickel, and vanadium to enhance properties like strength, toughness, corrosion resistance, and thermal stability [4–6]. The mechanical behaviour of steel varies significantly based on its composition and processing conditions, making it indispensable in industries such as automotive, aerospace, construction, and energy [7–11]. Given its wide range of applications, optimizing steel’s mechanical properties is crucial. These properties can be tailored through various strengthening mechanisms, including solid solution strengthening and thermomechanical processes such as heat treatment, cold working, and annealing [12]. Among these factors, grain size plays a key role in determining steel’s mechanical strength. The Hall-Petch relationship demonstrates that smaller grain sizes increase strength due to the increased number of grain boundaries acting as obstacles to dislocation motion [13]. However, grain growth is an inevitable process in metals, as polycrystalline materials naturally seek to reduce their total grain boundary energy through coarsening. In steels, this phenomenon is often associated with the growth of the austenite (γ) phase grains, which can significantly influence the final microstructure and mechanical performance. If left uncontrolled, excessive grain growth can lead to a decline in mechanical strength and overall performance. During the cooling process, austenite grain boundaries serve as nucleation sites for phases such as ferrite, cementite, bainite, and martensite. When the austenite grain size is small, the number of nucleation sites increases, leading to a finer grain structure in the final phase [14]. Therefore, a thorough understanding of grain growth kinetics particularly in the austenite phase is essential for optimizing the microstructure of steel and maintaining its desired mechanical properties [15,16]. The rate of austenite grain growth depends on factors such as temperature, time, initial grain size, and the presence of solute elements and second-phase particles that inhibit boundary motion. The kinetics of austenitic grain growth are generally described by the empirical Equation (1) for isothermal grain growth, which defines the evolution of grain size during isothermal heating [17,18].
where D represents the average grain diameter, D0 is the initial grain diameter, t denotes the soaking time at a given temperature, n and K are the growth exponent and rate constant, respectively. In addition to empirical formulations, the evolution of grain size in steels has been extensively studied through classical and computational approaches. Traditional theoretical models, such as the Burke–Turnbull parabolic growth law and Hillert’s mean-field theory, describe grain coarsening through curvature-driven grain boundary motion and provide the foundation for understanding thermal grain growth behaviour [19,20]. More advanced numerical techniques including Monte Carlo Potts simulations and phase-field modelling enable the description of complex mechanisms such as abnormal grain growth, solute drag, and particle-boundary interactions that cannot be captured by simple analytical equations [21,22]. These modelling approaches collectively form the basis of modern grain evolution theory and highlight the multi-scale nature of grain growth phenomena in steels. The growing field of data-driven microstructure prediction further builds on these foundations, motivating the development of machine learning (ML) based strategies for modelling grain evolution under diverse processing conditions. There have been various methods to control grain size in steel, including alloying strategies and advanced thermomechanical processing techniques [23]. One of the method is utilization microalloying elements like Nb, Ti, and V form fine precipitates like carbides, nitrides, and carbonitrides acting as effective grain boundary pinning agents that inhibit grain growth [24]. For instance, Zhu et al. [25] investigated the austenite grain growth for Nb and Al microalloyed 20MnCr steels. They found that NbC precipitates inhibit grain growth by stabilising grain boundaries more effectively than AlN precipitates. Xu et al. [26] investigated the effect of heating temperature and dwell time on austenite grain growth in SCM435 steel. A new model was proposed to accurately predict the average austenite grain size of SCM435 steel during heating, showing strong consistency with experimental data and improving upon the limitations of the classical Sellars model. Despite extensive experimental studies, predicting austenite grain growth behaviour remains challenging due to the complex interplay of composition, processing parameters, and microstructural interactions. Traditional empirical models often fail to capture these intricate dependencies, necessitating extensive experimental calibration for each steel grade. In this context, ML presents a powerful tool for uncovering hidden patterns within large experimental datasets and predicting microstructural evolution with high accuracy [27,28]. The goal of the ML model is to establish a functional relationship between inputs and outputs, represented as y = fmodel (xi). Various algorithms are employed to optimize the model’s parameters, ensuring that the predicted output fmodel (xi) closely approximates the true value [29,30]. By integrating experimental data with ML models, researchers can establish quantitative relationships between alloying elements, heat treatment parameters, and resulting grain structures. This data-driven approach has the potential to revolutionize steel processing by minimizing trial-and-error experimentation and optimizing manufacturing techniques for improved performance. To the best of our knowledge, there is no research on utilizing ML for predicting grain growth kinetics in steel based on thermomechanical processing conditions and alloy composition. Unlike previous studies that primarily focused on empirical models or mechanical property predictions, this work presents the first comprehensive ML-based framework aimed specifically at predicting austenitic grain growth kinetics in steels. By combining an extensive dataset covering a wide range of chemical compositions and thermomechanical parameters with advanced feature selection techniques and explainable artificial intelligence (XAI) methods, this study establishes a direct link between input parameters and grain size evolution. Furthermore, the integration of experimental validation through 316L stainless steel annealing tests ensures the robustness and real-world applicability of the proposed model. The results demonstrate that ML, particularly the XGBoost algorithm, can achieve highly accurate predictions (R2 = 0.9728), providing a transformative tool for accelerating steel design and optimizing processing routes with minimal experimental effort.
2. Methodology and material
This research focuses on uncovering the relationship between the chemical composition of steels and the thermomechanical processing conditions that govern grain growth kinetics and final grain size. To achieve this, a machine learning-based framework was developed, encompassing four sequential phases: comprehensive data acquisition, data preprocessing, model training, and rigorous performance validation, as illustrated in Fig 1.
2.1. Data acquisition
For the purposes of this study, the dataset was obtained by conducting a thorough survey of existing experimental literature, regarding steels with a wide range of compositions and processing conditions. The data was manually gathered from well-established scientific platforms, including but not limited to Elsevier, Springer, and Wiley [1,2,16,17,23,31–67]. The dataset consists of 1039 samples in total and comprehensive statistical information in terms of both inputs (features) and outputs is given in Table 1. Input variables can be divided into two main groups: chemical compositions of steel (e.g., Fe, C, Cr, Mn, Si, etc.) and thermomechanical parameters such as temperature (°C), holding time (h), initial grain size (D0, μm), final grain size (D, μm), growth exponent (n), rate constant (k) and % cold rolling (R). The final grain size (μm) is denoted as an output in the problem. The output variable D exhibits a wide distribution between 3.99 μm and 990 μm. The wide range of the output value D indicates that the problem could be quite difficult. This problem may require the consideration of multidimensional and potentially non-linear relationships. In general, the dataset is diverse in both chemical and physical parameters and allows for a statistically meaningful analysis.
2.2. Data preprocessing
In modelling the grain size of steels using machine learning, it is essential to account for a wide range of variables, including both chemical compositions and thermomechanical processing parameters such as temperature, time, and deformation. These variables are often collected from diverse sources and may contain missing values, inconsistencies, or differences in scale and units. Feeding such raw data directly into a model can lead to incorrect learning patterns or physically invalid predictions. Moreover, not all input variables have a significant influence on the final microstructure, and including irrelevant features may reduce model performance. Therefore, the data preprocessing comprising data cleaning, normalization and unit consistency is a critical step to ensure that the model produces reliable, interpretable, and scientifically meaningful results. This process directly enhances the generalizability and accuracy of data-driven predictions in materials science applications. In this context, normalization is applied to improve model robustness.
2.2.1. Normalization.
Normalization is a crucial preprocessing step for ensuring balanced model training, especially when input variables have different scales. In this study, Min-Max normalization was applied to rescale all features between 0 and 1, as shown in Equation (2) [68–70].
where is the actual value, and
and
are the minimum and maximum values of the feature, respectively. Normalization prevents models from being biased toward variables with larger magnitudes and enhances the performance of especially distance-based algorithms. This step is essential given the wide range of parameters such as temperature, holding time, and alloying element concentrations in the dataset.
2.3. Data splitting
After the normalisation process, the dataset was divided into 80% training and 20% testing in order to provide an objective evaluation of the model performance. This distinction makes it possible for ML algorithms to use only training data in the learning process and then evaluate the generalisation ability of the modelled structure on test data. In order to ensure that all algorithms operate on the same data split and to increase the comparability of the results, the random_state parameter is fixed in the data splitting process. Fixing this parameter guarantees experimental reproducibility by preventing variations due to randomness. Thus, by training and testing different algorithms on the same samples, the reliability of the performance metrics obtained and the consistency of the comparative analyses are increased.
2.4. Machine learning (ML) algorithms
In this study, various supervised ML algorithms are applied to predict the dependent variable D in the dataset. The methods used consist of both parametric and non-parametric algorithms, representing different modelling approaches. In this context, Adaboost [71], Decision Tree (DT) [72], Gaussian Process Regression (GPR) [73], K-Nearest Neighbors (KNN) [74], Linear Regression (LR) [75], Multi-Layer Perceptron (MLP) [76], Random Forest (RF) [77], Support Vector Regression (SVR) [78] and XGBoost [79] algorithms were used through the scikit-learn library. The selected ML algorithms were applied with default hyperparameter settings to obtain baseline model performances. This baseline model is intended to observe how the algorithms interact with the data under the default configurations. However, it is very useful to optimise the hyperparameter settings to improve model performance and to obtain better generalisation capability. Within this context, a systematic hyperparameter search was performed for each algorithm using the GridSearchCV method in the scikit-learn library. GridSearchCV selects the parameter set that provides the best performance by evaluating the specified parameter combinations with the cross-validation method. Thus, the configuration of each model is tailored specifically to the data set, reducing the risk of overfitting and providing more balanced and effective prediction models. Table 2 shows the hyperparameters obtained with the GridSearchCV approach.
2.5. Feature selection
In this study, feature selection was performed to enhance model performance, reduce computational cost, and improve interpretability. Three different methods were applied: Pearson correlation coefficient, CfsSubset, and ReliefF [80,81]. Pearson correlation coefficient, r, measures the degree and direction of the linear relationship between two variables and is defined as given in Equation 3.
Variables with strong correlations to the target were selected, while weakly correlated or irrelevant features were excluded. The CfsSubset method evaluates subsets of features based on their predictive ability and redundancy, aiming to select inputs that are highly correlated with the target but minimally correlated with each other [82]. Its fitness criterion is shown in Equation (4).
where k is the number of features, is the average feature-target correlation, and
is the average feature-feature correlation. Finally, the ReliefF algorithm, particularly effective for complex, nonlinear problems, assigns weights to features based on their ability to differentiate instances with similar or different outputs [83]. The weight update rule is given in Equation (5) [84].
where is the weight of feature A, H and M are the nearest neighbours from the same and different classes (adapted for continuous targets), and mmm is the number of random samples. These three methods collectively provide complementary insights, improving feature selection robustness for predicting grain growth kinetics.
2.6. Evaluation of metrics
The models are assessed using three distinct evaluation metrics: Mean Absolute Error (MAE), Root Mean Square Error (RMSE), and R-squared (R2) in the following equations [70].
Here ,
,
and n represent the actual value, the predicted value, mean value, and the number of observations, respectively.
2.7. Experimental studies
In this study, the austenitic stainless steel (316) is used to validate ML prediction. The chemical composition of the steels is given in the Table 3.
The steels sample were heat treated at 1100 °C for 30, 45, and 60 minutes and cooled in air in order to observe grain growth. As-received and heat-treated samples were metallographically prepared by sequential grinding with SiC papers of 180, 240, 320, 400, 600, 800, 1000, and 1200 grit, followed by final surface polishing using a 0.05 µm alumina suspension. Then, all samples were cleaned in an ultrasonic bath with ethanol for five minutes. The etching of the austenitic stainless steel was performed by using Aqua Regia. The observation of the microstructures was carried out with an optical microscope (Olympus). The optical micrographs, captured at 20 × magnification, were analysed with ImageJ software to determine average grain size for 316 austenitic stainless steels. The hardness of each sample was measured utilizing a microhardness tester (FutureTech FM-ARS7000) using 100 g of load. The results were reported as an average of five measurements.
3. Results and discussions
3.1. Performances of ML algorithms
In this section, the performances of ML algorithms on the prediction of the output value are analysed and discussed. After the normalisation process is applied to the dataset, the dataset is divided into 80% training and 20% testing in order to provide an objective evaluation of the model performance. Firstly, the selected ML algorithms were applied with default hyperparameter settings and baseline model performances were obtained. Table 4 shows the results obtained using the predefined parameters of the ML algorithms in the sklearn library.
According to the modelling results with predefined parameters, significant differences were observed in the prediction performance of different ML algorithms on the test data as given in Table 4. The highest performance was achieved by the XGBoost algorithm, which, with R2 = 0.9605, RMSE = 26.39 and MAE = 12.77. Following XGBoost, the RF algorithm also provided high prediction. The RF algorithm showed a remarkable success with R2 = 0.9424, RMSE = 31.89 and MAE = 14.03. These results show that ensemble-based methods (especially tree-based models) are suitable for the considered problem and have strong predictive capacities. However, classical regression and distance-based algorithms showed more limited performance. For example, although the LR model has a reasonable explanatory power with R2 = 0.8056, the RMSE and MAE values (58.61 and 39.87, respectively) are higher, indicating that the model does not adequately capture complex patterns. Similarly, SVR and GPR algorithms also showed limited performance with relatively low R2 values (0.7623 and 0.7682) and high error metric values. In this context, ensemble and network-based algorithms, which can better model non-linear relationships and input interactions, stand out. However, it is very useful to optimise the hyperparameter settings to improve model performance and to obtain better generalisation capability. Within this context, a systematic hyperparameter search was performed for each algorithm using the GridSearchCV method in the scikit-learn library. Table 5 shows the results obtained using the GridSearchCV method in the scikit-learn library. As a result of the hyperparameter optimisation performed with the GridSearchCV method, a general improvement in the performance of ML models was observed.
In particular, the XGBoost algorithm showed a significant performance improvement after hyperparameter adjustments, reaching R2 = 0.9728, RMSE = 21.89 and MAE = 10.97. These values indicate that the model is able to predict the output variable D with a very high accuracy and the error level is minimised. Similarly, the DT algorithm has also improved compared to its previous state with R2 = 0.9228, RMSE = 36.93 and MAE = 18.87 after optimisation. In contrast, some algorithms showed limited or no performance improvement. For example, performance metrics did not change despite hyperparameter optimisation in the GPR and LR models. Moreover, the MLP algorithm improved significantly after GridSearchCV, reaching R2 = 0.8632, indicating that the learning capacity of the model can be significantly increased with appropriate hyperparameters. In general, it has been shown that hyperparameter optimisation with GridSearchCV provides significant performance improvements especially in complex and parametric models and is a critical step in the model selection process. This reveals that algorithms can become significantly more efficient when configured with the right combinations of hyperparameters.
Fig 2 shows a boxplot plot to observe the absolute error between the prediction values of ML algorithms and the actual output values. The boxplot plot, which was created to examine the prediction performances obtained after hyperparameter optimisation in more detail, visually presents the absolute error distributions of ML algorithms in a comparative manner. In the Fig 2, especially the XGBoost algorithm has a low median error value and narrow interquartile range, indicating that this model produces forecasts with high accuracy and the error distribution is stable. Similarly, RF, DT and KNN models also show strong performance with relatively low error levels and limited outliers. On the other hand, the high number and value of outliers observed in the GPR algorithm shows that this model makes quite large errors in some cases and its predictions are not consistent. Overall, the findings in the Fig 2 are consistent with the numerical performance metrics and confirm that the XGBoost model is the most successful model in terms of both error level and stability of the error distribution.
Fig 3 shows the regression plots of the 4 most successful algorithms (XGBoost, RF, DT and KNN). Regression plots are very useful for comparing the accuracy of the models by visually presenting the relationship between predicted values and actual values. The four most successful algorithms, which are XGBoost, RF, DT and KNN models, were evaluated on the test data. The orange dots in the graphs represent the predictions and the red dashed line represents the ideal prediction state (actual = predicted). The XGBoost model showed a superior performance, with the point forecasts showing a distribution very close to the actual values and reaching a high coefficient of determination of R2 = 0.9728. The RF and DT models also have high explanatory power, the density of the predictions is distributed close to the linear. Although the KNN model exhibited a more dispersed structure than the other three models, it provided a high accuracy with R2 = 0.9029 and demonstrated reasonable generalisation success of the model. Overall, these graphs effectively visualise both the model performance and the error distribution, confirming that XGBoost performs well on this problem.
3.2. Effect of feature selection on the ML performances
Three different feature methods are applied to determine which inputs are more important for the regression problem. Firstly, the Pearson correlation analysis is performed to determine the linear relationships between the parameters analysed in the study. This method is preferred to quantitatively evaluate the direction and strength of the linear relationship between two continuous variables. The heat map given in Fig 4 visually presents the Pearson correlation coefficients between the variables in the dataset. Positive values of the correlation coefficient close to +1 indicate a strong linear relationship, while negative values close to −1 indicate a strong inverse relationship. Correlation coefficient values close to 0 indicate that there is no significant linear relationship between the variables.
Fig 4 shows the Pearson correlations between all variables and their relationship with the final grain size (D). As expected, the initial grain size (D0) has the strongest positive correlation with D (r = 0.76). This result clearly demonstrates that the final grain size of the initial microstructure is the most important determinant. In contrast, the holding time (h) (r = 0.18) and temperature (T) (r = 0.47) show only weak to moderate correlations with D. Therefore, these two variables alone should not be expected to strongly explain grain size; however, they still make measurable contributions to growth behaviour. The effect of chemical composition on D is generally weak. Among the alloying elements, tungsten (W), chromium (Cr), and molybdenum (Mo) have the highest positive correlation coefficients; however, these values are also quite low (r = 0.26, 0.17, and 0.15). Nevertheless, these trends are significant from a metallurgical perspective: W can increase grain boundary mobility by affecting grain boundary energy and diffusion, while Mo can alter grain growth at high temperatures through precipitation and solute drag effects [85–87]. However, low correlation coefficients indicate that the effects of these elements are not decisive on their own, but rather occur in conjunction with other variables. Although slight negative correlations were observed for iron (Fe), sulfur (S), and growth base (n), the very small values of these correlations (|r| < 0.1) indicate that their effects are negligible [88]. In general, it is understood that heating conditions, duration, and chemical composition must be evaluated together, and that grain size cannot be explained by a single parameter.
Fig 5 presents the ranking of the importance of the inputs in terms of their effect on the final grain size as a result of correlation analysis and ReliefF algorithm.
(a) Features ranked by correlation coefficient values. (b) Feature importance ranking obtained using the ReliefF algorithm.
The first 6 inputs with the highest correlation coefficient values and 3 inputs with the lowest correlation coefficient values were selected, totalling 9 inputs as seen in Fig 5(a). This input subset consists of μm, T, W, h, Cr, Mo, Co, Fe, S and n. According to the ReliefF algorithm, the input importance ranking is given in Fig 5(b). According to the ReliefF algorithm, the number of nearest neighbours (k) value was found to be 10. A threshold value greater than 0.0002 was determined on the values given in Fig 5(b) and input subset selection was performed. As a result, 10 inputs were selected by ReliefF algorithm. These are μm, T, h, P, lnk, S, n, C, Ni and Ti. As for CfsSubset technique, merit of best subset was found as to be 0.777. Seven features were automatically selected by CfsSubset. These attributes are T, h, μm, Fe, V, P and W. Table 6 shows the regression results obtained using feature selection methods. The performance results obtained after feature selection reveal that the inclusion of certain variables in the model leads to significant improvements in prediction accuracy. As a result of modelling using input sets selected by correlation coefficient, CfsSubset and ReliefF methods, the highest performance was generally obtained by the XGBoost algorithm. In particular, modelling with CfsSubset method using 7 inputs yielded R2 = 0.975, RMSE = 21.00 and MAE = 10.59, indicating that even a small number of carefully selected inputs can provide high accuracy. This has the advantage of reducing the complexity of the model, reducing the computational cost and preventing overlearning.
The RF and DT algorithms have also largely maintained their high performance after feature selection, and in some cases even improved it. On the other hand, the MLP algorithm performed worse than all three methods, indicating that MLP is more sensitive to the selected input combinations. In general, the ReliefF method provided more balanced results, especially in terms of DT and MLP algorithms. The findings show that model performance can be improved with appropriate feature selection, and this effect becomes even more pronounced when used with powerful algorithms such as XGBoost. In this context, feature selection makes an important contribution to both model optimisation and interpretability. When the performance of the KNN algorithm after feature selection is analysed, certain differences are observed in the results depending on the methods used. The R2 = 0.9087, RMSE = 40.15 and MAE = 18.98 values obtained with 9 inputs selected by the correlation coefficient method reveal that the KNN algorithm shows a very successful prediction performance. These results show that KNN can be a highly powerful local prediction model when fed only with meaningful inputs. However, when using the 7 inputs selected with the CfsSubset method, a decrease to R2 = 0.8842 was observed and the error metrics also deteriorated significantly (RMSE = 45.23; MAE = 25.14). This shows that the inputs selected with the CfsSubset method are not sufficiently compatible with the distance-based structure of the KNN model. On the other hand, achieving R2 = 0.902 with 10 inputs selected with the ReliefF method and maintaining a successful prediction performance with low error metrics (RMSE = 41.61, MAE = 20.34) indicate that this method provides more suitable input combinations for KNN model. In general, it can be said that the KNN algorithm is sensitive to feature selection and especially the correlation coefficient and ReliefF based selections positively affect the performance of this algorithm. This is evidenced by the fact that results close to these results are obtained with the KNN algorithm without feature selection.
3.3. Interpreting results with explainable artificial intelligence (XAI)
In this subsection, the SHAP analysis was performed on XGBoost since XGBoost was the most successful algorithm in the experiments. This analysis not only increases the explainability of the model but also reveals which inputs are really effective in a concrete way. Figs 6 and 7 demonstrate the graphs obtained from the SHAP analysis. These graphs show in detail the input effects of the XGBoost model on the output variable D at both global and local level. The bar graph in Fig 6 shows the average contribution (mean |SHAP| value) of the inputs to the model.
According to Fig 6, the three most effective inputs are T, μm and h, respectively. These three physical parameters are the variables that make the highest contribution to the output of the model on average, and it is understood that the model considers these factors primarily on the output variable. In particular, T and μm have very close SHAP contribution values, indicating that these two variables are effective with similar weights. The fact that chemical and structural properties such as lnk, n, Nb, Fe and Cr are among the following variables reveals that not only physical but also alloy composition and microstructure characteristics make a significant contribution to the model. The beeswarm plot given in Fig 7 reflects the direction and intensity of the input values on the prediction for each sample. The graph in Fig 7 visualises not only the magnitude but also the direction of the effect of the inputs on the model.
Here, pink tones represent high values and blue tones represent low values. For example, high values of variables such as T and μm generally have a positive effect on the forecast, while low values have a negative effect as seen in the Fig 7. This shows that the model tends to increase the D output with higher values of T and μm. Similarly, higher values of alloying elements such as Nb and Cr also contribute positively to the model, but their effects are more limited compared to the physical parameters. As a result, SHAP analyses reveal in detail not only which inputs are important but also how these inputs shape model decisions. In this respect, it is seen that the XAI approach is an indispensable component for model interpretability as well as model accuracy. The findings support that the data preprocessing and engineering information are selected correctly and the model makes decisions in accordance with the physical reality. When the results obtained by feature selection methods and SHAP-based explainability analysis are evaluated together, significant overlaps are observed between the variables that the model inputs importance in the learning process and the selected inputs. In particular, the μm, T and h inputs were selected by both Correlation Coefficient, CfsSubset and ReliefF methods. Also, the μm, T and h inputs stood out as the top three inputs that contributed the most to the SHAP analysis. This strongly confirms that these physical parameters have a determining effect on the output variable D of the model. Moreover, the inclusion of variables such as Fe, W, n, P, and S in the model through different feature selection approaches, along with their notable contributions in the SHAP analysis, indicates that both statistical and explainability-based analyses yielded consistent results. In contrast, features like lnk, C, Ni, and Ti, which were selected only by the ReliefF method, exhibited lower SHAP values, implying a relatively minor contribution to the model’s predictions on a sample-by-sample basis. In general, a high level of agreement was observed between the results of the feature selection methods and the findings of the SHAP analysis, which both increases model reliability and supports the accuracy of the feature selection strategies.
3.4. Experimental verification
Fig 8 illustrates the optical microstructure of the 316 steels after heat treatment at temperature of 1100 °C with a different holding times. As shown in Fig 8, these micrographs of steels enable relatively precise identification of the austenite grain boundaries. Initially, small and dispersed grains are observed, while as the annealing time increases, the grains become larger and the grain boundaries become more pronounced. This indicates that grain growth occurs with the activation of diffusion mechanisms at high temperature. The microstructure of steels shows the equiaxed grains. The average grain sizes of the austenite phase of 316 stainless steel are estimated 20 μm, 32 μm and 44 μm with increasing time, respectively. The obtained grain sizes were good agreement with literature [69]. The predicted and measured grain size of steels after heating process are shown in the Fig 8(e). The grain sizes were predicted by using XGBoost algorithm which exhibited best performances. As seen in Fig 8(e), the model is able to successfully predict grain growth, which shows the reliability of the prediction system developed or used.
Moreover, the Hall–Petch equation describes the relationship between a material’s hardness and its grain size as expressed:
in which H₀ refers to the inherent hardness of the alloy, d stands for the mean grain size, and ky represents the Hall–Petch constant [89]. The hardness results of samples heated at 1100 °C were found to be 162 HV, 160 HV and 151HV with increasing holding time, respectively. This behaviour is attributed to the fact that steels with finer grain sizes possess a greater number of grain boundaries, which act as barriers to dislocation motion. As a result, dislocations require more energy to cross these boundaries due to variations in atomic orientation. The trend observed follows the Hall-Petch relationship, where the hardness of annealed steel decreases as the grain size becomes larger [90].
3.5. 95% confidence interval estimation using quantile XGBoost
To estimate the predictive uncertainty of the XGBoost model, we used quantile regression with the “reg:quantileerror” objective. In addition to the standard model that gives point predictions, we trained two extra models to estimate the lower (τ = 0.025) and upper (τ = 0.975) bounds of a 95% prediction interval. These quantile models were trained using the same hyperparameters as the main regressor, with the quantile level defined by the quantile_alpha parameter. For each test sample, the three models together provide the median prediction and the uncertainty range
, which forms the 95% prediction interval. The accuracy of these quantile estimates was assessed using pinball loss. On the test set, the pinball loss values were 2.26 for the 2.5th percentile, 7.39 for the median (50th), and 4.62 for the 97.5th percentile. These values show that the quantile models provide a reasonable fit across the range of target values. Fig 9 illustrates the median XGBoost predictions along with the 95% prediction intervals and the corresponding experimental measurements. Experimental points are shown in blue, median predictions in red, and the shaded red region represents the uncertainty band. The interval successfully covers most of the experimental values and becomes wider in regions where the data show greater variability. This indicates that the quantile-based approach gives a meaningful picture of prediction uncertainty, which is important for evaluating the reliability of the model in microstructure design problems. The target variable ranges from about 4–990, which naturally affects quantile regression performance. The pinball loss at the median (τ = 0.5) is higher (7.3885) compared to the losses at the lower (2.2595) and upper (4.6231) quantiles. This suggests that the model has more difficulty predicting values near the center of the distribution, where the data are denser and more heterogeneous. In contrast, the lower pinball loss values at the extreme quantiles show that the model performs better at estimating the tails of the distribution. Overall, the consistent interval widths and relatively low pinball losses especially at the extreme quantiles confirm that the quantile XGBoost method captures prediction uncertainty effectively.
The quantile regression results for the experimental validation samples show a different pattern compared to the test set as see in Fig 10. The pinball loss values were 3.61 for the 2.5th quantile, 15.66 for the median (50th), and 1.24 for the 97.5th quantile. The much higher loss at the median indicates that the model makes larger errors in predicting the central value for these samples. This is not surprising, because the validation data come from a separate experimental group with different processing conditions. In other words, these samples lie outside the distribution of the training data. Even though the median prediction is less accurate, the 95% prediction interval remains meaningful. The relatively low pinball loss values at the lower and upper quantiles show that the model still captures the uncertainty bounds quite well. As shown in the Fig 10, most experimental values fall within these intervals. The interval becomes wider in regions where the response varies more or where the validation conditions differ strongly from the training data. This behaviour suggests that the quantile XGBoost model adjusts its uncertainty estimates appropriately when facing unseen or shifted data. In summary, although the median predictions are less precise for out-of-distribution samples, the quantile-based intervals remain reliable and well-calibrated. This confirms that quantile regression is a useful tool for quantifying uncertainty and evaluating prediction reliability under new experimental conditions.
4. Conclusions
In this study, a comprehensive ML framework was developed to predict austenitic grain growth kinetics in steels based on alloy composition and thermomechanical processing parameters. A dataset comprising 1039 experimentally collected samples, covering a wide range of chemical elements (such as C, Mn, Cr, Ni, Mo) and processing conditions (temperature, holding time, initial grain size), was used to train and validate the models. Among various ML algorithms evaluated, the XGBoost model exhibited the best predictive performance, achieving an R2 value of 0.9728 after hyperparameter optimization. Feature selection techniques, including Pearson correlation, CfsSubset, and ReliefF, identified temperature, initial grain size, and holding time as the most influential parameters on grain growth behaviour. Explainable AI (SHAP analysis) further enhanced the interpretability of the model outputs. Experimental validation was performed on 316L stainless steel samples heat-treated at 1100 °C. The predicted grain sizes showed strong agreement with the measured values, confirming the model’s reliability. Furthermore, microhardness measurements demonstrated a consistent decrease in hardness with increasing grain size, aligning with the Hall–Petch relationship. This correlation between grain size and hardness not only validates the grain growth predictions but also emphasizes the critical role of microstructure control in mechanical property optimization. Overall, this study represents a novel integration of machine learning, feature engineering, and experimental validation for predicting grain growth kinetics in steels. The proposed framework significantly reduces the need for extensive experimental efforts and offers a powerful tool for accelerating alloy design and process optimization. Future studies will focus on expanding the model by incorporating additional processing variables, such as cooling rate and strain, and extending its application to other alloy systems beyond steels.
Acknowledgments
The authors would like to acknowledge the valuable support provided by University of Birmingham, Marmara University and Ondokuz Mayıs University throughout this study. This work did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors.
References
- 1. Mochtar MA, Fatriansyah JF, Suhariadi I, Harefa VB. Numerical prediction of austenite grain growth characteristic of AISI O1 Alloy Steel. J Mater Eng Perform. 2022;32(7):3322–34.
- 2. An X, Tian Y, Wang H, Shen Y, Wang Z. Suppression of austenite grain coarsening by using Nb–Ti microalloying in high temperature carburizing of a gear steel. Adv Eng Mater. 2019;21(8).
- 3. Kisko A, Talonen J, Porter DA, Karjalainen LP. Effect of Nb microalloying on reversion and grain growth in a high-Mn 204Cu austenitic stainless steel. ISIJ Int. 2015;55(10):2217–24.
- 4. Liou H-Y, Pan Y-T, Hsieh R-I, Tsai W-T. Effects of alloying elements on the mechanical properties and corrosion behaviors of 2205 duplex stainless steels. J Mater Eng Perform. 2001;10(2):231–41.
- 5. Isfahany AN, Saghafian H, Borhani G. The effect of heat treatment on mechanical properties and corrosion behavior of AISI420 martensitic stainless steel. J Alloys Compound. 2011;509(9):3931–6.
- 6. Hwang KC, Lee S, Lee HC. Effects of alloying elements on microstructure and fracture properties of cast high speed steel rolls. Part I: Microstruct Anal Mater Sci Eng: A. 1998;254(1–2):296–304.
- 7. Sun W, Wang Q, Zhou Y, Wu J. Material and energy flows of the iron and steel industry: status quo, challenges and perspectives. Appl Energy. 2020;268:114946.
- 8.
Balakrishnan TS, Sultan MTH, Shahar FS, Saravanakumar YN, Chandran NK. Chapter eleven - Aerospace steel: properties, processing, and applications. In: Sultan MTH, Uthayakumar M, Korniejenko K, Mashinini PM, Najeeb MI, Krishnamoorthy RR, editors. Aerospace Materials. Elsevier; 2025. pp. 275–90.
- 9. Terentyev D, Puype A, Kachko O, Van Renterghem W, Henry J. Development of RAFM steel for nuclear applications with reduced manganese, silicon and carbon content. Nuclear Mater Energy. 2021;29:101070.
- 10. Baddoo NR. Stainless steel in construction: a review of research, applications, challenges and opportunities. J Construct Steel Res. 2008;64(11):1199–206.
- 11. Tisza M, Czinege I. Comparative study of the application of steels and aluminium in lightweight production of automotive parts. Int J Lightweight Mater Manufact. 2018;1(4):229–38.
- 12. Naghizadeh M, Mirzadeh H. Effects of grain size on mechanical properties and work-hardening behavior of AISI 304 austenitic stainless steel. Steel Res Int. 2019;90(10):1900153.
- 13. Park S-J, Kim K-S, Kang J-H, Kim S-J. Effects of carbon and nitrogen on Hall–Petch relationship in austenitic stainless steel. J Mater Res Technol. 2022;19:2960–4.
- 14. Jeon J, Seo N, Jung J-G, Kim H-S, Son SB, Lee S-J. Prediction and mechanism explain of austenite-grain growth during reheating of alloy steel using XAI. J Mater Res Technol. 2022;21:1408–18.
- 15. Graux A, Cazottes S, De Castro D, San Martín D, Capdevila C, Cabrera JM, et al. Precipitation and grain growth modelling in Ti-Nb microalloyed steels. Materialia. 2019;5:100233.
- 16. Moravec J, Novakova I, Sobotka J, Neumann H. Determination of grain growth kinetics and assessment of welding effect on properties of S700MC Steel in the HAZ of Welded Joints. Metals. 2019;9(6):707.
- 17. Moon J, Lee J, Lee C. Prediction for the austenite grain size in the presence of growing particles in the weld HAZ of Ti-microalloyed steel. Mater Sci Eng: A. 2007;459(1–2):40–6.
- 18. Xu Y, Tang D, Song Y, Pan X. Prediction model for the austenite grain growth in a hot rolled dual phase steel. Mater Design. 2012;36(6):275–8.
- 19. Burke J, Turnbull D. Recrystallization and grain growth. Progr Metal Phys. 1952;3:220–92.
- 20. Andersen I, Grong Ø. Analytical modelling of grain growth in metals and alloys in the presence of growing and dissolving precipitates—I. Normal grain growth. Acta Metallurgica et Materialia. 1995;43(7):2673–88.
- 21. Krill III CE, Chen L-Q. Computer simulation of 3-D grain growth using a phase-field model. Acta Materialia. 2002;50(12):3059–75.
- 22. Staublin P, Mukherjee A, Warren JA, Voorhees PW. Phase-field model for anisotropic grain growth. Acta Materialia. 2022;237:118169.
- 23. Liu Z, Bao Y, Wang M, Li X, Zeng F. Austenite grain growth of medium-carbon alloy steel with aluminum additions during heating process. Int J Miner Metall Mater. 2019;26(3):282–90.
- 24. Easton MA, Qian M, Prasad A, StJohn DH. Recent advances in grain refinement of light metals and alloys. Curr Opin Solid State Mater Sci. 2016;20(1):13–24.
- 25. Zhu Y, Fan S, Lian X, Min N. Effect of precipitated particles on austenite grain growth of Al- and Nb-microalloyed 20MnCr Gear Steel. Metals. 2024;14(4):469.
- 26. Xu D, Ji C, Zhao H, Ju D, Zhu M. A new study on the growth behavior of austenite grains during heating processes. Sci Rep. 2017;7(1):3968. pmid:28638061
- 27. Mobarak MH, Mimona MA, Islam MDA, Hossain N, Zohura FT, Imtiaz I, et al. Scope of machine learning in materials research—a review. Appl Surface Sci Adv. 2023;18:100523.
- 28. Wei J, Chu X, Sun X, Xu K, Deng H, Chen J, et al. Machine learning in materials science. InfoMat. 2019;1(3):338–58.
- 29. Fu H, Zhang H, Wang C, Yong W, Xie J. Recent progress in the machine learning-assisted rational design of alloys. Int J Miner Metall Mater. 2022;29(4):635–44.
- 30. Wang AYT, Murdock RJ, Kauwe SK, Oliynyk AO, Gurlo A, Brgoch J, et al. Machine learning for materials scientists: an introductory guide toward best practices. Chem Mater 2020;32(12);4954–65.
- 31. Field DM, Limmer KR, Hornbuckle BC. On the grain growth kinetics of a low density steel. Metals. 2019;9(9):997.
- 32. Mičian M, Frátrik M, Moravec J, Švec M. Determination of grain growth kinetics of S960MC steel. Materials. 2022;15(23):8539.
- 33. Du S, Li Y, Zheng Y. Kinetics of austenite grain growth during heating and its influence on hot deformation of LZ50 steel. J Mater Eng Perform. 2016;25(7):2661–9.
- 34. Huda Z, Zaharinie T, Aniszulaikha I, Almitani K. Austenite-grain-growth kinetics and mechanism in type 347H alloy steel for boiler tubes. Adv Mater Sci. 2023;23(3):79–97.
- 35. Radionova LV, Perevozchikov DV, Makoveckii AN, Eremin VN, Akhmedyanov AM, Rushchits SV. Grain growth during mechanical processing of austenitic stainless steel AISI 321. Metals. 2023;13(8):1421.
- 36. Saberi Tavakoli K, Abdi A, Sohrabi MJ, Kalhor A, Mirzadeh H. Hall-Petch analysis of fine-grained AISI 309Si stainless steel upon recrystallization and grain growth annealing. J Ultrafine Grain Nanostruct Mater. 2023;56(2):165–72.
- 37. Yue C, Zhang L, Liao S, Gao H. Kinetic analysis of the austenite grain growth in GCr15 steel. J Mater Eng Perform. 2009;19(1):112–5.
- 38. Melinia RK, Anwar MS, Wijaya RR, Siradj ES. Grain growth kinetics and hardness empirical model of 253 MA austenitic stainless steel after multi-pass cold rolling. Eng Appl Sci Res. 2021;49(3):425–32.
- 39.
Khzouz ER. Grain Growth Kinetics in Steels. Worcester Polytechnic Institute; 2011.
- 40.
Yıldız S, Yilmaz H, Yazır B, Evcil GE, Şimşir C. Determination of isothermal austenite grain growth model parameters for SAE 8620H carburizing steel. 2016.
- 41. Hossain Seikh A, Soliman MS, AlMajid A, Alhajeri K, Alshalfan W. Austenite grain growth kinetics in API X65 and X70 line-pipe steels during isothermal heating. Adv Mater Sci Eng. 2014;2014(1):1–8.
- 42. Mat MF, Manurung YHP, Olanrewaju Busari Y, Adenan MS, Sulaiman MS, Muhammad N, et al. Experimental analysis on grain growth kinetics of SS316L austenitic stainless steel. J Mech Eng. 2021;18(3):97–111.
- 43. Meng F, Wang J, Guo J, Fu K, Zhao M, Wang Z, et al. Growth behavior and kinetics of austenite grain in low-carbon high-strength steel with copper. Mater Res Express. 2021;8(9):096504.
- 44. Chamanfar A, Chentouf SM, Jahazi M, Lapierre-Boire L-P. Austenite grain growth and hot deformation behavior in a medium carbon low alloy steel. J Mater Res Technol. 2020;9(6):12102–14.
- 45. Vasilyev AA, Sokolov SF, Zhitelev PS, Sokolov DF, Kolbasnikov NG. A growth kinetics study and simulation for recrystallized ferrite grains under the annealing of automotive steel cold-rolled sheets. Steel Transl. 2019;49(11):806–12.
- 46. Wang L, Qian D, Guo J, Pan Y. Austenite grain growth behavior of AISI 4140 alloy steel. Adv Mech Eng. 2013;5:762890.
- 47. Zhang SS, Li MQ, Liu YG, Luo J, Liu TQ. The growth behavior of austenite grain in the heating process of 300M steel. Mater Sci Eng: A. 2011;528(15):4967–72.
- 48. Zhang Y, Li X, Liu Y, Liu C, Dong J, Yu L, et al. Study of the kinetics of austenite grain growth by dynamic Ti-rich and Nb-rich carbonitride dissolution in HSLA steel: In-situ observation and modeling. Mater Charact. 2020;169:110612.
- 49. Dong D, Chen F, Cui Z. Modeling of austenite grain growth during austenitization in a low alloy steel. J Mater Eng Perform. 2015;25(1):152–64.
- 50. Wang W, Wang L, Lyu P. Kinetics of austenite growth and bainite transformation during reheating and cooling treatments of high strength microalloyed steel produced by sub-rapid solidification. Int J Miner Metall Mater. 2022;30(2):354–64.
- 51. Maity J, Mondal DK. Isothermal grain growth of austenite in hypoeutectoid and hypereutectoid plain carbon steels. J Iron Steel Res Int. 2010;17(7):38–43.
- 52. Ji G, Gao XH, Liu ZG, Zhang K. In situ observation and modeling of austenite grain growth in a Nb–Ti-bearing high carbon steel. J Iron Steel Res Int. 2019;26(3):292–300.
- 53. Xu Y, Liu J, Zhao Y, Jiao Y. Austenite grain growth kinetics and mechanism of grain growth in 12Cr ultra-super-critical rotor steel. Philos Magaz. 2020;101(1):77–95.
- 54. Khalaj G, Yoozbashizadeh H, Khodabandeh A, Tamizifar M. Austenite grain growth modelling in weld heat affected zone of Nb/Ti microalloyed linepipe steel. Mater Sci Technol. 2014;30(4):424–33.
- 55. Li S-S, Liu Y-H, Song Y-L, Kong L-N, Li T-J, Zhang R-H. Austenitic grain growth behavior during austenization in an aluminum-alloyed 5% Cr-Mo-V Steel. steel research int. 2016;87(11):1450–60.
- 56. Li Z, Wen Z, Su F, Zhang R, Li Z. Austenite grain growth behavior of a GCr15 bearing steel cast billet in the homogenization heat treatment process. J Mater Res. 2016;31(14):2105–13.
- 57. Magalhães CHXM, Silva PH, Azzi PC, Faria GL. Effect of austenite grain size on phase transformation and Q&P design for a commercial CMnSi Steel. J Mater Eng Perform. 2023;33(13):6586–600.
- 58. Chen R, Zheng Z, Li J, Li N, Feng F. In situ investigation of grain evolution of 300M steel in isothermal holding process. Mater. 2018;11(10):1862.
- 59. Zhou XG, Li X, Zeng C-Y, Wu S-W, Liu ZY. Austenite grain growth and its equation in the austenitizing process for 700 MPa grade high-strength steel. Transac Indian Inst Metals 2023;76(11):3115–25.
- 60. Yuan J, Xiao Y, Min N, Li W, Zhao S. The influence of precipitate morphology on the growth of austenite grain in Nb-Ti-Al microalloyed steels. Materials (Basel). 2022;15(9):3176. pmid:35591509
- 61. An X, Cao W, Zhang X, Yu J. Suppress austenite grain coarsening by Nb alloying in high-temperature-pseudo-carburized bearing steel. Materials (Basel). 2024;17(12):2962. pmid:38930331
- 62. Park BC, Kim SD, Park I, Shin JH, Jang JH, Kang N. Effect of Nb on austenite grain growth in 10Cr-3Co-2W martensitic heat-resistant steel. Metals Mater Int. 2024;30(12):3311–9.
- 63. Wu D, Wang F, Cheng J, Li C. Effect of Nb and V on austenite grain growth behavior of the Cr-Mo-V steel for brake discs. High Temp Mater Process. 2018;37(9–10):899–907.
- 64. Prawoto Y, Jasmawati N, Sumeru K. Effect of prior austenite grain size on the morphology and mechanical properties of martensite in medium carbon steel. J Mater Sci Technol. 2012;28(5):461–6.
- 65. Tian S, Liu Z, Fu R, Wang X. Effect of niobium alloying on the austenite grain growth and mechanical properties of ultrahigh-strength stainless steel. Mater Res Express. 2022;9(2):026511.
- 66. Wang J, Nie P, Qiao S, Ojo OA, Yao C, Li Z, et al. Investigation on microstructural evolutions and mechanical properties of P92 steel during thermal processing. J Mater Eng Perform. 2018;27(9):4392–404.
- 67. Zheng X, Shi C, Zhu X, Li J, Xu H. Effect of solution treatment on grain growth and precipitates in electroslag remelted 15Cr-22Ni iron-base superalloy. Metall Mater Trans B. 2022;53(2):877–94.
- 68. Sharma V. A study on data scaling methods for machine learning. Int J Glob Acad Sci Res. 2022;1(1):31–42.
- 69. Aksu G, Güzeller Co, Eser MT. The effect of the normalization method used in different sample sizes on the success of artificial neural network model. Int J Assess Tools Educ. 2019;6(2):170–92.
- 70. Demirci S, Şahin DÖ, Demirci S, Gümüş A, Tünçay MM. Optimizing boride coating thickness on steel surfaces through machine learning: development, validation, and experimental insights. Appl Sci. 2025;15(5):2540.
- 71. Freund Y, Schapire RE. A decision-theoretic generalization of on-line learning and an application to boosting. J Comput Syst Sci. 1997;55(1):119–39.
- 72. Dabiri H, Farhangi V, Moradi MJ, Zadehmohamad M, Karakouzian M. Applications of decision tree and random forest as tree-based machine learning techniques for analyzing the ultimate strain of spliced and non-spliced reinforcement bars. Appl Sci. 2022;12(10):4851.
- 73.
Rasmussen CE, Williams CKI. Gaussian Processes for Machine Learning. The MIT Press; 2005.
- 74. Cover T, Hart P. Nearest neighbor pattern classification. IEEE Trans Inform Theory. 1967;13(1):21–7.
- 75. Bao Q, Zhang S, Guo J, Li Z, Zhang Z. Multivariate linear-regression variable parameter spatio-temporal zoning model for temperature prediction in steel rolling reheating furnace. J Process Control. 2023;123:108–22.
- 76. Johns JM, Burkes D. Development of multilayer perceptron networks for isothermal time temperature transformation prediction of U-Mo-X alloys. J Nuclear Mater. 2017;490:155–66.
- 77. Breiman L. Random forests. Mach Learn. 2001;45(1):5–32.
- 78. Barik S, Bhandari R, Mondal MK. Optimization of wire arc additive manufacturing process parameters for low-carbon steel and properties prediction by support vector regression model. Steel Res Int. 2024;95(1):2300369.
- 79. Song K, Yan F, Ding T, Gao L, Lu S. A steel property optimization model based on the XGBoost algorithm and improved PSO. Comput Mater Sci. 2020;174:109472.
- 80. Dai D, Xu T, Wei X, Ding G, Xu Y, Zhang J, et al. Using machine learning and feature engineering to characterize limited material datasets of high-entropy alloys. Comput Mater Sci. 2020;175:109618.
- 81. Liu Y, Mu Y, Chen K, Li Y, Guo J. Daily activity feature selection in smart homes based on Pearson correlation coefficient. Neural Process Lett. 2020;51(2):1771–87.
- 82. Sharma M. Improved autistic spectrum disorder estimation using Cfs subset with greedy stepwise feature selection technique. Int J Inf Tecnol. 2019;14(3):1251–61.
- 83. Urbanowicz RJ, Meeker M, La Cava W, Olson RS, Moore JH. Relief-based feature selection: introduction and review. J Biomed Inform. 2018;85:189–203. pmid:30031057
- 84. Almaraashi M. Investigating the impact of feature selection on the prediction of solar radiation in different locations in Saudi Arabia. Appl Soft Comput. 2018;66:250–63.
- 85. Ghomashchi MR. Quantitative microstructural analysis of M2 grade high speed steel during high temperature treatment. Acta Materialia. 1998;46(14):5207–20.
- 86. Rodríguez VM, López-Morelos VH, Nadimpalli VK, Bue Pedersen D, Ruiz A, Somers MAJ. Effect of Heat Treatment Processes on the Microstructure and Mechanical Properties of Spray‐Formed 440C Martensitic Stainless Steel. Steel Research Int. 2023;94(6).
- 87. Ji Y, Yang C, Xu T, Xu L, Li Y, Hu C, et al. Study on the characteristics and high-temperature dissolution mechanism of eutectic carbides in medium-alloy steel. J Mater Res Technol. 2024;31:276–86.
- 88. Kern M, Bernhard M, Scheiber D, Suhane A, Militzer M, Bernhard C. Influence of carbon and sulfur on austenite grain boundary mobility. Acta Materialia. 2025;288:120851.
- 89. Wang P, Siddiqui MF, Poletti MC, Enzinger N. Phase transformation and recrystallization of cold-rolled AISI 304L austenitic stainless steel during annealing. Mater Design. 2025;252:113738.
- 90. Lehto P, Remes H, Saukkonen T, Hänninen H, Romanoff J. Influence of grain size distribution on the Hall–Petch relationship of welded structural steel. Mater Sci Eng: A. 2014;592:28–39.