Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Chronic kidney disease prediction using boosting techniques based on clinical parameters

  • Shahid Mohammad Ganie,

    Roles Conceptualization, Data curation, Formal analysis, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing

    Affiliation AI Research Centre, School of Business, Woxsen University, Hyderabad, Telangana, India

  • Pijush Kanti Dutta Pramanik,

    Roles Formal analysis, Validation, Visualization, Writing – original draft, Writing – review & editing

    Affiliation School of Computing Science and Engineering, Galgotias University, Greater Noida, Uttar Pradesh, India

  • Saurav Mallik,

    Roles Validation, Writing – review & editing

    Affiliation Department of Environmental Health, Harvard T. H. Chan School of Public Health, Boston, MA, United States of America

  • Zhongming Zhao

    Roles Funding acquisition, Supervision, Writing – review & editing

    zhongming.zhao@uth.tmc.edu

    Affiliation Center for Precision Health, School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, TX, United States of America

Abstract

Chronic kidney disease (CKD) has become a major global health crisis, causing millions of yearly deaths. Predicting the possibility of a person being affected by the disease will allow timely diagnosis and precautionary measures leading to preventive strategies for health. Machine learning techniques have been popularly applied in various disease diagnoses and predictions. Ensemble learning approaches have become useful for predicting many complex diseases. In this paper, we utilise the boosting method, one of the popular ensemble learnings, to achieve a higher prediction accuracy for CKD. Five boosting algorithms are employed: XGBoost, CatBoost, LightGBM, AdaBoost, and gradient boosting. We experimented with the CKD data set from the UCI machine learning repository. Various preprocessing steps are employed to achieve better prediction performance, along with suitable hyperparameter tuning and feature selection. We assessed the degree of importance of each feature in the dataset leading to CKD. The performance of each model was evaluated with accuracy, precision, recall, F1-score, Area under the curve-receiving operator characteristic (AUC-ROC), and runtime. AdaBoost was found to have the overall best performance among the five algorithms, scoring the highest in almost all the performance measures. It attained 100% and 98.47% accuracy for training and testing sets. This model also exhibited better precision, recall, and AUC-ROC curve performance.

1. Introduction

Chronic kidney disease (CKD) has become very common across races [1], resulting in millions of deaths worldwide annually [2]. Proper diagnosis and timely treatment are major concerns in most developing countries. CKD mostly hits older people [3, 4], and by 2050, the number of people aged 65 years and above is estimated to increase to 1.5 billion from 703 million in 2019, with a more than double growth rate [5]. This will put a significant additional burden on healthcare services across the countries [6].

According to a study by the Center for Disease Control and Prevention, in 2017, approximately thirty million people in the U.S. alone were affected by CKD [7], which has been increased to 37 million in 2021 [8]. Moreover, most people are not aware of being infected by CKD. Traditionally, doctors confirm the CKD for any patient based on some clinical tests such as estimating glomerular filtration rate (GFR) from a filtration marker (e.g., serum creatinine or cystatin C) or through a urine test, detecting the presence of albumin and/or protein [911]. However, these tests may not always give accurate results, leading to the wrong diagnosis.

CKD can be mitigated to some extent if the possibility of it can be predicted beforehand for the suspected patients [12, 13]. This would allow healthcare professionals to deliver better services by embracing precautionary measures and early diagnosis and treatment. Machine learning algorithms have been popularly used in several disease diagnoses and predictions [1417]. For CKD prediction also, various such techniques have been explored [1822]. Machine learning algorithms are powerful for analysing large and complex datasets and identifying patterns and relationships that may not be apparent to human experts. In the context of CKD prediction, machine learning has the potential to improve accuracy and reduce costs by identifying early signs of disease progression and predicting the risk of developing CKD in at-risk populations.

However, traditional machine learning techniques suffer from some crucial limitations, including [23, 24]:

  • Overfitting, where the algorithm becomes too specialised to the training data and fails to generalise to new data.
  • Large, high-quality datasets are needed to train and validate the algorithms, which can be challenging to obtain in some clinical settings.
  • Training and evaluating machine learning algorithms may require considerable computational time and resources, especially for large datasets.
  • High dependency on the quality and quantity of data available for training. If the data is incomplete, biased, or otherwise of poor quality, the resulting algorithm will be inaccurate or may not work at all.
  • The machine learning algorithms can inadvertently incorporate biases present in the training data, leading to unfair or discriminatory outcomes.

Recently, ensemble learning techniques have shown great promise in improving the accuracy, robustness, and generalizability of predictive models, making them valuable in many fields, including healthcare, finance, marketing, social media analytics, etc. The ensemble learning approaches are gaining attention for disease prediction with higher accuracy [2531]. Among the ensemble learning techniques such as boosting, bagging, and stacking, boosting algorithms can reduce the training error (bias) and testing error (variance).

In this paper, we design a novel CKD prediction model using boosting algorithms. We aim to improve the performance of the disease prediction model over similar existing works. The contributions of this paper are summarised as follows.

  • Exploratory data analysis is performed to transform the considered dataset for better experimental usability.
  • Hyperparameter techniques, such as standardisation, normalisation, feature selection, and fine-tuning, are employed to achieve optimal results.
  • The attribution of existing dataset features to disease prediction is assessed.
  • Five boosting algorithms are individually applied to build the prediction model.
  • The prediction performances of the five boosting algorithms are evaluated and compared.
  • Our model achieved better accuracy and runtime than other machine learning-based CKD prediction models in method evaluation.

2. Related work

As mentioned above, machine learning has been extensively used for various disease diagnoses and predictions [17, 32, 33]. To improve the performance of these models, several machine learning techniques are combined to extract the advantages of each of them. This ensemble approach has gained acceptance and popularity after successful implementations for the prediction, detection, diagnosis, and prognosis of different diseases, such as heart disease [34, 35], breast cancer [36], skin disease [37], thyroid disease [38], myocardial infarction [39], Alzheimer’s disease [40], etc. For CKD prediction, several prediction techniques and models have already been proposed [41]. In the following, we briefly review some notable experiments for the diagnosis and prediction of CKD using ensemble learning techniques.

For CKD prediction, Kumar et al. [42] proposed an ensemble learning approach that comprises a support vector machine (SVM), decision tree, C4.5 decision tree, particle swarm optimisation ‐ multilayer perceptron (PSO-MLP), and artificial bee colony C4.5. The prediction process has two steps–i) in the first step, weak decision tree classifiers are obtained from C4.5, and ii) in the second step, the weak classifiers are combined with the weighted sum to get the final output from the classifier, attaining accuracy of 92.76%. Pal [43] developed a bagging ensemble method comprising a decision tree, SVM, and logistic regression to predict CKD. The best accuracy of 95.92% was achieved in the case of the decision tree. Hasan and Hasan [44] proposed an ensemble method for kidney disease diagnosis. They used adaptive boosting (AdaBoost), bootstrap aggregating, extra trees, gradient boosting, and random forest to build their prediction model. They performed tenfold cross-validation to validate the results. The highest accuracy of 99% was attained with adaptive boosting. For CKD detection, Wibawa et al. [45] developed an ensemble learning method that comprises three stages. In the first stage, base classifiers like Naive Bayes, SVM, and k nearest neighbour (kNN) were used. Correlation-based feature selection (CFS) was combined with the base classifiers mentioned above in the second stage. In the third stage, they used CFS with AdaBoost, achieving the highest accuracy of 98.01%. For CKD diagnosis, Jongbo et al. [1] built an ensemble learning model through bagging and random subspace based on three base classifiers–kNN, naïve Bayes, and decision tree. Data preprocessing was done to mitigate the missing value issue and data normalisation for scaling the independent variables within a certain range. The random subspace gained better performance than bagging in most performance measure metrics. It achieved an accuracy of 98.30% when combined with the decision tree method. To detect CKD, Ebiaredoh-Mienye et al. [46] combined the information-gain-based feature selection technique with the proposed cost-sensitive AdaBoost (C.S. AdaBoost), intending to save CKD screening time and cost. They trained the proposed C.S. AdaBoost with the reduced feature set, which attained a maximum accuracy of 99.8%. Emon et al. [47] used various boosting techniques to predict the risk of CKD progression among patients. The authors applied the principal component analysis (PCA) method to get the optimal feature set and attained the highest accuracy rate of 99.0% using random forest (R.F.). Ramaswamyreddy et al. [48] used wrapper methods along with bagging and boosting models to develop a CKD prediction model, attaining an accuracy of 99.0% with gradient boosting. However, the authors did not evaluate their model using other performance measure metrics.

3. Research methodology

This section briefly discusses the research steps followed and the ensemble learning techniques used in the experiment.

3.1 Research workflow

The workflow of the proposed work is shown in Fig 1. We performed exploratory data analysis on the considered dataset for better quality assessment. In this phase, missing values are identified and replaced using data imputation methods. The interquartile range (IQR) method is used to detect outliers present in the dataset. Some other required libraries are executed to check the corrupt data, if any, in the dataset. Also, standardisation, normalisation, feature selection, and tuning are made during the prediction model development process using five boosting algorithms. The dataset was split into training (60%) and test (40%) subsets. The results are assessed through various performance metrics.

thumbnail
Fig 1. The workflow of the proposed ensemble learning based CKD prediction.

https://doi.org/10.1371/journal.pone.0295234.g001

3.2 Boosting algorithms

Ensemble learning is a method that combines different traditional machine learning approaches to enhance the performance of the prediction model [49]. Various ensemble learning approaches are proposed [50, 51]. Boosting algorithm is one of the effective approaches in the ensemble learning family. In the literature, several boosting algorithms can be found [52, 53]. In this experiment, specifically for CKD prediction, we considered the following five ensemble learning based boosting algorithms:

XGBoost.

XGBoost (eXtreme gradient boosting) works by combining different kinds of decision trees (weak learners) to calculate the similarity scores independently [54]. It helps to overcome the problem of overfitting during the training phase by adapting the gradient descent and regularisation process. The mathematical formula for the XGBoost algorithm is shown in Eq 1. (1) where fθ(x) is XGBoost model with parameters θ,hm is the mth weak decision tree with parameters θm, and γm is the weight associated with mth tree. T denotes the number of decision trees, l denotes the loss function, and Rjm is an indicator function that returns 1 if x is in region Rjm, otherwise 0.

CatBoost.

CatBoost (categorical boosting) is faster than other boosting algorithms as it does not require the exploration of data preprocessing [55]. It is used to deal with high cardinality categorical variables. For low cardinality variables, one-hot encoding techniques are used for conversion. The objective function for the CatBoost algorithm is defined using Eq 2. (2) where y is the true label of the training set, f(x) is the predicted label, N is the number of training samples, l denotes the loss function, λ is the regularisation parameter used to penalise overfitting, P is the number of features and w is the weight associated with each feature of the dataset.

LightGBM.

LightGBM is an extension of a gradient boosting algorithm, capable of handling large datasets with less memory utilisation during the model evaluation process [56]. Gradient-based one-side sampling method is used for splitting the data samples, reducing the number of features in sparse datasets during training. The objective function for the LightGBM algorithm is defined using Eq 3. (3) where θ is a set of model parameters, N is the number of training samples, l denotes the loss function, yi is the true label of ith sample, is the predicted label for the model, fj is the jth decision tree, T is the number of trees, and ω is the regularisation term.

AdaBoost.

AdaBoost works by adjusting all the weights without prior knowledge of weak learners [57]. The weakness of all the base learners is measured by the estimator’s error rate while training the models. Decision tree stumps are widely used with the AdaBoost algorithm to solve classification and regression problems. The objective function for the AdaBoost algorithm is defined using Eq 4. (4) where H(xi) is the prediction of the classifier on the ith sample xi and yi is its corresponding true label in {-1, +1}and N denotes the number of training samples.

Gradient boosting.

In this method, the weak learners are trained sequentially, and all estimators are added one by one by adapting the weights [58]. The gradient boosting algorithm focuses on predicting the residual errors of previous estimators and tries to minimise the difference between the predicted and actual values. The objective function for the gradient boosting algorithm is written using Eq 5. (5) where F is the ensemble model, n is the number of training examples, yi is the true label of the ith sample, l denotes the loss function, and Fmathbf(xi) is the output of the ensemble model on example mathbf(xi).

4. Dataset collection and manipulation

We used the CKD data set (https://archive.ics.uci.edu/ml/datasets/chronic_kidney_disease), publicly available at the UCI machine learning repository, for the experiment. The dataset was collected from Apollo Hospitals, Managiri, India.

4.1 Dataset description

The dataset contains 400 instances and 25 attributes. The first 24 attributes are predicate/independent, and the last one is a dependent/target attribute. Among the attributes, 11 are numeric, and 14 are categorical. The attributes are described in Table 1. It represents the information about considered attributes, the description of attributes, their measurements, and the range values.

Table 2 describes the attribute information with their measures like count of records, mean, standard deviation (std), minimum (min) value, and maximum (max) value. For example, the blood pressure (bp) attribute has a count value of 400, mean 76.175, std 13.769, min 50, and max 180, respectively.

4.2 Data preprocessing

We performed some preprocessing on the considered CKD dataset to make the dataset most usable. The purpose was to transform the available raw data into a format easily understood by the ensemble learning algorithms. We conducted the following steps as data preprocessing:

  1. Identify and replace duplicate values.
  2. Identify and replace missing values.
  3. Detect and replace the outliers.
  4. Convert categorical variables to numerical values using one-hot encoding.
  5. Perform data transformation (-1 to 1) and scaling (0 to 1).

The results of the above steps are discussed below.

Class balancing.

The training dataset should be balanced of positive and negative instances to achieve reasonable prediction. From Fig 2(A), it can be observed that the considered dataset was highly biased toward the positive class, i.e., “patients having CKD” over the negative class, “patients not having CKD.” To minimise this difference, we used SMOTE to balance the dataset. From Fig 2(B), it can be observed that the resultant dataset is fairly balanced.

Exploratory data analysis.

We used different data visualisation tools to visualise and analyse the distribution of the data samples. Fig 3 shows the normally distributed histograms that group all the attributes of the considered dataset within the range values. Here, the X- and Y-axes describe the input attributes, and their corresponding values, respectively. Fig 4 plots the probability density using the kernel density estimation (KDE) method. The X- and Y-axes denote each attribute’s parameter value and probability density function, respectively. Fig 5 depicts the boxplot of all the considered attributes of the dataset. It provides a good indication of how the dispersion of values is spread out. To handle the outliers in the dataset, the IQR method was used.

Correlation coefficient analysis.

To identify and plot the relationship among the dataset attributes, we used the correlation coefficient analysis (CCA) method. A strong association/relationship between the set of independent and dependent attributes indicates a good-quality dataset. Fig 6 presents the CCA of the dataset attributes used in the experiment. The relationship range lies between +1 to -1 along the X- and Y-axes.

thumbnail
Fig 6. Correlation coefficient analysis of the dependent and independent attributes in the dataset.

https://doi.org/10.1371/journal.pone.0295234.g006

Data wrangling and cleaning.

To clean the dataset, we identified the missing values using the isnull() method and then calculated the percentage of null values present in the dataset. We used the data imputation methods (mean, median, fill, and original) to replace the null values. The missing values were replaced using the column’s mean, median, and mode. We used the IRQ method to detect the outliers and replace them using the Z-score method. The Z-score method shifts the distribution of all the data samples and makes the mean 0. Using data cleaning methods, we further checked for duplicate, inconsistent, and corrupt values in the dataset and neutralised them wherever applicable.

Data standardisation and normalisation.

We used the MinMaxScaler() for feature scaling. We scaled the data values using Eq 1 for standardisation and batch normalisation. Here, the data mean is set to 0 and the standard deviation to 6. (6) where N, X, xi, σ(x), xmin, and xmax denote the total sample in the data, ith attribute, the mean of the attributes, the sample variance of the attributes, the minimum value of the sample, and the maximum value of the sample, respectively.

5. Experiment, results, and discussion

In this section, we present the experimental details of this work and the obtained results by using the five boosting algorithms to predict CKD. We used 60% of the dataset to train the boosting algorithms and the rest 40% to test and validate their efficacy. The evaluations are extensively discussed in terms of accuracy, recall, precision, F1-score, micro-weighted, average-weighted, and AUC-ROC (area under curve-receiver operating characteristic) curve for each algorithm.

5.1 Hardware and software specifications

An HP Z60 workstation was used to carry out this research work. The hardware specification of the system is: Intel XEON 2.4 GHz CPU (12 core), 8 GB RAM, 1 T.B. hard disk, with Windows 10 pro-64-bit O.S. environment. As software requirements, we used the GUI-based Anaconda Navigator, the web-based computing platform Jupyter notebook, and Python as the programming language.

5.2 Feature importance

The feature importance is used to assess the contribution of an independent/predicate attribute that leads to CKD. Generally, not all attributes contribute to disease prediction. For instance, after running all five boosting algorithms on the original dataset, we found that the attributes–‘ane’, ‘appet’, ‘ba’, ‘cad’, ‘pc’, ‘pcc’, ‘pe’, ‘su’, and ‘wc’ have no role in CKD prediction. Hence, we eliminated these attributes from the dataset and kept only those that contributed at least for one algorithm, as shown in Fig 7.

thumbnail
Fig 7. Contributing features in CKD prediction for all boosting algorithms.

https://doi.org/10.1371/journal.pone.0295234.g007

We used the forward selection, a wrapper method, to calculate the feature importance [59]. A higher F-score of a feature indicates more importance of an attribute. For example, in Fig 7, it can be seen that the haemoglobin (hemo) attribute has the highest contribution in the CKD prediction for all the algorithms.

5.3 Hyperparameter tuning

We used the grid search method for hypermeter tuning to achieve optimality in the proposed model’s performance. By specifying a grid or a specified set of values for each hyperparameter, grid search enables methodically examining various combinations of hyperparameters. This ensures that all the options are tried to find the optimal values of the hyperparameters. The deterministic nature of grid search ensures consistency, i.e., it always yields the same outcomes when the same hyperparameters and data are used. This characteristic facilitates transparent testing and evaluation by making results simple to replicate and compare. One of the major advantages of grid search is that it is fairly straightforward to implement. Also, most of the machine learning frameworks and libraries provide built-in functions or modules for grid search. The best values of the hyperparameters found for each algorithm are shown in Table 3. The listed values for each parameter for the respective algorithm were found to be the best performers in our experiment.

thumbnail
Table 3. The optimal hyperparameters of boosting algorithms.

https://doi.org/10.1371/journal.pone.0295234.t003

5.4 Cross-validation scheme

Cross-validation is conducted to provide an unbiased evaluation of the prediction model. We performed the k-fold cross-validation to validate the performance of the proposed model on the training dataset. Here, we kept the value of k as 6. Based on the validation bias, the hyperparameters used in the experiment were tuned.

5.5 Performance evaluation

In this section, the performance of the proposed prediction model for the considered boosting algorithms is discussed in terms of different performance metrics.

5.5.1 Classification accuracy.

The classification performances of the algorithms are evaluated using a confusion matrix. The confusion matrices of all five boosting algorithms applied on the test dataset are shown in Fig 8. The left upper and the right lower boxes denote the correct predictions for the patients having (true positive) and not having (true negative) CKD, respectively. The right upper box and the left lower box indicate the number of incorrect predictions for patients having (false positive) and not having (false negative) CKD, respectively. The training and testing accuracies of all the boosting algorithms are shown in Fig 9. As per our experiment, on the test dataset, AdaBoost outperformed other algorithms by attaining the maximum accuracy rate for the training set of 100% and the test set of 98.47%, followed by LightGBM, gradient boost, XGBoost, and CatBoost at 99.73%, 99.21%, 97.23%, and 96.97%, respectively on the training set, and 97.96%, 97.46%, 95.93%, and 96.44%, respectively on the test set.

thumbnail
Fig 8. Confusion matrices of the prediction performance on the test set for all the five boosting algorithms.

https://doi.org/10.1371/journal.pone.0295234.g008

thumbnail
Fig 9. Training and testing accuracy statistics of all the boosting algorithms.

https://doi.org/10.1371/journal.pone.0295234.g009

5.5.2 Other measurements.

In addition to accuracy, we calculated the precision, recall, F1-score, and support of the five boosting algorithms on the test set, as shown in Figs 1013, respectively. In addition, the macro and weighted average were measured for both classes (0: no CKD, 1: CKD). As shown in those figures, AdaBoost produced the best precision in identifying the presence of CKD, while all algorithms identified the non-infection of CKD with equal precision. AdaBoost has a better recall and F1-score in confirming the absence of CKD. Regarding the case of support, i.e., the occurrence of class, AdaBoost performs slightly better than the other algorithms.

thumbnail
Fig 10. Precision comparison of five boosting algorithms on test set.

https://doi.org/10.1371/journal.pone.0295234.g010

thumbnail
Fig 11. Recall comparison of five boosting algorithms on test set.

https://doi.org/10.1371/journal.pone.0295234.g011

thumbnail
Fig 12. F1-score comparison of five boosting algorithms on test set.

https://doi.org/10.1371/journal.pone.0295234.g012

thumbnail
Fig 13. Support comparison of five boosting algorithms on test set.

https://doi.org/10.1371/journal.pone.0295234.g013

5.5.3 AUC-ROC curve.

The AUC-ROC curve was used to show the prediction ability of the boosting algorithms at different thresholds. It represents a false-positive rate (FPR) vs. a true-positive rate (TPR) along the x-axis and y-axis. A larger AUC-ROC area suggests the model’s ability to distinguish between 0’s and 1’s, leading to a better prediction. Also, an AUC value closer to 1 denotes a good separability measure, while in the case of an AUC of below 0.5, the model becomes ineffective in separating the classes, denoting the bad measure of disassociation. The AUC-ROC for the experiment is shown in Fig 14. It can be observed that AdaBoost performs best while XGBoost performs worst.

thumbnail
Fig 14. AUC-ROC curves for the experimented boosting algorithms.

https://doi.org/10.1371/journal.pone.0295234.g014

6. Comparative analysis

Table 4 presents a comparative analysis of the five boosting algorithms applied on the test dataset in terms of accuracy, misclassification rate, and runtimes. It can be observed that AdaBoost has the highest accuracy and least misclassification rate, but it has a slightly higher runtime than LightGBM and XGBoost.

thumbnail
Table 4. Comparative analysis of the considered algorithms performed on the test set.

https://doi.org/10.1371/journal.pone.0295234.t004

Since, in our experiment, we found AdaBoost to have the best overall performance in predicting CKD, we compared it with a few related research works in terms of accuracy, as shown in Table 5. The justification for achieving higher accuracy can be credited to the adopted procedures like data imputation for handling missing values, detection and replacing outliers, and effective data standardisation and normalisation.

thumbnail
Table 5. Comparison of the proposed work with existing similar works.

https://doi.org/10.1371/journal.pone.0295234.t005

7. Conclusion, limitations, and future directions

Diagnosis and prevention of chronic kidney disease have become challenging for healthcare professionals and other concerned authorities. It can be mitigated to some extent if it can be pre-diagnosed in well advance. In this paper, we attempted to predict CKD using an ensemble learning approach. Specifically, we used five boosting algorithms: XGBoost, CatBoost, LightGBM, AdaBoost, and gradient boosting. We employed different preprocessing techniques like the imputation method for handling missing values and min-max scalar and Z-score for data standardisation and normalisation. In addition, hyperparameter techniques like grid search were used to find the optimal parameter values. Furthermore, feature selection was carried out for each algorithm. AdaBoost emerged as the overall best performer in accuracy (99.17%), precision, recall, f1-score, and support in the experiment. AdaBoost also attained better results for AUC-ROC and misclassification rate. Comparing our proposed model with similar works, we found that our method outperformed others.

Though the proposed model performed relatively well, it has some obvious limitations. The size of the considered dataset is small, which may limit the prediction model’s performance in generic situations. It is observed that most of the features are having least contribution towards CKD. A more balanced dataset would lead to a better prediction model.

As an extension of this work, other ensemble learning techniques, like bagging, stacking, etc., can be explored to improve the results. Additionally, deep learning techniques can also be experimented with the exercised dataset. To validate the effectivity of the proposed model, additional and larger datasets are needed in future. Our proposed model can be applied to other disease datasets (e.g., diabetes) with common features. We expect more powerful disease prediction models to be developed and implemented in medical diagnosis and treatment.

References

  1. 1. Jongbo O. A., Adetunmbi A. O., Ogunrinde R. B. and Badeji-Ajisafe B., “Development of an ensemble approach to chronic kidney disease diagnosis,” Scientific African, vol. 8 (Article Id e00456), 2020.
  2. 2. Rhee C. M. and Kovesdy C. P., “Epidemiology: spotlight on CKD deaths-increasing mortality worldwide,” Nature Reviews Nephrology, vol. 11, pp. 199–200, 2015.
  3. 3. Ravani P., Quinn R., Fiocco M., Liu P., Al-Wahsh H., Lam N., Hemmelgarn B. R., Manns B. J., James M. T., Joanette Y. and Tonelli M., “Association of age with risk of kidney failure in adults with stage IV chronic kidney disease in Canada,” JAMA Network Open, vol. 3, no. 9, p. e2017150, 2020. pmid:32945876
  4. 4. Liu P., Quinn R. R., Lam N. N., Al-Wahsh H., Sood M. M., Tangri N., Tonelli M. and Ravani P., “Progression and regression of chronic kidney disease by age among adults in a population-based cohort in Alberta, Canada,” JAMA Network Open, vol. 4, no. 6, p. e2112828, 2021. pmid:34100938
  5. 5. “World Population Ageing 2019,” United Nations, Department of Economic and Social Affairs, Population Division, 31 December 2019. [Online]. Available: https://www.un.org/en/development/desa/population/publications/pdf/ageing/WorldPopulationAgeing2019-Report.pdf. [Accessed 24 September 2022].
  6. 6. Bello A. K., Levin A., Tonelli M., Okpechi I. G., Feehally J., Harris D., Jindal K., Salako B. L., Rateb A., Osman M. A., Qarni B., Saad S., Lunney M., Wiebe N., Ye F. and Johnson D. W., “Assessment of global kidney health care status,” JAMA, vol. 317, no. 18, pp. 1864–1881, 2017. pmid:28430830
  7. 7. Centers for Disease Control and Prevention, “National Chronic Kidney Disease Fact Sheet, 2017,” 2017.
  8. 8. Centers for Disease control and Prevention, “Chronic Kidney Disease in the United States, 2021,” 2021.
  9. 9. Levey A. S. and Coresh J., “Chronic kidney disease,” Lancet, vol. 379, no. 9811, pp. 165–180, 2012. pmid:21840587
  10. 10. Kovesdy C. P., “Epidemiology of chronic kidney disease: an update 2022,” Kidney International Supplements, vol. 12, no. 1, pp. 7–11, 2022. pmid:35529086
  11. 11. KDIGO, “Kidney Disease: Improving Global Outcomes (KDIGO) CKD Work Group KDIGO 2012 clinical practice guideline for the evaluation and management of chronic kidney disease,” Kidney International Supplements, vol. 3, no. 1, 2013.
  12. 12. Yuan Q., Zhang H., Deng T., Tang S., Yuan X., Tang W., Xie Y., Ge H., Wang X., Zhou Q. and Xiao X., “Role of Artificial Intelligence in Kidney Disease,” International Journal of Medical Sciences, vol. 17, no. 7, pp. 970–984, 2020. pmid:32308551
  13. 13. Chan L., Vaid A. and Nadkarni G. N., “Applications of machine learning methods in kidney disease: Hope or Hype?,” Current Opinion in Nephrology and Hypertension, vol. 29, no. 3, pp. 319–326, 2020. pmid:32235273
  14. 14. P. K. D. Pramanik, M. Mukhopadhyay and S. Pal, “Big data classification: applications and challenges,” in Artificial Intelligence and IoT: Smart Convergence for Eco Friendly Topography. Studies in Big Data, vol. 85, M. K. Geetha, J. A. Nehru and B. Sivaraman, Eds., Singapore, Springer, 2021, pp. 53–84.
  15. 15. Ganie S. M., Malik M. B. and Arif T., “Performance analysis and prediction of type 2 diabetes mellitus based on lifestyle data using machine learning approaches,” Journal of Diabetes & Metabolic Disorders, vol. 21, no. 1, pp. 339–352, 2022. pmid:35673418
  16. 16. Hazra A., Mandal S. K., Gupta A., Mukherjee A. and Mukherjee A., “Heart disease diagnosis and prediction using machine learning and data mining techniques: a review,” Advances in Computational Sciences and Technology, vol. 10, no. 7, pp. 2137–2159, 2017.
  17. 17. Nissa N., Jamwal S. and Mohammad S., “Early detection of cardiovascular disease using machine learning techniques an experimental study,” International Journal of Recent Technology and Engineering (IJRTE), vol. 9, no. 3, pp. 635–641, 2020.
  18. 18. Senan E. M., Al-Adhaileh M. H., Alsaade F. W., Aldhyani T. H. H., Alqarni A. A., Alsharif N., Uddin M. I., Alahmadi A. H., Jadhav M. E. and Alzahrani M. Y., “Diagnosis of chronic kidney disease using effective classification algorithms and recursive feature elimination techniques,” Journal of Healthcare Engineering, vol. 2021, 2021. pmid:34211680
  19. 19. Abdel-Fattah M. A., Othman N. A. and Goher N., “Predicting chronic kidney disease using hybrid machine learning based on Apache Spark,” Computational Intelligence and Neuroscience, vol. 2022, p. 9898831, 2022. pmid:35251161
  20. 20. Bai Q., Su C., Tang W. and Li Y., “Machine learning to predict end stage kidney disease in chronic kidney disease,” Scientific Reports, vol. 12 (Article number: 8377), 2022. pmid:35589908
  21. 21. Ilyas H., Ali S., Ponum M., Hasan O., Mahmood M. T., Iftikhar M. and Malik M. H., “Chronic kidney disease diagnosis using decision tree algorithms,” BMC Nephrology, vol. 22 (Article number: 273), 2021. pmid:34372817
  22. 22. Yashfi S. Y., Islam M. A., Pritilata N. Sakib T. Islam M. Shahbaaz and Pantho S. S., “Risk prediction of chronic kidney disease using machine learning algorithms,” in 11th International Conference on Computing, Communication and Networking Technologies (ICCCNT), Kharagpur, India, 2020.
  23. 23. Onyema E. M., Almuzaini K. K., Onu F. U., Verma D., Gregory U. S., Puttaramaiah M. and Afriyie R. K., “Prospects and Challenges of Using Machine Learning for Academic Forecasting,” Computational Intelligence and Neuroscience, vol. 2022 (Article ID 5624475), 2022. pmid:35909823
  24. 24. Ghassemi M., Naumann T., Schulam P., Beam A. L., Chen I. Y. and Ranganath R., “A Review of Challenges and Opportunities in Machine Learning for Health,” AMIA Joint Summits on Translational Science Proceedings, vol. 2020, pp. 191–200, 2020. pmid:32477638
  25. 25. Ganie S. M. and Malik M. B., “An ensemble machine learning approach for predicting type-II diabetes mellitus based on lifestyle indicators,” Healthcare Analytics, vol. 22 (Article number: 100092), 2022.
  26. 26. Naveen R. K. Sharma and Nair A. R., “Efficient breast cancer prediction using ensemble machine learning models,” in 4th International Conference on Recent Trends on Electronics, Information, Communication & Technology (RTEICT), Bangalore, India, 2019.
  27. 27. Oswald G. J. Sathwika and Bhattacharya A., “Prediction of cardiovascular disease (CVD) using ensemble learning algorithms,” in 5th Joint International Conference on Data Science & Management of Data (9th ACM IKDD CODS and 27th COMAD), Bangalore, India, 2022.
  28. 28. Mienye I. D., Sun Y. and Wang Z., “An improved ensemble learning approach for the prediction of heart disease risk,” Informatics in Medicine Unlocked, vol. 20 (100402), 2020.
  29. 29. Shanbhag G. A., Prabhu K. A., Subba Reddy N. V. and Rao B. A., “Prediction of lung cancer using ensemble classifiers,” Journal of Physics: Conference Series, vol. 2161 (012007), 2022.
  30. 30. Verma A. K., Pal S. and Tiwari B. B., “Skin disease prediction using ensemble methods and a new hybrid feature selection technique,” Iran Journal of Computer Science, vol. 3, pp. 207–216, 2020.
  31. 31. Jia P., Hu R. and Zhao Z., “Benchmark of embedding-based methods for accurate and transferable prediction of drug response,” Briefings in Bioinformatics, p. bbad098, 2023. pmid:36961310
  32. 32. Ganie S. M. and Malik M. B., “Comparative analysis of various supervised machine learning algorithms for the early prediction of type-II diabetes mellitus,” International Journal of Medical Engineering and Informatics, vol. 14, no. 6, pp. 473–483, 2022.
  33. 33. Shaikh F. J. and Rao D. S., “Prediction of cancer disease using machine learning approach,” Materialstoday: Proceedings, vol. 50 (Part 1), pp. 40–47, 2022.
  34. 34. Latha C. B. C. and Jeeva S. C., “Improving the accuracy of prediction of heart disease risk based on ensemble classification techniques,” Informatics in Medicine Unlocked, vol. 16 (100203), 2019.
  35. 35. Ali F., El-Sappagh S., Islam S. R., Kwak D., Ali A., Imran M. and Kwak K.-S, “A smart healthcare monitoring system for heart disease prediction based on ensemble deep learning and feature fusion,” Information Fusion, vol. 63, pp. 208–222, 2020.
  36. 36. Senthilkumar B., Zodinpuii D., Pachuau L., Chenkual S., Zohmingthanga J., Kumar N. c. S and Hmingliana L., “Ensemble modelling for early breast cancer prediction from diet and lifestyle,” IFAC-PapersOnLine, vol. 55, no. 1, pp. 429–435, 2022.
  37. 37. Verma A. K., Pal S. and Kumar S., “Comparison of skin disease prediction by feature selection using ensemble data mining techniques,” Informatics in Medicine Unlocked, vol. 16 (100202), 2019.
  38. 38. Yadav D. C. and Pal S., “Prediction of thyroid disease using decision tree ensemble method,” Human-Intelligent Systems Integration, vol. 2, pp. 89–95, 2020.
  39. 39. Hakim M. A., Jahan N., Zerin Z. A. and Farha A. B., “Performance evaluation and comparison of ensemble based bagging and boosting machine learning methods for automated early prediction of myocardial infarction,” in 12th International Conference on Computing Communication and Networking Technologies (ICCCNT), Kharagpur, India, 2021.
  40. 40. El-Sappagh S., Ali F., Abuhmed T., Singh J. and Alonso J. M., “Automatic detection of Alzheimer’s disease progression: An efficient information fusion approach with heterogeneous ensemble classifiers,” Neurocomputing, vol. 512, pp. 203–224, 2022.
  41. 41. Lim D. K. E., Boyd J. H., Thomas E., Chakera A., Tippaya S., Irish A., Manuel J., Betts K. and Robinson S., “Prediction models used in the progression of chronic kidney disease: A scoping review,” PLoS ONE, vol. 17, no. 7, p. e0271619, 2022. pmid:35881639
  42. 42. Komal Kumar N., Tulasi R. L. and Vigneswari D., “An ensemble multi-model technique for predicting chronic kidney disease,” International Journal of Electrical and Computer Engineering (IJECE), vol. 9, no. 2, pp. 1321–1326, 2019.
  43. 43. Pal S., “Chronic kidney disease prediction using machine learning techniques,” Biomedical Materials & Devices, 2022.
  44. 44. Hasan K. M. Z. and Hasan M. Z., “Performance evaluation of ensemble-based machine learning techniques for prediction of chronic kidney disease,” in Emerging Research in Computing, Information, Communication and Applications. Advances in Intelligent Systems and Computing, vol. 882, Shetty N., Patnaik L., Nagaraj H., Hamsavath P. and Nalini N., Eds., Singapore, Springer, 2019, pp. 415–426.
  45. 45. Wibawa M. S., Maysanjaya I. M. D. and Putra I. M. A. W., “Boosted classifier and features selection for enhancing chronic kidney disease diagnose,” in 5th International Conference on Cyber and IT Service Management (CITSM), Denpasar, Indonesia, 2017.
  46. 46. Ebiaredoh-Mienye S. A., Swart T. G., Esenogho E. and Mienye I. D., “A machine learning method with filter-based feature selection for improved prediction of chronic kidney disease,” Bioengineering, vol. 9, no. 8, p. 350, 2022. pmid:36004875
  47. 47. Emon M. U., Imran A. M., Islam R., Keya M. S., Zannat R. and Ohidujjaman , “Performance analysis of chronic kidney disease through machine learning approaches,” in 6th International Conference on Inventive Computation Technologies (ICICT), Coimbatore, India, 2021.
  48. 48. Ramaswamyreddy A., Shiva Prasad S., Rangarao K. V. and Saranya A., “Efficient datamining model for prediction of chronic kidney disease using wrapper methods,” International Journal of Informatics and Communication Technology, vol. 8, no. 2, pp. 63–70, 2019.
  49. 49. G. Brown, “Ensemble learning,” in Encyclopedia of Machine Learning, C. Sammut and G. I. Webb, Eds., Boston, MA, Springer, 2011, pp. 312–320.
  50. 50. Sagi O. and Rokach L., “Ensemble learning: a survey,” WIREs Data Mining and Knowledge Discovery, vol. 8, no. 4, p. e1249, 2018.
  51. 51. Zhang C. and Ma Y., Eds., Ensemble machine learning: methods and applications, New York: Springer, 2012.
  52. 52. Ferreira A. J. and Figueiredo M. A. T., “Boosting algorithms: a review of methods, theory, and applications,” in Ensemble Machine Learning, Zhang C. and Ma Y., Eds., Boston, MA, Springer, 2012, pp. 35–85.
  53. 53. Tanha J., Abdi Y., Samadi N., Razzaghi N. and Asadpour M., “Boosting methods for multi-class imbalanced data classification: an experimental review,” Journal of Big Data, vol. 7, no. 1, 2020.
  54. 54. Santhanam R., Uzir N., Raman S. and Banerjee S., “Experimenting XGBoost algorithm for prediction and classification of different datasets,” International Journal of Control Theory and Applications, vol. 9, no. 40, pp. 651–662, 2016.
  55. 55. Hancock J. T. and Khoshgoftaar T. M., “CatBoost for big data: an interdisciplinary review,” Journal of Big Data, vol. 7, no. 1 (Article No. 94), 2020. pmid:33169094
  56. 56. Machado M. R., Karray S. and de Sousa I. T., “LightGBM: an effective decision tree gradient boosting method to predict customer loyalty in the finance industry,” in International Conference on Computer Science & Education (ICCSE), Toronto, Canada, 2019.
  57. 57. Sevinc E., “An empowered AdaBoost algorithm implementation: A COVID-19 dataset study,” Computers & Industrial Engineering, vol. 165 (Article No. 107912), 2022. pmid:35013637
  58. 58. Aziz N., Akhir E. A. P., Aziz I. A., Jaafar J., Hasan M. H. and Abas A. N. C., “A study on gradient boosting algorithms for development of AI monitoring and prediction systems,” in International Conference on Computational Intelligence (ICCI), Malaysia, 2020.
  59. 59. Kohavi R. and John G. H., “Wrappers for feature subset selection,” Artificial Intelligence, vol. 97, no. 1–2, pp. 273–324, 1997.