Comparison of ensemble learning methods on diabetes disease classification with various datasets
- This project compares various ensemble learning techniques for the classification of diabetes disease. Ensemble methods combine multiple machine learning models to improve predictive performance and robustness.
- In this project, we explore and compare the effectiveness of popular ensemble algorithms, such as Random Forest, AdaBoost, Gradient Boosting, and more, in diagnosing diabetes based on three different datasets of relevant features.
- Key Features:
- Implementation of different ensemble methods for classification
- Evaluation and comparison of model performance using metrics like accuracy, precision, recall, and F1-score
- Jupyter notebooks with detailed explanations and visualizations
- Dataset used for experimentation
- Code for preprocessing, model training, and evaluation
- This project has already been published in JMASIF (Jurnal Masyarakat Informatika) with the title Perbandingan Metode Ensemble Learning pada Klasifikasi Penyakit Diabetes.
- Python
- Pandas
- Matplotlib
- Seaborn
- Scikit-learn
- xgboost
- lightgbm
- catboost
Diabetes is a medical condition characterized by elevated blood sugar levels. According to the World Health Organization (WHO), the number of diabetes cases increased from 108 million to 422 million between 1980 and 2014. Machine Learning offers methods like Ensemble Learning for diabetes classification. This study compares three Ensemble Learning techniques, namely Bagging, Boosting, and Stacking, using three datasets: Pima Indians Diabetes, Frankfurt Hospital Diabetes, and Sylhet Hospital Diabetes.
- Pima Indians Diabetes Database by UCI Machine Learning
- Frankfurt Hospital Diabetes Dataset by John
- Sylhet Hospital Diabetes Dataset by Ishan Dutta
- Data Preprocessing
- MinMaxScaler for each dataset (change range of data to to fall within 0 and 1)
- Data Exploration
- Feature Engineering
- Data Splitting
- 80% Training data
- 20% Testing data
- Model Building
- Model Training & Testing
- Model Evaluation
- Accuracy
- Precision
- Recall
- F1-score
Bagging | Boosting | Stacking |
---|---|---|
- Bagging
- Bagging
- Random Forest
- Extra Trees
- Boosting
- Adaptative Boosting
- Gradient Boosting
- Extreme Gradient Boosting
- Light Gradient Boosting
- Cat Boosting
- Stacking
- Stacked Generalization
- In general, all Boosting methods give the best results for all datasets, but specifically, the Light Gradient Boosting method gives the best results in most of the data (Dataset 2 & Dataset 3)
- L. M. Cendani, and A. Wibowo, "Perbandingan Metode Ensemble Learning pada Klasifikasi Penyakit Diabetes," JURNAL MASYARAKAT INFORMATIKA, vol. 13, no. 1, pp. 33 - 44, May. 2022. https://doi.org/10.14710/jmasif.13.1.42912
- Linggar Maretva Cendani - linggarmc@gmail.com
- Adi Wibowo - bowo.adi@live.undip.ac.id
This project is licensed under the MIT License - see the LICENSE file for details