Browsing by Author "Petronilla, Muthoni"
Now showing 1 - 2 of 2
- Results Per Page
- Sort Options
Item Comparative Performance of Machine Leaning Algorithms in Prediction of Cervical Cancer(IEEE xplore, 2021-10) Emmanuel, Ahishakiye; Waweru, Mwangi; Petronilla, Muthoni; Lawrence, Nderu; Ruth, WarioCervical cancer is among the most common types of cancer affecting women around the world despite the advances in prevention, screening, diagnosis, and treatment during the past decade. Cervical cancer can be treated if diagnosed in its early stages. Machine learning algorithms like multi-layer perceptron, decision trees, random forest, K-Nearest Neighbor, and Naïve-Bayes have been used for the prediction of cervical cancer to aid in its early diagnoses. In this study, we compare the performance of ensemble methods (AdaBoost, Stochastic Gradient Boosting, Random Forests, and Extra Trees), and classification algorithms (K-Nearest Neighbor and Support Vector Machine) in the prediction of cervical cancer basin g risk factors. Ensemble methods and classification algorithms were used during this study. Ensemble methods were selected because they combine several machine learning techniques into one model to decrease variance, bias, or improve performance while the classification methods were selected because our dataset was generally categorical and therefore could work well with our problem domain. Experimental results revealed that all the algorithms did not perform well on the “imbalanced” dataset. Experiments on balanced revealed an improved performance. The performance metrics used include Fl-score, Area Under Curve (AUC), and Recall. Extra Trees performed better than the rest when using the Fl-score metric, Stochastic Gradient Boosting and Random Forest performed better than the rest when using the AUC metric, K-Nearest Neighbors outperformed the rest using the recall metric, and Extra Trees had the best accuracy 0.96. The application of machine learning methods in the prediction of cervical cancer using risk factors may lead to early detection of the disease which can be treated if diagnosed early. Six algorithms have been considered in this study. The general performance reveals that ensemble methods performed better than classification methods using both imbalanced and balanced datasets.Item A comparative study of some pre-trained models on transfer learning approach in detection and classification of cassava leaf diseases(SSRN, 2022) Emmanuel, Ahishakiye; Waweru, Mwangi; Petronilla, Muthoni; Danison, Taremwa; Fredrick, KanobeCassava diseases affect cassava harvest posing the greatest danger to the food security and livelihoods of more than 200 million people. To identify cassava diseases, government professionals visit various sections of the country and visually score the plants by looking for disease indicators on the leaves. This procedure is notoriously subjective; it is not uncommon for specialists to differ on a plant's diagnosis. Automating the detection and classification of crop diseases could help professionals diagnose diseases more accurately and allow farmers in remote locations to monitor their crops without the help of specialists. Machine learning algorithms have been used in the early detection and classification of crop diseases. Motivated by the current developments and many influential studies in the field of deep learning and transfer learning models in the detection and classification of crop diseases, this study evaluates the performance of VGG16, VGG19, ResNet50, InceptionV3, DenseNet201, and MobileNetV2 in detection and classification of cassava leaf diseases. Fine-tuning of the hyperparameters was done during training to improve the accuracy of the models. Experimental results on the cassava dataset revealed that InceptionV3, DenseNet201, and MobileNetV2 models had high training accuracy but low validation accuracy with various epochs which means that they had issues with over fitting while ResNet50 had issues with underfitting. Moreso, VGG16 and VGG19 models performed well on both training and validation datasets, though VGG16 performed relatively well compared to VGG19.