PLoS One. 2025 Jun 18;20(6):e0325900. doi: 10.1371/journal.pone.0325900. eCollection 2025.

ABSTRACT

Thyroid hormones control crucial physiological activities, such as metabolism, oxidative stress, erythropoiesis, thermoregulation, and organ development. Hormonal imbalances may cause serious conditions like cognitive impairment, depression, and nervous system damage. Traditional diagnostic techniques, based on hormone level measurements (TSH, T3, FT4, T4, and FTI), are usually lengthy and laborious. This study uses machine learning (ML) algorithms and feature selection based on GA to improve the accuracy and efficiency of diagnosing thyroid disorders using the UCI thyroid dataset. Five ML algorithms-LR, RF, SVM, AB, and DT- were tested using two paradigms: (1) default classifiers and (2) hybrid GA-ML models- GA-RF, GA-LR, GA-SVM, GA-DT, and GA-AB. The data pre-processed included handling missing values, feature scaling, and correlation analysis. In this case, the performance metrics used for model evaluation are accuracy, F1 Score, sensitivity, specificity, precision, and Cohen’s Kappa with 80% of the dataset to train the model and the rest 20% used to test it. Among the non-hybrid models, RF achieved the highest accuracy, which was 93.93%. The hybrid GA-RF model outperformed all others, achieving a remarkable accuracy of 97.21%, along with superior metrics across all the evaluated parameters. These findings highlight the diagnostic potential of the GA-RF model in providing faster, more accurate, and reliable thyroid disorder detection. The research illustrated the potential of the hybrid GA-ML approaches to improving the clinical diagnostic process while proposing a strong and scalable approach towards thyroid disorder identification.

PMID:40531844 | DOI:10.1371/journal.pone.0325900