Optimizing SVM classification in medical datasets: A comparative study of swarm intelligence algorithms for feature selection and parameter tuning.

Gespeichert in:
Bibliographische Detailangaben
Titel: Optimizing SVM classification in medical datasets: A comparative study of swarm intelligence algorithms for feature selection and parameter tuning.
Autoren: Hussein, Safa Ali Abdo, Ahmad, R. Badlishah, Yaakob, Naimah, Khan, Abdul Ghani
Quelle: AIP Conference Proceedings; 2024, Vol. 3135 Issue 1, p1-10, 10p
Schlagwörter: SWARM intelligence, METAHEURISTIC algorithms, PARTICLE swarm optimization, FEATURE selection, MEDICAL coding
Abstract: Support Vector Machine (SVM) modelling approach focuses on dimensional reduction with a powerful generalization capability to solve problems such as non-linearity, and local extremes. However, the ability of SVM to learn and generalize depends on the collection of acceptable parameters that directly affect the output of the model. In addition, deploying relevant feature sets is also an issue in obtaining optimal classification. Thus, this work deploys eight swarm intelligence algorithms to improve the SVM classifier's accuracy by conducting two experiments: feature selection and tuning SVM parameters using the obtained feature set. The deployed swarm intelligence algorithms include Particle swarm optimization (PSO), Genetic Algorithm (GA), Firefly Algorithm (FA), Salp Swarm Algorithm (SSA), Bat Algorithm (BA), Sine Cosine Algorithm (SCA), Whale Optimization Algorithm (WOA), and Multi Verse Optimization (MVO). The undertaken experiments were on six medical benchmark datasets that are of various dimensions. Evaluation was based on two metrics: classification accuracy and size of feature set. This work's significance is to present the comparison and determine which swarm algorithm is the best to be used in a large size medical dataset. The experimental results support existing literature that noted the swarm intelligence are useful feature selection as well as optimizing SVM parameter while maintaining acceptable accuracy. [ABSTRACT FROM AUTHOR]
Copyright of AIP Conference Proceedings is the property of American Institute of Physics and its content may not be copied or emailed to multiple sites without the copyright holder's express written permission. Additionally, content may not be used with any artificial intelligence tools or machine learning technologies. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
Datenbank: Complementary Index
Beschreibung
Abstract:Support Vector Machine (SVM) modelling approach focuses on dimensional reduction with a powerful generalization capability to solve problems such as non-linearity, and local extremes. However, the ability of SVM to learn and generalize depends on the collection of acceptable parameters that directly affect the output of the model. In addition, deploying relevant feature sets is also an issue in obtaining optimal classification. Thus, this work deploys eight swarm intelligence algorithms to improve the SVM classifier's accuracy by conducting two experiments: feature selection and tuning SVM parameters using the obtained feature set. The deployed swarm intelligence algorithms include Particle swarm optimization (PSO), Genetic Algorithm (GA), Firefly Algorithm (FA), Salp Swarm Algorithm (SSA), Bat Algorithm (BA), Sine Cosine Algorithm (SCA), Whale Optimization Algorithm (WOA), and Multi Verse Optimization (MVO). The undertaken experiments were on six medical benchmark datasets that are of various dimensions. Evaluation was based on two metrics: classification accuracy and size of feature set. This work's significance is to present the comparison and determine which swarm algorithm is the best to be used in a large size medical dataset. The experimental results support existing literature that noted the swarm intelligence are useful feature selection as well as optimizing SVM parameter while maintaining acceptable accuracy. [ABSTRACT FROM AUTHOR]
ISSN:0094243X
DOI:10.1063/5.0214947