Optimal feature selection through reinforcement learning and fuzzy signature for improving classification accuracy
It is the main objective of feature selection to reduce the computational cost of a predictive model while increasing its performance. For feature selection, an exact search approach is used to evaluate all possible combinations of features in exponential time. The use of meta-heuristic algorithms i...
Saved in:
| Published in: | Multimedia tools and applications Vol. 84; no. 10; pp. 6931 - 6965 |
|---|---|
| Main Authors: | , , |
| Format: | Journal Article |
| Language: | English |
| Published: |
New York
Springer US
01.03.2025
Springer Nature B.V |
| Subjects: | |
| ISSN: | 1573-7721, 1380-7501, 1573-7721 |
| Online Access: | Get full text |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
| Summary: | It is the main objective of feature selection to reduce the computational cost of a predictive model while increasing its performance. For feature selection, an exact search approach is used to evaluate all possible combinations of features in exponential time. The use of meta-heuristic algorithms is another option that has some drawbacks. This paper converts the feature selection into a multi-armed bandit problem (MAB) by using the ε-Greedy algorithm. The
ε
parameter balances exploration and exploitation. In traditional
ε
-Greedy, the
ε
parameter is considered to be fixed. The creation of a good balance is impossible with a fixed parameter. A fuzzy signature approach is used to adaptively adjust the parameter of the ε-Greedy algorithm to take advantage of the exploration-exploitation trade-off inherent in the multi-armed bandit problem. In each episode, rewards are calculated based on correlations between features and objective functions (e. g., error classification, number of selected features, and redundancy). The calculated reward is used to adjust
ε
parameter adaptively and also is used to determine the number of iterations of the inner loop in the
ε
-Greedy algorithm. Adaptive adjustment allows the value of
ε
to be dynamic, adapting to the behavior of the environment. Lastly, different algorithms are compared, including Bat Algorithm (BA), Grasshopper Optimization Algorithm (GOA), Binary Monarch Butterfly Optimization (BMBO), Upper Confidence Bound (UCB), Stochastic Gradient Ascent (SGA), Greedy and classical
ε
-Greedy algorithms. As compared to BA, GOA, BMBO, UCB, SGA, Greedy, and classical
ε
-Greedy, the proposed algorithm improved the classification accuracy by 8.5%, 10.6%, 5.9%, 12.9%, 17.5%, 22.0%, and 3.2%, respectively. |
|---|---|
| Bibliography: | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 14 |
| ISSN: | 1573-7721 1380-7501 1573-7721 |
| DOI: | 10.1007/s11042-024-19069-z |