DMRL: A distributed multi-agent reinforcement learning algorithm for imbalanced classification

Traditional imbalanced classification methods rely on sampling or allocating different weights to different classes to improve the recognition rate for minority classes. However, these methods ignore the importance of adaptability, particularly as the degree of imbalance increases, resulting in sign...

Celý popis

Uloženo v:
Podrobná bibliografie
Vydáno v:Knowledge-based systems Ročník 327; s. 114101
Hlavní autoři: Ji, Yixin, Jing, Chao
Médium: Journal Article
Jazyk:angličtina
Vydáno: Elsevier B.V 09.10.2025
Témata:
ISSN:0950-7051
On-line přístup:Získat plný text
Tagy: Přidat tag
Žádné tagy, Buďte první, kdo vytvoří štítek k tomuto záznamu!
Popis
Shrnutí:Traditional imbalanced classification methods rely on sampling or allocating different weights to different classes to improve the recognition rate for minority classes. However, these methods ignore the importance of adaptability, particularly as the degree of imbalance increases, resulting in significant limitations when dynamically selecting the optimal classification strategy. To tackle this issue, we propose a distributed multi-agent reinforcement learning (DMRL) method for imbalanced classification problems, which models the classification problem as a multi-agent Markov decision-making process within the distributed computing scheme. Subsequently, there are three important schemes implemented in DMRL: 1) A multi-agent classification scheme based on improved double deep Q-network (MCSQ) that dynamically optimizes the imbalanced classification strategy through the reward function and importance weights; 2) A prioritized experience replay-based scheme for sampling agents experience (PERS) that uses prioritized experience replay to learn from important samples; 3) A distributed computing scheme based on the multi-agent centralized training and decentralized execution (CTDE) paradigm (DCSM) that combines distributed computing with multi-agent CTDE to improve learning efficiency. Finally, we perform our experiment on public datasets such as IMDB, Cifar-10, Fashion-Mnist, and Mnist using the various imbalanced ratios. Experimental results demonstrate that DMRL outperforms eight representative methods, with a maximum improvement of 8.9 % in G-mean and 10.7 % in F-measure when compared to the suboptimal method. Simultaneously, we study the impact of the value of the reward function, the number of agents and servers, and the effectiveness of prioritized experience replay on the performance of DMRL.
ISSN:0950-7051
DOI:10.1016/j.knosys.2025.114101