Research on the Robustness of Cooperative Multi-Agent Deep Reinforcement Learning: A Policy Regularization Method Based on Transfer Learning

Cooperative multi-agent deep reinforcement learning algorithm(c-MADRL)is proven vulnerable to adversarial attacks and its corresponding practical applications are subject to security risks, thus it is crucial for us to improve the robustness of c-MADRL. In this paper, we use a combination of policy...

Full description

Saved in:
Bibliographic Details
Published in:Chinese Control Conference pp. 5933 - 5939
Main Authors: Zan, Lixia, Zhu, Xiangbin
Format: Conference Proceeding
Language:English
Published: Technical Committee on Control Theory, Chinese Association of Automation 28.07.2025
Subjects:
ISSN:1934-1768
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Cooperative multi-agent deep reinforcement learning algorithm(c-MADRL)is proven vulnerable to adversarial attacks and its corresponding practical applications are subject to security risks, thus it is crucial for us to improve the robustness of c-MADRL. In this paper, we use a combination of policy regularization and transfer learning to improve the robustness of c-MADRL. Firstly, we use the regularization term based on the worst-case adversarial sample, that is, the adversarial attack that minimizes the expected reward to update the algorithm; Secondly, in order to improve the training efficiency, we employ the policy transfer method, which first performs regularization training on a leader agent, and then transfers the trained policy to other agents; Finally, we verify the performance of the algorithm before and after regularization by adversarial attacks. We have conducted experiments on the multiple-particle environment (MPE), the experimental results show that our methods can improve the robustness of the algorithm and raise defense performance against some adversarial attacks.
ISSN:1934-1768
DOI:10.23919/CCC64809.2025.11179405