论文标题
关于合作多代理增强学习的鲁棒性
On the Robustness of Cooperative Multi-Agent Reinforcement Learning
论文作者
论文摘要
在合作的多代理增强学习(C-MARL)中,代理商学会了合作采取行动,以最大程度地提高团队奖励。我们分析了C-MARL对能够攻击团队中一名代理商的对手的鲁棒性。通过操纵该代理商的观察能力,对手试图减少团队的总奖励。 攻击C-MARL的挑战是有三个原因:首先,很难估计团队的奖励或如何受到代理商错误预测的影响;其次,模型是不可差异的;第三,特征空间是低维的。因此,我们引入了一种新颖的攻击。攻击者首先通过强化学习训练政策网络,以找到错误的行动,它应该鼓励受害者采取行动。然后,对手使用有针对性的对抗例子来迫使受害者采取这一行动。 我们在Startcraft II多代理基准上的结果表明,C-MARL团队非常容易受到对其代理人观察之一的扰动的影响。通过攻击单个代理商,我们的攻击方法对整个团队奖励产生了很大的负面影响,从而将其从20降低到9.4。这导致球队的获胜率从98.9%下降到0%。
In cooperative multi-agent reinforcement learning (c-MARL), agents learn to cooperatively take actions as a team to maximize a total team reward. We analyze the robustness of c-MARL to adversaries capable of attacking one of the agents on a team. Through the ability to manipulate this agent's observations, the adversary seeks to decrease the total team reward. Attacking c-MARL is challenging for three reasons: first, it is difficult to estimate team rewards or how they are impacted by an agent mispredicting; second, models are non-differentiable; and third, the feature space is low-dimensional. Thus, we introduce a novel attack. The attacker first trains a policy network with reinforcement learning to find a wrong action it should encourage the victim agent to take. Then, the adversary uses targeted adversarial examples to force the victim to take this action. Our results on the StartCraft II multi-agent benchmark demonstrate that c-MARL teams are highly vulnerable to perturbations applied to one of their agent's observations. By attacking a single agent, our attack method has highly negative impact on the overall team reward, reducing it from 20 to 9.4. This results in the team's winning rate to go down from 98.9% to 0%.