Self-Inference Of Others' Policies For Homogeneous Agents In Cooperative Multi-Agent Reinforcement Learning
Qifeng Lin, Qing Ling
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 00:11:55
Multi-agent reinforcement learning (MARL) has been widely applied in various cooperative tasks, where multiple agents are trained to collaboratively achieve global goals. During the training stage of MARL, inferring policies of other agents is able to improve the coordination efficiency. However, most of the existing policy inference methods require each agent to model all other agents separately, which results in quadratic growth of resource consumption as the number of agents increases. In addition, inferring the policy of an agent solely from its observations and actions may lead to failure of agent modeling. To address this issue, we propose to let each agent infer the others' policies with its own model, given that the agents are homogeneous. This self-inference approach significantly reduces the computation and storage consumption, and guarantees the quality of agent modeling. Experimental results demonstrate effectiveness of the proposed approach.
Chairs:
Seung-Jun Kim