Stabilizing Multi-Agent Deep Reinforcement Learning By Implicitly Estimating Other Agents’ Behaviors
Yue Jin, Shuangqing Wei, Jian Yuan, Xudong Zhang, Chao Wang
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 14:12
Deep reinforcement learning (DRL) is able to learn control policies for many complicated tasks, but itâs power has not been unleashed to handle multi-agent circumstances. Independent learning, where each agent treats others as part of the environment and learns its own policy without considering othersâ policies is a simple way to apply DRL to multi-agent tasks. However, since agentsâ policies change as learning proceeds, from the perspective of each agent, the environment is non-stationary, which makes conventional DRL methods inefficient. To cope with this challenge, we propose a novel approach where each agent uses an implicit estimate of othersâ actions to guide its own policy learning. We demonstrate that given the implicit estimate of othersâ actions, each agent can learn its policy in a relatively stationary environment. Extensive experiments show that our method significantly alleviates the non-stationarity and outperforms the state-of-the-art in terms of both convergence speed and policy performance.