Skip to main content

CONVERGENCE ANALYSIS OF GRAPHICAL GAME-BASED NASH $Q-$LEARNING USING THE INTERACTION DETECTION SIGNAL OF $\mathcal{N}-$STEP RETURN

Yunkai Zhuang (Nanjing University); Shangdong Yang (Nanjing University of Posts and Telecommunications); Wenbin Li (Nanjing University); Yang Gao (Nanjing University)

  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00
07 Jun 2023

The graphical game provides an effective method for modeling different kinds of sparse interactions in multi-agent reinforcement learning. Most previous work on game abstraction lacks theoretical guarantees of convergence. In this paper, we adopt the $\mathcal{N}$-step return signal to detect interactions between agents and build the Markov graphical game based on it. We analyze that the solution of the Markov graphical game is an $\epsilon$-Nash equilibrium which guarantees the convergence of the proposed NSR-G$^2$NashQ algorithm theoretically. Also, we have done experiments in different multi-agent reinforcement learning tasks with both tabular and function approximation solutions. The results show the NSR-G$^2$NashQ algorithm accelerates the convergence of agents to the optimal policy.

More Like This

  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00
  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00
  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00