INVERSE REINFORCEMENT LEARNING WITH GRAPH NEURAL NETWORKS FOR IOT RESOURCE ALLOCATION
Guangchen Wang (The University of Sydney); Peng Cheng (La Trobe University); Zhuo Chen (CSIRO); Wei Xiang (La Trobe University); Branka Vucetic (University of Sydney); Yonghui Li (THE UNIVERSITY OF SYDNEY)
-
SPS
IEEE Members: $11.00
Non-members: $15.00
Massive Internet of Things (IoT) applications require efficient computing and communication resource allocation to streamline existing network operations. These strategies could be formulated as mixed-integer nonlinear programming (MINLP) problems, where the optimal branch-and-bound (B&B) with the full strong branching (FSB) variable selection policy features an extremely high complexity. We propose inverse reinforcement learning with graph neural networks (GNNIRL) to generate a new variable selection policy that closely matches the FSB variable selection. Without sacrificing the optimality, the GNNIRL can directly infer the variable selection with a significantly lower complexity, which is also verified by simulation.