DEEP REINFORCEMENT LEARNING FOR GREEN UAV-ASSISTED DATA COLLECTION
Abhishek Mondal (National Institute of Technology Silchar); Deepak Mishra (University of New South Wales, Sydney); Ganesh Prasad (National Institute of Technology Silchar); Ashraf Hossain (National Institute of Technology Silchar)
-
SPS
IEEE Members: $11.00
Non-members: $15.00
Due to high maneuverability and flexible deployment, unmanned aerial vehicles (UAVs) are emerging as an alternative for reliable wireless communications. The main challenge of integrating UAVs with cellular networks is their limited on-board energy capacity, which restricts their operation period. Hence, this article examines the energy-efficiency (EE) maximization under the constraint of UAV’s propulsion and data reception energy. Specifically, the formulated problem optimizes user associations with UAV or base station, their respective transmit power allocations, and UAV’s trajectory subject to the user data rate requirements. As this joint optimization problem is combinatorial and involves multiple variables, we have reduced it into an equivalent tractable form using the Markov decision process (MDP). Later we leverage deep reinforcement learning (DRL) framework based on deep deterministic policy gradient (DDPG) algorithm to learn UAV’s trajectory. The proposed green DRL algorithm improves total EE of the system by 15.63% compared to the benchmark particle swarm optimization.