Deep Reinforcement Learning Driven UAV-Assisted Edge Computing
Document Type
Article
Publication Date
12-15-2022
Abstract
Unmanned aerial vehicles (UAVs) are playing a critical role in provisioning instant connectivity and computational needs of Internet of Things Devices (IoTDs), especially in crisis and disaster management. In this work, we focus on optimizing trajectories of UAVs along which IoTDs are served with communication and computing resources in multiple time slots. The Quality of Experience (QoE) of an IoTD depends on its latency performance; we thus aim to maximize the average aggregate QoE of all IoTDs overall time slots. However, this is a nonconvex, nonlinear, and mixed discrete optimization problem, which is difficult to solve and obtain the optimal solution. We thus propose two deep reinforcement learning algorithms to solve this problem by considering UAV path planning, user assignment, bandwidth, and computing resource assignment. We compare the performance of our proposed algorithms through simulations with three baseline cases: 1) with fixed UAV locations; 2) without UAVs; and 3) the fixed UAV trajectories. We demonstrate that the deep reinforcement learning algorithms perform better than all baseline cases.
Identifier
85136116426 (Scopus)
Publication Title
IEEE Internet of Things Journal
External Full Text Location
https://doi.org/10.1109/JIOT.2022.3196842
e-ISSN
23274662
First Page
25449
Last Page
25459
Issue
24
Volume
9
Recommended Citation
Zhang, Liang; Jabbari, Bijan; and Ansari, Nirwan, "Deep Reinforcement Learning Driven UAV-Assisted Edge Computing" (2022). Faculty Publications. 2404.
https://digitalcommons.njit.edu/fac_pubs/2404