Computation-Efficient Offloading and Power Control for MEC in IoT Networks by Meta-Reinforcement Learning

Document Type

Article

Publication Date

5-1-2024

Abstract

Due to the proliferation of devices and the availability of computing servers, mobile-edge computing (MEC) has gained popularity in executing various computational tasks. MEC offers computing services at the network's edge, providing user equipment (UE) with reduced latency for their applications. However, determining a suitable offloading policy for UEs in MEC, considering wireless resource allocation and power management, is computationally demanding. Additionally, the problem is NP-hard, making it challenging to find an optimal solution within a reasonable time frame. In this work, we propose a meta-reinforcement learning (MRL)-based computational task offloading and power control mechanism for UEs in a resource-constrained environment of a MEC network to tackle the NP-hardness of the problem. We first develop an optimization problem to maximize UE computation efficiency by minimizing their power consumption for local computing and uplink transmission of UEs in the MEC network. We propose to use both binary offloading (full offloading or full local computing) and partial offloading schemes in the system. Our proposed MRL algorithm can figure out a suitable offloading policy for UEs within a short time. Unlike the traditional deep-reinforcement learning algorithms, our approach can resolve the issue of obtaining proper solutions in a new environment. Extensive simulation results prove the feasibility of our proposed work.

Identifier

85182946378 (Scopus)

Publication Title

IEEE Internet of Things Journal

External Full Text Location

https://doi.org/10.1109/JIOT.2024.3355023

e-ISSN

23274662

First Page

16722

Last Page

16730

Issue

9

Volume

11

This document is currently not available here.

Share

COinS