Jiang, Feibo and Dong, Li and Wang, Kezhi and Yang, Kun and Pan, Cunhua (2022) Distributed Resource Scheduling for Large-Scale MEC Systems: A Multi-Agent Ensemble Deep Reinforcement Learning with Imitation Acceleration. IEEE Internet of Things Journal, 9 (9). pp. 6597-6610. DOI https://doi.org/10.1109/jiot.2021.3113872
Jiang, Feibo and Dong, Li and Wang, Kezhi and Yang, Kun and Pan, Cunhua (2022) Distributed Resource Scheduling for Large-Scale MEC Systems: A Multi-Agent Ensemble Deep Reinforcement Learning with Imitation Acceleration. IEEE Internet of Things Journal, 9 (9). pp. 6597-6610. DOI https://doi.org/10.1109/jiot.2021.3113872
Jiang, Feibo and Dong, Li and Wang, Kezhi and Yang, Kun and Pan, Cunhua (2022) Distributed Resource Scheduling for Large-Scale MEC Systems: A Multi-Agent Ensemble Deep Reinforcement Learning with Imitation Acceleration. IEEE Internet of Things Journal, 9 (9). pp. 6597-6610. DOI https://doi.org/10.1109/jiot.2021.3113872
Abstract
In large-scale mobile edge computing (MEC) systems, the task latency and energy consumption are important for massive resource-consuming and delay-sensitive Internet of things devices (IoTDs). Against this background, we propose a distributed intelligent resource scheduling (DIRS) framework to minimize the sum of task latency and energy consumption for all IoTDs, which can be formulated as a mixed integer nonlinear programming. The DIRS framework includes centralized training relying on the global information and distributed decision making by each agent deployed in each MEC server. Specifically, we first introduce a novel multi-agent ensemble-assisted distributed deep reinforcement learning (DRL) architecture, which can simplify the overall neural network structure of each agent by partitioning the state space and also improve the performance of a single agent by combining decisions of all the agents. Secondly, we apply action refinement to enhance the exploration ability of the proposed DIRS framework, where the near-optimal state-action pairs are obtained by a novel Levy flight search. Finally, an imitation acceleration scheme is presented to pre-train all the agents, which can significantly accelerate the learning process of the proposed framework through learning the professional experience from a small amount of demonstration data. The simulation results in three typical scenarios demonstrate that the proposed DIRS framework is efficient and outperforms the existing benchmark schemes.
Item Type: | Article |
---|---|
Uncontrolled Keywords: | Multi-agent reinforcement learning; Distributed deep reinforcement learning; Imitation learning; Resource scheduling; LC)vy flight |
Divisions: | Faculty of Science and Health Faculty of Science and Health > Computer Science and Electronic Engineering, School of |
SWORD Depositor: | Unnamed user with email elements@essex.ac.uk |
Depositing User: | Unnamed user with email elements@essex.ac.uk |
Date Deposited: | 09 Feb 2022 12:29 |
Last Modified: | 30 Oct 2024 16:39 |
URI: | http://repository.essex.ac.uk/id/eprint/32239 |
Available files
Filename: Distributed_Resource_Scheduling_for_Large-Scale_MEC_Systems_A_Multi-Agent_Ensemble_Deep_Reinforcement_Learning_with_Imitation_Acceleration.pdf