Fayaz, Muhammad and Yi, Wenqiang and Liu, Yuanwei and Nallanathan, Arumugam (2021) Transmit Power Pool Design for Grant-Free NOMA-IoT Networks via Deep Reinforcement Learning. IEEE Transactions on Wireless Communications, 20 (11). pp. 7626-7641. DOI https://doi.org/10.1109/twc.2021.3086762
Fayaz, Muhammad and Yi, Wenqiang and Liu, Yuanwei and Nallanathan, Arumugam (2021) Transmit Power Pool Design for Grant-Free NOMA-IoT Networks via Deep Reinforcement Learning. IEEE Transactions on Wireless Communications, 20 (11). pp. 7626-7641. DOI https://doi.org/10.1109/twc.2021.3086762
Fayaz, Muhammad and Yi, Wenqiang and Liu, Yuanwei and Nallanathan, Arumugam (2021) Transmit Power Pool Design for Grant-Free NOMA-IoT Networks via Deep Reinforcement Learning. IEEE Transactions on Wireless Communications, 20 (11). pp. 7626-7641. DOI https://doi.org/10.1109/twc.2021.3086762
Abstract
Grant-free non-orthogonal multiple access (GF-NOMA) is a potential multiple access framework for short-packet internet-of-things (IoT) networks to enhance connectivity. However, the resource allocation problem in GF-NOMA is challenging due to the absence of closed-loop power control. We design a prototype of transmit power pool (PP) to provide open-loop power control. IoT users acquire their transmit power in advance from this prototype PP solely according to their communication distances. Firstly, a multi-agent deep Q-network (DQN) aided GF-NOMA algorithm is proposed to determine the optimal transmit power levels for the prototype PP. More specifically, each IoT user acts as an agent and learns a policy by interacting with the wireless environment that guides them to select optimal actions. Secondly, to prevent the Q-learning model overestimation problem, double DQN (DDQN) based GF-NOMA algorithm is proposed. Numerical results confirm that the DDQN based algorithm finds out the optimal transmit power levels that form the PP. Comparing with the conventional online learning approach, the proposed algorithm with the prototype PP converges faster under changing environments due to limiting the action space based on previous learning. The considered GF-NOMA system outperforms the networks with fixed transmission power, namely all the users have the same transmit power and the traditional GF with orthogonal multiple access techniques, in terms of throughput.
Item Type: | Article |
---|---|
Uncontrolled Keywords: | Double Q learning; grant-free NOMA; Internet of Things; multi-agent deep reinforcement learning; resource allocation |
Divisions: | Faculty of Science and Health Faculty of Science and Health > Computer Science and Electronic Engineering, School of |
SWORD Depositor: | Unnamed user with email elements@essex.ac.uk |
Depositing User: | Unnamed user with email elements@essex.ac.uk |
Date Deposited: | 04 Feb 2025 11:58 |
Last Modified: | 04 Feb 2025 11:58 |
URI: | http://repository.essex.ac.uk/id/eprint/40219 |
Available files
Filename: Transmit_Power_Pool_Design_for_Grant-Free_NOMA-IoT_Networks_via_Deep_Reinforcement_Learning.pdf
Licence: Creative Commons: Attribution 4.0