D2D Mobile Relaying Meets NOMA-Part II: A Reinforcement Learning Perspective

5Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.

Abstract

Structureless communications such as Device-to-Device (D2D) relaying are undeniably of paramount importance to improving the performance of today's mobile networks. Such a communication paradigm requires a certain level of intelligence at the device level, thereby allowing it to interact with the environment and make proper decisions. However, decentralizing decision-making may induce paradoxical outcomes, resulting in a drop in performance, which sustains the design of self-organizing yet efficient systems. We propose that each device decides either to directly connect to the eNodeB or get access via another device through a D2D link. In the first part of this article, we describe a biform game framework to analyze the proposed self-organized system's performance, under pure and mixed strategies. We use two reinforcement learning (RL) algorithms, enabling devices to self-organize and learn their pure/mixed equilibrium strategies in a fully distributed fashion. Decentralized RL algorithms are shown to play an important role in allowing devices to be self-organized and reach satisfactory performance with incomplete information or even under uncertainties. We point out through a simulation the importance of D2D relaying and assess how our learning schemes perform under slow/fast channel fading.

Cite

CITATION STYLE

APA

Driouech, S., Sabir, E., Ghogho, M., & Amhoud, E. M. (2021). D2D Mobile Relaying Meets NOMA-Part II: A Reinforcement Learning Perspective. Sensors (Basel, Switzerland), 21(5). https://doi.org/10.3390/s21051755

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free