Multi-agent deep reinforcement learning to enable dynamic TDD in a multi-cell environment

Boutiba, Karim; Bagaa, Miloud; Ksentini, Adlen
IEEE Transactions on Mobile Computing, 15 September 2023

Dynamic Time Division Duplex (D-TDD) is a promising solution to address newly emerging 5 G and 6 G services characterized by asymmetric and dynamic uplink (UL) and downlink (DL) traffic demands. However, there are two major issues: (i) determining the TDD scheme (i.e., the number of slots devoted to UL and DL) to meet the dynamic traffic demands of the Users Equipment (UE); (ii) cross-link interference between cells that use different TDD schemes. The 3GPP standard neither specifies algorithms or solutions to derive the TDD configuration nor solves the cross-link interference. To fill this gap, we model the dynamic TDD problem in 5 G NR as a linear programming problem. Then, we design Multi-Agent Deep Reinforcement Learning based 5 G RAN TDD Pattern (MADRP), a fully decentralized solution based on the Multi-Agent Deep Reinforcement Learning (MADRL) approach. Based on the simulation results, the algorithm effectively prevents buffer overflows, avoids cross-link interference, and adapts to changes in the traffic pattern, ensuring its versatility. We compared our solution with the optimal solution and different static TDD configurations.. We found that MADRP outperforms the static TDD configurations. We finally discuss the algorithm's limitations in terms of the number of cells, traffic variance, and cross-link interference probability.


DOI
Type:
Journal
Date:
2023-09-15
Department:
Systèmes de Communication
Eurecom Ref:
7438
Copyright:
© 2023 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.

PERMALINK : https://www.eurecom.fr/publication/7438