Improving Scalability of 6G Network Automation with Distributed Deep Q-Networks

Sayantini Majumdar, Leonardo Goratti, Riccardo Trivisonno, Georg Carle

Publikation: Beitrag in FachzeitschriftKonferenzartikelBegutachtung

3 Zitate (Scopus)

Abstract

In recent years, owing to the architectural evolution of 6G towards decentralization, distributed intelligence is being studied extensively for 6G network automation. Distributed intelligence, based on Reinforcement Learning (RL), particularly Q-Learning (QL), has been proposed as a potential direction. The distributed framework consists of independent QL agents, attempting to reach their own individual objectives. The agents need to learn using a sufficient number of training steps before they converge to the optimal performance. After convergence, they can take reliable management actions. However, the scalability of QL could be severely hindered, particularly in the convergence time - when the number of QL agents increases. To overcome the scalability issue of QL, in this paper, we explore the potentials of the Deep Q-Network (DQN) algorithm, a function approximation-based method. Results show that DQN outperforms QL by at least 37% in terms of convergence time. In addition, we highlight that DQN is prone to divergence, which, if solved, could rapidly advance distributed intelligence for 6G.

OriginalspracheEnglisch
Seiten (von - bis)1265-1270
Seitenumfang6
FachzeitschriftProceedings - IEEE Global Communications Conference, GLOBECOM
DOIs
PublikationsstatusVeröffentlicht - 2022
Veranstaltung2022 IEEE Global Communications Conference, GLOBECOM 2022 - Virtual, Online, Brasilien
Dauer: 4 Dez. 20228 Dez. 2022

Fingerprint

Untersuchen Sie die Forschungsthemen von „Improving Scalability of 6G Network Automation with Distributed Deep Q-Networks“. Zusammen bilden sie einen einzigartigen Fingerprint.

Dieses zitieren