Multi-Agent Dynamic Resource Allocation in 6G in-X Subnetworks with Limited Sensing Information

In this paper, we investigate dynamic resource selection in dense deployments of the recent 6G mobile in-X subnetworks (inXSs). We cast resource selection in inXSs as a multi-objective optimization problem involving maximization of the minimum capacity per inXS while minimizing overhead from intra-s...

Full description

Bibliographic Details
Main Authors: Ramoni Adeogun, Gilberto Berardinelli
Format: Article
Language:English
Published: MDPI AG 2022-07-01
Series:Sensors
Subjects:
Online Access:https://www.mdpi.com/1424-8220/22/13/5062
_version_ 1797441799837450240
author Ramoni Adeogun
Gilberto Berardinelli
author_facet Ramoni Adeogun
Gilberto Berardinelli
author_sort Ramoni Adeogun
collection DOAJ
description In this paper, we investigate dynamic resource selection in dense deployments of the recent 6G mobile in-X subnetworks (inXSs). We cast resource selection in inXSs as a multi-objective optimization problem involving maximization of the minimum capacity per inXS while minimizing overhead from intra-subnetwork signaling. Since inXSs are expected to be autonomous, selection decisions are made by each inXS based on its local information without signaling from other inXSs. A multi-agent Q-learning (MAQL) method based on limited sensing information (SI) is then developed, resulting in low intra-subnetwork SI signaling. We further propose a rule-based algorithm termed Q-Heuristics for performing resource selection based on similar limited information as the MAQL method. We perform simulations with a focus on joint channel and transmit power selection. The results indicate that: (1) appropriate settings of Q-learning parameters lead to fast convergence of the MAQL method even with two-level quantization of the SI, and (2) the proposed MAQL approach has significantly better performance and is more robust to sensing and switching delays than the best baseline heuristic. The proposed Q-Heuristic shows similar performance to the baseline greedy method at the 50th percentile of the per-user capacity and slightly better at lower percentiles. The Q-Heuristic method shows high robustness to sensing interval, quantization threshold and switching delay.
first_indexed 2024-03-09T12:31:18Z
format Article
id doaj.art-0bb6749013a143e683c5f16a545d257a
institution Directory Open Access Journal
issn 1424-8220
language English
last_indexed 2024-03-09T12:31:18Z
publishDate 2022-07-01
publisher MDPI AG
record_format Article
series Sensors
spelling doaj.art-0bb6749013a143e683c5f16a545d257a2023-11-30T22:29:00ZengMDPI AGSensors1424-82202022-07-012213506210.3390/s22135062Multi-Agent Dynamic Resource Allocation in 6G in-X Subnetworks with Limited Sensing InformationRamoni Adeogun0Gilberto Berardinelli1Department of Electronic Systems, Aalborg University, 9220 Aalborg, DenmarkDepartment of Electronic Systems, Aalborg University, 9220 Aalborg, DenmarkIn this paper, we investigate dynamic resource selection in dense deployments of the recent 6G mobile in-X subnetworks (inXSs). We cast resource selection in inXSs as a multi-objective optimization problem involving maximization of the minimum capacity per inXS while minimizing overhead from intra-subnetwork signaling. Since inXSs are expected to be autonomous, selection decisions are made by each inXS based on its local information without signaling from other inXSs. A multi-agent Q-learning (MAQL) method based on limited sensing information (SI) is then developed, resulting in low intra-subnetwork SI signaling. We further propose a rule-based algorithm termed Q-Heuristics for performing resource selection based on similar limited information as the MAQL method. We perform simulations with a focus on joint channel and transmit power selection. The results indicate that: (1) appropriate settings of Q-learning parameters lead to fast convergence of the MAQL method even with two-level quantization of the SI, and (2) the proposed MAQL approach has significantly better performance and is more robust to sensing and switching delays than the best baseline heuristic. The proposed Q-Heuristic shows similar performance to the baseline greedy method at the 50th percentile of the per-user capacity and slightly better at lower percentiles. The Q-Heuristic method shows high robustness to sensing interval, quantization threshold and switching delay.https://www.mdpi.com/1424-8220/22/13/50626Greinforcement learningin-X subnetworksresource allocationQ-learningindustrial control
spellingShingle Ramoni Adeogun
Gilberto Berardinelli
Multi-Agent Dynamic Resource Allocation in 6G in-X Subnetworks with Limited Sensing Information
Sensors
6G
reinforcement learning
in-X subnetworks
resource allocation
Q-learning
industrial control
title Multi-Agent Dynamic Resource Allocation in 6G in-X Subnetworks with Limited Sensing Information
title_full Multi-Agent Dynamic Resource Allocation in 6G in-X Subnetworks with Limited Sensing Information
title_fullStr Multi-Agent Dynamic Resource Allocation in 6G in-X Subnetworks with Limited Sensing Information
title_full_unstemmed Multi-Agent Dynamic Resource Allocation in 6G in-X Subnetworks with Limited Sensing Information
title_short Multi-Agent Dynamic Resource Allocation in 6G in-X Subnetworks with Limited Sensing Information
title_sort multi agent dynamic resource allocation in 6g in x subnetworks with limited sensing information
topic 6G
reinforcement learning
in-X subnetworks
resource allocation
Q-learning
industrial control
url https://www.mdpi.com/1424-8220/22/13/5062
work_keys_str_mv AT ramoniadeogun multiagentdynamicresourceallocationin6ginxsubnetworkswithlimitedsensinginformation
AT gilbertoberardinelli multiagentdynamicresourceallocationin6ginxsubnetworkswithlimitedsensinginformation