Multi-Agent Dynamic Resource Allocation in 6G in-X Subnetworks with Limited Sensing Information
In this paper, we investigate dynamic resource selection in dense deployments of the recent 6G mobile in-X subnetworks (inXSs). We cast resource selection in inXSs as a multi-objective optimization problem involving maximization of the minimum capacity per inXS while minimizing overhead from intra-s...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2022-07-01
|
Series: | Sensors |
Subjects: | |
Online Access: | https://www.mdpi.com/1424-8220/22/13/5062 |
_version_ | 1797441799837450240 |
---|---|
author | Ramoni Adeogun Gilberto Berardinelli |
author_facet | Ramoni Adeogun Gilberto Berardinelli |
author_sort | Ramoni Adeogun |
collection | DOAJ |
description | In this paper, we investigate dynamic resource selection in dense deployments of the recent 6G mobile in-X subnetworks (inXSs). We cast resource selection in inXSs as a multi-objective optimization problem involving maximization of the minimum capacity per inXS while minimizing overhead from intra-subnetwork signaling. Since inXSs are expected to be autonomous, selection decisions are made by each inXS based on its local information without signaling from other inXSs. A multi-agent Q-learning (MAQL) method based on limited sensing information (SI) is then developed, resulting in low intra-subnetwork SI signaling. We further propose a rule-based algorithm termed Q-Heuristics for performing resource selection based on similar limited information as the MAQL method. We perform simulations with a focus on joint channel and transmit power selection. The results indicate that: (1) appropriate settings of Q-learning parameters lead to fast convergence of the MAQL method even with two-level quantization of the SI, and (2) the proposed MAQL approach has significantly better performance and is more robust to sensing and switching delays than the best baseline heuristic. The proposed Q-Heuristic shows similar performance to the baseline greedy method at the 50th percentile of the per-user capacity and slightly better at lower percentiles. The Q-Heuristic method shows high robustness to sensing interval, quantization threshold and switching delay. |
first_indexed | 2024-03-09T12:31:18Z |
format | Article |
id | doaj.art-0bb6749013a143e683c5f16a545d257a |
institution | Directory Open Access Journal |
issn | 1424-8220 |
language | English |
last_indexed | 2024-03-09T12:31:18Z |
publishDate | 2022-07-01 |
publisher | MDPI AG |
record_format | Article |
series | Sensors |
spelling | doaj.art-0bb6749013a143e683c5f16a545d257a2023-11-30T22:29:00ZengMDPI AGSensors1424-82202022-07-012213506210.3390/s22135062Multi-Agent Dynamic Resource Allocation in 6G in-X Subnetworks with Limited Sensing InformationRamoni Adeogun0Gilberto Berardinelli1Department of Electronic Systems, Aalborg University, 9220 Aalborg, DenmarkDepartment of Electronic Systems, Aalborg University, 9220 Aalborg, DenmarkIn this paper, we investigate dynamic resource selection in dense deployments of the recent 6G mobile in-X subnetworks (inXSs). We cast resource selection in inXSs as a multi-objective optimization problem involving maximization of the minimum capacity per inXS while minimizing overhead from intra-subnetwork signaling. Since inXSs are expected to be autonomous, selection decisions are made by each inXS based on its local information without signaling from other inXSs. A multi-agent Q-learning (MAQL) method based on limited sensing information (SI) is then developed, resulting in low intra-subnetwork SI signaling. We further propose a rule-based algorithm termed Q-Heuristics for performing resource selection based on similar limited information as the MAQL method. We perform simulations with a focus on joint channel and transmit power selection. The results indicate that: (1) appropriate settings of Q-learning parameters lead to fast convergence of the MAQL method even with two-level quantization of the SI, and (2) the proposed MAQL approach has significantly better performance and is more robust to sensing and switching delays than the best baseline heuristic. The proposed Q-Heuristic shows similar performance to the baseline greedy method at the 50th percentile of the per-user capacity and slightly better at lower percentiles. The Q-Heuristic method shows high robustness to sensing interval, quantization threshold and switching delay.https://www.mdpi.com/1424-8220/22/13/50626Greinforcement learningin-X subnetworksresource allocationQ-learningindustrial control |
spellingShingle | Ramoni Adeogun Gilberto Berardinelli Multi-Agent Dynamic Resource Allocation in 6G in-X Subnetworks with Limited Sensing Information Sensors 6G reinforcement learning in-X subnetworks resource allocation Q-learning industrial control |
title | Multi-Agent Dynamic Resource Allocation in 6G in-X Subnetworks with Limited Sensing Information |
title_full | Multi-Agent Dynamic Resource Allocation in 6G in-X Subnetworks with Limited Sensing Information |
title_fullStr | Multi-Agent Dynamic Resource Allocation in 6G in-X Subnetworks with Limited Sensing Information |
title_full_unstemmed | Multi-Agent Dynamic Resource Allocation in 6G in-X Subnetworks with Limited Sensing Information |
title_short | Multi-Agent Dynamic Resource Allocation in 6G in-X Subnetworks with Limited Sensing Information |
title_sort | multi agent dynamic resource allocation in 6g in x subnetworks with limited sensing information |
topic | 6G reinforcement learning in-X subnetworks resource allocation Q-learning industrial control |
url | https://www.mdpi.com/1424-8220/22/13/5062 |
work_keys_str_mv | AT ramoniadeogun multiagentdynamicresourceallocationin6ginxsubnetworkswithlimitedsensinginformation AT gilbertoberardinelli multiagentdynamicresourceallocationin6ginxsubnetworkswithlimitedsensinginformation |