Deep Multi-Agent Reinforcement Learning With Minimal Cross-Agent Communication for SFC Partitioning

Network Function Virtualization (NFV) decouples network functions from the underlying specialized devices, enabling network processing with higher flexibility and resource efficiency. This promotes the use of virtual network functions (VNFs), which can be grouped to form a service function chain (SF...

Full description

Bibliographic Details
Main Authors: Angelos Pentelas, Danny De Vleeschauwer, Chia-Yu Chang, Koen De Schepper, Panagiotis Papadimitriou
Format: Article
Language:English
Published: IEEE 2023-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10107385/
_version_ 1797838071330242560
author Angelos Pentelas
Danny De Vleeschauwer
Chia-Yu Chang
Koen De Schepper
Panagiotis Papadimitriou
author_facet Angelos Pentelas
Danny De Vleeschauwer
Chia-Yu Chang
Koen De Schepper
Panagiotis Papadimitriou
author_sort Angelos Pentelas
collection DOAJ
description Network Function Virtualization (NFV) decouples network functions from the underlying specialized devices, enabling network processing with higher flexibility and resource efficiency. This promotes the use of virtual network functions (VNFs), which can be grouped to form a service function chain (SFC). A critical challenge in NFV is SFC partitioning (SFCP), which is mathematically expressed as a graph-to-graph mapping problem. Given its NP-hardness, SFCP is commonly solved by approximation methods. Yet, the relevant literature exhibits a gradual shift towards data-driven SFCP frameworks, such as (deep) reinforcement learning (RL). In this article, we initially identify crucial limitations of existing RL-based SFCP approaches. In particular, we argue that most of them stem from the centralized implementation of RL schemes. Therefore, we devise a cooperative deep multi-agent reinforcement learning (DMARL) scheme for decentralized SFCP, which fosters the efficient communication of neighboring agents. Our simulation results (i) demonstrate that DMARL outperforms a state-of-the-art centralized double deep <inline-formula> <tex-math notation="LaTeX">$Q$ </tex-math></inline-formula> -learning algorithm, (ii) unfold the fundamental behaviors learned by the team of agents, (iii) highlight the importance of information exchange between agents, and (iv) showcase the implications stemming from various network topologies on the DMARL efficiency.
first_indexed 2024-04-09T15:34:56Z
format Article
id doaj.art-df15bfa3754243bf8681528473d29de3
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-04-09T15:34:56Z
publishDate 2023-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-df15bfa3754243bf8681528473d29de32023-04-27T23:00:53ZengIEEEIEEE Access2169-35362023-01-0111403844039810.1109/ACCESS.2023.326957610107385Deep Multi-Agent Reinforcement Learning With Minimal Cross-Agent Communication for SFC PartitioningAngelos Pentelas0https://orcid.org/0000-0002-8502-7872Danny De Vleeschauwer1https://orcid.org/0000-0002-0718-8048Chia-Yu Chang2https://orcid.org/0000-0002-7020-8451Koen De Schepper3https://orcid.org/0000-0003-2839-7006Panagiotis Papadimitriou4https://orcid.org/0000-0001-5005-8866Nokia Bell Laboratories, Antwerp, BelgiumNokia Bell Laboratories, Antwerp, BelgiumNokia Bell Laboratories, Antwerp, BelgiumNokia Bell Laboratories, Antwerp, BelgiumDepartment of Applied Informatics, University of Macedonia, Thessaloniki, GreeceNetwork Function Virtualization (NFV) decouples network functions from the underlying specialized devices, enabling network processing with higher flexibility and resource efficiency. This promotes the use of virtual network functions (VNFs), which can be grouped to form a service function chain (SFC). A critical challenge in NFV is SFC partitioning (SFCP), which is mathematically expressed as a graph-to-graph mapping problem. Given its NP-hardness, SFCP is commonly solved by approximation methods. Yet, the relevant literature exhibits a gradual shift towards data-driven SFCP frameworks, such as (deep) reinforcement learning (RL). In this article, we initially identify crucial limitations of existing RL-based SFCP approaches. In particular, we argue that most of them stem from the centralized implementation of RL schemes. Therefore, we devise a cooperative deep multi-agent reinforcement learning (DMARL) scheme for decentralized SFCP, which fosters the efficient communication of neighboring agents. Our simulation results (i) demonstrate that DMARL outperforms a state-of-the-art centralized double deep <inline-formula> <tex-math notation="LaTeX">$Q$ </tex-math></inline-formula> -learning algorithm, (ii) unfold the fundamental behaviors learned by the team of agents, (iii) highlight the importance of information exchange between agents, and (iv) showcase the implications stemming from various network topologies on the DMARL efficiency.https://ieeexplore.ieee.org/document/10107385/Multi-agent reinforcement learningnetwork function virtualizationself-learning orchestration
spellingShingle Angelos Pentelas
Danny De Vleeschauwer
Chia-Yu Chang
Koen De Schepper
Panagiotis Papadimitriou
Deep Multi-Agent Reinforcement Learning With Minimal Cross-Agent Communication for SFC Partitioning
IEEE Access
Multi-agent reinforcement learning
network function virtualization
self-learning orchestration
title Deep Multi-Agent Reinforcement Learning With Minimal Cross-Agent Communication for SFC Partitioning
title_full Deep Multi-Agent Reinforcement Learning With Minimal Cross-Agent Communication for SFC Partitioning
title_fullStr Deep Multi-Agent Reinforcement Learning With Minimal Cross-Agent Communication for SFC Partitioning
title_full_unstemmed Deep Multi-Agent Reinforcement Learning With Minimal Cross-Agent Communication for SFC Partitioning
title_short Deep Multi-Agent Reinforcement Learning With Minimal Cross-Agent Communication for SFC Partitioning
title_sort deep multi agent reinforcement learning with minimal cross agent communication for sfc partitioning
topic Multi-agent reinforcement learning
network function virtualization
self-learning orchestration
url https://ieeexplore.ieee.org/document/10107385/
work_keys_str_mv AT angelospentelas deepmultiagentreinforcementlearningwithminimalcrossagentcommunicationforsfcpartitioning
AT dannydevleeschauwer deepmultiagentreinforcementlearningwithminimalcrossagentcommunicationforsfcpartitioning
AT chiayuchang deepmultiagentreinforcementlearningwithminimalcrossagentcommunicationforsfcpartitioning
AT koendeschepper deepmultiagentreinforcementlearningwithminimalcrossagentcommunicationforsfcpartitioning
AT panagiotispapadimitriou deepmultiagentreinforcementlearningwithminimalcrossagentcommunicationforsfcpartitioning