Data-Driven Joint Beam Selection and Power Allocation for Multiple Target Tracking

For the problem of joint beam selection and power allocation (JBSPA) for multiple target tracking (MTT), existing works tend to allocate resources only considering the MTT performance at the current tracking time instant. However, in this way, it cannot guarantee the long-term MTT performance in the...

Full description

Bibliographic Details
Main Authors: Yuchun Shi, Hao Zheng, Kang Li 
Format: Article
Language:English
Published: MDPI AG 2022-03-01
Series:Remote Sensing
Subjects:
Online Access:https://www.mdpi.com/2072-4292/14/7/1674
_version_ 1827622573625573376
author Yuchun Shi
Hao Zheng
Kang Li 
author_facet Yuchun Shi
Hao Zheng
Kang Li 
author_sort Yuchun Shi
collection DOAJ
description For the problem of joint beam selection and power allocation (JBSPA) for multiple target tracking (MTT), existing works tend to allocate resources only considering the MTT performance at the current tracking time instant. However, in this way, it cannot guarantee the long-term MTT performance in the future. If the JBSPA not only considers the tracking performance at the current tracking time instant but also at the future tracking time instant, the allocation results are theoretically able to enhance the long-term tracking performance and the robustness of tracking. Motivated by this, the JBSPA is formulated as a model-free Markov decision process (MDP) problem, and solved with a data-driven method in this article, i.e., deep reinforcement learning (DRL). With DRL, the optimal policy is given by learning from the massive interacting data of the DRL agent and environment. In addition, in order to ensure the information prediction performance of target state in maneuvering target scenarios, a data-driven method is developed based on Long-short term memory (LSTM) incorporating the Gaussian mixture model (GMM), which is called LSTM-GMM for short. This method can realize the state prediction by learning the regularity of nonlinear state transitions of maneuvering targets, where the GMM is used to describe the target motion uncertainty in LSTM. Simulation results have shown the effectiveness of the proposed method.
first_indexed 2024-03-09T11:28:17Z
format Article
id doaj.art-32ea725cf9be450580f0713c4cca15b9
institution Directory Open Access Journal
issn 2072-4292
language English
last_indexed 2024-03-09T11:28:17Z
publishDate 2022-03-01
publisher MDPI AG
record_format Article
series Remote Sensing
spelling doaj.art-32ea725cf9be450580f0713c4cca15b92023-11-30T23:57:20ZengMDPI AGRemote Sensing2072-42922022-03-01147167410.3390/rs14071674Data-Driven Joint Beam Selection and Power Allocation for Multiple Target TrackingYuchun Shi0Hao Zheng1Kang Li 2School of Computer Science and Technology, Zhoukou Normal University, Zhoukou 466001, ChinaNational Laboratory of Radar Signal Processing, Xidian University, Xi’an 710071, ChinaNational Laboratory of Radar Signal Processing, Xidian University, Xi’an 710071, ChinaFor the problem of joint beam selection and power allocation (JBSPA) for multiple target tracking (MTT), existing works tend to allocate resources only considering the MTT performance at the current tracking time instant. However, in this way, it cannot guarantee the long-term MTT performance in the future. If the JBSPA not only considers the tracking performance at the current tracking time instant but also at the future tracking time instant, the allocation results are theoretically able to enhance the long-term tracking performance and the robustness of tracking. Motivated by this, the JBSPA is formulated as a model-free Markov decision process (MDP) problem, and solved with a data-driven method in this article, i.e., deep reinforcement learning (DRL). With DRL, the optimal policy is given by learning from the massive interacting data of the DRL agent and environment. In addition, in order to ensure the information prediction performance of target state in maneuvering target scenarios, a data-driven method is developed based on Long-short term memory (LSTM) incorporating the Gaussian mixture model (GMM), which is called LSTM-GMM for short. This method can realize the state prediction by learning the regularity of nonlinear state transitions of maneuvering targets, where the GMM is used to describe the target motion uncertainty in LSTM. Simulation results have shown the effectiveness of the proposed method.https://www.mdpi.com/2072-4292/14/7/1674multiple target trackingjoint beam selection and power allocationdeep reinforcement learning
spellingShingle Yuchun Shi
Hao Zheng
Kang Li 
Data-Driven Joint Beam Selection and Power Allocation for Multiple Target Tracking
Remote Sensing
multiple target tracking
joint beam selection and power allocation
deep reinforcement learning
title Data-Driven Joint Beam Selection and Power Allocation for Multiple Target Tracking
title_full Data-Driven Joint Beam Selection and Power Allocation for Multiple Target Tracking
title_fullStr Data-Driven Joint Beam Selection and Power Allocation for Multiple Target Tracking
title_full_unstemmed Data-Driven Joint Beam Selection and Power Allocation for Multiple Target Tracking
title_short Data-Driven Joint Beam Selection and Power Allocation for Multiple Target Tracking
title_sort data driven joint beam selection and power allocation for multiple target tracking
topic multiple target tracking
joint beam selection and power allocation
deep reinforcement learning
url https://www.mdpi.com/2072-4292/14/7/1674
work_keys_str_mv AT yuchunshi datadrivenjointbeamselectionandpowerallocationformultipletargettracking
AT haozheng datadrivenjointbeamselectionandpowerallocationformultipletargettracking
AT kangli datadrivenjointbeamselectionandpowerallocationformultipletargettracking