Data-Driven Joint Beam Selection and Power Allocation for Multiple Target Tracking
For the problem of joint beam selection and power allocation (JBSPA) for multiple target tracking (MTT), existing works tend to allocate resources only considering the MTT performance at the current tracking time instant. However, in this way, it cannot guarantee the long-term MTT performance in the...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2022-03-01
|
Series: | Remote Sensing |
Subjects: | |
Online Access: | https://www.mdpi.com/2072-4292/14/7/1674 |
_version_ | 1827622573625573376 |
---|---|
author | Yuchun Shi Hao Zheng Kang Li |
author_facet | Yuchun Shi Hao Zheng Kang Li |
author_sort | Yuchun Shi |
collection | DOAJ |
description | For the problem of joint beam selection and power allocation (JBSPA) for multiple target tracking (MTT), existing works tend to allocate resources only considering the MTT performance at the current tracking time instant. However, in this way, it cannot guarantee the long-term MTT performance in the future. If the JBSPA not only considers the tracking performance at the current tracking time instant but also at the future tracking time instant, the allocation results are theoretically able to enhance the long-term tracking performance and the robustness of tracking. Motivated by this, the JBSPA is formulated as a model-free Markov decision process (MDP) problem, and solved with a data-driven method in this article, i.e., deep reinforcement learning (DRL). With DRL, the optimal policy is given by learning from the massive interacting data of the DRL agent and environment. In addition, in order to ensure the information prediction performance of target state in maneuvering target scenarios, a data-driven method is developed based on Long-short term memory (LSTM) incorporating the Gaussian mixture model (GMM), which is called LSTM-GMM for short. This method can realize the state prediction by learning the regularity of nonlinear state transitions of maneuvering targets, where the GMM is used to describe the target motion uncertainty in LSTM. Simulation results have shown the effectiveness of the proposed method. |
first_indexed | 2024-03-09T11:28:17Z |
format | Article |
id | doaj.art-32ea725cf9be450580f0713c4cca15b9 |
institution | Directory Open Access Journal |
issn | 2072-4292 |
language | English |
last_indexed | 2024-03-09T11:28:17Z |
publishDate | 2022-03-01 |
publisher | MDPI AG |
record_format | Article |
series | Remote Sensing |
spelling | doaj.art-32ea725cf9be450580f0713c4cca15b92023-11-30T23:57:20ZengMDPI AGRemote Sensing2072-42922022-03-01147167410.3390/rs14071674Data-Driven Joint Beam Selection and Power Allocation for Multiple Target TrackingYuchun Shi0Hao Zheng1Kang Li 2School of Computer Science and Technology, Zhoukou Normal University, Zhoukou 466001, ChinaNational Laboratory of Radar Signal Processing, Xidian University, Xi’an 710071, ChinaNational Laboratory of Radar Signal Processing, Xidian University, Xi’an 710071, ChinaFor the problem of joint beam selection and power allocation (JBSPA) for multiple target tracking (MTT), existing works tend to allocate resources only considering the MTT performance at the current tracking time instant. However, in this way, it cannot guarantee the long-term MTT performance in the future. If the JBSPA not only considers the tracking performance at the current tracking time instant but also at the future tracking time instant, the allocation results are theoretically able to enhance the long-term tracking performance and the robustness of tracking. Motivated by this, the JBSPA is formulated as a model-free Markov decision process (MDP) problem, and solved with a data-driven method in this article, i.e., deep reinforcement learning (DRL). With DRL, the optimal policy is given by learning from the massive interacting data of the DRL agent and environment. In addition, in order to ensure the information prediction performance of target state in maneuvering target scenarios, a data-driven method is developed based on Long-short term memory (LSTM) incorporating the Gaussian mixture model (GMM), which is called LSTM-GMM for short. This method can realize the state prediction by learning the regularity of nonlinear state transitions of maneuvering targets, where the GMM is used to describe the target motion uncertainty in LSTM. Simulation results have shown the effectiveness of the proposed method.https://www.mdpi.com/2072-4292/14/7/1674multiple target trackingjoint beam selection and power allocationdeep reinforcement learning |
spellingShingle | Yuchun Shi Hao Zheng Kang Li Data-Driven Joint Beam Selection and Power Allocation for Multiple Target Tracking Remote Sensing multiple target tracking joint beam selection and power allocation deep reinforcement learning |
title | Data-Driven Joint Beam Selection and Power Allocation for Multiple Target Tracking |
title_full | Data-Driven Joint Beam Selection and Power Allocation for Multiple Target Tracking |
title_fullStr | Data-Driven Joint Beam Selection and Power Allocation for Multiple Target Tracking |
title_full_unstemmed | Data-Driven Joint Beam Selection and Power Allocation for Multiple Target Tracking |
title_short | Data-Driven Joint Beam Selection and Power Allocation for Multiple Target Tracking |
title_sort | data driven joint beam selection and power allocation for multiple target tracking |
topic | multiple target tracking joint beam selection and power allocation deep reinforcement learning |
url | https://www.mdpi.com/2072-4292/14/7/1674 |
work_keys_str_mv | AT yuchunshi datadrivenjointbeamselectionandpowerallocationformultipletargettracking AT haozheng datadrivenjointbeamselectionandpowerallocationformultipletargettracking AT kangli datadrivenjointbeamselectionandpowerallocationformultipletargettracking |