Multi-Level Fusion Temporal–Spatial Co-Attention for Video-Based Person Re-Identification
A convolutional neural network can easily fall into local minima for insufficient data, and the needed training is unstable. Many current methods are used to solve these problems by adding pedestrian attributes, pedestrian postures, and other auxiliary information, but they require additional collec...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2021-12-01
|
Series: | Entropy |
Subjects: | |
Online Access: | https://www.mdpi.com/1099-4300/23/12/1686 |
_version_ | 1797504886986768384 |
---|---|
author | Shengyu Pei Xiaoping Fan |
author_facet | Shengyu Pei Xiaoping Fan |
author_sort | Shengyu Pei |
collection | DOAJ |
description | A convolutional neural network can easily fall into local minima for insufficient data, and the needed training is unstable. Many current methods are used to solve these problems by adding pedestrian attributes, pedestrian postures, and other auxiliary information, but they require additional collection, which is time-consuming and laborious. Every video sequence frame has a different degree of similarity. In this paper, multi-level fusion temporal–spatial co-attention is adopted to improve person re-identification (reID). For a small dataset, the improved network can better prevent over-fitting and reduce the dataset limit. Specifically, the concept of knowledge evolution is introduced into video-based person re-identification to improve the backbone residual neural network (ResNet). The global branch, local branch, and attention branch are used in parallel for feature extraction. Three high-level features are embedded in the metric learning network to improve the network’s generalization ability and the accuracy of video-based person re-identification. Simulation experiments are implemented on small datasets PRID2011 and iLIDS-VID, and the improved network can better prevent over-fitting. Experiments are also implemented on MARS and DukeMTMC-VideoReID, and the proposed method can be used to extract more feature information and improve the network’s generalization ability. The results show that our method achieves better performance. The model achieves 90.15% Rank1 and 81.91% mAP on MARS. |
first_indexed | 2024-03-10T04:10:47Z |
format | Article |
id | doaj.art-1044c09cd5a44d0b9841cba1f8946eea |
institution | Directory Open Access Journal |
issn | 1099-4300 |
language | English |
last_indexed | 2024-03-10T04:10:47Z |
publishDate | 2021-12-01 |
publisher | MDPI AG |
record_format | Article |
series | Entropy |
spelling | doaj.art-1044c09cd5a44d0b9841cba1f8946eea2023-11-23T08:11:41ZengMDPI AGEntropy1099-43002021-12-012312168610.3390/e23121686Multi-Level Fusion Temporal–Spatial Co-Attention for Video-Based Person Re-IdentificationShengyu Pei0Xiaoping Fan1School of Automation, Central South University, Changsha 410075, ChinaSchool of Automation, Central South University, Changsha 410075, ChinaA convolutional neural network can easily fall into local minima for insufficient data, and the needed training is unstable. Many current methods are used to solve these problems by adding pedestrian attributes, pedestrian postures, and other auxiliary information, but they require additional collection, which is time-consuming and laborious. Every video sequence frame has a different degree of similarity. In this paper, multi-level fusion temporal–spatial co-attention is adopted to improve person re-identification (reID). For a small dataset, the improved network can better prevent over-fitting and reduce the dataset limit. Specifically, the concept of knowledge evolution is introduced into video-based person re-identification to improve the backbone residual neural network (ResNet). The global branch, local branch, and attention branch are used in parallel for feature extraction. Three high-level features are embedded in the metric learning network to improve the network’s generalization ability and the accuracy of video-based person re-identification. Simulation experiments are implemented on small datasets PRID2011 and iLIDS-VID, and the improved network can better prevent over-fitting. Experiments are also implemented on MARS and DukeMTMC-VideoReID, and the proposed method can be used to extract more feature information and improve the network’s generalization ability. The results show that our method achieves better performance. The model achieves 90.15% Rank1 and 81.91% mAP on MARS.https://www.mdpi.com/1099-4300/23/12/1686video-based person re-identificationmulti-level fusiontemporal–spatial co-attentionknowledge evolution |
spellingShingle | Shengyu Pei Xiaoping Fan Multi-Level Fusion Temporal–Spatial Co-Attention for Video-Based Person Re-Identification Entropy video-based person re-identification multi-level fusion temporal–spatial co-attention knowledge evolution |
title | Multi-Level Fusion Temporal–Spatial Co-Attention for Video-Based Person Re-Identification |
title_full | Multi-Level Fusion Temporal–Spatial Co-Attention for Video-Based Person Re-Identification |
title_fullStr | Multi-Level Fusion Temporal–Spatial Co-Attention for Video-Based Person Re-Identification |
title_full_unstemmed | Multi-Level Fusion Temporal–Spatial Co-Attention for Video-Based Person Re-Identification |
title_short | Multi-Level Fusion Temporal–Spatial Co-Attention for Video-Based Person Re-Identification |
title_sort | multi level fusion temporal spatial co attention for video based person re identification |
topic | video-based person re-identification multi-level fusion temporal–spatial co-attention knowledge evolution |
url | https://www.mdpi.com/1099-4300/23/12/1686 |
work_keys_str_mv | AT shengyupei multilevelfusiontemporalspatialcoattentionforvideobasedpersonreidentification AT xiaopingfan multilevelfusiontemporalspatialcoattentionforvideobasedpersonreidentification |