Margin-Based Modal Adaptive Learning for Visible-Infrared Person Re-Identification

Visible-infrared person re-identification (VIPR) has great potential for intelligent transportation systems for constructing smart cities, but it is challenging to utilize due to the huge modal discrepancy between visible and infrared images. Although visible and infrared data can appear to be two d...

Full description

Bibliographic Details
Main Authors: Qianqian Zhao, Hanxiao Wu, Jianqing Zhu
Format: Article
Language:English
Published: MDPI AG 2023-01-01
Series:Sensors
Subjects:
Online Access:https://www.mdpi.com/1424-8220/23/3/1426
_version_ 1797623199004884992
author Qianqian Zhao
Hanxiao Wu
Jianqing Zhu
author_facet Qianqian Zhao
Hanxiao Wu
Jianqing Zhu
author_sort Qianqian Zhao
collection DOAJ
description Visible-infrared person re-identification (VIPR) has great potential for intelligent transportation systems for constructing smart cities, but it is challenging to utilize due to the huge modal discrepancy between visible and infrared images. Although visible and infrared data can appear to be two domains, VIPR is not identical to domain adaptation as it can massively eliminate modal discrepancies. Because VIPR has complete identity information on both visible and infrared modalities, once the domain adaption is overemphasized, the discriminative appearance information on the visible and infrared domains would drain. For that, we propose a novel margin-based modal adaptive learning (MMAL) method for VIPR in this paper. On each domain, we apply triplet and label smoothing cross-entropy functions to learn appearance-discriminative features. Between the two domains, we design a simple yet effective marginal maximum mean discrepancy (M<inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><msup><mrow></mrow><mn>3</mn></msup></semantics></math></inline-formula>D) loss function to avoid an excessive suppression of modal discrepancies to protect the features’ discriminative ability on each domain. As a result, our MMAL method could learn modal-invariant yet appearance-discriminative features for improving VIPR. The experimental results show that our MMAL method acquires state-of-the-art VIPR performance, e.g., on the RegDB dataset in the visible-to-infrared retrieval mode, the rank-1 accuracy is 93.24% and the mean average precision is 83.77%.
first_indexed 2024-03-11T09:25:21Z
format Article
id doaj.art-806e0f935bb347f8a72afc7b70b3d9f2
institution Directory Open Access Journal
issn 1424-8220
language English
last_indexed 2024-03-11T09:25:21Z
publishDate 2023-01-01
publisher MDPI AG
record_format Article
series Sensors
spelling doaj.art-806e0f935bb347f8a72afc7b70b3d9f22023-11-16T18:00:55ZengMDPI AGSensors1424-82202023-01-01233142610.3390/s23031426Margin-Based Modal Adaptive Learning for Visible-Infrared Person Re-IdentificationQianqian Zhao0Hanxiao Wu1Jianqing Zhu2College of Information Science and Engineering, Huaqiao University, Xiamen 361021, ChinaCollege of Information Science and Engineering, Huaqiao University, Xiamen 361021, ChinaCollege of Engineering, Huaqiao University, Quanzhou 362021, ChinaVisible-infrared person re-identification (VIPR) has great potential for intelligent transportation systems for constructing smart cities, but it is challenging to utilize due to the huge modal discrepancy between visible and infrared images. Although visible and infrared data can appear to be two domains, VIPR is not identical to domain adaptation as it can massively eliminate modal discrepancies. Because VIPR has complete identity information on both visible and infrared modalities, once the domain adaption is overemphasized, the discriminative appearance information on the visible and infrared domains would drain. For that, we propose a novel margin-based modal adaptive learning (MMAL) method for VIPR in this paper. On each domain, we apply triplet and label smoothing cross-entropy functions to learn appearance-discriminative features. Between the two domains, we design a simple yet effective marginal maximum mean discrepancy (M<inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><msup><mrow></mrow><mn>3</mn></msup></semantics></math></inline-formula>D) loss function to avoid an excessive suppression of modal discrepancies to protect the features’ discriminative ability on each domain. As a result, our MMAL method could learn modal-invariant yet appearance-discriminative features for improving VIPR. The experimental results show that our MMAL method acquires state-of-the-art VIPR performance, e.g., on the RegDB dataset in the visible-to-infrared retrieval mode, the rank-1 accuracy is 93.24% and the mean average precision is 83.77%.https://www.mdpi.com/1424-8220/23/3/1426deep learningmaximum mean discrepancyvisible-infrared person re-identification
spellingShingle Qianqian Zhao
Hanxiao Wu
Jianqing Zhu
Margin-Based Modal Adaptive Learning for Visible-Infrared Person Re-Identification
Sensors
deep learning
maximum mean discrepancy
visible-infrared person re-identification
title Margin-Based Modal Adaptive Learning for Visible-Infrared Person Re-Identification
title_full Margin-Based Modal Adaptive Learning for Visible-Infrared Person Re-Identification
title_fullStr Margin-Based Modal Adaptive Learning for Visible-Infrared Person Re-Identification
title_full_unstemmed Margin-Based Modal Adaptive Learning for Visible-Infrared Person Re-Identification
title_short Margin-Based Modal Adaptive Learning for Visible-Infrared Person Re-Identification
title_sort margin based modal adaptive learning for visible infrared person re identification
topic deep learning
maximum mean discrepancy
visible-infrared person re-identification
url https://www.mdpi.com/1424-8220/23/3/1426
work_keys_str_mv AT qianqianzhao marginbasedmodaladaptivelearningforvisibleinfraredpersonreidentification
AT hanxiaowu marginbasedmodaladaptivelearningforvisibleinfraredpersonreidentification
AT jianqingzhu marginbasedmodaladaptivelearningforvisibleinfraredpersonreidentification