Illumination-aware hallucination-based domain adaptation for thermal pedestrian detection

<p>Thermal imagery is emerging as a viable candidate for 24-7, all-weather pedestrian detection owning to thermal sensors&rsquo; robust performance for pedestrian detection under different weather and illumination conditions. Despite the promising results obtained from combining visible (R...

Full description

Bibliographic Details
Main Authors: Xie, Q, Cheng, T-Y, Dai, Z, Tran, V, Trigoni, A, Markham, A
Format: Journal article
Language:English
Published: IEEE 2023
_version_ 1797112021225832448
author Xie, Q
Cheng, T-Y
Dai, Z
Tran, V
Trigoni, A
Markham, A
author_facet Xie, Q
Cheng, T-Y
Dai, Z
Tran, V
Trigoni, A
Markham, A
author_sort Xie, Q
collection OXFORD
description <p>Thermal imagery is emerging as a viable candidate for 24-7, all-weather pedestrian detection owning to thermal sensors&rsquo; robust performance for pedestrian detection under different weather and illumination conditions. Despite the promising results obtained from combining visible (RGB) and thermal cameras in multi-spectral fusion techniques, the complex synchronization requirements, including alignment and calibration of sensors, impede their deployment in real-world scenarios. In this paper, we introduce a novel approach for domain adaptation to enhance the performance of pedestrian detection based solely on thermal images. Our proposed approach involves several stages. Firstly, we use both thermal and visible images as input during the training phase. Secondly, we leverage a thermal-to-visible hallucination network to generate feature maps that are similar to those generated by the visible branch. Finally, we design a transformer-based multi-modal fusion module to integrate the hallucinated visible and thermal information more effectively. The thermal-to-visible hallucination network acts as domain adaptation, allowing us to obtain pseudo-visual and thermal features using solely thermal input. Based on the experimental results, it is observed the mean average precision (mAP) increases by 4.72% and the miss rate decreases by 7.56% on the KAIST dataset when compared to the baseline model.</p>
first_indexed 2024-03-07T08:18:27Z
format Journal article
id oxford-uuid:419524a7-d50d-4983-a2f2-3472951ac614
institution University of Oxford
language English
last_indexed 2024-03-07T08:18:27Z
publishDate 2023
publisher IEEE
record_format dspace
spelling oxford-uuid:419524a7-d50d-4983-a2f2-3472951ac6142024-01-19T07:03:55ZIllumination-aware hallucination-based domain adaptation for thermal pedestrian detectionJournal articlehttp://purl.org/coar/resource_type/c_dcae04bcuuid:419524a7-d50d-4983-a2f2-3472951ac614EnglishSymplectic ElementsIEEE2023Xie, QCheng, T-YDai, ZTran, VTrigoni, AMarkham, A<p>Thermal imagery is emerging as a viable candidate for 24-7, all-weather pedestrian detection owning to thermal sensors&rsquo; robust performance for pedestrian detection under different weather and illumination conditions. Despite the promising results obtained from combining visible (RGB) and thermal cameras in multi-spectral fusion techniques, the complex synchronization requirements, including alignment and calibration of sensors, impede their deployment in real-world scenarios. In this paper, we introduce a novel approach for domain adaptation to enhance the performance of pedestrian detection based solely on thermal images. Our proposed approach involves several stages. Firstly, we use both thermal and visible images as input during the training phase. Secondly, we leverage a thermal-to-visible hallucination network to generate feature maps that are similar to those generated by the visible branch. Finally, we design a transformer-based multi-modal fusion module to integrate the hallucinated visible and thermal information more effectively. The thermal-to-visible hallucination network acts as domain adaptation, allowing us to obtain pseudo-visual and thermal features using solely thermal input. Based on the experimental results, it is observed the mean average precision (mAP) increases by 4.72% and the miss rate decreases by 7.56% on the KAIST dataset when compared to the baseline model.</p>
spellingShingle Xie, Q
Cheng, T-Y
Dai, Z
Tran, V
Trigoni, A
Markham, A
Illumination-aware hallucination-based domain adaptation for thermal pedestrian detection
title Illumination-aware hallucination-based domain adaptation for thermal pedestrian detection
title_full Illumination-aware hallucination-based domain adaptation for thermal pedestrian detection
title_fullStr Illumination-aware hallucination-based domain adaptation for thermal pedestrian detection
title_full_unstemmed Illumination-aware hallucination-based domain adaptation for thermal pedestrian detection
title_short Illumination-aware hallucination-based domain adaptation for thermal pedestrian detection
title_sort illumination aware hallucination based domain adaptation for thermal pedestrian detection
work_keys_str_mv AT xieq illuminationawarehallucinationbaseddomainadaptationforthermalpedestriandetection
AT chengty illuminationawarehallucinationbaseddomainadaptationforthermalpedestriandetection
AT daiz illuminationawarehallucinationbaseddomainadaptationforthermalpedestriandetection
AT tranv illuminationawarehallucinationbaseddomainadaptationforthermalpedestriandetection
AT trigonia illuminationawarehallucinationbaseddomainadaptationforthermalpedestriandetection
AT markhama illuminationawarehallucinationbaseddomainadaptationforthermalpedestriandetection