Scale-Sensitive Feature Reassembly Network for Pedestrian Detection
Serious scale variation is a key challenge in pedestrian detection. Most works typically employ a feature pyramid network to detect objects at diverse scales. Such a method suffers from information loss during channel unification. Inadequate sampling of the backbone network also affects the power of...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2021-06-01
|
Series: | Sensors |
Subjects: | |
Online Access: | https://www.mdpi.com/1424-8220/21/12/4189 |
_version_ | 1797529693716480000 |
---|---|
author | Xiaoting Yang Qiong Liu |
author_facet | Xiaoting Yang Qiong Liu |
author_sort | Xiaoting Yang |
collection | DOAJ |
description | Serious scale variation is a key challenge in pedestrian detection. Most works typically employ a feature pyramid network to detect objects at diverse scales. Such a method suffers from information loss during channel unification. Inadequate sampling of the backbone network also affects the power of pyramidal features. Moreover, an arbitrary RoI (region of interest) allocation scheme of these detectors incurs coarse RoI representation, which becomes worse under the dilemma of small pedestrian relative scale (PRS). In this paper, we propose a novel scale-sensitive feature reassembly network (SSNet) for pedestrian detection in road scenes. Specifically, a multi-parallel branch sampling module is devised with flexible receptive fields and an adjustable anchor stride to improve the sensitivity to pedestrians imaged at multiple scales. Meanwhile, a context enhancement fusion module is also proposed to alleviate information loss by injecting various spatial context information into the original features. For more accurate prediction, an adaptive reassembly strategy is designed to obtain recognizable RoI features in the proposal refinement stage. Extensive experiments are conducted on CityPersons and Caltech datasets to demonstrate the effectiveness of our method. The detection results show that our SSNet surpasses the baseline method significantly by integrating lightweight modules and achieves competitive performance with other methods without bells and whistles. |
first_indexed | 2024-03-10T10:17:24Z |
format | Article |
id | doaj.art-3f9554682fc14669a67c8fbb7d510d7f |
institution | Directory Open Access Journal |
issn | 1424-8220 |
language | English |
last_indexed | 2024-03-10T10:17:24Z |
publishDate | 2021-06-01 |
publisher | MDPI AG |
record_format | Article |
series | Sensors |
spelling | doaj.art-3f9554682fc14669a67c8fbb7d510d7f2023-11-22T00:42:01ZengMDPI AGSensors1424-82202021-06-012112418910.3390/s21124189Scale-Sensitive Feature Reassembly Network for Pedestrian DetectionXiaoting Yang0Qiong Liu1School of Software Engineering, South China University of Technology, Guangzhou 510006, ChinaSchool of Software Engineering, South China University of Technology, Guangzhou 510006, ChinaSerious scale variation is a key challenge in pedestrian detection. Most works typically employ a feature pyramid network to detect objects at diverse scales. Such a method suffers from information loss during channel unification. Inadequate sampling of the backbone network also affects the power of pyramidal features. Moreover, an arbitrary RoI (region of interest) allocation scheme of these detectors incurs coarse RoI representation, which becomes worse under the dilemma of small pedestrian relative scale (PRS). In this paper, we propose a novel scale-sensitive feature reassembly network (SSNet) for pedestrian detection in road scenes. Specifically, a multi-parallel branch sampling module is devised with flexible receptive fields and an adjustable anchor stride to improve the sensitivity to pedestrians imaged at multiple scales. Meanwhile, a context enhancement fusion module is also proposed to alleviate information loss by injecting various spatial context information into the original features. For more accurate prediction, an adaptive reassembly strategy is designed to obtain recognizable RoI features in the proposal refinement stage. Extensive experiments are conducted on CityPersons and Caltech datasets to demonstrate the effectiveness of our method. The detection results show that our SSNet surpasses the baseline method significantly by integrating lightweight modules and achieves competitive performance with other methods without bells and whistles.https://www.mdpi.com/1424-8220/21/12/4189pedestrian detectionscale variationfeature fusionRoI featureroad scene |
spellingShingle | Xiaoting Yang Qiong Liu Scale-Sensitive Feature Reassembly Network for Pedestrian Detection Sensors pedestrian detection scale variation feature fusion RoI feature road scene |
title | Scale-Sensitive Feature Reassembly Network for Pedestrian Detection |
title_full | Scale-Sensitive Feature Reassembly Network for Pedestrian Detection |
title_fullStr | Scale-Sensitive Feature Reassembly Network for Pedestrian Detection |
title_full_unstemmed | Scale-Sensitive Feature Reassembly Network for Pedestrian Detection |
title_short | Scale-Sensitive Feature Reassembly Network for Pedestrian Detection |
title_sort | scale sensitive feature reassembly network for pedestrian detection |
topic | pedestrian detection scale variation feature fusion RoI feature road scene |
url | https://www.mdpi.com/1424-8220/21/12/4189 |
work_keys_str_mv | AT xiaotingyang scalesensitivefeaturereassemblynetworkforpedestriandetection AT qiongliu scalesensitivefeaturereassemblynetworkforpedestriandetection |