Scale-Sensitive Feature Reassembly Network for Pedestrian Detection

Serious scale variation is a key challenge in pedestrian detection. Most works typically employ a feature pyramid network to detect objects at diverse scales. Such a method suffers from information loss during channel unification. Inadequate sampling of the backbone network also affects the power of...

Full description

Bibliographic Details
Main Authors: Xiaoting Yang, Qiong Liu
Format: Article
Language:English
Published: MDPI AG 2021-06-01
Series:Sensors
Subjects:
Online Access:https://www.mdpi.com/1424-8220/21/12/4189
_version_ 1797529693716480000
author Xiaoting Yang
Qiong Liu
author_facet Xiaoting Yang
Qiong Liu
author_sort Xiaoting Yang
collection DOAJ
description Serious scale variation is a key challenge in pedestrian detection. Most works typically employ a feature pyramid network to detect objects at diverse scales. Such a method suffers from information loss during channel unification. Inadequate sampling of the backbone network also affects the power of pyramidal features. Moreover, an arbitrary RoI (region of interest) allocation scheme of these detectors incurs coarse RoI representation, which becomes worse under the dilemma of small pedestrian relative scale (PRS). In this paper, we propose a novel scale-sensitive feature reassembly network (SSNet) for pedestrian detection in road scenes. Specifically, a multi-parallel branch sampling module is devised with flexible receptive fields and an adjustable anchor stride to improve the sensitivity to pedestrians imaged at multiple scales. Meanwhile, a context enhancement fusion module is also proposed to alleviate information loss by injecting various spatial context information into the original features. For more accurate prediction, an adaptive reassembly strategy is designed to obtain recognizable RoI features in the proposal refinement stage. Extensive experiments are conducted on CityPersons and Caltech datasets to demonstrate the effectiveness of our method. The detection results show that our SSNet surpasses the baseline method significantly by integrating lightweight modules and achieves competitive performance with other methods without bells and whistles.
first_indexed 2024-03-10T10:17:24Z
format Article
id doaj.art-3f9554682fc14669a67c8fbb7d510d7f
institution Directory Open Access Journal
issn 1424-8220
language English
last_indexed 2024-03-10T10:17:24Z
publishDate 2021-06-01
publisher MDPI AG
record_format Article
series Sensors
spelling doaj.art-3f9554682fc14669a67c8fbb7d510d7f2023-11-22T00:42:01ZengMDPI AGSensors1424-82202021-06-012112418910.3390/s21124189Scale-Sensitive Feature Reassembly Network for Pedestrian DetectionXiaoting Yang0Qiong Liu1School of Software Engineering, South China University of Technology, Guangzhou 510006, ChinaSchool of Software Engineering, South China University of Technology, Guangzhou 510006, ChinaSerious scale variation is a key challenge in pedestrian detection. Most works typically employ a feature pyramid network to detect objects at diverse scales. Such a method suffers from information loss during channel unification. Inadequate sampling of the backbone network also affects the power of pyramidal features. Moreover, an arbitrary RoI (region of interest) allocation scheme of these detectors incurs coarse RoI representation, which becomes worse under the dilemma of small pedestrian relative scale (PRS). In this paper, we propose a novel scale-sensitive feature reassembly network (SSNet) for pedestrian detection in road scenes. Specifically, a multi-parallel branch sampling module is devised with flexible receptive fields and an adjustable anchor stride to improve the sensitivity to pedestrians imaged at multiple scales. Meanwhile, a context enhancement fusion module is also proposed to alleviate information loss by injecting various spatial context information into the original features. For more accurate prediction, an adaptive reassembly strategy is designed to obtain recognizable RoI features in the proposal refinement stage. Extensive experiments are conducted on CityPersons and Caltech datasets to demonstrate the effectiveness of our method. The detection results show that our SSNet surpasses the baseline method significantly by integrating lightweight modules and achieves competitive performance with other methods without bells and whistles.https://www.mdpi.com/1424-8220/21/12/4189pedestrian detectionscale variationfeature fusionRoI featureroad scene
spellingShingle Xiaoting Yang
Qiong Liu
Scale-Sensitive Feature Reassembly Network for Pedestrian Detection
Sensors
pedestrian detection
scale variation
feature fusion
RoI feature
road scene
title Scale-Sensitive Feature Reassembly Network for Pedestrian Detection
title_full Scale-Sensitive Feature Reassembly Network for Pedestrian Detection
title_fullStr Scale-Sensitive Feature Reassembly Network for Pedestrian Detection
title_full_unstemmed Scale-Sensitive Feature Reassembly Network for Pedestrian Detection
title_short Scale-Sensitive Feature Reassembly Network for Pedestrian Detection
title_sort scale sensitive feature reassembly network for pedestrian detection
topic pedestrian detection
scale variation
feature fusion
RoI feature
road scene
url https://www.mdpi.com/1424-8220/21/12/4189
work_keys_str_mv AT xiaotingyang scalesensitivefeaturereassemblynetworkforpedestriandetection
AT qiongliu scalesensitivefeaturereassemblynetworkforpedestriandetection