A Parallel Convolutional Neural Network for Pedestrian Detection

Pedestrian detection is a crucial task in many vision-based applications, such as video<br />surveillance, human activity analysis and autonomous driving. Recently, most of the existing<br />pedestrian detection frameworks only focus on the detection accuracy or model parameters.<br /...

Full description

Bibliographic Details
Main Authors: Mengya Zhu, Yiquan Wu
Format: Article
Language:English
Published: MDPI AG 2020-09-01
Series:Electronics
Subjects:
Online Access:https://www.mdpi.com/2079-9292/9/9/1478
id doaj-8ccb8090890f46f3b398c14aa6c7b582
record_format Article
spelling doaj-8ccb8090890f46f3b398c14aa6c7b5822020-11-25T03:11:49ZengMDPI AGElectronics2079-92922020-09-0191478147810.3390/electronics9091478A Parallel Convolutional Neural Network for Pedestrian DetectionMengya Zhu0Yiquan Wu1College of Electronic and Information Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 211106, ChinaCollege of Electronic and Information Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 211106, ChinaPedestrian detection is a crucial task in many vision-based applications, such as video<br />surveillance, human activity analysis and autonomous driving. Recently, most of the existing<br />pedestrian detection frameworks only focus on the detection accuracy or model parameters.<br />However, how to balance the detection accuracy and model parameters, is still an open problem for<br />the practical application of pedestrian detection. In this paper, we propose a parallel, lightweight<br />framework for pedestrian detection, named ParallelNet. ParallelNet consists of four branches, each<br />of them learns different high-level semantic features. We fused them into one feature map as the<br />final feature representation. Subsequently, the Fire module, which includes Squeeze and Expand<br />parts, is employed for reducing the model parameters. Here, we replace some convolution modules<br />in the backbone with Fire modules. Finally, the focal loss is led into the ParallelNet for end-to-end<br />training. Experimental results on the Caltech–Zhang dataset and KITTI dataset show that:<br />Compared with the single-branch network, such as ResNet and SqueezeNet, ParallelNet has<br />improved detection accuracy with fewer model parameters and lower Giga Floating Point<br />Operations (GFLOPs).https://www.mdpi.com/2079-9292/9/9/1478pedestrian detectionconvolutional neural networkfeature extractionFire modulefocal loss
collection DOAJ
language English
format Article
sources DOAJ
author Mengya Zhu
Yiquan Wu
spellingShingle Mengya Zhu
Yiquan Wu
A Parallel Convolutional Neural Network for Pedestrian Detection
Electronics
pedestrian detection
convolutional neural network
feature extraction
Fire module
focal loss
author_facet Mengya Zhu
Yiquan Wu
author_sort Mengya Zhu
title A Parallel Convolutional Neural Network for Pedestrian Detection
title_short A Parallel Convolutional Neural Network for Pedestrian Detection
title_full A Parallel Convolutional Neural Network for Pedestrian Detection
title_fullStr A Parallel Convolutional Neural Network for Pedestrian Detection
title_full_unstemmed A Parallel Convolutional Neural Network for Pedestrian Detection
title_sort parallel convolutional neural network for pedestrian detection
publisher MDPI AG
series Electronics
issn 2079-9292
publishDate 2020-09-01
description Pedestrian detection is a crucial task in many vision-based applications, such as video<br />surveillance, human activity analysis and autonomous driving. Recently, most of the existing<br />pedestrian detection frameworks only focus on the detection accuracy or model parameters.<br />However, how to balance the detection accuracy and model parameters, is still an open problem for<br />the practical application of pedestrian detection. In this paper, we propose a parallel, lightweight<br />framework for pedestrian detection, named ParallelNet. ParallelNet consists of four branches, each<br />of them learns different high-level semantic features. We fused them into one feature map as the<br />final feature representation. Subsequently, the Fire module, which includes Squeeze and Expand<br />parts, is employed for reducing the model parameters. Here, we replace some convolution modules<br />in the backbone with Fire modules. Finally, the focal loss is led into the ParallelNet for end-to-end<br />training. Experimental results on the Caltech–Zhang dataset and KITTI dataset show that:<br />Compared with the single-branch network, such as ResNet and SqueezeNet, ParallelNet has<br />improved detection accuracy with fewer model parameters and lower Giga Floating Point<br />Operations (GFLOPs).
topic pedestrian detection
convolutional neural network
feature extraction
Fire module
focal loss
url https://www.mdpi.com/2079-9292/9/9/1478
work_keys_str_mv AT mengyazhu aparallelconvolutionalneuralnetworkforpedestriandetection
AT yiquanwu aparallelconvolutionalneuralnetworkforpedestriandetection
AT mengyazhu parallelconvolutionalneuralnetworkforpedestriandetection
AT yiquanwu parallelconvolutionalneuralnetworkforpedestriandetection
_version_ 1724652745031417856