Next Article in Journal
An Efficient Conformal Stacked Antenna Array Design and 3D-Beamforming for UAV and Space Vehicle Communications
Next Article in Special Issue
Unsupervised Trademark Retrieval Method Based on Attention Mechanism
Previous Article in Journal
Enhanced Accuracy of Airborne Volcanic Ash Detection Using the GEOKOMPSAT-2A Satellite
Previous Article in Special Issue
Compressing Deep Networks by Neuron Agglomerative Clustering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multipath Fusion Strategy Based Single Shot Detector †

1
Department of Computer Science, University of Liverpool, Liverpool L69 7ZX, UK
2
School of Advanced Technology, Xi’an Jiaotong-Liverpool University, Suzhou 215123, China
3
School of Computing, Edinburgh Napier University, Edinburgh EH11 4BN, UK
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper: Qu, S.; Huang, K.; Hussain, A.; Goulermas, Y. MPSSD: Multi-Path Fusion Single Shot Detector. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019.
Submission received: 7 December 2020 / Revised: 27 January 2021 / Accepted: 9 February 2021 / Published: 15 February 2021

Abstract

:
Object detection has wide applications in intelligent systems and sensor applications. Compared with two stage detectors, recent one stage counterparts are capable of running more efficiently with comparable accuracy, which satisfy the requirement of real-time processing. To further improve the accuracy of one stage single shot detector (SSD), we propose a novel Multi-Path fusion Single Shot Detector (MPSSD). Different from other feature fusion methods, we exploit the connection among different scale representations in a pyramid manner. We propose feature fusion module to generate new feature pyramids based on multiscale features in SSD, and these pyramids are sent to our pyramid aggregation module for generating final features. These enhanced features have both localization and semantics information, thus improving the detection performance with little computation cost. A series of experiments on three benchmark datasets PASCAL VOC2007, VOC2012, and MS COCO demonstrate that our approach outperforms many state-of-the-art detectors both qualitatively and quantitatively. In particular, for input images with size 512 × 512, our method attains mean Average Precision (mAP) of 81.8% on VOC2007 t e s t , 80.3% on VOC2012 t e s t , and 33.1% mAP on COCO t e s t - d e v 2015.

1. Introduction

Object detection has been a fundamental task in computer vision which aims at localizing objects via bounding boxes and assigning a certain class to each of them. It has been widely adopted on intelligent systems as a crucial component and applied on specific purposes, such as pedestrian detection, face detection, and text detection [1]. Recent deep learning based detectors [2,3,4,5,6,7] have seen great achievements compared with traditional methods. Inspired by these works, we propose a novel efficient yet accurate deep detector. We extend previous work [8] substantially by presenting comprehensive related work, performing a series of additional experiments with more ablation studies, comments on inference time, and failure cases analysis which better show the effectiveness of our method.
Most state-of-the-art object detectors fall into two categories, one-stage detectors and two-stage ones. Two-stage methods are proposed earlier, especially for Faster R-CNN (Regions with CNN features) [4] and R-FCN (Regions with Fully Convolutional Networks) [7]. On the first stage, regions with high probability that contain foregrounds are extracted as proposals. On the second stage, these proposals are sent to a network for classification and bounding-box regression. While archiving high accuracy on benchmark datasets, such as PASCAL VOC [9] and MS COCO [10], two-stage approaches usually run slowly because of the high computation cost of the region proposal generation. Another one-stage branch (e.g., Single Shot Detector (SSD) [6], You Only Look Once (YOLO) [5]) employs proposal-free pipeline, thus it needs few computation resources. These detectors usually adopt fully convolutional architecture and calculate class confidence and regression results directly on predefined boxes. One-stage detectors could run efficiently with accuracy slightly inferior to two-stage ones, thus they are favorable for sensors system due to the crucial need of real-time inference.
The one-stage approach SSD detects objects directly from multiscale features. This feature pyramid consists of last layers of backbone and adjacent convolutional layers. Shallow layers within this pyramid are used to detect small scale objects and deep ones responsible for larger objects. Although efficient, such scheme behaves badly on small objects. Since shallow layers always learn localization information while deep layers have more semantics information [11]. The semantics are crucial for detecting small objects, thus exploiting shallow layers alone is not enough for small scales detection.
To address this issue, some recent works [12,13] introduce a top-down pyramid structure. In order to pass semantics information to shallow layers, they upsample deeper layers before combining them with lower ones. Although borrowing some semantics from deep layers, there is only one feature pyramid engaged. We believe one single pyramid is still not informative enough for accurate detection. Thus, we propose a multipath model which consists of several feature pyramids in order to learn the most informative representations. Although novel in deep object detector, the strategy of multipath is widely applied in computer vision areas [14,15]. Inspired by Feature Fusion Single Shot Multibox Detector (FSSD) [16], we use Feature Fusion Module (FFM) to obtain fused features from base pyramid and obtain our multipath feature pyramids. Then we generate our final features by sending these pyramid features to our Pyramid Aggregation Module (PAM). At last, these informative multiscale features are fed into detection heads for final processing. We conduct extensive experiments on challenging datasets PASCAL VOC and MS COCO, and the results show that our algorithm is better than most stage-of-the-art one-stage object detectors. Below are our main contributions:
  • We propose a multipath fusion strategy to enhance the feature pyramid in single shot detector;
  • Our feature fusion module and pyramid aggregation module are introduced that proves able to fuse information from base pyramid and generate our informative pyramid efficiently;
  • Extensive experiments on PASCAL VOC 2007 & 2012 and MS COCO 2015 show our proposed Multi-Path fusion Single Shot Detector (MPSSD) can outperform other one-stage detectors consistently.

2. Related Work

We present the related work by dividing it into three parts. We will first describe the deep learning based object detector, followed by an introduction of single shot detector branch. Finally, we will discuss the feature enhancement in deep detectors.

2.1. Deep Object Detector

Traditional object detection methods always rely on hand-crafted features [17,18,19,20,21]. Histogram of Oriented Gradients (HOG) [18] is a representative feature descriptor which can be calculated on densely uniformed cells. Deformable Part-based Model (DPM) is an improved version of HOG which follows the “divide and conquer” scheme to detect an image on different parts during inference [21]. Traditional methods are limited by the hand-crafted features and less efficient computation resources. Recently, thanks to the great achievements of deep convolutional neural networks on computer vision [22,23,24,25,26,27,28,29], many deep learning based detectors are proposed with superior performance compared with conventional methods. R-CNN [2] introduces the idea of region proposals, which are regions with high probabilities to include objects. They extract these proposals at first and send them to CNN for further prediction. This two-stage scheme becomes popular due to its superiority over conventional hand-crafted methods on both accuracy and speed. Fast R-CNN [3] accelerates the training of R-CNN through a novel multitask loss to train the classification and bounding box regression simultaneously. To alleviate the heavy computation cost of proposals generation process in these algorithms, faster R-CNN [4] designs a learning-based region proposal network (RPN) to generate proposals efficiently. R-FCN [7] adopt fully-convolutional networks, and they propose the position sensitive RoI Pooling (PSRoI) to replace the RoI Pooling in faster R-CNN to further improving the accuracy and efficiency. To solve the extreme foreground-background classes imbalance in training, focal loss [30] as an variant of the standard cross entropy loss is proposed to learn more hard examples during training.

2.2. Single Shot Detector

Apart from region proposal based detectors, another branch termed as one stage detectors abandons the procedure of proposals extraction in order to achieve faster inference. Among these methods, YOLO [5] probably presents the first one-stage detector which applies successfully classification and bounding box regression directly on each predefined image grid. The drop of proposal generation results in a very high speed, but the accuracy is relatively low. YOLO’s improved versions [31,32] have focused more on this problem, especially for small objects detection.
Another one-stage detector is called SSD [6] that exploits a multiscale fashion to predict objects with various scales, and this improves the performance of one-stage detector significantly. We adopt SSD as our base model since it satisfies the trade-off between speed and accuracy. At first, adopting VGG16 as backbone, SSD modifies the last fully connected layers into convolutional version. For an input image with size 300 × 300, they extract layers conv4_3 with size 38 × 38 and conv_fc7 19 × 19 from backbone. Then, they add several convolution layers conv8_2, conv9_2, conv10_2 and conv11_2 to extract features with size 10 × 10, 5 × 5, 3 × 3 and 1 × 1 respectively. The extracted features establish the detection pyramid in SSD, and each feature is responsible for detecting a particular scale. The shallow layer, i.e., conv4_3, is to detect relative small objects in the image, while the deep layer, i.e., conv9_2, is used to detect large objects. On each feature, they design a number of defined boxes and each box is assumed to include foregrounds. The adjacent head block of each feature is to predict the offsets and classification confidences for these boxes. Finally, a postprocessing called non-maximum suppression (NMS) is used to refine these predicted results.

2.3. Deep Feature Enhancement

The quality of feature representations is important for object detection. In order to improve the accuracy of deep detectors, many recent works study how to build more informative features. Feature Pyramid Network (FPN) [12] adopts the top-down feature pyramid design to add lower level features with higher ones. This straightforward design passes higher level semantics to lower levels, thus improving the detection performance especially on small objects. Deconvolutional Single Shot Detector (DSSD) [13] applies the same strategy into vanilla SSD architecture and archieves better performance, but their deconvolution design would hurt the high-efficient property of the original SSD. HyperNet [33] and ION [34] combine the hierarchical features into one layer before the prediction, aiming for building the most representative feature which enjoys both local and global information. In terms of feature enhancement in one-stage detectors, inspired by FPN, StairNet [35] improves the nearest neighbor upsampling with a newly introduced top-down feature combining module and attains even better results. RFBNet [36] proposes a novel Receptive Field Block (RFB) with groups of dilated convolution design, and this module is able to build high-quality representatives through enlarging the receptive field of input features. Some other SSD based methods try to enhance the base features through fusion of different layers with their novel fusion modules [16,37,38]. In this paper, we propose a new multipath fusion strategy to generate multilevel feature pyramids. We illustrate that our SSD based object detector could outperform many counterparts both qualitatively and quantitatively on three benchmark datasets.

3. Methodology

Before introducing our novel architecture and modules, we revisit one-stage detector SSD and its variant FSSD. We elaborate our methodology in detail and then offer interpretations accordingly.

3.1. Deep Feature Pyramid

Single shot multibox object detector often extracts features from one backbone network and adds several convolution layers to obtain feature pyramid. We denote the feature of the l t h layer is x l R H × W × C (where H, W, and C refer to height, width, channel respectively). The feature pyramid in SSD for multiscale detection is defined as:
X D e t e c t i o n = { x k , x k + 1 , . . . , x L } ,
where x l is used to predict objects within a certain range of scale. Specifically, the shallowest feature x k responsible for detecting small objects within inputs, and larger objects are detected in deeper layers. However, as the feature goes deeper, the more semantics it contains, which could provide global information. Shallow features missing this global guidance would lead to misdetection. Thus, SSD has poor accuracy on relatively small objects.
To resolve this problem, many investigations focus on building more informative representations. For instance, the straightforward way is adding higher level semantics to low level local features [13]. However, this strategy may need much heavy computation owning to the complicated element-wise operations between adjacent features. An efficient method is proposed by FSSD [16] to fuse low level and high level information. They concatenate several features and generate the following pyramid the same as SSD. The whole process can be shown as below:
X D e t e c t i o n = { x k , x k + 1 , . . . , x L } ,
and each level feature is computed by:
x l = F ( { x k , x k + 1 , . . . , x k + n } ) ,
where F indicates the operation of feature fusion and pyramid generation, and the number of base feature from SSD to be fused is ( n + 1 ) . We can find in (3) only base features from k to k + n are used to generate final feature pyramid, while the rest of the base features with rich semantics are discarded. We argue that these features can also be used to generate final pyramid through a dedicated transformation. Thus, we design a multipath routine to maximize the usage of base features. Firstly, we generate several groups of features on base pyramid through feature fusion modules. Then, we pass these pyramids into our aggregation module for final fusion. The detection results are predicted on the final enhanced pyramid:
X D e t e c t i o n = { x k , x k + 1 , . . . , x L } ,
and each feature is expressed by:
x L = A ( { F 1 ( { x k , x k + 1 , . . . , x k + 1 + n } ) } , . . . , { F M ( { x k + M 1 , x k + M , . . . , x k + M + n } ) } ) ,
where M is the number of pyramid paths. We apply several feature fusion modules, which is represented by F m and pyramid aggregation module indicated by A. More details about these two modules can be found in Section 3.3 and Section 3.4.
The generated final feature pyramid is rich in both localization and semantic information at each level. For instance, the last layer of this pyramid is fused from same size features in each path, and these features are enhanced through nonlinear transformations of different level base features. Our fusion strategy could generate features rich in local and global clues and it is lightweight to be inserted in vanilla SSD model.

3.2. MPSSD Architecture

As shown in Figure 1, our architecture is based on single-stage design in vanilla SSD. The backbone network we use is VGG16. Same as that in SSD, we replace last fc6 and fc7 layers with convolution layers for a fully convolutional fashion. Other base features are generated from the following several convolution layers. These base features are then sent to two modules for enhancement. The enhanced feature pyramid with both local and global information is connected with detection heads for prediction. We illustrate the modules in the following sections.

3.3. Feature Fusion Module

In order to better integrate various features with different scales, FSSD engages an efficient fusion method. Inspired by the success, we also exploit this module so as to generate our novel multipath feature pyramids. The structure is shown in Figure 2. For an input image of size 300 × 300, we choose layers from conv_fc7, conv6_2 and conv7_2 to fuse. Since these features are in different sizes (size of conv_fc7 is 19 × 19, conv6_2 is 10 × 10, conv7_2 is 5 × 5), smaller features conv6_2 and conv7_2 are upsampled to the same size of conv_fc7 before fusion. We also utilize 1 × 1 convolution before upsampling to reduce the channels into a particular number, and different channel numbers are set in our FFMs. Next, we concatenate these same size features in channel dimension. The feature pyramid is then generate through several 3 × 3 conv + Rectified Linear Unit (ReLU) layers. For multipath pyramids, we repeat this process several times.

3.4. Pyramid Aggregation Module

We devise a novel pyramid aggregation module to transform multipath pyramid from feature fusion modules into final detection pyramid. This module is illustrated in the bottom of Figure 1.
The whole process can be divided into two steps. Firstly, we aim to arrange these features into a single pyramid for further detection, thus same scale features are concatenated along the channel dimension. It is noticeable that each path pyramid has a different number of features. On the second step, we adopt attention block in Squeeze-and-Excitation Networks (SENet) [39] to enhance the concatenated features. This attention block is implemented through squeeze and excitation steps. We show its procedure in the following. Firstly on squeeze step, the global average pooling is used to generate z R C for input X R H × W × C :
z c = 1 H × W i = 1 H j = 1 W X c ( i , j ) .
Then, the excitation step generates the attention activation feature s by
s = F e x ( z , W ) = σ ( W 2 δ ( W 1 z ) ) ,
where σ and δ refer to ReLU and sigmoid operation respectively, W 1 R C × C and W 2 R C × C , we adopt C = 1 16 C in our experiments. The last stage outputs the result X ˜ R H × W × C by
X ˜ c = F s c a l e ( X , s c ) = s c · X c .

4. Results and Discussion

We conduct extensive experiments on three benchmarks PASCAL VOC2007, PASCAL VOC2012, and MS COCO. As for the evaluation metrics, in VOC, predicted boxes that have Intersection over Union (IoU) with the ground truth higher than 0.5 are defined as positive results. In COCO, the metrics are split into different parts based on different Intersection over Union (IoU) settings. Our implementation is based on Pytorch 0.4.0. Backbone network VGG16 is pretrained on ImageNet [40] with modification into fully convolutional version. As for the training settings, we keep them mostly the same as the settings in SSD for fair comparison. On the other hand, we adopt the network initial method from [41].

4.1. Dataset

We applied two popular datasets, PASCAL VOC and MS COCO, during training and testing. The PASCAL VOC dataset contains images collected from the Internet. The 2007 and 2012 versions are two most used in object detection. They are categorized into 20 classes since 2007 based on four groups: person, animal, vehicle, and indoor. MS COCO is a newly established dataset for object detection, segmentation, and captioning. Compared with VOC, MS COCO adopts annotations not only for the bounding box, but for each object it is labeled with instance segmentation for more precise localization. This dataset is more challenging and with more than 200,000 images categorized into 80 classes, and there are more labeled objects per image compared with PASCAL VOC. Table 1 shows the detailed statistics of these datasets. For intuitive observations, we also show some examples from these two datasets in Figure 3.

4.2. PASCAL VOC 2007

All the experiments on PASCAL VOC apply the popular split, which uses the union of VOC2007 t r a i n v a l and VOC2012 t r a i n v a l as the training data, and VOC2007 t e s t which contains 4952 images for testing. For training on inputs with scale 300 × 300 , the batch size is set at 24 on single NVIDIA 1080Ti GPU. The initial learning rate is 4 × 10 3 . The learning rate decreases to 4 × 10 4 at 150 epoch and to 4 × 10 5 at 200 epoch. We finish our training on 250 epoch.
Table 2 shows the results on VOC2007. Our model outperforms vanilla SSD significantly. As for 300 × 300 inputs, by embedding our algorithm into SSD, the mAP (mean Average Precision) improves from 77.5% to 80.3%, while increases from 79.5% to 81.8% on 512 × 512 inputs. We also archive better performance compared with DSSD with ResNet101 [22] as backbone. We find FSSD with similar fusion strategy but with less features is inferior to our method. Based on the mAP, we archive 1.5% higher on 300 × 300 and 0.9% higher on 512 × 512, and this show our multipath design is effective. Compared with top-down design, such as StairNet, the 1.5 point higher accuracy proves our architecture helps to learn more powerful features. Our method also reaches better performance compared with these two-stage algorithms.

4.3. PASCAL VOC 2012

We adopt the same training setting as that in our VOC2007 experiments to evaluate our model on dataset VOC2012 t e s t , which contains 10,991 images in total. From Table 3, our method achieved the same results as the evaluation on VOC2007 dataset. For both input resolutions, our model improves significantly from the SSD baseline. Compared with other single-stage methods, our MPSSD reached the best performance under the same training settings. It is noteworthy that FSSD obtained better results via adopting additional training dataset (MS COCO). These results again show the advantages of our proposed method.

4.4. MS COCO

We also conduct experiments on a more challenging dataset, MS COCO, for further evaluation. For the training and validation split, we adopt t r a i n v a l 35 k for training, m i n i v a l for validation, and t e s t - d e v 2015 for testing. For input on 300 × 300, our batch size is 16 on single NVIDIA 1080Ti GPU. The learning rate is initialized at at 2 × 10 3 with schedule decay on epoch 90 and 120 by the factor of 10. The training ends at 150 epoch.
Shown in Table 4, the evaluation metrics are acquired through official server on CodaLab. For 300 × 300 input images, our model outperforms SSD in a large margin, mAP improves from 25.1% to 27.5%. Compared with FSSD, our model obtains 0.4% higher accuracy, which shows our effectiveness. For 512 × 512 input images, our model archives 33.1% accuracy, which is 4.3% and 1.3% higher than SSD and FSSD, respectively. Our method is only 0.1% inferior to DSSD513, and we believe this mainly because they adopt a more powerful backbone ResNet101. Specifically, we can find ours works best on small scale detection among these algorithms.

4.5. Inference Time Analysis

We take into consideration the inference time of our algorithms because of the real-time requirement of sensor system. To compare fairly, for 4952 images in VOC 2007 t e s t with size 300 × 300, we average their total inference time without considering the postprocessing (NMS) step. All the results are evaluated on one NVIDIA 1080Ti GPU with batch size set at 1, and these methods are trained under the same settings.
We analyze how our newly introduced modules affect the inference speed. The results are shown in Figure 4. Compared with vanilla SSD, our methods could reach a much higher accuracy with little sacrifice of inference time. In addition, in contrast to FSSD, our multiple design of FFMs and Pyramid Aggregation Module (PAM) would not spend too much computation with better performance. We also compared one-stage baseline Faster RCNN, and our method could again achieve higher accuracy with much less time cost. Since other one-stage methods are based on Faster RCNN baseline, we do not make the comparisons for simplicity. Finally, we test how the number of Feature Fusion Module (FFM) affects the detect time. From Figure 4, we find that one more FFM could reduce the inference speed a little, and thus it is essential to select the optimal parameters.

4.6. Qualitative Evaluation

In this part, we show some qualitative results on VOC 2007 t e s t . Seen from each example in Figure 5, the result of SSD is in the upper row and ours is in the lower row. Boxes with classification confidences higher than 0.6 are chosen and each category is labeled on the top of box.
For the first example, the potted plant with low illumination is misdetected in SSD, while our model could handle this illumination variance and detect it successfully. The second example shows our model works better on detecting small objects, such as cows with this image. The following three examples indicate our model outperforms baseline on occlusion objects detection. For the last example, the sofa is neglected by SSD, while it is detected by our model. This proves our learned semantics help to infer objects under certain environment.

4.7. Ablation Study on Feature Fusion Modules

To show the effectiveness of our module, we conduct experiments on different settings of feature fusion modules. We conduct all the experiments on VOC 2007 t e s t with 300 × 300 inputs but we should bear in mind that a similar conclusion can also be made in the other datasets.
We run experiments with different numbers of FFMs and different choices of channel numbers. As seen from Table 5, the results on the first column come from vanilla SSD. Through adding FFMs, we could find the accuracy improves a lot progressively. The model with three FFMs and with channel equals to 128 on third FFM reaches the highest accuracy, and we choose this as our final model.

4.8. Failure Cases Analysis

In this part, we analyze the shortcomings of our method. Figure 6 shows some examples with misdetection or incorrect classification from COCO 2015 t e s t . In Figure 6a, the wine glass held by the man is not detected, and we conjecture that the failure is due to the object being tilted with a certain degree, which is not the same as its normal state. On the contrary, the wine glass held by the woman vertically is well-recognized. This case has shown that our model is unable to detect objects with abnormal positions. In Figure 6b, the top keyboard is classified into laptop incorrectly. Although the object “laptop” consists of “keyboard” and “screen”, this incorrect classification is probably because the model is incapable of learning the distinguishable features of these two objects. Finally, in Figure 6c, the green signpost behind the man with yellow coat is misdetected. Since this object is in abnormal aspect ratio (tall and slender), this kind of shape is hard to be recognized mainly because of the weakness of anchor design scheme. We adopt vanilla SSD anchor design which may not be able to cover objects with arbitrary shape, especially for objects with large aspect ratio.

4.9. Future Work

We consider three aspects for the future improvements. First, to improve the accuracy of our current model, we believe further optimizing of feature pyramid could reach a more accurate detection. Precisely speaking, the current method builds a fixed number of pyramids, which may only optimize on low resolution inputs, i.e., 300 × 300. With rapid improvements on camera devices, high resolution detection is highly required and thus we need to redesign an optimal strategy. Second, to speed up the inference time of our model, we could focus on alleviating the tedious computation on detection head. The predictions on defined boxes are redundant due to the original design in SSD, and this hurts the speed a great deal. Finally, we will extend the proposed model in various other domains of intelligent systems and sensor applications. As a one-shot detector both efficiently and accurately, the proposed multipath fusion strategy based algorithm could find its applications extensively.

5. Conclusions

We devise a novel SSD based one stage detector. Different from prior methods, we propose a multipath fusion strategy to fully utilize the local and global information from different layers. In order to aggregate this information, we first use feature fusion modules to generate feature pyramids, followed by pyramid aggregation module to fuse and enhance relative features. The generated pyramid is suitable for detection in multiscale manner with rich information. Quantitative and qualitative experiments on PASCAL VOC and MS COCO show our MPSSD outperforms vanilla SSD significantly. While for other single stage counterparts, our method archives comparable results without hurting efficiency.

Author Contributions

Conceptualization, S.Q. and K.H.; Formal analysis, S.Q. and K.H.; Funding acquisition, K.H.; Investigation, A.H.; Methodology, S.Q. and K.H.; Project administration, S.Q. and K.H.; Resources, Y.G.; Software, S.Q.; Supervision, A.H. and Y.G.; Validation, S.Q.; Visualization, S.Q.; Writing—original draft, S.Q.; Writing—review and editing, K.H., A.H. and Y.G. All authors have read and agreed to the published version of the manuscript.

Funding

The work was partially supported by the following: National Natural Science Foundation of China under no. 61876155; The Natural Science Foundation of the Jiangsu Higher Education Institutions of China under no. 17KJD520010; Suzhou Science and Technology Program under no. SYG201712, SZS201613; Natural Science Foundation of Jiangsu Province BK20181189; Key Program Special Fund in XJTLU under no. KSF-A-01, KSF-P-02, and KSF-E-26.

Data Availability Statement

Acknowledgments

The authors would like to thank the guest editors and the anonymous reviewers for their work and time on the publication of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, L.; Ouyang, W.; Wang, X.; Fieguth, P.; Chen, J.; Liu, X.; Pietikäinen, M. Deep learning for generic object detection: A survey. Int. J. Comput. Vis. 2020, 128, 261–318. [Google Scholar]
  2. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 24–27 June 2014; pp. 580–587. [Google Scholar]
  3. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  4. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 779–788. [Google Scholar]
  6. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  7. Dai, J.; Li, Y.; He, K.; Sun, J. R-fcn: Object detection via region-based fully convolutional networks. arXiv 2016, arXiv:1605.06409. [Google Scholar]
  8. Qu, S.; Huang, K.; Hussain, A.; Goulermas, Y. MPSSD: Multi-Path Fusion Single Shot Detector. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019; pp. 1–6. [Google Scholar]
  9. Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef] [Green Version]
  10. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar]
  11. Ghodrati, A.; Diba, A.; Pedersoli, M.; Tuytelaars, T.; Van Gool, L. Deepproposal: Hunting objects by cascading deep convolutional layers. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 13–16 December 2015; pp. 2578–2586. [Google Scholar]
  12. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  13. Fu, C.Y.; Liu, W.; Ranga, A.; Tyagi, A.; Berg, A.C. DSSD: Deconvolutional single shot detector. arXiv 2017, arXiv:1701.06659. [Google Scholar]
  14. Eigen, D.; Fergus, R. Predicting depth, surface normals and semantic labels with a common multi-scale convolutional architecture. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 13–16 December 2015; pp. 2650–2658. [Google Scholar]
  15. Ochoa-Zezzatti, A.; Mejía, J.; Contreras-Masse, R.; Martínez, E.; Hernández, A. Ethnic Characterization in Amalgamated People for Airport Security Using a Repository of Images and Pigeon-Inspired Optimization (PIO) Algorithm for the Improvement of Their Results. In Applications of Hybrid Metaheuristic Algorithms for Image Processing; Springer: Berlin/Heidelberg, Germany, 2020; pp. 105–119. [Google Scholar]
  16. Li, Z.; Zhou, F. FSSD: Feature Fusion Single Shot Multibox Detector. arXiv 2017, arXiv:1712.00960. [Google Scholar]
  17. Csurka, G.; Dance, C.; Fan, L.; Willamowski, J.; Bray, C. Visual categorization with bags of keypoints. In Workshop on Statistical Learning in Computer Vision; 004; ECCV: Prague, Gzech Republic; Volume 1, pp. 1–2. Available online: https://people.eecs.berkeley.edu/~efros/courses/AP06/Papers/csurka-eccv-04.pdf (accessed on 15 February 2021).
  18. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  19. Wang, X.; Han, T.X.; Yan, S. An HOG-LBP human detector with partial occlusion handling. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 32–39. [Google Scholar]
  20. Perronnin, F.; Sánchez, J.; Mensink, T. Improving the fisher kernel for large-scale image classification. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2010; pp. 143–156. [Google Scholar]
  21. Felzenszwalb, P.F.; Girshick, R.B.; McAllester, D. Cascade object detection with deformable part models. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 2241–2248. [Google Scholar]
  22. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  23. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  24. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  25. Huang, K.; Hussain, A.; Wang, Q.F.; Zhang, R. Deep Learning: Fundamentals, Theory and Applications; Springer: Berlin/Heidelberg, Germany, 2019; Volume 2. [Google Scholar]
  26. Jiang, C.; Xiao, J.; Xie, Y.; Tillo, T.; Huang, K. Siamese network ensemble for visual tracking. Neurocomputing 2018, 275, 2892–2903. [Google Scholar] [CrossRef]
  27. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  28. Zhong, G.; Yan, S.; Huang, K.; Cai, Y.; Dong, J. Reducing and stretching deep convolutional activation features for accurate image classification. Cogn. Comput. 2018, 10, 179–186. [Google Scholar] [CrossRef]
  29. Yang, X.; Huang, K.; Zhang, R.; Goulermas, J.Y. A Novel Deep Density Model for Unsupervised Learning. Cogn. Comput. 2019, 11, 778–788. [Google Scholar] [CrossRef]
  30. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  31. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  32. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  33. Kong, T.; Yao, A.; Chen, Y.; Sun, F. Hypernet: Towards accurate region proposal generation and joint object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 845–853. [Google Scholar]
  34. Bell, S.; Lawrence Zitnick, C.; Bala, K.; Girshick, R. Inside-outside net: Detecting objects in context with skip pooling and recurrent neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2874–2883. [Google Scholar]
  35. Woo, S.; Hwang, S.; Kweon, I.S. Stairnet: Top-down semantic aggregation for accurate one shot detection. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1093–1102. [Google Scholar]
  36. Liu, S.; Huang, D. Receptive field block net for accurate and fast object detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 385–400. [Google Scholar]
  37. Pang, Y.; Wang, T.; Anwer, R.M.; Khan, F.S.; Shao, L. Efficient featurized image pyramid network for single shot detector. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 7336–7344. [Google Scholar]
  38. Zhao, Q.; Sheng, T.; Wang, Y.; Tang, Z.; Chen, Y.; Cai, L.; Ling, H. M2det: A single-shot object detector based on multi-level feature pyramid network. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 9259–9266. [Google Scholar]
  39. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  40. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  41. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 13–16 December 2015; pp. 1026–1034. [Google Scholar]
  42. Hariharan, B.; Arbeláez, P.; Bourdev, L.; Maji, S.; Malik, J. Semantic contours from inverse detectors. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 991–998. [Google Scholar]
Figure 1. Our model pipeline. From top to bottom, each part indicates: Single Shot Detector (SSD) base feature pyramid, multipath feature pyramids generated through Feature Fusion Modules, final feature pyramid out from pyramid aggregation module.
Figure 1. Our model pipeline. From top to bottom, each part indicates: Single Shot Detector (SSD) base feature pyramid, multipath feature pyramids generated through Feature Fusion Modules, final feature pyramid out from pyramid aggregation module.
Sensors 21 01360 g001
Figure 2. Design of feature fusion module. Top are base features. Three features are selected to be fused. These features are computed through 1 × 1 conv and upsampled into same scale and channel. Following by concatenation to form the aggregation feature. Bottom shows the generated new features from the aggregation.
Figure 2. Design of feature fusion module. Top are base features. Three features are selected to be fused. These features are computed through 1 × 1 conv and upsampled into same scale and channel. Following by concatenation to form the aggregation feature. Bottom shows the generated new features from the aggregation.
Sensors 21 01360 g002
Figure 3. Some object detection examples in Pascal VOC (upper) and MS COCO (lower).
Figure 3. Some object detection examples in Pascal VOC (upper) and MS COCO (lower).
Sensors 21 01360 g003
Figure 4. Comparison on detect time of each image. In the figure, 3FFMs indicates the model which adopts 3 feature fusion modules, while 4FFMs applies 4 feature fusion modules.
Figure 4. Comparison on detect time of each image. In the figure, 3FFMs indicates the model which adopts 3 feature fusion modules, while 4FFMs applies 4 feature fusion modules.
Sensors 21 01360 g004
Figure 5. Some detection results from VOC 2007 t e s t . Upper: SSD. Lower: Ours. See more details in Section 4.6. Adapted from [8].
Figure 5. Some detection results from VOC 2007 t e s t . Upper: SSD. Lower: Ours. See more details in Section 4.6. Adapted from [8].
Sensors 21 01360 g005
Figure 6. Some failure cases. The results (ac) are all from MS COCO 2015 t e s t .
Figure 6. Some failure cases. The results (ac) are all from MS COCO 2015 t e s t .
Sensors 21 01360 g006
Table 1. Dataset statistics of PASCAL Visual Object Classes (VOC) & Microsoft Common Objects in COntext (MS COCO).
Table 1. Dataset statistics of PASCAL Visual Object Classes (VOC) & Microsoft Common Objects in COntext (MS COCO).
NameTrain ImagesValidation ImagesTest ImagesCategoryTotal Annotated Objects
VOC 2007 [9]250125104952209963
VOC 2012 [9]5717582310,9912011,530
MS COCO 2015 [10]82,78341,620125,436802,500,000
Table 2. PASCAL VOC 2007 t e s t results. Faster R-CNN (Regions with CNN features), HyperNet, Inside-Outside Net (ION) and R-FCN (Regions with Fully Convolutional Networks) adopted inputs with minimum size 600, while StariNet kept with 300 × 300. Data from [8].
Table 2. PASCAL VOC 2007 t e s t results. Faster R-CNN (Regions with CNN features), HyperNet, Inside-Outside Net (ION) and R-FCN (Regions with Fully Convolutional Networks) adopted inputs with minimum size 600, while StariNet kept with 300 × 300. Data from [8].
MethodBackbonemAP (%)AeroBikeBirdBoatBottleBusCarCatChairCowTableDogHorseMbikePerson
Faster [4]VGG1673.276.579.070.965.552.183.184.786.452.081.965.784.884.677.576.7
Faster [4]ResNet10176.479.880.776.268.355.985.185.389.856.787.869.488.388.980.978.4
HyperNet [33]VGG1676.377.483.375.069.162.483.187.487.457.179.871.485.185.180.079.1
ION [34]VGG1676.579.279.277.469.855.785.284.289.857.578.573.887.885.981.375.3
R-FCN [7]ResNet10180.579.987.281.572.069.886.888.589.867.088.174.589.890.679.981.2
SSD300 [6]VGG1677.579.583.976.069.650.587.085.788.160.381.577.086.187.583.979.4
DSSD321 [13]ResNet10178.681.984.980.568.453.985.686.288.961.183.578.786.788.786.779.7
StairNet [35]VGG1678.881.385.477.872.159.286.486.887.562.785.776.084.188.486.178.8
FSSD300 [16]VGG1678.8---------------
Ours300VGG1680.383.586.779.074.459.987.487.486.564.988.177.986.688.087.280.9
SSD512 [6]VGG1679.584.885.181.573.057.887.888.387.463.585.473.286.286.783.982.5
DSSD513 [13]ResNet10180.092.186.680.368.758.284.385.094.663.385.965.693.088.587.886.4
FSSD512 [16]VGG1680.9---------------
Ours512VGG1681.887.988.481.776.765.188.588.887.966.989.076.986.189.987.484.6
Table 3. PASCAL VOC 2012 t e s t results from official evaluation server. For training data: “07+12”: union of VOC2007 t r a i n v a l and VOC2012 t r a i n v a l ; “07++12”: union of VOC2007 t r a i n v a l and t e s t and VOC2012 t r a i n v a l ; “07+12+S” union of VOC2007 trainval and VOC2012 t r a i n v a l , plus SBD segmentation labels [42]; “07++12+COCO”: union of “07++12” and COCO t r a i n v a l 35 k .
Table 3. PASCAL VOC 2012 t e s t results from official evaluation server. For training data: “07+12”: union of VOC2007 t r a i n v a l and VOC2012 t r a i n v a l ; “07++12”: union of VOC2007 t r a i n v a l and t e s t and VOC2012 t r a i n v a l ; “07+12+S” union of VOC2007 trainval and VOC2012 t r a i n v a l , plus SBD segmentation labels [42]; “07++12+COCO”: union of “07++12” and COCO t r a i n v a l 35 k .
MethodBackboneDatamAP (%)
Faster [4]VGG1607++1270.4
Faster [4]ResNet10107++1273.8
HyperNet [33]VGG1607++1271.4
ION [34]VGG1607+12+S76.4
R-FCN [7]ResNet10107++1277.6
SSD300 [6]VGG1607++1272.4
DSSD321 [13]ResNet10107++1276.3
StairNet [35]VGG1607++1276.4
FSSD300 [16]VGG1607++12++COCO82.0
Ours300VGG1607++1277.7
SSD512 [6]VGG1607++1274.9
DSSD513 [13]ResNet10107++1280.0
FSSD512 [16]VGG1607++12++COCO84.2
Ours512VGG1607++1280.3
Table 4. MS COCO t e s t - d e v 2015 results. Note: The “+++” includes box refinement, context, and multiscale testing. Data from [8].
Table 4. MS COCO t e s t - d e v 2015 results. Note: The “+++” includes box refinement, context, and multiscale testing. Data from [8].
MethodBackboneData SplitAvg. Precision, IoU:Avg. Precision, Area
0.5:0.950.50.75SML
Faster [4]VGG16trainval21.942.7----
Faster+++ [22]ResNet101trainval34.955.7----
ION [34]VGG16train23.643.223.66.424.138.3
R-FCN [7]ResNet101trainval29.251.5-10.332.443.3
SSD300 [6]VGG16trainval35k25.143.125.86.625.941.4
DSSD321 [13]ResNet101trainval35k28.046.129.27.428.147.6
FSSD300 [16]VGG16trainval35k27.147.727.88.729.242.2
Ours300VGG16trainval35k27.547.828.510.230.241.0
SSD512 [6]VGG16trainval35k28.848.530.310.931.843.5
DSSD513 [13]ResNet101trainval35k33.253.335.213.035.451.1
FSSD512 [16]VGG16trainval35k31.852.833.514.235.145.0
Ours512VGG16trainval35k33.153.835.117.136.045.7
Table 5. Ablation study on VOC 2007 t e s t . In the table, ffm3_c indicates the channel number of third feature fusion module.
Table 5. Ablation study on VOC 2007 t e s t . In the table, ffm3_c indicates the channel number of third feature fusion module.
1 FFM
2 FFMs
4 FFMs
3 FFMs with ffm3_c = 256
3 FFMs with ffm3_c = 128 (final model)
mAP(%)77.579.079.780.179.880.3
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Qu, S.; Huang, K.; Hussain, A.; Goulermas, Y. A Multipath Fusion Strategy Based Single Shot Detector. Sensors 2021, 21, 1360. https://0-doi-org.brum.beds.ac.uk/10.3390/s21041360

AMA Style

Qu S, Huang K, Hussain A, Goulermas Y. A Multipath Fusion Strategy Based Single Shot Detector. Sensors. 2021; 21(4):1360. https://0-doi-org.brum.beds.ac.uk/10.3390/s21041360

Chicago/Turabian Style

Qu, Shuyi, Kaizhu Huang, Amir Hussain, and Yannis Goulermas. 2021. "A Multipath Fusion Strategy Based Single Shot Detector" Sensors 21, no. 4: 1360. https://0-doi-org.brum.beds.ac.uk/10.3390/s21041360

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop