Next Article in Journal
Adaptive Selection of Truncation Radius in Calderon’s Method for Direct Image Reconstruction in Electrical Capacitance Tomography
Next Article in Special Issue
Authentication and Delegation for Operating a Multi-Drone System
Previous Article in Journal
A Lightweight Three-Factor Authentication and Key Agreement Scheme in Wireless Sensor Networks for Smart Homes
Previous Article in Special Issue
Parallel Transmission of Distributed Sensor Based on SCTP and TCP for Heterogeneous Wireless Networks in IoT
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Design and Analysis of a Lightweight Context Fusion CNN Scheme for Crowd Counting

1
College of Information, Mechanical and Electrical Engineering, Shanghai Normal University, Shanghai 201418, China
2
DS Information Technology Co., Ltd., Shanghai 200032, China
3
Department of Mathematics and Computer Science, Northeastern State University, Tahlequah, OK 74464, USA
*
Author to whom correspondence should be addressed.
Submission received: 11 April 2019 / Revised: 27 April 2019 / Accepted: 28 April 2019 / Published: 29 April 2019

Abstract

:
Crowd counting, which is widely used in disaster management, traffic monitoring, and other fields of urban security, is a challenging task that is attracting increasing interest from researchers. For better accuracy, most methods have attempted to handle the scale variation explicitly. which results in huge scale changes of the object size. However, earlier methods based on convolutional neural networks (CNN) have focused primarily on improving accuracy while ignoring the complexity of the model. This paper proposes a novel method based on a lightweight CNN-based network for estimating crowd counting and generating density maps under resource constraints. The network is composed of three components: a basic feature extractor (BFE), a stacked à trous convolution module (SACM), and a context fusion module (CFM). The BFE encodes basic feature information with reduced spatial resolution for further refining. Various pieces of contextual information are generated through a short pipeline in SACM. To generate a context fusion density map, CFM distills feature maps from the above components. The whole network is trained in an end-to-end fashion and uses a compression factor to restrict its size. Experiments on three highly-challenging datasets demonstrate that the proposed method delivers attractive performance.

Graphical Abstract

1. Introduction

Automatic crowd analysis is one of the more promising solutions for urban security services. Given the increasing attention of city managers to security issues, researchers have attempted to analyze issues related to various foundational aspects, such as crowd counting (estimating the head count) [1,2], density estimation [3,4], behavior analysis [5], pedestrian tracking [6], and anomaly detection [7]. However, because of the presence of various complexities in a real scene, these techniques are far from optimal. In this paper, we further explore the task of estimating the crowd count from a single image under resource constraints.
In addition to the common challenges of messy scenes and changeable light, large variations in scale are a key factor limiting the accuracy of most proposed methods for crowd counting. Depending on the camera perspective or the real gaps between heads, areas with heads range from hundreds of pixels to a few pixels. Several CNN-based models [8,9,10,11] that are adaptive to multi-scale blobs have been proposed, and the general trend has been to make deeper and more complicated networks for higher accuracy. In the recent literature, adaption to different density domains has become a first-order question, with less attention paid to an appropriately right-sized network.
The size of these improved networks is expanding because of the special components of multi-scale or contextual features. On the other hand, some patch-based networks [11,12] fuse multi-level contextual information by cropping the original image into several patches or rescaling the input image several times. While these methods restrict the size of networks, the computation in the whole feedforward process is still increased several fold. Overall, all of the above methods compromise accuracy for speed. By contrast, a single-column network [13], which is always synonymous with higher speed and fewer trainable parameters, shows potential for crowd counting.
Under a practical scenario of crowd counting, the size is of equal importance as the accuracy level. Several limits of the platform impose several rigorous restrictions on the time and space complexity of the method: e.g., complex methods may slow down the whole system for crowd counting, and methods that require less memory could be feasible to deploy on contemporary field-programmable gate arrays (FPGAs) or some other edge computing devices.
In this paper, we propose a lightweight CNN network that maintains a good balance between accuracy and size. Our framework follows a single-column style in its backbone and extracts low-level features with reduced spatial resolution in the basic feature extractor (BFE). In particular, skipping connections and bottleneck layers are used in the architecture to increase the efficiency of the feature extractor. Considering the importance of the receptive field size and the size of the network, we replace traditional convolution with à trous convolution, which enlarges the size of the receptive field without reducing spatial resolution in the following layers. One benefit of this strategy is that feature maps in different stages can be integrated without up-sampling from different spatial resolutions. Thus, several à trous convolutional layers form the stacked à trous convolution module (SACM). At the stage of context fusion in the context fusion module (CFM), several short branches connect to different layers. To estimate the density map for crowd counting, the network integrates multi-level information.
Experiments on three main benchmarks, ShanghaiTech [8], UCF_CC_50 [3], and WorldExpo’10 [14], demonstrate the effectiveness of the proposed method. Compared with these methods, the CNN-based network in our method strikes a good balance between size and a series of metrics. This most lightweight network for crowd counting shows attractive performance in terms of feedforward speed and estimated accuracy.
The remainder of the paper is organized as follows. In Section 2, previous works for crowd counting and density maps are reviewed. Section 3 presents the architecture and configuration of our method. In Section 4, experimental results on three main datasets are introduced. Finally, we conclude our work in Section 5.

2. Related Work

With the development of deep learning, CNN-based solutions combined with a new technique have shown performance far beyond that of traditional methods. Loy et al. [15] categorized existing traditional crowd counting approaches into: (1) detection-based approaches, (2) regression-based approaches, and (3) density estimation-based approaches. This classification standard is also applicable to approaches based on neural networks.
The vast majority of initial detection-based approaches typically use a sliding window-based solution to detect objects and count the number of people in the vision [16]. Early works on still images were usually performed either in the monolithic style or using parts-based detection [17] by training a classifier (such as support vector machines [18] and random forest [19]) to extract low-level handcrafted features, e.g., Haar wavelets [20] and histograms of oriented gradients [21]. In the fashion of deep-learning-based feature extraction, CNN-based architectures such as Faster-RCNN [22] and YOLO [23] were introduced to tackle the problem. Although these approaches have achieved progress for sparse scenes, poor performance occurs for dense crowds due to hidden objects and background clutter.
In contrast to the complex task of locating all objects in the scene and counting them in detection-based approaches, regression-based approaches attempt to count by regressing a mapping between features extracted from local image patches and their counts [24,25,26]. While these approaches are successful in addressing the issues of occlusion and clutter, most ignore significant spatial location information as they are regressed on the global count.
To tackle the above issues, density maps may be an appropriate expression mode that characterizes the spatial and numerical information of crowds. Lempitsky et al. [27] proposed to learn a linear mapping between local patch features and corresponding object density maps, thereby incorporating spatial details in the learning process. Another solution of learning a non-linear mapping by random forest regression was subsequently proposed by Pham et al [28]. In addition to these methods, a number of approaches based on density maps have been proposed recently [29,30].
Researchers have attempted to use CNN-based approaches to generate density maps or corresponding counts [14,31,32,33] due to the success of CNNs in numerous computer vision tasks. However, due to the limitations of these methods, large-scale variations in crowd scenes may be a notable problem.
Many CNNs based on a multi-column-like architecture (e.g., CrowdNet [34], MCNN [8], and Hydra-CNN [11]) capture a combination of high-level and low-level semantic information to predict the density map for a given scene. To combat the weakness of context fusion, Sam et al. proposed a patch-based architecture called Switch-CNN [12] to exploit local crowd density variation within a scene by sending patches in certain columns, while MoC [35] performs a late selection with a gating CNN in the architecture. CP-CNN [9] goes further by incorporating local and global contextual information from the whole scene image. Sam et al. developed the idea of an expert network group in their new work, and IG-CNN [36] makes the child regressors specialized.
Because optimizing a multi-column network with huge parameters and a complex structure is a different task, the single-column architecture [2,37,38] shows relative advantages. A cascaded multi-task architecture [4] or other effective ideas [1,39,40] improve the performance of the single-column architecture. Counterintuitively, a single-column network with fewer parameters could outperform MCNN on a specific dataset in [13]. Inspired by that result, CSRNet [13], which is powered by the VGG network [41], surpassed the previous state-of-the-art crowd counting solution. Such ideas were considered with pre-trained parameters in [9,12,13], and the methods performed better, but with higher complexity. Kang and Chan [10] applied an image pyramid with an attention map to fuse various scales of objects adaptively. Although its light backbone is attractive, the extra processes of feedforward still slow down their network. Another important idea is to address the issue of inconsistent estimation across different scaled inputs in [42]. Adversarial information, such as adversarial loss and a scale-consistency regularizer, helps the generation network boost density estimation performance.

3. Method

Recent works have experimentally introduced new architectures and techniques to challenge previous state-of-the-art works. Most of these efforts have focused on metrics rather than the complexity of their models. For network design, a larger or deeper model is not necessarily better. In other words, the complex engineering problems surrounding us always restrict the application of these methods, and a simple and inexpensive model is almost always easy to deploy.

3.1. Architecture

Most proposed CNN-based methods have complicated architectures or a vast number of trainable parameters, which increase the difficulty of model training or make the model feedforward process more time-consuming. For example, some models (e.g., CSRNet [13] and CP-CNN [9]) inherit many parameters from the VGG network [41] for transfer learning. Although these parameters may be frozen during model training, they still dilate the whole model and slow down the speed of this method. Multi-column or multi-branch approaches are synonymous with huge floating-point operations. More attention should be paid to the right-sized architecture. In this paper, the single-column network achieves good performance with less computation and fewer trainable parameters.
Due to the lack of external perspective information about crowds and the presence of multiple scales in an image, it is difficult to estimate crowd counts under a single standard or generate a high-quality density map without spatial distribution information. Conceptually, CNNs need larger receptive fields to recognize the same kind of crowd in a close shot than in a vista. In other words, fusing contextual details or dilating receptive fields is the key factor for a successful crowd counting method. Further-more, context information and multi-scale feature representation play an active part in distinguishing crowd from other objects. Nevertheless, the huge amount of computation cost and parameters restrict the deployment of such applications.
In the following subsections, we design the architecture under the guidance of three key strategies:
  • Single-column style
  • A minimum sufficient receptive size
  • Context embedding

3.1.1. Backbone

In recent years, different types of efficient convolutional neural network architectures for eliminating the redundancy in convolution kernels have been published. SqueezeNet [43] achieves AlexNet-level accuracy on ImageNet with 50× fewer parameters. Another notable work in this direction is the DarkNet Reference network [23], which is optimized for not only parameter count, but also inference speed. MobileNet [44], which uses depth-wise convolution for spatial convolutions, exceeds AlexNet’s performance with only 1.32 million parameters. This idea is extended to pointwise group convolution along with channel shuffling by ShuffleNet [45]. Furthermore, SqueezeNext [46], whose design was guided by considering previous architectures, achieves better classification accuracy with fewer parameters, but avoids using depth-wise-separable convolutions, which are inefficient on some mobile processor platforms. The difference between the performance of these works should not attract too much attention. In a real scenario, better automation capabilities and ease of implementation are more efficient.
Although the above works have studied convolution in-depth, additional complexity may be injected into our experimental comparison. Considering the design of previous works for crowd counting, slightly fine-tuning the architecture composed of traditional convolutional layers may cause impartial competition among such works. Due to our modular design, it is also easy to enhance our model based on the components proposed in the above works for better accuracy or better real-time ability.
The design philosophy of the proposed model comes from Tiny Darknet, which is the front-end of the famous object detection network YOLO [23]. As shown in Figure 1, the backbone is repeatedly stacked with a block composed of a 1 × 1 bottleneck layer and a 3 × 3 ordinary convolutional layer. In particular, the compression factor ( α ) and the 1 × 1 filter kernel in the bottleneck layer contribute to limiting the quantity of parameters in this lightweight architecture. In other words, the compression factor ( α ) and the filters of the unit ( f i l t e r s ) jointly determine the number of filters in each layer. A parameter-free shortcut connection is also used to help all information pass through the blocks.
In this paper, the proposed backbone was YOLO style. To further reduce computation, the max-pooling layer was also replaced with a convolutional layer with a stride of 2. However, modern approaches usually compromise spatial resolution to achieve faster inference speed, which leads to poor performance. Due to the importance of the density map resolution and the hard work required to remedy the loss of spatial details, the backbone only reduced the spatial resolution of feature maps twice to obtain the 1/4 feature map. In Table 1, the configuration of the proposed network is presented. The parameters in the convolutional layers are denoted as “(kernel size), (number of filters), (stride), (dilation rate)”. Unless otherwise stated, the stride and dilation rate were all set to 1 by default.
A basic feature extractor (BFE) consisting of three convolutional blocks was employed, which is a fast down-sampling strategy. Following the above extractor, a transition layer was used to compress the depth of the feature maps. This layer provided a “soft landing” to reduce information loss from channel truncation.
Another critical component of our design was the stacked à trous convolution module (SACM). Similar to the component reported in [13], à trous convolution helps the network enlarge the receptive field without extra trainable parameters or any loss of spatial resolution of features. For example, the receptive size of the last layer in the first à trous block {117} is almost three-times larger than the receptive size of the layer above this block {37} in Table 1. In addition, the main idea of à trous convolution is to insert “holes” in basic convolutional kernels, thus enabling employing them easily without other by-products.
Only three à trous convolutional layers with the same dilation rate were deployed in [13], which had the same number of parameters as traditional convolution (same filter size, rate = 1). In contrast to CSRNet, we grouped together three layers as a block with an increasing dilation rate 2, 3, 5 to avoid the “gridding” problem in [47]. At the same time, the receptive sizes of these feature maps increased more efficiently. To ensure an appropriate receptive size of the network, a SACM composed of two consecutive blocks was deployed in this paper.
The final context fusion layer used ReLU activation without batch normalization, while the other convolutional layers were followed with a batch normalization layer and leaky ReLU activation (the slope of the negative section was 0.1).

3.1.2. Context Fusion Module

Several context fusion architectures have been presented: the multi-column architecture was an initial attempt to fuse local context, while CP-CNN [9] and Hydra CNN [11] have further applications. In [10], the classic image pyramid also helped the FCN handle context fusion attributed to multiple scales.
Inspired by the success of spatial pyramid pooling [48], which fuses information of various arbitrary scales, à trous spatial pyramid pooling (ASPP) effectively resamples features at different scales to classify regions of arbitrary scale accurately and efficiently. Various pooling layers (e.g., max pooling) are widely used in current image-to-image models because they can multiply the receptive field in addition to providing other benefits, despite reducing the spatial resolution of feature maps. Attenuation of spatial resolution is a serious problem that leads to loss of information in image-to-image tasks.
In many proposed models [47,49], deconvolution or other upsampling components, which certainly add extra complexity to the original model, are used to recover the resolution. In particular, à trous convolution enlarges the receptive field without loss of spatial details or increasing the number of trainable parameters. This means that CNNs can appear wider with fewer layers, and little attention will be paid to handling feature resolution at different contextual scales.
Multiple parallel à trous convolutional layers with different sampling rates are used in ASPP to generate different scale information in separate branches. The above outputs are joined for final feature fusion. However, the parallel branches in ASSP slow down the inference speed.
In the context fusion module (CFM) we proposed, different contextual features were extracted from different parts of the backbone for further isolated distillation. In other words, various contextual information will not be fused until the last few layers. The experiments in Section 4.2 demonstrate that this design not only improved context fusion, but also reduced the number of parameters and increased the inference speed.
Figure 2 presents the completed scheme of the proposed network for crowd counting. The network extracted common feature maps with reduced spatial resolution using BFE. To ensure the diversity of the contextual features, SACM enlarged the gap between the receptive sizes of feature maps. CFM distilled different contextual information and fused it to estimate the density map.

3.2. Training Details

In our proposed method, the Adam optimizer was used to update the parameters of our network with a fixed learning rate during training. The batch size was set to 32, and different fixed learning rates were set for different datasets. For the sake of discussion, the specific setups for different datasets are given in Section 4.3.

3.2.1. Data Augmentation

In the training dataset, we trained our network on images of a single scale. S × S patches (parameter S depends on the dataset) were cropped randomly from each image (N patches per image), and horizontal flipping generated the other N patches; consequently, we obtained 2 N patches in total from a single image. If the shorter side of the image was shorter than S pixels, the image should be rescaled to fit the above requirement.

3.2.2. Ground Truth

The ground truth in the task of crowd counting was the density map, which demonstrated the total number of the crowd and the spatial distribution information simultaneously. If there was a head at pixel x i , we represented it as a delta function δ ( x x i ) . The ground truth was generated by blurring each head annotation using a normalized Gaussian kernel G σ i ( x ) . Hence, an image with Nheads labeled can be defined as follows:
F ( x ) = i = 1 N δ ( x x i ) G σ i ( x ) .
In the fixed Gaussian kernel, σ i is constant for a single image, whose value is defined by the area of each person head. Unfortunately, geometric distortion caused by the perspective effect leads to the variation on the area of the head and limits the quality of the crowd density map.
To tackle the above problem, geometry-adaptive kernels [8] were used to generate the ground truth for highly congested scenes. The parameter σ i of the Gaussian kernel was determined by the average distance d ¯ i of the k nearest neighbors, which can be represented as:
σ i = β d ¯ i .
where β and k are hyperparameters. In our experiment, we followed the configuration in [8], where β = 0.3 and k = 5 .

3.2.3. Object Function

The loss function for training the model was the Euclidean distance, which is defined as follows:
L ( Θ ) = 1 N i = 1 N F ( X i ; Θ ) F i G T 2 2 .
where N is the size of the training batch and F ( X i ; Θ ) is the density map generated by the proposed model with trainable parameters Θ . L ( Θ ) denotes the loss between the estimated density map F ( X i ; Θ ) and the ground truth F i G T .

4. Experimental Results and Discussions

Considering practical applications of our algorithm, we trained and evaluated our network on an Xeon(R) CPU E5-1650 machine with an NVIDIA GeForce GTX 1080 Ti using Tensorflow framework [50]. In addition, the network can run at 28.2 FPS when the image size is resized to 1024 × 768 .

4.1. Metrics

The mean absolute error (MAE) and root mean squared error (RMSE) were used to compare between different methods. These metrics are defined as follows:
M A E = 1 N i = 1 N C i C i G T .
R M S E = 1 N i = 1 N C i C i G T 2 .
where N is the number of images in the evaluation and C i represents the estimated count for the ith image, which is the integral value of the corresponding density map. C i G T is the ground truth count.
Because various platforms use different optimization programs, it is difficult to determine which model really has the faster inference speed. In this paper, floating point operations per second (FLOPS) was introduced to measure runtime speed. In particular, only the computation of convolution was considered, and batch normalization, bias adding operation, and other operations were excluded.

4.2. Ablation Study and Comparisons

The effectiveness of the configurations of the proposed architecture will be evaluated in this subsection. The ablation study was performed on the ShanghaiTech Part A dataset [8], which is a subset of the ShanghaiTech dataset that includes 482 images of congested scenes and 241,667 annotated heads.
Three configurations are shown in Table 2: A. the completed model; B. the cropped model excluding the context fusion module (CFM); and C. the model that does not compress the depth of feature maps in its bottlenecks (compression rate was set to one). In Table 2, the performance of Config.A is much better than that of Config.B. Obviously, the CFM boosted the performance, which had little impact on calculation. While Config.C was similar to Config.A on both metrics (MAE and MSE), its larger size may slow down the whole model. Consequently, we selected Config.A for further discussion.
Several classic architectures of networks for crowd counting are listed in Table 3. To compare the performance of each method, Table 3 enumerates the number of parameters in the networks, the FLOPS (floating point operations per second) of the main layers, and the two metrics (MAE and RMSE). FLOPS is a de facto standard for computation. In our analysis, the size of the input image for networks was set to 448 × 448 px. Due to the specific design, multi-column networks (such as MCNN [8], CP-CNN [9], and CrowdNet [34]) are adaptive to scale variations. Although each branch of these networks is short, extra computations are still generated. For example, two networks (MCNN and ours) with similar parameters have completely different computations, such that the FLOPS of MCNN were almost double that of ours. CSRNet [13] is a classic single-column network, but the huge number of parameters from the VGG network slow down the whole method greatly. By contrast, our network had almost 125 × fewer parameters than CSRNet. Although many state-of-the-art methods (such as CP-CNN [9], Switching-CNN [12], IG-CNN [36], ACSCP [42], and CSRNet [13]) benefiting from such a transfer learning method perform better, they are very slow for images of high resolution. FCN-7c-3s [10] applies the image pyramid with the attention map to fuse various scales of objects. Its light backbone is attractive, but the feedforward process will run several times per prediction. Consequently, the total FLOPS of FCN-7c-3s were 7 × greater than ours. Our network only achieved the second-best performance on MAE and RMSE among the listed networks, but its lowest FLOPS and its lightweight backbone make it more flexible with respect to resource constraints.
Many crowd counting works suffered from semantic mixtures, which make the networks regard other blobs (trees, buildings, and other background) as the crowd. After an analysis of these results, we believe that some kinds of context fusion can aid us in achieving better accuracy. The work in [9] presented the same insight. In Figure 3, the outputs of CFM are visualized, and the red bounding boxes mark the same spatial region in different maps. In these maps, the color from red to blue reflects the response from strong to weak. Although a poor response was observed in some feature maps, the corrected density map was still generated by the other substitutes. The research in [13] showed that the multi-column design performed well against the original intention of the MCNN design for learning different features for each column. In contrast to previous works, CFM can finish homogeneous works in the early stage and then focus on expert specialization works. The component also reduces the amount of computation while ensuring those of different features are learned for each column.

4.3. Performance Evaluation

4.3.1. ShanghaiTech Dataset

The ShanghaiTech dataset [8] has two separate subsets: Part A with 482 images (300 images were for training, and the rest were for testing) and Part B with 719 images (400 images were for training, and the rest were for testing). In Part A, images with different resolutions depict highly-congested scenes. By contrast, images with a fixed resolution in Part B were taken from sparse crowd scenes. Following [8], the geometry-adaptive kernel was used to generate ground truth density maps for Part A, and the fixed Gaussian kernel ( σ = 15 ) was used for Part B. In this subsection, we train our network on the data augmentation in which 20 patches of 448 × 448 px are generated from each image. The learning rates in these subsets were 1 × 10 3 and 1 × 10 4 , respectively.
As shown in Table 4, we evaluated and compared our method with eight other methods on these two subsets separately. In this paper, we further divided all CNN-based methods into two categories: VGG-powered methods and other methods. VGG-powered methods inherited most of their parameters from the VGG network [41], and feature maps generated by extractors consisting of these parameters were used directly in their methods. Although VGG-powered methods have challenged state-of-the-art methods constantly in recent years (such as CP-CNN [9], Switching-CNN [12], IG-CNN [36], ACSCP [42], and CSRNet [13]), we did not compare them with other methods directly due to their huge number of parameters and the benefits of prior knowledge transferred from other tasks. Therefore, the most state-of-the-art results in these two categories are highlighted in bold separately.
Table 4 demonstrates that our method outperformed all methods except VGG-powered methods on ShanghaiTech Part A. On Part B, our method achieved the third-best performance, slightly poorer than the top two. In particular, our network had the fewest parameters and FLOPS: 125 × fewer parameters and 40 × fewer FLOPS than CSRNet, which achieved the best performance. The proposed network won over FCN-7c-3s by a small margin, but 7 × fewer FLOPS relieved the pressure on the computational needs. Density maps generated by the proposed method on the validation dataset are shown in Figure 4 and Figure 5.

4.3.2. UCF_CC_50 Dataset

The UCF_CC_50 dataset [3] contains 50 images of extremely-dense crowds. The dataset exhibits a large variance in crowd counts, with a range of 84–543. Following the standard protocol in [3], a five-fold cross-validation was performed to validate the performance of the proposed method.
The geometry-adaptive kernel was also performed in these highly-congested scenes for ground-truth density map generation. In consideration of the specific properties of this dataset, 448 × 448 patches were cropped, and a total of 60 samples was generated from each image. During the training time, the learning rate was set to 1 × 10 3 . The results are listed in Table 5, which demonstrated the improved performance of our method. Our network achieved comparable MAE and RMSE scores, with current state-of-the-art MAE values, but high RMSE values. These results demonstrated that our model had low bias and high variance, which means that the model was not robust enough for this dataset. Despite this quality, fewer parameters and the least FLOPS among the top four methods make our method more flexible for practical implementation. Figure 6 illustrates the density maps obtained using the proposed method.

4.3.3. The WorldExpo’10 Dataset

The WorldExpo’10 dataset, which was first introduced by Zhang et al. [14], includes 3980 annotated frames from 1132 annotated video sequences captured by 108 surveillance cameras. All frames in the dataset were divided into two sets: the training set contained 3380 frames, and the testing set had 600 frames from five different scenes. The region of interest (ROI) was provided for each scene. Following the evaluation protocol mentioned in [14], a human-like Gaussian kernel was adopted, and only the crowd in the ROI was taken into consideration.
In this subsection, our network was trained on the data augmentation in which 40 patches of 448 × 448 px were generated from each image. Considering five different crowd scenes in this dataset, the learning rate was set from 1 × 10 3 to 1 × 10 5 in the specific scene (from the crowded scene to the sparse scene). The results are listed in Table 6. The average performance (Avg. MAE) of VGG-powered models was still dominant. In particular, our model achieved top rankings and even surpassed VGG-powered models in several scenes. In Sce.2, more complex models (such as CP-CNN and Switching-CNN) achieved better performance, while this scene was the most complex scene among the five scenes. The results of these experiments demonstrated that the complexity of the model should be adapted to the specific task. Figure 7 illustrates the density maps obtained using the proposed method.

5. Conclusions

In this paper, we proposed a lightweight CNN-based crowd counting network for better performance in crowd counting. Accuracy and complexity have gradually become unbalanced in previous research. Benefiting from a specially-designed BFE, SACM, and CFM, the proposed network not only achieved excellent performance with the three main datasets, but also was the smallest network among the recent series reported in the literature. Although VGG-powered networks had better rankings, the smaller size and shorter feedforward time made our method more favorable.
Future work will cover the extension of the proposed method to more robust performance, as well as the deployment of our method with realistic scenes. The performance on different platforms will also be investigated. Finally, domain adaption problems caused by limited expensive datasets will be further studied.

Author Contributions

Conceptualization, Y.Y. and J.H.; methodology, Y.Y.; validation, W.D. and Y.Y.; investigation, J.H.; resources, Y.Y.; data curation, Y.Y.; writing—original draft preparation, Y.Y.; writing—review and editing, N.X. and J.H.; visualization, Y.Y.; supervision, W.D. and J.H.; project administration, J.H.; funding acquisition, W.D.

Funding

This work was supported in part by the Artificial Intelligence Development and Innovation Project of Shanghai (No. 2018-RGZN-01013).

Acknowledgments

The authors are grateful for the free copyright sharing of the datasets mentioned in this paper, as well as previous awesome inspiring works on crowd counting.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huang, S.; Li, X.; Zhang, Z.; Wu, F.; Gao, S.; Ji, R.; Han, J. Body Structure Aware Deep Crowd Counting. IEEE Trans. Image Process. 2018, 27, 1049–1059. [Google Scholar] [CrossRef] [PubMed]
  2. Hu, Y.; Chang, H.; Nian, F.; Wang, Y.; Li, T. Dense Crowd Counting from Still Images with Convolutional Neural Networks. J. Vis. Commun. Image Represent. 2016, 38, 530–539. [Google Scholar] [CrossRef]
  3. Idrees, H.; Saleemi, I.; Seibert, C.; Shah, M. Multi-source Multi-scale Counting in Extremely Dense Crowd Images. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 2547–2554. [Google Scholar] [CrossRef]
  4. Sindagi, V.A.; Patel, V.M. CNN-Based cascaded multi-task learning of high-level prior and density estimation for crowd counting. In Proceedings of the IEEE International Conference on Advanced Video and Signal Based Surveillance, Lecce, Italy, 29 August–1 September 2017; pp. 1–6. [Google Scholar] [CrossRef]
  5. Herath, S.; Harandi, M.; Porikli, F. Going deeper into action recognition: A survey. Image Vis. Comput. 2017, 60, 4–21. [Google Scholar] [CrossRef]
  6. Smeulders, A.W.M.; Chu, D.M.; Cucchiara, R.; Calderara, S.; Dehghan, A.; Shah, M. Visual Tracking: An Experimental Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1442–1468. [Google Scholar] [CrossRef] [PubMed]
  7. Li, T.; Chang, H.; Wang, M.; Ni, B.; Hong, R.; Yan, S. Crowded Scene Analysis: A Survey. IEEE Trans. Circuits Syst. Video Technol. 2015, 25, 367–386. [Google Scholar] [CrossRef]
  8. Zhang, Y.; Zhou, D.; Chen, S.; Gao, S.; Ma, Y. Single-Image Crowd Counting via Multi-Column Convolutional Neural Network. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 589–597. [Google Scholar] [CrossRef]
  9. Sindagi, V.A.; Patel, V.M. Generating High-Quality Crowd Density Maps Using Contextual Pyramid CNNs. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1879–1888. [Google Scholar] [CrossRef]
  10. Kang, D.; Chan, A. Crowd Counting by Adaptively Fusing Predictions from an Image Pyramid. In Proceedings of the British Machine Vision Conference (BMVC), Newcastle upon Tyne, UK, 3–6 September 2018. [Google Scholar]
  11. Oñoro, D.; J López-Sastre, R. Towards Perspective-Free Object Counting with Deep Learning. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 615–629. [Google Scholar]
  12. Sam, D.B.; Surya, S.; Babu, R.V. Switching Convolutional Neural Network for Crowd Counting. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4031–4039. [Google Scholar] [CrossRef]
  13. Li, Y.; Zhang, X.; Chen, D. CSRNet: Dilated Convolutional Neural Networks for Understanding the Highly Congested Scenes. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; Volume 1802. [Google Scholar] [CrossRef]
  14. Cong, Z.; Hongsheng, L.; Wang, X.; Xiaokang, Y. Cross-scene crowd counting via deep convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 833–841. [Google Scholar] [CrossRef]
  15. Chen, C.L.; Chen, K.; Gong, S.; Xiang, T. Crowd Counting and Profiling: Methodology and Evaluation. In Modeling, Simulation and Visual Analysis of Crowds: A Multidisciplinary Perspective; Springer: New York, NY, USA, 2013; pp. 347–382. [Google Scholar]
  16. Dollar, P.; Wojek, C.; Schiele, B.; Perona, P. Pedestrian Detection: An Evaluation of the State of the Art. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 743–761. [Google Scholar] [CrossRef] [PubMed]
  17. Li, M.; Zhang, Z.; Huang, K.; Tan, T. Estimating the number of people in crowded scenes by MID based foreground segmentation and head-shoulder detection. In Proceedings of the 2008 19th International Conference on Pattern Recognition, Tampa, FL, USA, 8–11 December 2008; pp. 1–4. [Google Scholar] [CrossRef]
  18. Liu, S.; Zhai, S.; Li, C.; Tang, J. An effective approach to crowd counting with CNN-based statistical features. In Proceedings of the 2017 International Smart Cities Conference (ISC2), Wuxi, China, 14–17 September 2017; pp. 1–5. [Google Scholar] [CrossRef]
  19. Gall, J.; Yao, A.; Razavi, N.; Gool, L.V.; Lempitsky, V. Hough Forests for Object Detection, Tracking, and Action Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 2188–2202. [Google Scholar] [CrossRef] [PubMed]
  20. Viola, P.; Jones, M.J. Robust Real-Time Face Detection. Int. J. Comput. Vis. 2004, 57, 137–154. [Google Scholar] [CrossRef]
  21. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar] [CrossRef]
  22. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  23. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  24. Chan, A.B.; Vasconcelos, N. Bayesian Poisson regression for crowd counting. In Proceedings of the IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 545–551. [Google Scholar] [CrossRef]
  25. Ryan, D.; Denman, S.; Fookes, C.; Sridharan, S. Crowd Counting Using Multiple Local Features. In Proceedings of the 2009 Digital Image Computing: Techniques and Applications, Melbourne, VIC, Australia, 1–3 December 2009; pp. 81–88. [Google Scholar] [CrossRef]
  26. Chen, K.; Loy, C.C.; Gong, S.; Xiang, T. Feature Mining for Localised Crowd Counting. In Proceedings of the British Machine Vision Conference, Surrey, UK, 3–7 September 2012. [Google Scholar] [CrossRef]
  27. Lempitsky, V.S.; Zisserman, A. Learning To Count Objects in Images. In Proceedings of the 23rd International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 6–9 December 2010; Volume 1, pp. 1324–1332. [Google Scholar] [CrossRef]
  28. Pham, V.; Kozakaya, T.; Yamaguchi, O.; Okada, R. COUNT Forest: CO-Voting Uncertain Number of Targets Using Random Forest for Crowd Density Estimation. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 3253–3261. [Google Scholar] [CrossRef]
  29. Wang, Y.; Zou, Y. Fast visual object counting via example-based density estimation. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 3653–3657. [Google Scholar] [CrossRef]
  30. Xu, B.; Qiu, G. Crowd density estimation based on rich features and random projection forest. In Proceedings of the 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016; pp. 1–8. [Google Scholar] [CrossRef]
  31. Wang, C.; Zhang, H.; Yang, L.; Liu, S.; Cao, X. Deep People Counting in Extremely Dense Crowds. In Proceedings of the 23rd ACM International Conference on Multimedia, Brisbane, Australia, 26–30 October 2015; pp. 1299–1302. [Google Scholar] [CrossRef]
  32. Walach, E.; Wolf, L. Learning to Count with CNN Boosting. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 660–676. [Google Scholar]
  33. Shang, C.; Ai, H.; Bai, B. End-to-end crowd counting via joint learning local and global count. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 1215–1219. [Google Scholar] [CrossRef]
  34. Boominathan, L.; Kruthiventi, S.S.S.; Babu, R.V. CrowdNet: A Deep Convolutional Network for Dense Crowd Counting. In Proceedings of the 24th ACM International Conference on Multimedia, Amsterdam, The Netherlands, 15–19 October 2016; pp. 640–644. [Google Scholar] [CrossRef]
  35. Kumagai, S.; Hotta, K.; Kurita, T. Mixture of counting CNNs. Mach. Vis. Appl. 2018, 29, 1119–1126. [Google Scholar] [CrossRef]
  36. Sam, D.B.; Sajjan, N.N.; Babu, R.V.; Srinivasan, M. Divide and Grow: Capturing Huge Diversity in Crowd Images with Incrementally Growing CNN. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3618–3626. [Google Scholar] [CrossRef]
  37. Marsden, M.; McGuiness, K.; Little, S.; O’Connor, N. Fully Convolutional Crowd Counting On Highly Congested Scenes. In Proceedings of the 12th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISAPP), Porto, Portugal, 27 February–1 March 2017; Volume 5, pp. 27–33. [Google Scholar] [CrossRef]
  38. Zhao, Z.; Li, H.; Zhao, R.; Wang, X. Crossing-Line Crowd Counting with Two-Phase Deep Neural Networks. In European Conference on Computer Vision; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer: Cham, Switzerland, 2016; pp. 712–726. [Google Scholar]
  39. Marsden, M.; McGuinness, K.; Little, S.; Connor, N.E.O. ResnetCrowd: A residual deep learning architecture for crowd counting, violent behaviour detection and crowd density level classification. In Proceedings of the IEEE International Conference on Advanced Video and Signal Based Surveillance, Lecce, Italy, 29 August–1 September 2017; pp. 1–7. [Google Scholar] [CrossRef]
  40. Li, J.; Yang, H.; Chen, L.; Li, J.; Zhi, C. An end-to-end generative adversarial network for crowd counting under complicated scenes. In Proceedings of the IEEE International Symposium on Broadband Multimedia Systems and Broadcasting, Cagliari, Italy, 7–9 June 2017; pp. 1–4. [Google Scholar] [CrossRef]
  41. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  42. Shen, Z.; Xu, Y.; Ni, B.; Wang, M.; Hu, J.; Yang, X. Crowd Counting via Adversarial Cross-Scale Consistency Pursuit. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5245–5254. [Google Scholar] [CrossRef]
  43. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50× fewer parameters and <0.5 MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  44. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  45. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. ShuffleNet: An Extremely Efficient Convolutional Neural Network forMobile Devices. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6848–6856. [Google Scholar] [CrossRef]
  46. Gholami, A.; Kwon, K.; Wu, B.; Tai, Z.; Yue, X.; Jin, P.; Zhao, S.; Keutzer, K. SqueezeNext: Hardware-Aware Neural Network Design. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 1719–171909. [Google Scholar] [CrossRef]
  47. Wang, P.; Chen, P.; Yuan, Y.; Liu, D.; Huang, Z.; Hou, X.; Cottrell, G. Understanding Convolution for Semantic Segmentation. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1451–1460. [Google Scholar] [CrossRef]
  48. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed]
  49. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention (MICCAI); Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  50. Martín, A.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems. arXiv 2016, arXiv:1603.04467. [Google Scholar]
Figure 1. The architecture of the basic unit in our backbone. (a) is the layer organization in which a unit is stacked repeatedly and the shortcut connection aids information transfer, and (b) is the scheme of the unit, which is composed of a 1 × 1 bottleneck layer and a 3 × 3 ordinary convolutional layer. The compression factor ( α ) and the filters of the Unit ( f i l t e r s ) jointly determine the number of filters in each layer.
Figure 1. The architecture of the basic unit in our backbone. (a) is the layer organization in which a unit is stacked repeatedly and the shortcut connection aids information transfer, and (b) is the scheme of the unit, which is composed of a 1 × 1 bottleneck layer and a 3 × 3 ordinary convolutional layer. The compression factor ( α ) and the filters of the Unit ( f i l t e r s ) jointly determine the number of filters in each layer.
Sensors 19 02013 g001
Figure 2. Overview of the proposed network for density estimation. The network extracts common feature maps with reduced spatial resolution using basic feature extractor (BFE). To ensure a diversity of contextual features, the stacked à trous convolution module (SACM) enlarges the gap between the receptive sizes of the feature maps. The context fusion module (CFM) distills different contextual information and fuses it to estimate the density map.
Figure 2. Overview of the proposed network for density estimation. The network extracts common feature maps with reduced spatial resolution using basic feature extractor (BFE). To ensure a diversity of contextual features, the stacked à trous convolution module (SACM) enlarges the gap between the receptive sizes of the feature maps. The context fusion module (CFM) distills different contextual information and fuses it to estimate the density map.
Sensors 19 02013 g002
Figure 3. Analysis of the context fusion module (CFM). The outputs of CFM are visualized, and the red bounding boxes mark the same spatial region in different maps. In the feature maps, the color from red to blue reflects the response from strong to weak. (a) Input image, (b) estimation result, (c) first output of CFM, (d) second output of CFM, and (e) third output of CFM.
Figure 3. Analysis of the context fusion module (CFM). The outputs of CFM are visualized, and the red bounding boxes mark the same spatial region in different maps. In the feature maps, the color from red to blue reflects the response from strong to weak. (a) Input image, (b) estimation result, (c) first output of CFM, (d) second output of CFM, and (e) third output of CFM.
Sensors 19 02013 g003
Figure 4. Density estimation results on the ShanghaiTech Part A [8]. (a) Input images, (b) ground truth density maps, and (c) estimation results.
Figure 4. Density estimation results on the ShanghaiTech Part A [8]. (a) Input images, (b) ground truth density maps, and (c) estimation results.
Sensors 19 02013 g004
Figure 5. Density estimation results on ShanghaiTech Part B [8]. (a) Input images, (b) ground truth density maps, and (c) estimation results.
Figure 5. Density estimation results on ShanghaiTech Part B [8]. (a) Input images, (b) ground truth density maps, and (c) estimation results.
Sensors 19 02013 g005
Figure 6. Density estimation results on the UCF_CC_50 [3]. (a) Input images, (b) ground truth density maps, and (c) estimation results.
Figure 6. Density estimation results on the UCF_CC_50 [3]. (a) Input images, (b) ground truth density maps, and (c) estimation results.
Sensors 19 02013 g006
Figure 7. Density estimation results on the WorldExpo’10 dataset [14]. (a) Input images, (b) ground truth density maps, and (c) estimation results.
Figure 7. Density estimation results on the WorldExpo’10 dataset [14]. (a) Input images, (b) ground truth density maps, and (c) estimation results.
Sensors 19 02013 g007
Table 1. Configuration of the proposed fully-convolutional network. The parameters in the convolutional layers are denoted as “(kernel size), (number of filters), (stride), (dilation rate)”. Unless otherwise stated, the stride and dilation rate are all set to 1 by default.
Table 1. Configuration of the proposed fully-convolutional network. The parameters in the convolutional layers are denoted as “(kernel size), (number of filters), (stride), (dilation rate)”. Unless otherwise stated, the stride and dilation rate are all set to 1 by default.
Layer NameOutput SizeConfiguration
Conv Block 1 448 × 448 3 × 3,16
Conv Block 2 224 × 224 3 × 3,32 , s t r i d e 2
1 × 1,16 3 × 3,32
Conv Block 3 112 × 112 3 × 3,64 , s t r i d e 2
1 × 1 , 32 3 × 3 , 64 × 2
Transition 112 × 112 1 × 1,32 3 × 3,32
Atrous Block 1 112 × 112 1 × 1,16 3 × 3,32 , r a t e = 2
1 × 1,16 3 × 3,32 , r a t e = 3
1 × 1,16 3 × 3,32 , r a t e = 5
Atrous Block 2 112 × 112 1 × 1,16 3 × 3,32 , r a t e = 2
1 × 1,16 3 × 3,32 , r a t e = 3
1 × 1,16 3 × 3,32 , r a t e = 5
Distillation 112 × 112 1 × 1,16 3 × 3,32
1 × 1,1
Context Fusion 112 × 112 1 × 1,1
Table 2. Comparison of the proposed networks with different configurations on the ShanghaiTech Part A dataset. Config., configuration.
Table 2. Comparison of the proposed networks with different configurations on the ShanghaiTech Part A dataset. Config., configuration.
MethodsParametersFLOPSMAERMSE
Config.A130.40 k2040 m78.5126.4
Config.B119.60 k1911 m92.1142.0
Config.C223.46 k3389 m78.1129.2
Table 3. Comparison of the networks on the ShanghaiTech Part A dataset.
Table 3. Comparison of the networks on the ShanghaiTech Part A dataset.
MethodsParametersFLOPSMAERMSE
MCNN [8]127.95 k5384 m110.2185.9
CSRNet [13]16.26 m82,889 m68.2115.0
FCN-7c-3s [10]150.92 k14,327 m80.6126.7
Ours130.40 k2040 m78.5126.4
Table 4. Comparison of networks on the ShanghaiTech dataset.
Table 4. Comparison of networks on the ShanghaiTech dataset.
Part APart B
MethodMAERMSEMAERMSE
MCNN [8]110.2173.226.441.3
Cascaded-MTL [4]101.3152.420.031.1
FCN-7c-3s [10]80.6126.710.218.3
Ours78.5126.412.822.1
Switching-CNN [12]90.4135.021.633.4
CP-CNN [9]73.6106.420.130.1
CSRNet [13]68.2115.010.616.0
IG-CNN [36]72.5118.213.621.1
ACSCP [42]75.7102.717.227.4
Table 5. Comparison of networks on the UCF_CC_50 dataset.
Table 5. Comparison of networks on the UCF_CC_50 dataset.
MethodMAERMSE
Onoro et al. [11]465.7371.8
MCNN [8]377.6509.1
Cascaded-MTL [4]322.8341.4
Ours299.1391.8
Switching-CNN [12]318.1439.2
CP-CNN [9]295.8320.9
IG-CNN [36]291.4349.4
ACSCP [42]291.0404.6
Table 6. Comparison of the networks on the WorldExpo’10 dataset.
Table 6. Comparison of the networks on the WorldExpo’10 dataset.
MethodSce.1Sce.2Sce.3Sce.4Sce.5Avg.
Zhang et al. [14]9.814.114.322.23.712.9
MCNN [8]3.420.612.913.08.111.6
FCN-7c-3s [10]2.516.512.220.52.910.9
Ours2.427.07.98.43.19.76
Switching-CNN [12]4.214.914.218.74.311.2
CP-CNN [9]2.914.710.510.45.88.9
CSRNet [13]2.911.58.616.63.48.6
IG-CNN [36]2.616.110.1520.27.611.3
ACSCP [42]2.814.059.68.12.97.5

Share and Cite

MDPI and ACS Style

Yu, Y.; Huang, J.; Du, W.; Xiong, N. Design and Analysis of a Lightweight Context Fusion CNN Scheme for Crowd Counting. Sensors 2019, 19, 2013. https://0-doi-org.brum.beds.ac.uk/10.3390/s19092013

AMA Style

Yu Y, Huang J, Du W, Xiong N. Design and Analysis of a Lightweight Context Fusion CNN Scheme for Crowd Counting. Sensors. 2019; 19(9):2013. https://0-doi-org.brum.beds.ac.uk/10.3390/s19092013

Chicago/Turabian Style

Yu, Yang, Jifeng Huang, Wen Du, and Naixue Xiong. 2019. "Design and Analysis of a Lightweight Context Fusion CNN Scheme for Crowd Counting" Sensors 19, no. 9: 2013. https://0-doi-org.brum.beds.ac.uk/10.3390/s19092013

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop