Next Article in Journal
Prescribed-Time Convergent Adaptive ZNN for Time-Varying Matrix Inversion under Harmonic Noise
Previous Article in Journal
Monoscopic Phase Measuring Deflectometry Simulation and Verification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

YOLOv3_ReSAM: A Small-Target Detection Method

1
School of Computer Science and Engineering, Xi’an Technological University, Xi’an 710021, China
2
New Network and Testing Control National and Local Joint Engineering Laboratory, Xi’an 710021, China
3
Ordnance Science and Technology College, Xi’an Technological University, Xi’an 710021, China
*
Author to whom correspondence should be addressed.
Submission received: 24 March 2022 / Revised: 25 April 2022 / Accepted: 25 April 2022 / Published: 20 May 2022
(This article belongs to the Section Computer Science & Engineering)

Abstract

:
Small targets in long-distance aerial photography have the problems of small size and blurry appearance, and traditional object detection algorithms face great challenges in the field of small-object detection. With the collection of massive data in the information age, traditional object detection algorithms have been gradually replaced by deep learning algorithms and have an advantage. In this paper, the YOLOV3-Tiny backbone network is augmented by using the pyramid structure of image features to achieve multi-level feature fusion prediction. In order to eliminate the loss of spatial feature information and hierarchical information caused by pooling operations in convolution processes and multi-scale operations in multi-layer structures, a spatial attention mechanism based on residual structure is proposed. At the same time, the idea of reinforcement learning is introduced to guide bounding box regression on the basis of the rough positioning of the native boundary regression strategy, and the variable IoU calculation method is used as the evaluation index of the reward function, and the boundary regression model based on the reward mechanism is proposed for fine adjustment. The VisDrone2019 data set was selected as the experimental data support. Experimental results show that the mAP value of the improved small-object detection model is 33.15%, which is 11.07% higher than that of the native network model, and the boundary regression accuracy is improved by 23.74%.

1. Introduction

With the advantage of having a large perspective at high altitude, security patrol UAVs can eliminate blind spots in human patrols, quickly and efficiently extract feature information, automatically identify and locate targets in images, quickly capture suspicious non-cooperative targets and use them to accurately locate and identify targets in border patrol and border reconnaissance, saving a lot of manpower and time costs [1]. However, long-distance shooting images are different from still images in natural scenes, with high background complexity, small target size and blurry appearance, and target detection technology based on traditional methods has low accuracy for small-target detection and weak adaptability to targets in complex backgrounds, which are prone to leakage of alarms and false alarms, so the aerial small-target detection technology applied to low-altitude UAV platforms has important research significance. Traditional methods usually use server-side processing and send processing results to the on-board terminal, in which a large number of network requests consume a lot of computer resources while maintaining good network conditions. With the rise of artificial intelligence technology, convolutional neural network models are constantly changing, and models are constantly optimizing parameter redundancy, which can be directly deployed on airborne devices with limited computing resources. Therefore, under the premise of ensuring the correct rate, how to greatly reduce the number of parameters and calculations of the network model has practical application value.

2. Related Work

Aiming at the problem that the target scale in the UAV aerial image changes greatly and the proportion of small targets is much higher than that in the natural scene image [2,3], the convolutional neural network is effectively integrated into the target detection algorithm of aerial images [4], but the early convolutional neural network model is complex and the target detection efficiency is low. By constantly adjusting the network structure, the size of the anchor frame and the resolution of the feature map, it can adapt to small-target detection [5]. Although the convolutional neural network has a certain generalization ability, the convolution and pooling operations in the network make it difficult to detect small targets. Yoshihashi et al. in [6] proposed a joint framework that uses spatio-temporal information to detect and track small flying objects simultaneously. Their deep learning network consisted of four modules: convolutional, ConvLSTM, intercorrelation and fully connected layer. Learning motion patterns from multi-frame images with ConvLSTM improves inspection performance; however, these networks are computationally very expensive [7]. To achieve feature extraction at different scales and improve the performance of small-target detection, the bottom features and top features with rich semantic information are fused [8], but for targets of different sizes, the generalization learning ability is low. For this optimization structure, the effect of small-target detection is improved to a certain extent by using horizontal connection [9,10] to strengthen feature propagation.
Inspired by the human visual system, many researchers have introduced attention modules into convolutional neural networks to improve target detection performance in recent years. Attention modules are mainly divided into channel attention [11], spatial attention [12] and channel and spatial mixed attention [13]. Because the imaging angle of low-altitude UAV aerial photography is different from that of natural scene images, small targets usually exist in long-distance images, and the backgrounds are complex. The unique shooting angle and larger width also bring noise interference to the detection task. Therefore, the spatial attention module [14,15,16] is usually introduced into the target detection task to weight the characteristics of the target area, and the feature extraction network can selectively focus on the target area containing important information, suppress other irrelevant information and reduce the influence of background on the detection results so as to improve the detection performance of the model for small targets. The attention module is introduced to gradually extract the target region through the cascaded attention model and generate the convolution features of attention perception to guide the feature learning process, highlight the features related to the target and suppress the interference of the background [17].
While improving the accuracy of target detection, researchers are also constantly exploring to improve the regression efficiency of target bounding boxes and the positioning accuracy of bounding boxes. The YOLO detection framework [18] divides the input image into grids of different scales and then predicts two frames for each grid, each frame containing the confidence of the target and the probability of the target category and finally removes the less likely target window and redundant window to obtain the final detection result. YOLO introduces the regression idea to transform the target detection problem into a regression problem, which can greatly improve the detection speed. However, because it only uses grids with different scales for regression, it cannot accurately locate the target, so the accuracy is reduced. With the wide application of deep reinforcement learning, a target location algorithm based on deep reinforcement learning can be used [19]. The agent can perform a series of simple deformation actions on the bounding box according to the learned strategy by regarding the entire picture as an environment, and an agent is introduced to learn a top-down search strategy for the bounding box. The agent can perform a series of simple deformation actions on the bounding box according to the learned strategy, and the deep reinforcement learning algorithm based on multi-agent cooperation is used to learn the optimal strategy of cooperative target location [20]. The algorithm can effectively improve target location accuracy and locate the target accurately according to this contextual information.

3. Materials and Methods

This paper counts the use of the target detection algorithm represented by YOLO and the target algorithm represented by RCNN in the VisDrone 2019 Challenge, of which the target detection algorithm represented by YOLO has a small amount of operation and occupies an advantage in real-time detection, but the detection accuracy is low, and poor use only accounts for 4% of the use of the entire participating team, but this article considers its lightweight characteristics while taking into account the performance, so the YOLO series model is selected as the backbone model. Among them, the YOLOv3-Tiny network model inherits the superior performance advantages of the YOLOv3 series model in end-to-end real-time object detection and has the advantages of both being lightweight and real-time detection. However, there are still deficiencies in detection performance, and there is a semantic missing problem for the extraction of fine-grained target features which needs to be adjusted and optimized by the model structure to make up for the lack of detection performance. In this paper, the structure of the backbone network is modified as shown in Figure 1.

3.1. Expand the Image Feature Pyramid Structure

The common target detection network model uses top-level features for prediction, as shown in Figure 2. The semantic information of top-level features is rich, but the fine-grained features of small targets are ignored. For small targets, the detection efficiency is not high, and only the low-level feature semantic information is used for fine-grained feature extraction, which is relatively accurate. YOLOv3-Tiny network follows the idea of different scale predictions in YOLOv3. After multiple rounds of upsampling and fusion of features to the last layer, the features of the last layer are predicted to form an image feature pyramid structure model. YOLOv3-Tiny network is fused with 13 ∗ 13 small-scale YOLO layer and 26 ∗ 26 mesoscale layer, respectively. According to the definition of the International Society of Optical Engineering (SPIE), the target with an imaging area less than 80 pixels in the image with a small target of 256 ∗ 256, that is, if the target size is less than 0.12% of the original image, it can be considered as a small target. In this paper, taking 416 416 images as the input, the target with an imaging area less than 200 pixels is considered as a small target. YOLOv3-Tiny network is fused with 13 ∗ 13 small-scale YOLO layer and 26 ∗ 26 mesoscale layer, respectively, and expands 52 ∗ 52 large-scale layer on the original basis for multi-scale feature fusion prediction.
Figure 2 shows the common structure of the original target detection algorithm. The top-level feature prediction is adopted. After multiple rounds of upsampling and fusion of features to the top-level, the top-level feature is used for prediction. After improving the YOLOv3-Tiny network, the image feature pyramid structure is realized, as shown in Figure 3. Using the multi-level and multi-scale characteristics of this structure, the visual saliency region can be accurately captured without any prior knowledge, but the rescaling operation caused by the step-by-step operation of the maximum pool layer and convolution layer will change the receptive field of the feature map, cause the loss of spatial information hierarchy, affect the accuracy of pixel-level classification and reduce the performance of the saliency prediction network.

3.2. Spatial Attention Module Is Injected Based on the Residual Structure

According to the improved YOLOv3-Tiny network based on the image feature pyramid structure, the rescaling operation caused by multi-level and multi-scale operation will change the receptive field of the feature map, cause the loss of spatial information hierarchy, affect the accuracy of pixel-level classification and reduce the performance of significance prediction network. In this regard, the YOLOv3-Tiny network adds an attention module based on the residual structure to accurately locate the target features in space.
In the target detection data set, the proportion of small-target pixels is small. Adding spatial domain attention can enhance the compactness between different regions by modeling the correlation between different regions in the image so as to enhance the characteristic response of small targets. At the same time, the channel attention module and spatial attention module are tested on the PascalVOC2007 data set. The results show that the SE module represented by the channel attention module improves the accuracy by 1% and increases the amount of calculation by 2% and speed loss by 10%, while the SAM module represented by the spatial attention module improves the accuracy by 0.5% and only increases the amount of calculation by 0.1%. The specific SAM attention module model is shown in Figure 4.
Each pixel in the feature map is fed back through maximum pooling and average pooling, and the response weight of each pixel is obtained. Gradient backpropagation is carried out according to the maximum weight value, and finally, the region with the largest response in the feature map is obtained, which is regarded as the visual significance region. The time complexity of model training will affect the convergence ability of the network model. Through the residual structure, we can speed up the calculation of the weight of the spatial attention module and shorten the training time.
The spatial attention module is directly added to the original network structure vertically. The addition of the new module will increase the network depth. The deeper the network, the more abstract the features extracted and the more semantic the information. However, simply increasing the number of network layers will cause the gradient to disappear, so the residual network is added to deepen the network horizontally, as shown in Figure 5 and Figure 6. It can obtain more semantic information about features and solve the problem of gradient dispersion in gradient descent.
When the chain derivative is used to obtain the partial derivative of a parameter, the result of the partial derivative is always small in the deep network, resulting in too-small parameter updates. Therefore, the Loss value is smaller in the process of backpropagation by calculating the gradient in the training process, which increases the time complexity of model training and affects the convergence ability of the network model. The Loss function is shown in Equation (1), and the gradient calculation is shown in Equation (2).
L o s s = F ( X L , W L , b L )
L o s s X L = F ( X L , W L , b L ) X L
where X L represents the input value of the current layer, W L represents the weight value of the current layer and b L represents the offset term of the current layer. In the back-propagation process, the gradient of input X 1 is calculated as shown in Equation (3).
X L X 1 = F N ( X L N , W L N , b L N ) X L * * F 2 ( X L 2 , W L 2 , b L 2 ) X 1
By calculating the gradient, it can be seen that if one of the derivatives is very small, the gradient may become smaller and smaller after multiple multiplications, which is often called gradient dissipation. With the deepening of the network, for the deep network, the gradient value is very small when it is transferred from the deep layer close to the output to the shallow layer close to the input so that the shallow layer cannot be updated effectively. When the residual block is added, the gradient is calculated as shown in Equation (4).
X L X l = X l + F ( X l , W l , b l ) X l = 1 + F ( X L , W L , b L ) X L
At this time, the related problems of gradient descent are solved by the above methods. Moreover, after adding the fast residual, the fitting of function F(x) = 0 will be easier than F(x) = x and the change of parameters will be more stable. When the residual learning is in the shallow layer, the learning is linear superposition, and when it reaches the deep layer, F(x) will tend to zero, realizing the identity mapping so as to avoid the increase in layers affecting the learning results. At this time, on the premise of not affecting the learning results, we can increase the number of network layers, extract more fine-grained features and obtain more small-target feature information.
Considering that the model is more lightweight, the weight response can still be obtained for each pixel in the feature map by replacing the pooling operation with a 1 ∗ 1 convolution layer, and finally, the maximum value of regional weight can be obtained. The specific structure is shown in Figure 7.

3.3. K-Means Algorithm Improves the Prior Size of the Target Prediction Candidate Box

The improved YOLOv3-Tiny model follows the idea of a preset anchor box, sets 9 anchor boxes as target prediction candidate boxes and clusters the target types in the data set through the clustering algorithm to obtain the scale size of the target preselection box. As a classical clustering algorithm, the k-means algorithm is widely used because of its simplicity and efficiency. The k-means algorithm is sensitive to the selection of initial seed points, and the selection of initial seed points has a great impact on the final clustering results and running time, and completely random selection may lead to slow convergence of the algorithm. Optimized for random initialization of seed points in the k-means algorithm. The basic idea of the improved algorithm is to select points far away from each other as seed points. The steps are as follows:
  • Step 1: Randomly select an object from the input object set as the first cluster center;
  • Step 2: Calculate the distance between each object and the nearest cluster center;
  • Step 3: Select a new object as the new cluster center. The principle of selection is that the points with large distance are more likely to be selected as the cluster center;
  • Step 4: Repeat steps (2) and (3) until K clustering centers are selected;
  • Step 5: Finally, the k clustering centers are used as the initial clustering centers of the standard k-means algorithm.

3.4. The Goal Frame Regression Strategy Model Based on the Reward Module Is Established

For windows, four-dimensional vectors (x, y, w, h) are generally used to represent the coordinates of the center point and the width and height of the window, respectively. For Figure 8, the red box P represents the original proposal candidate target box, the blue box G represents the target box predicted by the boundary box regression algorithm, the green box G represents the ground truth real target box of the target, the red circle represents the center point of the selected candidate target box and the green circle represents the center point of the real target box. The blue circle represents the center point of the target box predicted by the selected bounding box regression algorithm. Our goal is to find a relationship so that the input original window P is mapped to a regression window G closer to the real window G .
Reinforcement learning provides a formal framework to solve the problem of what strategies agents (agent) adopt to maximize cumulative rewards in the environment. Regarding the entire image as an environment, the agent deforms the bounding box through concentrated actions. The goal is to make the bounding box tightly surround the target [21]. In order to build a complete reinforcement learning system, specific actions, states and reward functions are defined for target positioning tasks. According to the idea of reinforcement learning, the goal frame regression strategy model based on the reward module is established as follows [22].
Action set A: The set of actions an agent can take in order to achieve its goals. This article defines nine actions to search for the candidate area, including eight actions to deform the candidate area and one action to terminate the search. The specific definition of the action set is as follows: A: {Rightward, leftward, up, down, bigger, smaller, wider, higher, stop}. Each action is proportional to the size of the current bounding box α. If a discrete change to its size is made, the terminated action represents that the agent thinks it has found the target.
Status set s: It represents the agent’s understanding of the information of the current environment. The state is defined as a tuple: s = (o, h). O is a feature vector of the current observation area, that is, the four-dimensional vector of the target frame, which is composed of the YOLOv3-Tiny_ReSAM network is used to extract, and h is a fixed size vector, representing the action history taken by the agent, and the output is a vector with dimension 9, representing nine actions.
Reward function R: It represents the judgment of the environment on the quality of the selected action in this state and is used to guide the agent to learn the optimal strategy according to the current state. When the agent takes action a, when entering the next state s’ from state s, a reward R a ( a , s s ) is given by the environment to agent. The reward signal defines what actions are close to the target and what actions are far away from the target for the agent in the current state, that is, whether the actions taken in the current state contribute to the positioning of the target. The definition of the reward function is as follows.
R a ( a , s s ) = s i g n ( D I o U ( b , g , c ) D I o U ( b , g , c ) )
where DIoU is the intersection deformation ratio IoU between target g and boundary box b:
D I o U ( b , g , c ) = I o U ( b , g ) ρ 2 ( b , g ) c 2
where the center points of the prediction frame and the real frame are denoted by b and g and the Euclidean distance between the two center points is denoted by ρ . c is the diagonal distance of the minimum closure region that can contain both the prediction frame and the real frame. According to the action set, state set and reward function, the Berman equation is used to iteratively update the function:
Q ( s , a ) = r + γ max a Q ( s , a )
where s is the current state, a is the currently selected action, R is the immediate reward, γ represents the discount coefficient, s’ represents the next state and a’ represents the next action. After several iterations, the action is terminated. Finally, the experimental results will be obtained by comparing the accuracy of the target frame established by the model with that of the original algorithm.

4. Experimental Setup

4.1. Test the Target Detection Accuracy of the Improved Network Structure

Based on the above, YOLOv3-Tiny_ReSAM network structure and the original YOLOv3-Tiny network structure are tested against the open-source data set of VisionDrone2019 UAV aerial photography, and the following test results are obtained. After the improvement, the loss value for training 20,000 iterations is reduced to 18 and the loss value for training 100,000 epochs iterations is reduced to 30, as shown in Figure 9.
The test set of VisionDrone2019 UAV aerial photography open source data set is tested and the recall rate results are obtained. The recall rate reflects the number ratio of various targets successfully detected in the whole data test set. The test results of 300 iterations are drawn as shown in Figure 10. The results show that the missed detection rate of the improved network model for various targets decreases significantly. With the increase in the number of iterations, according to the curve trend, the recall rate increases and the missed detection rate decreases.
The open-source data set of VisionDrone2019 UAV aerial photography is tested and the average accuracy results are obtained. The average accuracy reflects the number ratio of various targets correctly identified in the whole data test set. By comparing two different network model tests, the average accuracy (AP) and mean average precision(mAP) of various targets are obtained, as shown in Table 1. Among them, the mAP0.5 value of the test result obtained by the YOLOv3-Tiny network model due to the small structure of the model is low and the mAP0.5 value of the YOLOv3_ReSAM network model test result is in the middle position, which is slightly lower than the mAP0.5 value of the YOLOv4 network model and the improved TPH_YOLOv5 network model. The improved YOLOv3_ReSAM network model has poor detection effect in two types of target detection, pedestrian and people, and has a significant advantage in the four types of targets of car, bus, motor and tricycle. Table 2 shows the results of the VisionDrone2019-Challenge competition. The detection model proposed in this paper is only 4.22% lower than the first mAP, as shown in Table 2.
The accuracy of various targets is further analyzed according to Table 1 to obtain the recognition accuracy of each target in the training process, as shown in Figure 11. The improved network model has a fast convergence speed and high accuracy.
The open-source test set images of VisionDrone2019 UAV aerial photography are selected for a visual test and the following test results are obtained. See Figure 12 for details.

4.2. Improved K-Means Clustering Algorithm and Gave a Priori Box Size Resultse

By default, the YOLOv3 algorithm is set to 3 anchor boxes per unit. A larger number of anchor boxes can better cover the target, that is, the average IOU of each anchor box and its assigned target real boundary box is larger. Therefore, 9 cluster centers are set, corresponding to 3 anchor boxes per unit. Based on the open-source data set of aerial photography of VisionDrone2019, the following test results are obtained through the k-means clustering model, as shown in Table 3.

4.3. Improved Boundary Regression Algorithm and Gave Accuracy Test Results

The boundary regression strategy based on the reward module tests the accuracy of the boundary box. The comparison curve of the iterative test IoU of the boundary regression algorithm before and after the improvement is shown in Figure 13. According to the curve, it can be clearly observed that under the improved boundary regression strategy, the average IOU of the boundary box accuracy (taking IoU value as the criterion) is increased from 0.485 to 0.597.
The results shown in Figure 14 show that the default boundary regression algorithm of native YOLOv3 depends on the size of the boundary prediction candidate box and has poor migration ability for small targets, as shown in Figure 14a,c. On the basis of the original algorithm, fine adjustment is made on the basis of the original coarse positioning through the boundary regression strategy based on the reward module. The specific test results are shown in Figure 14b,d. Note that Figure 14 is an enlarged image in the same scale after cutting the original test image after the experiment.

5. Conclusions

In this paper, the parallel structure of the feature pyramid is expanded and the shallow features are integrated so as to extract more fine-grained feature information of small targets in the shallow network model and predict them in the top-level feature map. In the multi-level parallel feature pyramid structure, the loss of spatial information hierarchy due to multi-level and multi-scale operations is compensated for the loss of accuracy of pixel-level classification by injecting spatial attention mechanism, and the response of small target features is strengthened while weakening background features. Experimental results show that the average accuracy mAP of the improved network model is 11.07% higher than that of the native network model, and the average recall rate of a single image is stable at about 45%. However, the background of low-altitude UAV detection of small targets on the ground is complex, the target features are blurred at long distances and the network model is huge and it is difficult to achieve real-time detection, which is still being continuously improved, but there is still no ideal solution.
A boundary regression strategy based on the reward mechanism is also established, and the reinforcement learning idea is introduced on the basis of the rough positioning of boundary regression in the native network model. Experimental results show that the finely adjusted boundary regression results are 23.74% higher than those of the native boundary regression algorithm. However, in the process of boundary regression, the precision and speed of fine adjustment depend to a large extent on the rough positioning basis of the native boundary regression algorithm. If the bounding box regression error is too large under the native boundary regression strategy, it will take a lot of time to adjust in the later refinement adjustment process. At the same time, the boundary regression refinement adjustment strategy leads to an increase in the amount of additional operations, which will delay the forward inference speed of the model, thereby affecting the real-time nature of the target detection algorithm. Therefore, in later research, it is necessary to optimize the native boundary regression strategy but also to dynamically optimize the reward function according to the coarse positioning accuracy of the fine adjustment strategy.

Author Contributions

Investigation, B.L.; Software, H.L.; Supervision, B.L.; Visualization, H.W.; Data curation, S.W. All authors have read and agreed to the published version of the manuscript.

Funding

Shaanxi Provincial Natural Science Basic Research Program Project (2019JM-603).

Data Availability Statement

The data that support the findings of this study are openly available in VisDrone-Dataset at https://github.com/VisDrone/VisDrone-Dataset (23 March 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, H.; Liu, B.; Liu, Z.; Shi, J.; Wang, P. LSS target threat estimation and capture prediction trajectory simulation. In Proceedings of the 2020 7th International Conference on Information Science and Control Engineering (ICISCE), Changsha, China, 18–20 December 2020; pp. 806–813. [Google Scholar] [CrossRef]
  2. Chen, T.; Zhang, Y.; Zhang, S.; Shao, Q. Research on small target detection algorithm based on deep learning network. Electron. Design Eng. 2020, 28, 84–87. [Google Scholar]
  3. Azimi, S.M.; Vig, E.; Bahmanyar, R.; Körner, M.; Reinartz, P. Towards multi-class object detection in unconstrained remote sensing imagery. In Proceedings of the Asian Conference on Computer Vision, Perth, Australia, 2–6 December 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 150–165. [Google Scholar]
  4. Sevo, I.; Avramovic, A. Convolutional neural net-work based automatic object detection on aerial images. IEEE Geosci. Remote Sens. Lett. 2016, 13, 740–744. [Google Scholar] [CrossRef]
  5. Sommer, L.W.; Schuchert, T.; Beyerer, J. Fast deep vehicle detection in aerial images. In Proceedings of the WACV 2017: 2017 IEEE Winter Conference on Applications of Computer Vision, Santa Rosa, CA, USA, 24–31 March 2017; IEEE Computer Society: Washington, DC, USA, 2017; pp. 311–319. [Google Scholar]
  6. Yoshihashi, R.; Trinh, T.T.; Kawakami, R.; You, S.; Iida, M.; Naemura, T. Differentiating Objects by Motion: Joint Detection and Tracking of Small Flying Objects. arXiv 2017, arXiv:1709.04666. [Google Scholar]
  7. Alqaysi, H.; Fedorov, I.; Qureshi, F.Z.; O’Nils, M. A temporal boosted yolo-based model for birds detection around wind farms. J. Imaging 2021, 7, 227. [Google Scholar] [CrossRef] [PubMed]
  8. Lin, S.; Dollar, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; IEEE Computer Society: Washington, DC, USA, 2017; pp. 2117–2125. [Google Scholar]
  9. Yang, X.; Sun, H.; Fu, K.; Yang, J.; Sun, X.; Yan, M.; Guo, Z. Automatic ship detection in remote sensing images from google earth of complex scenes based on multiscale rotation dense feature pyramid networks. Remote Sens. 2018, 10, 132. [Google Scholar] [CrossRef] [Green Version]
  10. Wang, J.; Ding, J.; Guo, H.; Chen, W.; Pan, T.; Yang, W. Mask OBB: A semantic attention-based mask oriented bounding box representation for multi-category object detection in aerial images. Remote Sens. 2019, 11, 2930. [Google Scholar] [CrossRef] [Green Version]
  11. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  12. Jaderberg, M.; Simonyan, K.; Zisserman, A.; Kavukcuoglu, K. Spatial transformer networks. In Proceedings of the Neural Information Processing Systems 2015, Montreal, QC, Canada, 7–12 December 2015; pp. 2017–2025. [Google Scholar]
  13. Woo, S.; Park, J.; Lee, J.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  14. Zhu, Y.; Zhao, C.; Guo, H.; Wang, J.; Zhao, X.; Lu, H. Attention CoupleNet: Fully Convolutional Attention Coupling Network for Object Detection. IEEE Trans. Image Process. 2019, 28, 113–126. [Google Scholar] [CrossRef] [PubMed]
  15. Li, H.; Li, C.; An, J.; Ren, J. Remote sensing image target detection based on improved convolutional neural network with attention module. Chin. J. Image Graph. 2019, 24, 1400–1408. [Google Scholar]
  16. Xiang, Y.; Wang, Q.; Li, X.; Yu, M.; Jiang, H.; Gao, J.; Liu, Z.; Yu, R. Multi-Attention Object Detection Model in Remote Sensing Images Based on Multi-Scale. IEEE Access 2019, 7, 94508–94519. [Google Scholar]
  17. Zhu, Y.; Zhao, C.; Wang, J.; Zhao, X.; Wu, Y.; Lu, H. CoupleNet: Coupling Global Structure with Local Parts for Object Detection. In Proceedings of the 2019 IEEE International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; IT: Venice, DC, USA, 2017; pp. 4146–4154. [Google Scholar]
  18. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  19. Caicedo, J.C.; Lazebnik, S. Active object localization with deep reinforcement learning. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 2488–2496. [Google Scholar]
  20. Kong, X.; Xin, B.; Wang, Y.; Hua, Y. Collaborative deep reinforcement learning for joint object search. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; IEEE Computer Society: Washington, DC, USA, 2017; pp. 1695–1704. [Google Scholar]
  21. Yao, H.-G.; Zhang, W.; Yang, H.-Q.; Yu, J. Joint regression object localization based on deep reinforcement learning. Acta Autom. Sin. 2020, 41, 1–10. [Google Scholar] [CrossRef]
  22. Hamid, R.; Nathan, T.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized Intersection over Union: A Metric and A Loss for Bounding Box Regression. In Proceedings of the 2019 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; IEEE Computer Society: Washington, DC, USA, 2019; pp. 658–666. [Google Scholar]
  23. Samyal, A.S.; Hans, S. Analysis and Adaptation of YOLOv4 for Object Detection in Aerial Images. arXiv 2022, arXiv:2203.10194. [Google Scholar]
  24. Zhu, X.; Lyu, S.; Wang, X.; Zhao, Q. TPH-YOLOv5: Improved YOLOv5 Based on Transformer Prediction Head for Object Detection on Drone-captured Scenarios. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 2778–2788. [Google Scholar]
Figure 1. YOLOv3-Tiny_ReSAM network structure (the green box is the improvement part).
Figure 1. YOLOv3-Tiny_ReSAM network structure (the green box is the improvement part).
Electronics 11 01635 g001
Figure 2. Top-level feature prediction structure.
Figure 2. Top-level feature prediction structure.
Electronics 11 01635 g002
Figure 3. Image feature pyramid prediction structure.
Figure 3. Image feature pyramid prediction structure.
Electronics 11 01635 g003
Figure 4. Structure of spatial attention module SAM.
Figure 4. Structure of spatial attention module SAM.
Electronics 11 01635 g004
Figure 5. Residual block structure.
Figure 5. Residual block structure.
Electronics 11 01635 g005
Figure 6. Adding SAM based on the residual structure.
Figure 6. Adding SAM based on the residual structure.
Electronics 11 01635 g006
Figure 7. Add attention mechanism based on SAM module (the red box is the improved part).
Figure 7. Add attention mechanism based on SAM module (the red box is the improved part).
Electronics 11 01635 g007
Figure 8. Boundary box regression(The red box P represents the original proposal candidate target box, the blue box G represents the target box predicted by the bounding box regression algorithm, the green box G represents the ground truth target box of the target, the red circle represents the center point of the candidate frame, the green circle represents the center point of the real target frame, and the blue circle represents the center point of the selected bounding box regression algorithm predicts the target box).
Figure 8. Boundary box regression(The red box P represents the original proposal candidate target box, the blue box G represents the target box predicted by the bounding box regression algorithm, the green box G represents the ground truth target box of the target, the red circle represents the center point of the candidate frame, the green circle represents the center point of the real target frame, and the blue circle represents the center point of the selected bounding box regression algorithm predicts the target box).
Electronics 11 01635 g008
Figure 9. Loss curves during training: (a) Loss value for training 100,000 iterations. (b) Loss value for training 20,000 iterations.
Figure 9. Loss curves during training: (a) Loss value for training 100,000 iterations. (b) Loss value for training 20,000 iterations.
Electronics 11 01635 g009
Figure 10. Recall rate curve.
Figure 10. Recall rate curve.
Electronics 11 01635 g010
Figure 11. (a) The result curve of various target recognition accuracy values under YOLOv3-Tiny network model. (b) The right figure is the result curve of various target recognition accuracy values under YOLOv3-Tiny_ReSAM network model.
Figure 11. (a) The result curve of various target recognition accuracy values under YOLOv3-Tiny network model. (b) The right figure is the result curve of various target recognition accuracy values under YOLOv3-Tiny_ReSAM network model.
Electronics 11 01635 g011
Figure 12. Visual test of small-target detection.
Figure 12. Visual test of small-target detection.
Electronics 11 01635 g012
Figure 13. IoU curve.
Figure 13. IoU curve.
Electronics 11 01635 g013
Figure 14. Boundary regression visualization test.
Figure 14. Boundary regression visualization test.
Electronics 11 01635 g014
Table 1. Average precision of various targets under different network models (Bold is an improved model based on YOLOv3-Tiny).
Table 1. Average precision of various targets under different network models (Bold is an improved model based on YOLOv3-Tiny).
VisDrone2019-DET-TestmAP
PedestrianPeopleBicycleCarVanTrunkTricycleAwing-
Tricycle
BusMotor
YOLOv3_Tiny15.52%15.66%25.19%80.21%43.83%25.64%16.75%14.22%34.85%38.96%22.08%
YOLOv3_ReSAM17.28%15.78%28.70%84.31%53.77%42.96%30.48%24.38%62.35%61.57%33.15%
YOLOv4 [23]44.03%28.18%23.14%84.67%52.67%54.31%31.416%26.72%65.36%45.87%35.72%
TPH_YOLOv5 [24]29%16.75%15.69%68.94%49.79%45.16%27.33%24.72%61.80%30.90%37.32%
Table 2. VisionDrone2019-Challenge competition results(The first was the model that won first place in the VisionDrone2019-Challenge competition, the eighth is an improved model based on YOLOv3-Tiny).
Table 2. VisionDrone2019-Challenge competition results(The first was the model that won first place in the VisionDrone2019-Challenge competition, the eighth is an improved model based on YOLOv3-Tiny).
RankingModel and MethodmAP (%)
1DPNetV337.37
2SMPNet35.98
3DBNet35.73
4CDNet34.19
5ECascade R-CNN34.09
6FPAFS-CenterNet32.34
7DOHR-RetinalNet21.68
8YOLOv3-Tiny-ReSAM33.15%
Table 3. Sizes of multi-scale target candidate boxes.
Table 3. Sizes of multi-scale target candidate boxes.
Serial Number13 ∗ 13 Prediction Frame (w, h)26 ∗ 26 Prediction Frame (w, h)52 ∗ 52 Prediction Frame (w, h)Accuracy
12, 34, 123, 721, 1813, 1812, 1133, 4327, 1826, 2631.24%
21, 34, 52, 68, 93, 116, 1617, 1512, 2630, 3839.79%
32, 96, 511, 1013, 2833, 2645, 1261, 3458, 109112, 10358.63%
41, 32, 84, 57, 1817, 154.1213.2831.3978, 5548.51%
55, 710, 1112, 1320, 1014, 1822, 1156, 3345, 67106, 13760.99%
67, 1812, 1117, 1513, 2921, 1833, 4372, 6352, 121116, 12467.88%
71, 43, 66, 74, 127, 1812, 1121, 1813, 3033, 4230.08%
82, 46, 89, 711, 1016, 1529, 1630, 4940, 5683, 6455.39%
96, 513, 910, 1521, 1319, 3334, 1574, 5529, 89124, 13560.78%
107, 313, 1412, 1113, 1116, 2333, 4171, 6873, 75127, 11867.63%
116, 1011, 1215, 1516, 1220, 3436, 3979, 5459, 88121, 13365.24%
122, 410, 59, 814, 1223, 1939, 2445, 5767, 81113, 9761.27%
132, 34, 79, 910, 1517, 1418, 2630, 2137, 4386, 4343.45%
142.86.610.915.1320.1130.2433.4154, 4976, 5149.64%
1510, 1316, 3033, 2330, 6162, 4559, 119116, 90156, 198373, 32643.64%
Note: Among them, the marked red is the target candidate box size with the highest accuracy under the VisionDrone2019 UAV aerial photography open-source data set and the marked blue is the default target candidate box size of the native network under the VisionDrone2019 UAV aerial photography open-source data set.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, B.; Luo, H.; Wang, H.; Wang, S. YOLOv3_ReSAM: A Small-Target Detection Method. Electronics 2022, 11, 1635. https://0-doi-org.brum.beds.ac.uk/10.3390/electronics11101635

AMA Style

Liu B, Luo H, Wang H, Wang S. YOLOv3_ReSAM: A Small-Target Detection Method. Electronics. 2022; 11(10):1635. https://0-doi-org.brum.beds.ac.uk/10.3390/electronics11101635

Chicago/Turabian Style

Liu, Bailin, Huan Luo, Haotong Wang, and Shaoxu Wang. 2022. "YOLOv3_ReSAM: A Small-Target Detection Method" Electronics 11, no. 10: 1635. https://0-doi-org.brum.beds.ac.uk/10.3390/electronics11101635

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop