Next Article in Journal
Design of a Standard and Programmatically Accessible Interface for Smart Meters to Allow Monitoring Automation of the Energy Consumed by the Execution of Computer Software
Previous Article in Journal
Investigating Lifelong Learners’ Continuing Learning Intention Moderated by Affective Support in Online Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Infrared Small-Target Detection Using Multidirectional Local Difference Measure Weighted by Entropy

Faculty of Artificial Intelligence in Education, Central China Normal University, Wuhan 430079, China
*
Author to whom correspondence should be addressed.
Sustainability 2023, 15(3), 1902; https://0-doi-org.brum.beds.ac.uk/10.3390/su15031902
Submission received: 27 October 2022 / Revised: 31 December 2022 / Accepted: 10 January 2023 / Published: 19 January 2023

Abstract

:
Detecting small targets from infrared remote sensing images is still a challenging task. In this research, we propose a multidirectional local difference measure weighted by entropy (MDLDE) to detect small targets from infrared images with messy backgrounds. First, a new multidirectional local difference measure is proposed to suppress the clutter background. Then, the entropy, which captures the overall heterogeneity between the target and the background, is utilized to enhance the target. Lastly, an adaptive threshold was adopted to segment the target region from the background. The designed MDLDE could effectively enhance the target and simultaneously suppress the background clutter. Experimental results on six datasets indicate that the proposed method outperformed other state-of-the-art methods in terms of the signal-to-clutter ratio gain (SCRG), background suppression factor (BSF), and receiver operating characteristic (ROC) curves.

1. Introduction

In infrared search and tracking (IRST) systems, infrared (IR) small-target detection is an active research field that occupies a crucial position and is broadly applied to a variety of fields, including early-warning, precision-guidance, and missile-tracking systems [1,2,3]. However, due to the far imaging distance and atmospheric transmissions, IR targets are small and have no concrete shape or texture features. In most instances, targets are drowned in heavy clutter and noise [4,5]. Additionally, target detection is usually deteriorated by background edges and high-brightness backgrounds. All those unfavorable factors render IR target detection more challenging [6,7,8].
Over the past several decades, researchers have been committed to studying IR small-target detection and have designed all kinds of algorithms [9,10,11,12,13,14,15]. Existing IR small-target detection methods are mainly divided into the following two categories: the first is single-frame detection, which is suitable for high-speed motion detection platforms and has faster detection speed, so it has more practical applications. The other is sequential detection. Unlike the former, sequential detection has limitations in stationary targets and rapidly changing backgrounds. In addition, it is usually based on single-frame detection such as the top-hat filter [16], max-mean/max-median filter [17], and wavelet transformation [18]. These methods are classified as traditional single-frame detection methods. However, they deliver dramatically declining performance when the background is messy.
Among numerous single-frame detection methods, methods that integrate the human visual system (HVS) are paid much attention by many researchers. The HVS indicates that human vision is more focused on contrast than brightness. As the most important characteristic of HVS, the contrast mechanism has been broadly adopted in IR small-target detectors [19]. For instance, Chen et al. [20] considered the central cell’s maximal gray value and calculated the ratio to the mean gray value of background cells to enhance the local contrast. Unlike the above method, Wei et al. [21] took into account the gray mean of the central patch instead of the maximal value and calculated the difference between the central patch and the eight background patches in the corresponding directions. Furthermore, Xiong et al. [22] also fully used directional information and proposed to detect targets in the gradient vector field. HB-MLCM [23] and MLHM [24] further enhanced the target area by designing an improved high-boost filter and calculating intrapatch homogeneity, respectively. However, in these methods, the size of the supposed background region in the sliding window increases proportionally with the growth of the target cell, which causes problems such as expansion effects and background clutter enhancement, thus affecting the detection results with false alarms and misdetections.
Recently, some novel methods for improving the structure of sliding windows have been proposed, such as TLLCM [25], in which study a trilayer filter was developed to measure the local contrast. Aghaziyarati et al. [26] designed a novel multiscale sliding window with only a one-pixel-wide background and claimed that it could compensate for three weak points of the regular methods. Wu et al. [27] proposed a novel double-neighborhood window. In addition, many infrared small-target detection methods integrating weighting functions have emerged [28,29,30,31]. For instance, Deng et al. [28] employed local entropy to weigh multiscale gray differences. Liu et al. [30] compared the gray value of the pixel in the central area with its neighborhood pixels to calculate the probability that the pixel belongs to targets. The above operations increase the robustness of the algorithm to a certain extent. However, the detection performance of those methods worsens when targets are submerged in a complex background with heavy clutter. In this paper, to robustly detect small targets from IR images with cluttered backgrounds, a novel algorithm named multidirectional local difference measure weighted by entropy (MDLDE) is proposed.
Our contribution is twofold:
(1)
A multidirectional local difference measure is proposed that can adaptively enhance a target despite variance in its size.
(2)
The overall heterogeneity measured by entropy is used for weighting directional local difference measures to enhance the robustness of the proposed method.
This paper is organized as follows. In Section 2, we describe the proposed MDLDE method in detail. The effectiveness of MDLDE is verified on six datasets in Section 3. In Section 4, the paper is concluded.

2. Methodology

2.1. Multidirectional Local Difference Measure

The flowchart of MDLDE is given in Figure 1, where the detected targets are marked in a red box. First, a multidirectional local difference measure is proposed to suppress the background. Second, the Renyi entropy is calculated to enhance the target. Both take a multiscale approach, and the saliency map is obtained. Lastly, the target segmentation separates the target from the saliency map.
To characterize the contrast between the central and surrounding areas, we adopted the sliding window given by MPCM [21] that divides the local area into the central patch (denoted as T) and eight surrounding background patches (denoted as B i , i = 1 , 2 , , 8 ). Then, the dissimilarity between patches, along with different directions, is defined as follows:
d i = d ( T , B i ) × ( T , B i + 4 ) , ( i = 1 , 2 , , 4 ) ,
where × stands for the multiplication operation, and d ( T , B i ) = m T m B i , ( i = 1 , 2 , , 8 ) . m T and m B i represent the mean values of the patch T and patch B i .
The local difference at a given scale is defined as follows:
D 1 = min i = 1 , 2 , , 4 d i .
Subsequently, to deal with the problems caused by the proposed background increasing proportionally to the target cell in the complex background and inspired by the idea of background-pixel fixation in the cumulative directional derivatives of [26], a novel multiscale convolutional kernel with eight directions is designed. As shown in Figure 2, the size of each kernel on the scale n was set to ( n + 2 ) × ( n + 2 ) , which was composed of an n × n target window (surrounded by a bold box in Figure 2) and a single-pixel-wide surrounding edge. Then, the pixel values along the given direction in the target window were set to 1 except for the central one. In contrast, the pixel at the directional far end of the kernel was set as the opposite cumulative number.
In the proposed multiscale convolutional kernel, the target window is supposed to change with the target size. Meanwhile, the size of its surrounding background is fixed to one pixel wide, which is less sensitive to background size. When the target window slides into the target area and is just the right size for the target, it is obvious that the pixel value of the target area is significantly greater than that of the surrounding background, so the target can be enhanced to some extent. Moreover, setting the central pixel of the convolutional kernel as 0 can suppress pixel-size noise with high brightness (PNHB). The multiscale convolutional kernel with eight directions takes into account the directional information of the local region; thus, sharp edges can be eliminated.
Generally, infrared targets are brighter or darker than the corresponding surrounding environment. Most methods only consider the positive contrast between the target and its surroundings to enhance a target. The negative contrast generated by a dark target is treated as the background. To tackle the negative contrast values caused by dark targets, it is necessary to analyze the types of targets in the scene, so a weighting coefficient E ( x ) in three application scenarios is constructed as follows: When there are only bright targets in the scene,
E ( x ) = 1 x 0 0 x < 0 .
When there are only dark targets in the scene,
E ( x ) = 1 x < 0 0 x 0 .
When there are both dark and bright targets in the scene,
E ( x ) = 1 x 0 1 x < 0 .
In practical application, the application of the E ( x ) coefficient needs to be selected according to the type of scene. Then, to calculate the difference between the target and the fixed background in the i-th direction, we calculated the cumulative directional derivatives with
K i = I F i × E ( I F i ) , ( i = 1 , 2 , , 8 ) ,
where F i is the cumulative directional filter of the i-th direction, I is the original image, and ∗ is the convolutional operation. The minimal value of K i is
D 2 = min i = 1 , 2 , , 8 K i .
Lastly, the multidirectional local difference measure can be defined as follows:
D = D 1 × D 2 .
With D, our method is more robust to the variance in size between the target and its local surrounding textural structures, thus notably reducing the false alarm rate (FAR).

2.2. Entropy-Based Target Enhancement

Directional local difference measure generally emphasizes the minimal differences between the central patch and its directional neighborhoods while overlooking the overall heterogeneity between the target and its surrounding background.
Regarding the consideration above, we introduced Rényi entropy to weight the local difference measure D. Rényi entropy is represented as follows:
H α ( x ) = 1 1 α log 2 i = 1 n p i α .
When α = 2 , entropy was more sensitive to local brightness variation, so entropy can be defined as follows:
H = log 2 i = T , B P i 2 , P i = m i i = T , B m i ,
where m i represents the mean value of the pixel in the i-th ( i = T , B ; T = t a r g e t , B = b a c k g r o u n d ) region. Figure 3 illustrates the background edge (EB), high-brightness background (HB), and target region (T) in a real IR image and their corresponding entropy values. When the central region belongs to the target region, its diversity with the background is remarkable. Thus, entropy is small. Otherwise, when the central region is part of the background, the variation in gray values is slight, thus leading to a larger entropy value.
With H, the heterogeneity of the central region and its whole surrounding background is described. By enhancing our multidirectional local difference measure with entropy, we can obtain the saliency map C as follows:    
C = D H .
For each scale, the sliding window scans the whole image from the top left to the bottom right. Lastly, the maximal C of different scales is taken as the contrast value in the final saliency map (MDLDE map).
The computational process of MDLDE is given in Algorithm 1, where L is the largest scale, and C l represents the saliency map on the l-th scale. p 1 and q 1 stand for the number of rows and columns of the input image, respectively.
Algorithm 1 MDLDE
Input: Input Image
Output: MDLDE map
   1: for  l = 1 : L  do
   2:    Compute multidirectional local difference measure map D according to Equation (8).
   3:    Compute entropy-based target enhancement map H according to Equation (10).
   4:    Obtain the saliency map C l according to Equation (11).
   5: end for
   6: for p = 1 : p 1 do
   7:    for  q = 1 : q 1   do
   8:      C ^ p , q = max l = 1 , 2 , , L C l ( p , q )
   9:     Replace the current central pixel value with the obtained C ^ p , q .
 10:    end for
 11: end for
 12: return MDLDE map

2.3. Target Segmentation

The greater the contrast value in the saliency map is, the greater the probability of it belonging to the target. Aiming at segmenting the target from the image, the adaptive threshold of the proposed algorithm is given as follows:
T h = μ + k × δ ,
where k is an adjustable parameter, μ and δ stand for the mean value and standard deviation of the MDLDE map, respectively. If the pixel’s value on the saliency map is greater than T h , it can be extracted as if it belongs to the target region.

2.4. Algorithm Analysis

In addition to the target itself, other elements in the infrared image, such as background, pixel-size noise with high brightness, and background edge, all affect the effect of target detection to a certain extent. However, it is important to suppress these elements while enhancing the target. Next, we discuss the proposed algorithm from the following different scenarios.
(1)
If the pixel belongs to the target (T): the target is significantly heterogeneous from the background and is usually brighter (bright target) or darker (dark target) than the background. Then, we can find that
D T > 0 , H T < 1 .
Thus,
M D L D E T > 0 .
In the multidirectional convolutional kernel, the central pixel of the target window is set to 0, which means that the current pixel does not participate in the calculation. However, since the neighboring pixel value of the current pixel is usually greater than the pixel value of the background area, the value of the multidirectional local difference is greater than 0. Moreover, entropy can further enhance the target area. Therefore, the target area is effectively enhanced.
(2)
If the pixel belongs to the background (B): the gray value of the background pixel is usually equal to that of the neighborhoods, so
D B 0 , H B > H T .
Thus,
M D L D E B 0 .
(3)
If the pixel belongs to pixel-size noise with high brightness (PNHB):
D 2 P N H B 0 , D P N H B 0 , H P N H B > H T ,
thus
M D L D E P N H B 0 .
When the pixel is the central pixel of the target window, the pixel is not calculated in the multidirectional convolutional kernel, so PNHB is equal to 0 in the multidirectional local difference measure and can be eliminated eventually.
(4)
If the pixel belongs to the background edge (EB):
D E B 0 , H E B > H T ,
thus
M D L D E E B 0 .
The above analysis shows that MDLDE can effectively enhance the target while suppressing other elements in the infrared image.

3. Experimental Results

3.1. Experimental Data

To validate the detection performance of the MDLDE, infrared image sequences of various scenes, including sky, sea–sky, ground, and architectural scenes, were adopted in our experiments. These scenes constituted six image sequences denoted as Sequences 1–6. The six datasets’ details are shown in Table 1. Meanwhile, Figure 4 shows examples of six real infrared image sequences.
Since the size of an infrared small target is usually in the range of 3 × 3–9 × 9, the number of scales L was set to 4, and the size of each scale was chosen as n = 3, 5, 7, 9 in our experiments. To validate the proposed method, seven advanced algorithms, namely, LCM [20], MPCM [21], RLCM [32], AADCDD [26], HWLCM [33], DNGM [27], and NTFRA [34], were used to compare with the proposed MDLDE. The parameters of all comparison methods were set with reference to their own articles.
All the experiments were conducted with MATLAB R2019b on a computer with a 3.8 GHz Intel Core i7 CPU and 8 GB RAM.

3.2. Evaluation Metrics

To evaluate the detection capability of algorithms in different scenes, BSF, SCRG, and ROC curves were adopted as the evaluation metrics.
BSF and SCRG were used to measure the degree of background suppression and the validity of target enhancement, which are defined as follows:
B S F = δ i n δ o u t + w ,
S C R G = S C R o u t S C R i n , S C R = P t μ b δ b + w ,
where δ i n and δ o u t denote the standard deviation of the raw image and the saliency map, respectively. P t stands for the gray peak of the target. μ b and δ b are the average grayscales and standard deviation of the background. w is an adjustable coefficient to avoid dividing by 0, which was set to 0.001 in this paper. S C R i n and S C R o u t represent the SCR values of the raw image and the saliency map, respectively.
By plotting the probability of detection (PD) and FAR under different threshold settings, the ROC curve can be created. PD and FAR can be represented as follows:
P D = n u m b e r o f d e t e c t e d t r u e t a r g e t s t o t a l n u m b e r o f r e a l t a r g e t s × 100 % ,
F A R = n u m b e r o f d e t e c t e d f a l s e t a r g e t s n u m b e r o f t e s t e d f r a m e s × 100 % .

3.3. Experimental Results and Comparisons

The 3D mesh view of example images and their saliency maps derived from different algorithms is shown in Figure 5. Since most infrared small-target detection methods can only be applied to bright targets, the E ( x ) of the proposed method is set only to detect the bright targets in all the six image sequences. The saliency maps of all algorithms were normalized to [0 1] in the experiment. Figure 5a1–a6 show the 3D mesh view of the raw image. As shown in Figure 5b1–b6, LCM could not suppress the background well in all six image sequences. It enhanced the target region to some extent, but was still not significant enough. In Sequence 1, MPCM significantly enhanced the hole-shaped dark areas in thick clouds. Except Sequence 3, HWLCM and RLCM also obtained strong residual clutter in the other five sequences, thus leading to low contrast, which is an obstacle to separating the target from the background. For Sequence 2, in the scene with heavy background clutter and a low signal clutter ratio, MPCM generated much background clutter in the saliency map. As shown in Figure 5h2, NTFRA did not enhance the dim target since the saliency map was flat. AADCDD significantly enhanced the nontarget region in Sequence 5. Our algorithm had strong capabilities for background suppression while enhancing the target, so it could achieve better results in real infrared image sequences. In the scene with only bright targets, all algorithms could enhance the target region to different degrees, but when dark targets appeared in the scene (see Figure 5a3–i3), except MPCM, most of the algorithms could not detect the dark target. In this case, we also set E(x) of the scene with bright and dark targets to enhance the target area in the corresponding environment. Overall, the proposed method performed best, enhancing the target while effectively suppressing the background.
The ROC curves of different algorithms in various scenes are presented in Figure 6. For Sequence 1, there were heavy clouds in the background. When FAR was 0, NTFRA, AADCDD, and MDLDE algorithms could detect all targets, and LCM and RLCM had slightly higher FAR values. In Sequence 2, because the target in the scene was dim and had continuous high-brightness noise in some images, none of the other seven methods could detect all the targets, but our method had a 100% PD while obtaining a low FAR. In Sequence 3, MPCM achieved superior detection results because it can detect dark and bright targets. Sequence 4 mainly consisted of several single frames of sky-scene images, including sea–sky, ground, and architectural scenes. In Sequence 4, the proposed method performed best. RLCM, AADCDD, DNGM, and MPCM also achieved good results, while the HWLCM, LCM, and NTFRA algorithms had higher FAR when the PD was relatively low. The target brightness in Sequence 5 was low, and the proposed method could achieve the highest detection rate and the lowest false alarm rate, which has significant advantages compared with other methods. In Sequence 6, the ROC curves of the proposed method were in the upper left corner, indicating better detection performance. Figure 7 shows the area under the curve(AUC) of different algorithms in six sequences. The figure shows that the proposed method achieved the maximal AUC value in Sequences 1, and 4–6, and obtained a suboptimal value in Sequence 2. The experimental results of ROC curve indicate that the proposed algorithm is more robust in a variety of scenarios.
Table 2 and Table 3 show the SCRG and BSF values of different algorithms, respectively. The maximal SCRG and BSF values are marked in bold and are underlined in the table. In Sequence 2, NTFRA achieved the maximal BSF value, while the SCRG value was 0 because NTFRA cannot effectively detect dim targets, and the saliency map of NTFRA is a plane with a value of 0. The proposed algorithm obtained the maximal SCRG values in Sequences 1 and 4–6, and suboptimal values in Sequence 2. Meanwhile, it obtained the maximal BSF values in Sequences 1, 3, 5, and 6, and suboptimal values in Sequences 2 and 4. The performance of the proposed method in BSF and SCRG showed that it had superior background suppression and background enhancement.
Table 4 presents the time consumption of different algorithms. Compared with LCM, MPCM, AADCDD, and DNGM, the proposed method requires more running time, but the difference is relatively small, which means that the time consumption of the proposed method is at the same level as those of LCM, MPCM, and AADCDD. The detection speed of RLCM, HWLCM, and NTFRA is slower, requiring more time. The running time of the proposed algorithms is primarily constituted by the following two parts: multidirectional local difference measure and entropy computation. Theoretically, the computational complexity of the MDLDE algorithm is approximately O ( L M N C 2 ) , where M and N represent the number of rows and columns in the raw image, respectively. L is the scale number. C is the size of each scale. LCM, MPCM, and AADCDD had the same time complexity as the proposed method. The time complexity of RLCM and HWLCM is O ( L C 2 log C 2 M N ) . The time complexity of NTFRA and DNGM is O ( n 1 n 2 n 3 log ( n 1 n 2 n 3 ) + i = 1 3 min ( n i , n i + 1 ) ) and O ( M N C 2 ) , respectively. In conclusion, the proposed algorithm has medium computational complexity.

4. Conclusions

In this paper, an IR small-target detection method based on the local difference measure combined with entropy was proposed. First, the minimal difference between the central patch and its directional neighborhoods was calculated with a novel multiscale convolutional kernel. Second, the overall heterogeneity between the target and its surrounding background was computed via entropy, and the saliency map was then obtained. Finally, an adaptive threshold was adopted to extract the target region from the saliency map. By utilizing this new detection model, we could suppress clutters and noise and effectively enhance the targets. Extensive experimental results illustrate that our algorithm could robustly detect IR small targets under complex backgrounds.
However, the proposed algorithm has limitations in time efficiency. Potential further improvement may include applying a GPU to accelerate the procedures.

Author Contributions

Conceptualization, Y.W. and H.Y.; methodology & software, H.Y. and L.L.; writing—original draft preparation, H.Y. and L.L.; visualization, L.L. and D.C.; writing—review & editing, Y.W., H.Y., L.L. and M.T.; project administration, Y.W. and M.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under grant 62277029, and the Fundamental Research Funds for the Central Universities under grants CCNU22JC025, CCNU22JC011, CCNU20ZN002, CCNU20TD005.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gao, C.; Meng, D.; Yang, Y.; Wang, Y.; Zhou, X.; Hauptmann, A.G. Infrared patch-image model for small target detection in a single image. IEEE Trans. Image Process. 2013, 22, 4996–5009. [Google Scholar] [CrossRef] [PubMed]
  2. Qin, Y.; Li, B. Effective infrared small target detection utilizing a novel local contrast method. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1890–1894. [Google Scholar] [CrossRef]
  3. Deng, H.; Sun, X.; Liu, M.; Ye, C.; Zhou, X. Entropy-based window selection for detecting dim and small infrared targets. Pattern Recognit. 2017, 61, 66–77. [Google Scholar] [CrossRef]
  4. Zhu, H.; Ni, H.; Liu, S.; Xu, G.; Deng, L. Tnlrs: Target-aware non-local low-rank modeling with saliency filtering regularization for infrared small target detection. IEEE Trans. Image Process. 2020, 29, 9546–9558. [Google Scholar] [CrossRef] [PubMed]
  5. Deng, H.; Sun, X.; Zhou, X. A multiscale fuzzy metric for detecting small infrared targets against chaotic cloudy/sea-sky backgrounds. IEEE Trans. Cybern. 2018, 49, 1694–1707. [Google Scholar] [CrossRef]
  6. Yao, S.; Chang, Y.; Qin, X. A coarse-to-fine method for infrared small target detection. IEEE Geosci. Remote Sens. Lett. 2018, 16, 256–260. [Google Scholar] [CrossRef]
  7. Zhu, H.; Liu, S.; Deng, L.; Li, Y.; Xiao, F. Infrared small target detection via low-rank tensor completion with top-hat regularization. IEEE Trans. Geosci. Remote Sens. 2019, 58, 1004–1016. [Google Scholar] [CrossRef]
  8. Dai, Y.; Wu, Y.; Zhou, F.; Barnard, K. Attentional local contrast networks for infrared small target detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9813–9824. [Google Scholar] [CrossRef]
  9. Kim, S.; Yang, Y.; Lee, J.; Park, Y. Small target detection utilizing robust methods of the human visual system for IRST. J. Infrared Millim. Terahertz Waves 2009, 30, 994–1011. [Google Scholar] [CrossRef]
  10. Shao, X.; Fan, H.; Lu, G.; Xu, J. An improved infrared dim and small target detection algorithm based on the contrast mechanism of human visual system. Infrared Phys. Technol. 2012, 55, 403–408. [Google Scholar] [CrossRef]
  11. Moradi, S.; Moallem, P.; Sabahi, M.F. Fast and robust small infrared target detection using absolute directional mean difference algorithm. Signal Process. 2020, 177, 107727. [Google Scholar] [CrossRef]
  12. Dai, Y.; Wu, Y.; Song, Y. Infrared small target and background separation via column-wise weighted robust principal component analysis. Infrared Phys. Technol. 2016, 77, 421–430. [Google Scholar] [CrossRef]
  13. Dai, Y.; Wu, Y.; Zhou, F.; Barnard, K. Asymmetric contextual modulation for infrared small target detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 950–959. [Google Scholar]
  14. Hou, Q.; Wang, Z.; Tan, F.; Zhao, Y.; Zheng, H.; Zhang, W. RISTDnet: Robust infrared small target detection network. IEEE Geosci. Remote Sens. Lett. 2021, 19, 7000805. [Google Scholar] [CrossRef]
  15. Yang, L.; Yan, P.; Li, M.; Zhang, J.; Xu, Z. Infrared Small Target Detection Based on a Group Image-Patch Tensor Model. IEEE Geosci. Remote. Sens. Lett. 2022, 19, 7003505. [Google Scholar] [CrossRef]
  16. Bai, X.; Zhou, F. Analysis of new top-hat transformation and the application for infrared dim small target detection. Pattern Recognit. 2010, 43, 2145–2156. [Google Scholar] [CrossRef]
  17. Deshpande, S.D.; Er, M.H.; Venkateswarlu, R.; Chan, P. Max-mean and max-median filters for detection of small targets. In Proceedings of the Signal and Data Processing of Small Targets 1999, Denver, CO, USA, 18–23 July 1999; Volume 3809, pp. 74–83. [Google Scholar]
  18. Li, L.; Tang, Y.Y. Wavelet-hough transform with applications in edge and target detections. Int. J. Wavelets Multiresolution Inf. Process. 2006, 4, 567–587. [Google Scholar] [CrossRef]
  19. Han, J.; Ma, Y.; Zhou, B.; Fan, F.; Liang, K.; Fang, Y. A robust infrared small target detection algorithm based on human visual system. IEEE Geosci. Remote Sens. Lett. 2014, 11, 2168–2172. [Google Scholar]
  20. Chen, C.P.; Li, H.; Wei, Y.; Xia, T.; Tang, Y.Y. A local contrast method for small infrared target detection. IEEE Trans. Geosci. Remote Sens. 2013, 52, 574–581. [Google Scholar] [CrossRef]
  21. Wei, Y.; You, X.; Li, H. Multiscale patch-based contrast measure for small infrared target detection. Pattern Recognit. 2016, 58, 216–226. [Google Scholar] [CrossRef]
  22. Xiong, B.; Huang, X.; Wang, M. Local gradient field feature contrast measure for infrared small target detection. IEEE Geosci. Remote Sens. Lett. 2020, 18, 553–557. [Google Scholar] [CrossRef]
  23. Shi, Y.; Wei, Y.; Yao, H.; Pan, D.; Xiao, G. High-boost-based multiscale local contrast measure for infrared small target detection. IEEE Geosci. Remote Sens. Lett. 2017, 15, 33–37. [Google Scholar] [CrossRef]
  24. Nie, J.; Qu, S.; Wei, Y.; Zhang, L.; Deng, L. An infrared small target detection method based on multiscale local homogeneity measure. Infrared Phys. Technol. 2018, 90, 186–194. [Google Scholar] [CrossRef]
  25. Han, J.; Moradi, S.; Faramarzi, I.; Liu, C.; Zhang, H.; Zhao, Q. A local contrast method for infrared small-target detection utilizing a tri-layer window. IEEE Geosci. Remote Sens. Lett. 2019, 17, 1822–1826. [Google Scholar] [CrossRef]
  26. Aghaziyarati, S.; Moradi, S.; Talebi, H. Small infrared target detection using absolute average difference weighted by cumulative directional derivatives. Infrared Phys. Technol. 2019, 101, 78–87. [Google Scholar] [CrossRef]
  27. Wu, L.; Ma, Y.; Fan, F.; Wu, M.; Huang, J. A Double-Neighborhood Gradient Method for Infrared Small Target Detection. IEEE Geosci. Remote Sens. Lett. 2021, 18, 1476–1480. [Google Scholar] [CrossRef]
  28. Deng, H.; Sun, X.; Liu, M.; Ye, C.; Zhou, X. Small infrared target detection based on weighted local difference measure. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4204–4214. [Google Scholar] [CrossRef]
  29. Nasiri, M.; Chehresa, S. Infrared small target enhancement based on variance difference. Infrared Phys. Technol. 2017, 82, 107–119. [Google Scholar] [CrossRef]
  30. Liu, J.; He, Z.; Chen, Z.; Shao, L. Tiny and dim infrared target detection based on weighted local contrast. IEEE Geosci. Remote Sens. Lett. 2018, 15, 1780–1784. [Google Scholar] [CrossRef]
  31. Han, J.; Moradi, S.; Faramarzi, I.; Zhang, H.; Zhao, Q.; Zhang, X.; Li, N. Infrared small target detection based on the weighted strengthened local contrast measure. IEEE Geosci. Remote Sens. Lett. 2020, 18, 1670–1674. [Google Scholar] [CrossRef]
  32. Han, J.; Liang, K.; Zhou, B.; Zhu, X.; Zhao, J.; Zhao, L. Infrared small target detection utilizing the multiscale relative local contrast measure. IEEE Geosci. Remote Sens. Lett. 2018, 15, 612–616. [Google Scholar] [CrossRef]
  33. Du, P.; Hamdulla, A. Infrared small target detection using homogeneity-weighted local contrast measure. IEEE Geosci. Remote Sens. Lett. 2019, 17, 514–518. [Google Scholar] [CrossRef]
  34. Kong, X.; Yang, C.; Cao, S.; Li, C.; Peng, Z. Infrared small target detection via nonconvex tensor fibered rank approximation. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5000321. [Google Scholar] [CrossRef]
Figure 1. Flowchart of MDLDE.
Figure 1. Flowchart of MDLDE.
Sustainability 15 01902 g001
Figure 2. Eight convolutional kernels in different directions with scale n = 5. (ah) F1–F8.
Figure 2. Eight convolutional kernels in different directions with scale n = 5. (ah) F1–F8.
Sustainability 15 01902 g002
Figure 3. Different types of components in real infrared images and their corresponding entropy values (where the vertical axis represents the entropy value, and the horizontal and vertical axes represent regions where different elements are located). EB, background edge; HB, high-brightness background; T1, T2, and T3, target regions; E, entropy value of the central pixel.
Figure 3. Different types of components in real infrared images and their corresponding entropy values (where the vertical axis represents the entropy value, and the horizontal and vertical axes represent regions where different elements are located). EB, background edge; HB, high-brightness background; T1, T2, and T3, target regions; E, entropy value of the central pixel.
Sustainability 15 01902 g003
Figure 4. Example frames of six sequences. (af) Sequences 1–6.
Figure 4. Example frames of six sequences. (af) Sequences 1–6.
Sustainability 15 01902 g004
Figure 5. Three-dimensional mesh view of example images and saliency maps of different algorithms: (a1a6) 3D mesh view of the raw image, (b1b6) LCM, (c1c6) MPCM, (d1d6) RLCM, (e1e6) AADCDD, (f1f6) HWLCM, (g1g6) DNGM, (h1h6) NTFRA, (i1i6) proposed method. Red box, detected bright target; green box, detected dark target; yellow box, location of the target in (h2).
Figure 5. Three-dimensional mesh view of example images and saliency maps of different algorithms: (a1a6) 3D mesh view of the raw image, (b1b6) LCM, (c1c6) MPCM, (d1d6) RLCM, (e1e6) AADCDD, (f1f6) HWLCM, (g1g6) DNGM, (h1h6) NTFRA, (i1i6) proposed method. Red box, detected bright target; green box, detected dark target; yellow box, location of the target in (h2).
Sustainability 15 01902 g005
Figure 6. ROC curves of different algorithms on six sequences. (af) Sequences 1–6.
Figure 6. ROC curves of different algorithms on six sequences. (af) Sequences 1–6.
Sustainability 15 01902 g006
Figure 7. AUC of different methods on six sequences. (af) Seqs. 1–6.
Figure 7. AUC of different methods on six sequences. (af) Seqs. 1–6.
Sustainability 15 01902 g007
Table 1. Details of the experimental data.
Table 1. Details of the experimental data.
SeqFramesSizeTarget NumberTarget TypeImage Detial
170256 × 201bright targetSky scene
High-brightness background
260320 × 2401bright targetLow SCR, dim target
PNHB
320576 × 7683bright/dark targetSea surface, sea–sky lines
One dark target
4184Multiple1bright targetSky, sea–sky, ground,
and architectural scenes
5200320 × 2561bright targetSky scene
Thick clouds
6200320 × 2561bright targetSky scene
Heavy clutters
Table 2. SCRG values of different algorithms.
Table 2. SCRG values of different algorithms.
SeqLCMMPCMRLCMAADCDDHWLCMDNGMNTFREProposed
10.4911.861.0822.461.0545.583.6765.80
20.465.191.1018.941.78104.970.0030.50
32.09112.432.8039.374.5544.6890.3853.13
40.4312.902.2726.411.3552.4110.2747.49
51.7281.0328.09123.4225.66198.5748.44245.21
61.82204.0946.72239.9428.58245.14114.96295.19
Table 3. BSF values of different algorithms.
Table 3. BSF values of different algorithms.
SeqLCMMPCMRLCMAADCDDHWLCMDNGMNTFRAProposed
1125.111867.961080.782609.87705.773419.441481.263679.54
258.23895.20393.011951.16577.412057.8516,225.852249.11
3458.1113,356.345373.6217,332.803170.6017,928.357205.8220,156.38
4191.292871.361504.943648.151815.584471.421330.144115.80
5263.016028.281902.477556.931861.019600.543927.3111,797.75
6223.358803.592546.3810,480.111570.3011,037.576135.2012,725.94
Table 4. Time consumption of different algorithms (in seconds).
Table 4. Time consumption of different algorithms (in seconds).
SeqLCMMPCMRLCMAADCDDHWLCMDNGMNTFREProposed
10.34070.33643.10520.31683.15610.33042.25120.3526
20.36020.36034.82370.34224.85770.35723.33010.3526
30.55750.614729.82280.503328.02300.647510.33160.8310
40.40920.41373.52120.39703.59060.4053.36930.4509
50.42080.37614.80080.36104.85890.39392.97280.4353
60.37970.38294.73860.36464.76860.38963.06960.4353
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yao, H.; Liu, L.; Wei, Y.; Chen, D.; Tong, M. Infrared Small-Target Detection Using Multidirectional Local Difference Measure Weighted by Entropy. Sustainability 2023, 15, 1902. https://0-doi-org.brum.beds.ac.uk/10.3390/su15031902

AMA Style

Yao H, Liu L, Wei Y, Chen D, Tong M. Infrared Small-Target Detection Using Multidirectional Local Difference Measure Weighted by Entropy. Sustainability. 2023; 15(3):1902. https://0-doi-org.brum.beds.ac.uk/10.3390/su15031902

Chicago/Turabian Style

Yao, Huang, Liping Liu, Yantao Wei, Di Chen, and Mingwen Tong. 2023. "Infrared Small-Target Detection Using Multidirectional Local Difference Measure Weighted by Entropy" Sustainability 15, no. 3: 1902. https://0-doi-org.brum.beds.ac.uk/10.3390/su15031902

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop