Next Article in Journal
Hybrid Genetic Algorithm and Tabu Search for Solving Preventive Maintenance Scheduling Problem for Cogeneration Plants
Previous Article in Journal
Adaptive Neural Network Prescribed Time Control for Constrained Multi-Robotics Systems with Parametric Uncertainties
Previous Article in Special Issue
On Hybrid Numbers with Gaussian Leonardo Coefficients
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PDE-Constrained Scale Optimization Selection for Feature Detection in Remote Sensing Image Matching

1
PipeChina Network Corporation Eastern Oil Storage and Transportation Co., Ltd., Xuzhou 221008, China
2
School of Sciences, Southwest Petroleum University, Chengdu 610500, China
3
Institute of Artificial Intelligence, Southwest Petroleum University, Chengdu 610500, China
*
Author to whom correspondence should be addressed.
Submission received: 8 May 2024 / Revised: 4 June 2024 / Accepted: 11 June 2024 / Published: 17 June 2024

Abstract

:
Feature detection and matching is the key technique for remote sensing image processing and related applications. In this paper, a PDE-constrained optimization model is proposed to determine the scale levels advantageous for feature detection. A variance estimation technique is introduced to treat the observation optical images polluted by additive zero-mean Gaussian noise and determine the parameter of a nonlinear scale space governed by the partial differential equation. Additive Operator Splitting is applied to efficiently solve the PDE constraint, and an iterative algorithm is proposed to approximate the optimal subset of the original scale level set. The selected levels are distributed more uniformly in the total variation sense and helpful for generating more accurate and robust feature points. The experimental results show that the proposed method can achieve about a 30% improvement in the number of correct matches with only a small increase in time cost.

1. Introduction

Remote sensing refers to the detection of targets through various sensors mounted on different aircraft, and the obtained remote sensing image data often carry a large amount of information, requiring the use of different efficient processing techniques for practical purposes [1,2]. Image matching is an important technical foundation in remote sensing image processing, and the effectiveness of registration often directly affects the subsequent image processing results [3,4,5,6].
The work related to image matching can be traced back to P.E. Anuta’s work on cross correlation image detection [3] and Brown’s summary of previous work [4], which has since produced numerous influential works. For image pairs with significant deformation, Flusher proposed an adaptive mapping method that segments the image into multiple sub-blocks, calculates the similarity between the sub-blocks, matches them using the similarity of the sub-blocks, and registers the original image using the coordinate relationship between the sub-blocks [7]. Studholme proposed a normalized mutual information method that overcomes the sensitivity of traditional mutual information methods to the size of overlapping parts between images and achieves good matching results [8]. David Lowe proposed the classic Scale Invariant Feature Transform (SIFT) by combining the scale space pyramid and Gaussian kernel to improve the features invariance and detection rate of feature points [9,10]. The SIFT feature maintains the same changes in image grayscale, angle, scale, etc., and performs well in natural image matching. Some techniques based on SIFT have been developed over the past years. Ke and Sukthankar introduced principal component analysis to replace the weighted histograms and reduce the dimension of features to 20 [11]. Speeded-up robust features (SURFs) were presented based on a Hessian matrix and image convolutions to faster the computing [12]. Dellinger et al. proposed an improved SIFT registration algorithm for Synthetic Aperture Radar (SAR) images called SAR-SIFT [13]. This algorithm uses a new gradient calculation method to effectively suppress the noise impact of SAR images. Alcantarilla et al. proposed a novel multiscale 2D feature detection and description algorithm in nonlinear scale spaces, named KAZE [14,15]. The results revealed a step forward in performance both in detection and description compared to some traditional methods. Gong et al. utilized two matching strategies, feature matching and regional matching, and then proposed a rough-to-accurate registration method, which first uses the SIFT algorithm to coarsely register the image and then uses mutual information methods to evaluate the initial matching points and find accurate matching point pairs [16]. Wu et al. improved the Random Sample Consistency (RANSAC) algorithm and proposed using a fast iterative strategy to filter feature point pairs to obtain more accurate transformation parameters in order to achieve the accurate matching of remote sensing images [17]. In order to effectively solve the problem of significant changes in image grayscale leading to the failure of traditional registration algorithms, Ma et al. proposed a Position Scale and Orientation Scale Invariant Feature Transform (PSO-SIFT) algorithm based on the SIFT algorithm, which uses a new gradient definition to calculate feature point descriptors [6]. Accurate matching between remote sensing image pairs with significant grayscale changes has been achieved. Wu et al. proposed the Particle Swarm Optimization Sample Consistency (PSOSAC) algorithm, which uses particle swarm optimization to optimize the model parameters solved using RANSAC based on the RANSAC algorithm to obtain more accurate matching results [18]. There are some other techniques that have been also introduced to achieve better performances [19,20,21,22,23,24].
Most of these researches are based on a Gaussian Scale Space and have shown success in many problems. However, it is still to be further explored in more complex scenes, such as heterogeneous image matching, multi-modal image matching, and medical image matching [25,26,27]. A remote sensing image is much different from a nature image, not only in acquisition and transport but also in information presentation and vision characteristics [1,28,29]. It is necessary to develop more suitable and efficient feature describing methods for remote sensing image matching. Scale selection techniques can be found in various image processing tasks. Olson presented a method where the scale of the filtering and feature detection is varied locally according to the distance to the scene pixel estimated through stereoscopy [30]. Energy models inspired from Markov random fields and segmentation quality measures can be introduced to determine scales for different purposes [31,32]. PDE-constrained optimization has generated a great deal of attention in many fields [33,34]. In related problems, a certain functional is aimed to be optimized with one or more PDEs serving as constraints.
This paper aimed to present a novel scale optimization selection technique for local feature describing in remote sensing image matching. A nonlinear partial differential equation is configured with the noise estimation results to generate a continuous nonlinear scale space. Then, a PDE-constrained model is proposed to determine the optimized scale levels. It can be numerically solved based on Additive Operator Splitting (AOS) and subset selection. After some calculation steps similar to those of KAZE and SIFT are performed on the scale space, the feature points can be found for remote sensing image matching.
The rest of this paper is arranged as follows. In Section 2, a linear scale space and nonlinear scale space are introduced. In Section 3, an evolution partial differential equation configured with noise estimation is applied to describe a nonlinear scale space. And then, a PDE-constrained model is presented to optimally select the scale levels. In Section 4, the several experiments implemented in this study to verify the accuracy and efficiency of proposed algorithm are discussed.

2. Fundamentals and Basis

2.1. Linear Scale Space

Scale space theory has been introduced in many models, and it can help to distinguish the features at different levels. After a scale parameter is added to the related model, a multi-scale representation sequence can be computed following a change in the parameter. Many tasks can be achieved by analyzing or processing the sequence, such as extracting features.
The scale space method can simulate the process of people observing objects from near to far, and the continuous change in scale can effectively describe the essential features of the image. The Scale Invariant Feature Transform (SIFT) uses a Gaussian pyramid to represent the linear scale space, which means to convolve a serial of Gaussian kernel functions with the original image function to obtain smoothed images at different scales [10,35]. It can be formulated as
I ( x , t ) = I ( x ) g t ( x ) = I ( x ) 1 4 π t 2 exp | x | 2 4 t 2 , ( x , t ) Ω × Γ .
Here, t denotes the scale parameter, and x denotes the pixel position. Ω R 2 denotes the image domain, and Γ denotes the time interval. I ( x ) is the original image, and I ( x , t ) means the related linear scale space. The value of t represents the degree of smoothness of the original image. The larger the value, the higher the degree of smoothness, and then, different scale image features can be obtained through some specific steps, which are the following: calculating the Difference of Gaussian (DoG), searching for extreme points, determining the key points (interesting points) and the main direction, and generating feature descriptors.
The discrete difference of Gaussian space can be denoted as
D o G ( x , t ) = I ( x , t + Δ t ) I ( x , t ) .
Here, Δ t can be preset or determined according to the image data. The first two rows in Figure 1 show different Gaussian kernels and the corresponding smoothed images. The last row shows the difference between two neighbor images. With the scale parameter t increasing, the original image will be gradually smoothed, and the feature information will be redundant. A downsampling technique is adopted in SIFT to obtain a more compact version (small-sized) for later calculations.
Based on the scale space and the difference space, the original SIFT can be summarized as two main steps (see [10,36]).

2.1.1. Calculating the DoG and Determining the Key Points

K = x P , 2 D o G x 2 1 D o G x < 1 2 | D o G ( x ) | < 0.03 T r ( H ) 2 | H | < ( r + 1 ) 2 r .
Here, H means the Hessian of the DoG, and P = { arg max z B D o G ( z ) } denotes a set of extreme points found in some basis subareas of the image domain. For example, a subarea can be set to a circle-centered pixel x with radius d, that is, B ( x , d ) . The second derivative of the DoG is actually the Hessian in a 3D space (spatial 2D and scale 1D), and it is different from the Hessian in the 2D spatial space. ∧ denotes the Boolean operator AND. T r denotes the trace operator. r is a threshold used to determine good key points.

2.1.2. Orientation Assignment and Generating Feature Descriptors

F e a t s S I F T ( x ) = histogram R θ I ( B i , j ) , θ d 4 × 4 × 8
θ ( B ) = arg min θ d histogram ( B ( x ) , θ d ) .
Here, θ d , d = 1 , 2 , , 8 means the preset eight direction angles. B i , j is a subdomain of B ( x ) , which is the neighborhood centered keypoint x. R θ denotes the rotate matrix with angle θ . ∇ denotes the gradient operator. The image is divided into 4 × 4 sub-regions, and eight direction statistics are calculated on each corresponding gradient field. Then, the statistics on 16 sub-regions are collected to form a feature vector with dimensions 4 × 4 × 8 .

2.2. Nonlinear Scale Space

Stable feature points at different scales can be extracted on the previously established linear scale space. However, while smoothing noise, Gaussian convolution also smooths image details and edges, affecting the accuracy and uniqueness of feature points. Some methods have been introduced to improve this defect in the past years and achieved good results in applications [11,14,19,37,38]. The KAZE features are extracted from a nonlinear scale space that is generated using a related partial differential equation. And it has been validated to have a comparable efficiency on some problems in image denoising and restoration. In general, a nonlinear scale space can be denoted as the solution of a nonlinear partial differential equation:
I t ( x , t ) = div c | I | · I
Here, c | I | means the diffusion velocity, and the equation will degenerate into a linear version if c degenerates into a constant. Unlike the linear scale space, there is no explicit formulation of the nonlinear scale space but an implicit formulation described using the above Equation (5).
It is clear that the continuous nonlinear scale space is determined using the diffusion velocity and evolution time. There are lots of different configurations for them. For the velocity, there are several useful formulations, such as the following [14]:
c 1 ( s ) = exp s 2 b 2 , c 2 ( s ) = 1 1 + s 2 b 2 , c 3 ( s ) = 1 , s = 0 1 exp 3.315 s 8 / b 8 , s > 0
As there is no analytical solution for Equation (5), it is necessary to introduce some numerical technique such as Additive Operator Splitting (AOS) to approximate the solution at a different scale t. The discretization of the equation can be denoted as
I n + 1 I n τ = l = 1 m A l ( I n ) I n + 1
and an efficient semi-implicit scheme can be obtained as
I n + 1 = 1 τ l = 1 m A l ( I n ) 1 I n .
Here, A l is a matrix that encodes the image conductivities for each dimension, and this scheme is advantageous for large time steps. After taking some steps similar to those in SIFT, the image features can be generated.
Figure 2 shows three nonlinear scale spaces based on the velocity c 2 ( s ) with different values of k. It can be found that the total variation decreased at different rates. And if such a space is divided according to a fixed mode (such as uniformly or proportionally in the scale parameter), the differences between two adjacent parts are various and difficult to be balanced. It is not advantageous to detect the image feature points and achieve a satisfactory performance in image matching.

3. Proposed Method

It can be found that the continuous nonlinear scale space I ( x , t ) is absolutely determined through the velocity and evolution time. Different diffusion velocities are very different in smoothing or preserving the potential features, so a proper velocity should be selected for each image. Of course, the noise level is also an important factor affecting the determination. The number of scale levels is another important parameter affecting the feature point detection. It is easy to know that if more scale levels are set, a higher computation cost is required though less feature points may be lost. Even if a proper number of levels is set, it is still difficult to determine which levels should be set. Though some experience and recommend selection methods can be taken from previous works, we still tried to present an optimization model to solve these difficulties.
The surroundings in remote sensing images is a complex variable, and the signal is susceptible to being polluted. There are also many changes in the type and intensity of image features; a flexible pattern of an alternating scale space is advantageous for the detection of more characteristics, which benefits later applications.
It is important to know some information about noise and feature intensity before the parameters of the scale space are determined. Here, we introduce a filter to compute a low-density part of the image [39]:
n ^ = I h = I 1 2 1 2 4 2 1 2 1
Then, the noise mean, noise variance, and Signal-to-Noise Ratio (SNR) can be estimated by
μ e s t = Mean ( n ^ ) , σ e s t 2 = Var ( n ^ ) , S N R e s t = 10 log 10 [ I n ^ ] 2 n ^ 2 .
Based on these, a new parameter configuration is formulated as (11), for Equation (5) to generate a continuous nonlinear scale space:
b = μ e s t · S N R 0 S N R e s t .
Here, S N R 0 means a reference SNR value that can be preset using some prior knowledge of the image data.
However, it is still difficult to determine which scale levels should be selected to compute the feature descriptor for the interesting points. In this paper, it is assumed that the interesting points are uniformly distributed in the scale space, so we tried to divide the scale space into several parts in which the changed amount of some metric is equal. Then, a related PDE-constrained optimization problem can be expressed as
( t 1 , t 2 , , t N ) = arg min t 1 , t 2 , , t N p = 0 N Ene ( I ( x , t p + 1 ) Ene ( I ( x , t p ) ) 2
s . t . I t ( x , t ) = div I 1 + S N R e s t μ e s t S N R 0 2 | I | 2 , ( x , t ) Ω × ( 0 , T ) I ( x , 0 ) = I 0 ( x ) , x Ω , I n ( x , t ) = 0 , x Ω .
Here, Ene ( I ) = | | I | | 2 2 denotes the image energy needed to measure the variant of I ( x , t ) . The parameters t 0 = 0 and t N + 1 = T are preset. The initial condition means that the scale space should be generated from the observed image I 0 . div denotes the divergence operator. This model tries to determine N scale levels t 1 , t 2 , , t N that uniformly divide the scale space I ( x , t ) , t [ t 0 , t N + 1 ] .
To find the optimal solution, the continuous scale space I ( x , t ) must be solved from Equation (13), which requires a fair amount of calculation. To reduce the cost for solving the whole scale space, a discrete version is introduced to approximate the scale space, and then the model (12) can be rewritten as
( k 1 , k 2 , , k N ) = arg min k 1 , k 2 , , k N p = 0 N Ene ( I k p + 1 ) Ene ( I k p ) 2
s . t . I n + 1 = 1 τ l = 1 m A l ( I n ) 1 I n
where τ denotes a small enough time step, and (15) means to solve Equation (13) solved using the AOS scheme.
Attributed to the reference step size τ , the original optimized scale levels t 1 , t 2 , , t N are approximated by k 1 τ , k 2 τ , , k N τ . The variables to be solved are changed from continuous type to integer type, and then the cost will be greatly reduced.
However, it is still not easy to properly choose N levels from the level set with about T τ elements, and this is similar to a subset selection problem to some extent. To efficiently solve such a problem, an iterative approximation is introduced to the proposed algorithm, Algorithm 1.
After setting the initial image I 0 = I ( x , 0 ) , maximal scale level T, maximal allowed error ε , maximal iteration number m a x I t e r , initial solution k 1 0 , k 2 0 , , k N 0 and then applying Algorithm 1, optimal scale levels k 1 , , k N will be determined, and the nonlinear scale space can be uniformly divided into several parts in terms of total variation. As shown in Figure 3, the nonlinear scale space of a remote sensing image is divided by five optimal scale levels, and the differences between two adjacent scales are almost uniform.
Algorithm 1 Optimal Scale Iterative Approximation
Require:  I 0 = I ( x , 0 ) , T, τ , ε , m a x I t e r , initial solution k 1 0 , k 2 0 , , k N 0 .
Ensure: Optimal solution k 1 , k 2 , , k N
1:
for  i = 1 to N do
2:
       k i = k i 0 ;
3:
end for
4:
E 0 = | | I 0 | | , k 0 = 0 , k N + 1 = [ T τ ] , Δ E = i n f , j = 0 ;
5:
while  Δ E < ε and j < m a x I t e r  do
6:
      for  i = 1 to N + 1  do
7:
             I k i = 1 τ ( k i k i 1 ) l = 1 m A l ( I k i 1 ) 1 I k i 1 ;
8:
             E i = | | I k i | | ;
9:
      end for
10:
      Δ E = i = 1 N + 1 ( E i E i 1 ) 2 N + 1 , j = j + 1 ;
11:
     for  i = 1 to N do
12:
            k i = k i [ ( k i + 1 k i 1 ) E i ( E 0 + i Δ E ) E i + 1 E i 1 ] ;
13:
     end for
14:
end while

4. Experiments and Results

In this study, we arranged several experiments to illustrate the performance and advantages of the proposed method. In real-world remote sensing, the same scene is often observed from different positions/viewpoint withs different surroundings. Subsequently, it may be more difficult to detect the feature points and search the matches. Here, we focused on remote sensing images collected with optical sensors. For a comprehensive comparison, the number of feature points was used to evaluate the cost. The match number and correct number are used to evaluate the efficiency. We also present the time cost of the comparison methods.
In the first experiment, we prepared one remote sensing images and generated four pairs for the test. One of each pair of images was the original image, and the other was generated by rotating and re-scaling the original image. The detected feature points and matching results are shown in Figure 4 and Figure 5. The detailed data (including time cost) can be found in Table 1 and Table 2.
For the first column of Figure 4, the rotation angle was 25 π 180 , and the scaling factor was 0.75 . For the next three columns, the scaling factor was fixed, and the rotation angle was set to be 55 π 180 , 85 π 180 , and 115 π 180 , respectively. The rows indicate different methods: SIFT, SURF, BRISK, KAZE, and the proposed method.
It can be found that SIFT finds the most feature points but few matches. Surf finds fewer feature points but obtains more matches than SIFT ol average. BRISK finds the least matches, but there is no error. KAZE can find more correct matches than the three previously mentioned methods but less than the proposed method. In general, because of the optimized scale space, the proposed method achieves the best performance with cost and efficiency being well balanced.
In the second experiment, we tried to test the performances of the different methods under a scaling scene. The rotation angle was fixed to be 25 π 180 . The original image was scaled using different factor separately to generate four image pairs. The detected feature points and matching results are shown in Figure 6 and Figure 7. The detailed data (including time cost) can be found in Table 3 and Table 4.
The columns correspond to different scaling factors ( 55 % , 65 % , 75 % , 85 % ), and the rows correspond to different methods (SIFT, SURF, BRISK, KAZE, proposed). It can be found that SIFT finds more feature points but fewer matches. Surf finds more matches than SIFT but shows less accuracy, especially when the scaling factor is smaller. BRISK finds the second most feature points but is not advantageous in terms of the number of correct matches. KAZE and the proposed method still find the most correct matches, but KAZE holds no obvious advantage in the match number compared with the other three methods. Our method finds not as many feature points but the most correct matches.
In the third experiment, the rotation angle was fixed to be 25 π 180 , and the scale factor was fixed to be 75 % . Gaussian noise was added to the original image to prepare one image of the image pair, and then, the polluted image was rotated and scaled to generate the other image of the pair. The detected feature points and matching results are shown in Figure 8 and Figure 9. The detailed data (including time cost) can be found in Table 5 and Table 6.
The columns correspond to different noise intensities (SNR = 20 db, 18 db, 16 db, 14 db), and the rows correspond to different methods (SIFT, SURF, BRISK, KAZE, proposed).
As shown in Figure 9, SIFT and BRISK find the most key points in each pair, but there is no advantage in finding correct matches. SURF finds the least key points, but there is no obvious disadvantage in finding correct matches. KAZE finds a few more key points than SURF but achieves a similar correct number of matches. The proposed method finds key points a little more than KAZE but achieves the largest correct number of matches.
In the fourth experiment, the rotation angle is fixed to be 55 π 180 and the scale factor is fixed to be 65 % . Four optical remote sensing images are prepared to be scaled and rotated. The feature detection and matching identify are performed on four image pairs. The results are shown in Figure 10 and Figure 11, Table 7 and Table 8.
Similarly, the columns are corresponding to different remote sensing image (rs1, rs2, rs3, rs4) and rows are related to different methods (SIFT, SURF, BRISK, KAZE, proposed).
As shown in Figure 11, SIFT finds the most feature points in each pair but no more matches than other methods. BRISK finds the second most feature points and achieve the least correct match number in average. SURF finds more feature points and correct matches than KAZE. The proposed method finds key points a little more than KAZE and achieves the biggest correct number.
From above experiments, the proposed method shows the significantly best comprehensive performance exceed other four.

5. Conclusions

In this paper, a constrained optimization model is introduce to determine some scale slices of a nonlinear scale space for remote sensing image feature detection and matching. With noise estimation introduced to determine the parameter values, a partial differential equation is applied to generate a continuous nonlinear scale space advantageous to represent the important features. The continuous scale space is approximated by a discrete one by Additive Operator Splitting solving the PDE. Then an discrete model is present to fast approach the optimal scale levels for feature detection and image matching. Four experiments are arranged to test the performance of proposed method from views of rotation, scaling, noise and the mixed version. The results show the accuracy and efficiency of the propose method. About 30% improvement in correct matches number with only a small increase in time cost can be found in the proposed method results against the others. In the future work, more efforts will be made to improve the proposed method and develop more efficient and robust method for complex scenes.

Author Contributions

Conceptualization, Y.P. and B.Z.; resources, F.Q.; data curation, Y.P.; writing—original draft preparation, F.Q. and B.Z.; writing—review and editing, Y.P. and F.Q.; visualization, B.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported in part by PipeChina Select the Best Candidates to Undertake Key Research Projects (WZXGL202106).

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

Authors Yunchao Peng and Feng Qi were employed by the company PipeChina Network Corporation Eastern Oil Storage and Transportation Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. All authors have read and agreed to the published version of the manuscript.

References

  1. Li, X.; Zhang, L.; Du, B.; Zhang, L.; Shi, Q. Iterative reweighting heterogeneous transfer learning framework for supervised remote sensing image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 2022–2035. [Google Scholar] [CrossRef]
  2. Li, Z.; Yue, J.; Fang, L. Adaptive regional multiple features for large-scale high-resolution remote sensing image registration. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
  3. Anuta, P.E. Spatial registration of multispectral and multitemporal digital imagery using fast Fourier transform techniques. IEEE Trans. Geosci. Electron. 1970, 8, 353–368. [Google Scholar] [CrossRef]
  4. Brown, L.G. A survey of image registration techniques. ACM Comput. Surv. (CSUR) 1992, 24, 325–376. [Google Scholar] [CrossRef]
  5. Zhang, X.; Zhou, Y.; Qiao, P.; Lv, X.; Li, J.; Du, T.; Cai, Y. Image Registration Algorithm for Remote Sensing Images Based on Pixel Location Information. Remote Sens. 2023, 15, 436. [Google Scholar] [CrossRef]
  6. Ma, W.; Wen, Z.; Wu, Y.; Jiao, L.; Gong, M.; Zheng, Y.; Liu, L. Remote sensing image registration with modified SIFT and enhanced feature matching. IEEE Geosci. Remote Sens. Lett. 2016, 14, 3–7. [Google Scholar] [CrossRef]
  7. Flusser, J. An adaptive method for image registration. Pattern Recognit. 1992, 25, 45–54. [Google Scholar] [CrossRef]
  8. Studholme, C.; Hill, D.L.; Hawkes, D.J. An overlap invariant entropy measure of 3D medical image alignment. Pattern Recognit. 1999, 32, 71–86. [Google Scholar] [CrossRef]
  9. Lowe, D.G. Object recognition from local scale-invariant features. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; Volume 2, pp. 1150–1157. [Google Scholar]
  10. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  11. Ke, Y.; Sukthankar, R. PCA-SIFT: A more distinctive representation for local image descriptors. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2004), Washington, DC, USA, 27 June–2 July 2004; Volume 2, p. II. [Google Scholar]
  12. Bay, H.; Tuytelaars, T.; Van Gool, L. Surf: Speeded up robust features. In Proceedings of the European Conference on Computer Vision, Graz, Austria, 7–13 May 2006; Springer: Berlin/Heidelberg, Germany, 2006; pp. 404–417. [Google Scholar]
  13. Dellinger, F.; Delon, J.; Gousseau, Y.; Michel, J.; Tupin, F. SAR-SIFT: A SIFT-like algorithm for SAR images. IEEE Trans. Geosci. Remote Sens. 2014, 53, 453–466. [Google Scholar] [CrossRef]
  14. Alcantarilla, P.F.; Bartoli, A.; Davison, A.J. KAZE features. In Proceedings of the Computer Vision—ECCV 2012: 12th European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; Proceedings, Part VI 12. Springer: Berlin/Heidelberg, Germany, 2012; pp. 214–227. [Google Scholar]
  15. Ma, J.; Jiang, X.; Fan, A.; Jiang, J.; Yan, J. Image matching from handcrafted to deep features: A survey. Int. J. Comput. Vis. 2021, 129, 23–79. [Google Scholar] [CrossRef]
  16. Gong, M.; Zhao, S.; Jiao, L.; Tian, D.; Wang, S. A novel coarse-to-fine scheme for automatic image registration based on SIFT and mutual information. IEEE Trans. Geosci. Remote Sens. 2013, 52, 4328–4338. [Google Scholar] [CrossRef]
  17. Wu, Y.; Ma, W.; Gong, M.; Su, L.; Jiao, L. A novel point-matching algorithm based on fast sample consensus for image registration. IEEE Geosci. Remote Sens. Lett. 2014, 12, 43–47. [Google Scholar] [CrossRef]
  18. Wu, Y.; Miao, Q.; Ma, W.; Gong, M.; Wang, S. PSOSAC: Particle swarm optimization sample consensus algorithm for remote sensing image registration. IEEE Geosci. Remote Sens. Lett. 2017, 15, 242–246. [Google Scholar] [CrossRef]
  19. Leutenegger, S.; Chli, M.; Siegwart, R. BRISK: Binary robust invariant scalable keypoints. In Proceedings of the 2011 IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011; pp. 2548–2555. [Google Scholar]
  20. Dou, J.; Qin, Q.; Tu, Z. Robust image matching based on the information of SIFT. Optik 2018, 171, 850–861. [Google Scholar] [CrossRef]
  21. Li, Y.; Li, Q.; Liu, Y.; Xie, W. A spatial-spectral SIFT for hyperspectral image matching and classification. Pattern Recognit. Lett. 2018, 127, 18–26. [Google Scholar] [CrossRef]
  22. He, Y.; Deng, G.; Wang, Y.; Wei, L.; Yang, J.; Li, X.; Zhang, Y. Optimization of SIFT algorithm for fast-image feature extraction in line-scanning ophthalmoscope. Optik 2018, 152, 21–28. [Google Scholar] [CrossRef]
  23. Chen, Z.; Sun, S.K. A Zernike moment phase-based descriptor for local image representation and matching. IEEE Trans. Image Process. 2009, 19, 205–219. [Google Scholar] [CrossRef] [PubMed]
  24. Zhou, B.; Duan, X.; Ye, D.; Wei, W.; Woźniak, M.; Damaševičius, R. Heterogeneous Image Matching via a Novel Feature Describing Model. Appl. Sci. 2019, 9, 4792. [Google Scholar] [CrossRef]
  25. Klare, B.; Jain, A.K. Heterogeneous Face Recognition: Matching NIR to Visible Light Images. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010. [Google Scholar]
  26. Wang, N.; Li, J.; Tao, D.; Li, X.; Gao, X. Heterogeneous image transformation. Pattern Recognit. Lett. 2013, 34, 77–84. [Google Scholar] [CrossRef]
  27. Peng, C.; Gao, X.; Wang, N.; Li, J. Graphical representation for heterogeneous face recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 301–312. [Google Scholar] [CrossRef] [PubMed]
  28. Wang, T.; Kemao, Q.; Seah, H.S.; Lin, F. A flexible heterogeneous real-time digital image correlation system. Opt. Lasers Eng. 2018, 110, 7–17. [Google Scholar] [CrossRef]
  29. Lin, G.; Fan, C.; Zhu, H.; Miu, Y.; Kang, X. Visual feature coding based on heterogeneous structure fusion for image classification. Inf. Fusion 2017, 36, 275–283. [Google Scholar] [CrossRef]
  30. Olson, C.F. Adaptive-scale filtering and feature detection using range data. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 983–991. [Google Scholar] [CrossRef]
  31. Lezoray, O.; Charrier, C. Color image segmentation using morphological clustering and fusion with automatic scale selection. Pattern Recognit. Lett. 2009, 30, 397–406. [Google Scholar] [CrossRef]
  32. Sun, J.; Xu, Z. Scale selection for anisotropic diffusion filter by Markov random field model. Pattern Recognit. 2010, 43, 2630–2645. [Google Scholar] [CrossRef]
  33. Biegler, L.T.; Ghattas, O.; Heinkenschloss, M.; van Bloemen Waanders, B. Large-scale PDE-constrained optimization: An introduction. In Large-Scale PDE-Constrained Optimization; Springer: Berlin/Heidelberg, Germany, 2003; pp. 3–13. [Google Scholar]
  34. Mang, A.; Gholami, A.; Davatzikos, C.; Biros, G. PDE-constrained optimization in medical image analysis. Optim. Eng. 2018, 19, 765–812. [Google Scholar] [CrossRef]
  35. Witkin, A.P. Scale-space filtering. In Readings in Computer Vision; Elsevier: Amsterdam, The Netherlands, 1987; pp. 329–332. [Google Scholar]
  36. Zhou, B.; Duan, X.M.; Wei, W.; Ye, D.J.; Woźniak, M.; Damaševičius, R. An adaptive local descriptor embedding zernike moments for image matching. IEEE Access 2019, 7, 183971–183984. [Google Scholar] [CrossRef]
  37. Fergus, R.; Perona, P.; Zisserman, A. Object class recognition by unsupervised scale-invariant learning. In Proceedings of the 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Madison, WI, USA, 18–20 June 2003; pp. 264–271. [Google Scholar]
  38. Jolliffe, I. Principal Component Analysis; Springer: Berlin/Heidelberg, Germany, 2011. [Google Scholar]
  39. Immerkaer, J. Fast noise variance estimation. Comput. Vis. Image Underst. 1996, 64, 300–302. [Google Scholar] [CrossRef]
Figure 1. Gaussian scale space.
Figure 1. Gaussian scale space.
Mathematics 12 01882 g001
Figure 2. Several nonlinear scale spaces.
Figure 2. Several nonlinear scale spaces.
Mathematics 12 01882 g002
Figure 3. Optimized scale levels.
Figure 3. Optimized scale levels.
Mathematics 12 01882 g003
Figure 4. Feature detection and matching under different rotations.
Figure 4. Feature detection and matching under different rotations.
Mathematics 12 01882 g004
Figure 5. Feature numbers and match numbers under different rotations.
Figure 5. Feature numbers and match numbers under different rotations.
Mathematics 12 01882 g005
Figure 6. Feature detection and matching under different scales.
Figure 6. Feature detection and matching under different scales.
Mathematics 12 01882 g006
Figure 7. Feature numbers and match numbers under different scales.
Figure 7. Feature numbers and match numbers under different scales.
Mathematics 12 01882 g007
Figure 8. Feature detection and matching under different noise levels.
Figure 8. Feature detection and matching under different noise levels.
Mathematics 12 01882 g008
Figure 9. Feature numbers and match numbers under different noise levels.
Figure 9. Feature numbers and match numbers under different noise levels.
Mathematics 12 01882 g009
Figure 10. Feature detection and matching for different images.
Figure 10. Feature detection and matching for different images.
Mathematics 12 01882 g010
Figure 11. Features number and matches number of different remote sensing images.
Figure 11. Features number and matches number of different remote sensing images.
Mathematics 12 01882 g011
Table 1. Feature numbers under different rotations.
Table 1. Feature numbers under different rotations.
MethodKeypoint Number 1 (1000)Keypoint Number 2 (1000)
0000 25 π 180 55 π 180 85 π 180 115 π 180
SIFT1.1461.1461.1461.1460.6410.6450.6240.638
SURF0.1700.1700.1700.1700.0890.0990.0910.084
BRISK0.8130.8130.8130.8130.2640.2710.2520.273
KAZE0.3030.3030.3030.3030.1680.1660.1640.172
Proposed0.4020.4020.4020.4020.2420.2400.2380.248
Table 2. Match numbers under different rotations.
Table 2. Match numbers under different rotations.
MethodMatches NumberCorrect NumberTotal Time Cost (s)
25 π 180 55 π 180 85 π 180 115 π 180 25 π 180 55 π 180 85 π 180 115 π 180 25 π 180 55 π 180 85 π 180 115 π 180
SIFT19212220191918161.0341.1411.1141.115
SURF22163926201437190.0140.0160.0160.015
BRISK11151811111518110.4180.4410.4390.464
KAZE30252333292423310.0630.0620.0640.067
Proposed50444456484342530.0610.0660.0640.078
Table 3. Feature numbers under different scales.
Table 3. Feature numbers under different scales.
MethodKeypoint Number 1 (1000)Keypoint Number 2 (1000)
100 % 100 % 100 % 100 % 55 % 65 % 75 % 85 %
SIFT1.1461.1461.1461.1460.3030.4210.6060.798
SURF0.1700.1700.1700.1700.0620.0780.0850.105
BRISK0.8130.8130.8130.8130.1110.1990.2630.410
KAZE0.3030.3030.3030.3030.1100.1290.1730.194
Proposed0.4020.4020.4020.4020.1430.1880.2460.283
Table 4. Match numbers under different scales.
Table 4. Match numbers under different scales.
MethodMatches NumberCorrect NumberTotal Time Cost (s)
55 % 65 % 75 % 85 % 55 % 65 % 75 % 85 % 55 % 65 % 75 % 85 %
SIFT8918328817290.7551.0181.1201.224
SURF1416193581118300.0130.0170.0160.016
BRISK127162912716290.4660.4490.4550.406
KAZE521303551829320.0500.0640.0640.069
Proposed12385156113549520.0560.0670.0620.076
Table 5. Featur numbers under different noise levels.
Table 5. Featur numbers under different noise levels.
MethodKeypoint Number 1 (1000)Keypoint Number 2 (1000)
----20 db18 db16 db14 db
SIFT1.0010.9690.9450.8730.8010.8100.7800.753
SURF0.1790.1920.2130.2190.1250.1180.1240.134
BRISK0.9961.2191.6362.1550.4400.4420.4970.570
KAZE0.3030.3070.3030.3010.1990.1970.1950.188
Proposed0.4120.4200.4140.3980.2920.2850.2950.293
Table 6. Match numbers under different noise levels.
Table 6. Match numbers under different noise levels.
MethodMatches NumberCorrect NumberTotal Time Cost (s)
20 db18 db16 db14 db20 db18 db16 db14 db20 db18 db16 db14 db
SIFT25222322252123220.9570.9670.9480.922
SURF35333937283026320.0170.0150.0150.016
BRISK35242223352422230.3920.3750.3830.395
KAZE28292935262726340.0570.0550.0560.055
Proposed41465249384447470.0560.0580.0570.058
Table 7. Features number of different remote sensing images.
Table 7. Features number of different remote sensing images.
MethodKeypoint Number 1 (1000)Keypoint Number 2 (1000)
rs1rs2rs3rs4rs1rs2rs3rs4
SIFT1.5231.6041.8961.4960.6310.6980.6860.663
SURF0.2410.1570.1880.1760.0950.0490.0380.012
BRISK0.7960.6970.9720.6210.2520.2080.2880.133
KAZE0.1630.0840.120.0530.0510.0180.0150.008
Proposed0.4130.2550.3490.1940.1280.070.0770.026
Table 8. Matches number of different remote sensing images.
Table 8. Matches number of different remote sensing images.
MethodMatches NumberCorrect NumberTotal Time Cost (s)
rs1rs2rs3rs4rs1rs2rs3rs4rs1rs2rs3rs4
SIFT20682175821.0350.9110.9630.915
SURF241310621121060.0140.0160.0120.012
BRISK10325103250.3670.3620.3680.359
KAZE13542135420.0440.0430.0440.043
Proposed272027826202780.0510.0460.0490.045
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Peng, Y.; Zhou, B.; Qi, F. PDE-Constrained Scale Optimization Selection for Feature Detection in Remote Sensing Image Matching. Mathematics 2024, 12, 1882. https://0-doi-org.brum.beds.ac.uk/10.3390/math12121882

AMA Style

Peng Y, Zhou B, Qi F. PDE-Constrained Scale Optimization Selection for Feature Detection in Remote Sensing Image Matching. Mathematics. 2024; 12(12):1882. https://0-doi-org.brum.beds.ac.uk/10.3390/math12121882

Chicago/Turabian Style

Peng, Yunchao, Bin Zhou, and Feng Qi. 2024. "PDE-Constrained Scale Optimization Selection for Feature Detection in Remote Sensing Image Matching" Mathematics 12, no. 12: 1882. https://0-doi-org.brum.beds.ac.uk/10.3390/math12121882

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop