Next Article in Journal
Time, Spatial, and Descriptive Features of Pedestrian Tracks on Set of Visualizations
Next Article in Special Issue
Building Virtual 3D City Model for Smart Cities Applications: A Case Study on Campus Area of the University of Novi Sad
Previous Article in Journal
Modelling Housing Rents Using Spatial Autoregressive Geographically Weighted Regression: A Case Study in Cracow, Poland
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hierarchical Point Matching Method Based on Triangulation Constraint and Propagation

1
School of Geomatics, Liaoning Technical University, Fuxin 123000, China
2
Collaborative Innovation Institute of Geospatial Information Service, Liaoning Technical University, Fuxin 123000, China
3
Combat Data Laboratory, Joint Logistic Support Force of PLA, Wuhan 430010, China
4
Research Institute of Smart Cities, School of Architecture and Urban Planning, Shenzhen University, Shenzhen 518060, China
*
Author to whom correspondence should be addressed.
ISPRS Int. J. Geo-Inf. 2020, 9(6), 347; https://0-doi-org.brum.beds.ac.uk/10.3390/ijgi9060347
Submission received: 24 April 2020 / Revised: 23 May 2020 / Accepted: 24 May 2020 / Published: 26 May 2020
(This article belongs to the Special Issue Virtual 3D City Models)

Abstract

:
Reliable image matching is the basis of image-based three-dimensional (3D) reconstruction. This study presents a quasi-dense matching method based on triangulation constraint and propagation as applied to different types of close-range image matching, such as illumination change, large viewpoint, and scale change. The method begins from a set of sparse matched points that are used to construct an initial Delaunay triangulation. Edge-to-edge matching propagation is then conducted for the point matching. Two types of matching primitives from the edges of triangles with areas larger than a given threshold in the reference image, that is, the midpoints of edges and the intersections between the edges and extracted line segments, are used for the matching. A hierarchical matching strategy is adopted for the above-mentioned primitive matching. The points that cannot be matched in the first stage, specifically those that failed in a gradient orientation descriptor similarity constraint, are further matched in the second stage. The second stage combines the descriptor and the Mahalanobis distance constraints, and the optimal matching subpixel is determined according to an overall similarity score defined for the multiple constraints with different weights. Subsequently, the triangulation is updated using the newly matched points, and the aforementioned matching is repeated iteratively until no new matching points are generated. Twelve sets of close-range images are considered for the experiment. Results reveal that the proposed method has high robustness for different images and can obtain reliable matching results.

1. Introduction

Image matching is a process of identifying corresponding features on different images and is an essential step in image processing, such as three-dimensional (3D) reconstruction, image fusion, and target tracking [1,2,3]. Numerous research papers were published in the field of image matching [4,5,6,7,8,9,10]. Existing image matching approaches can be classified into three categories according to the number of matching results: (i) feature-based sparse matching, which is focused on the construction of a descriptor and is widely used in initial matching to calculate projection parameters; (ii) quasi-dense stereo matching algorithm based on region growing, which is focused on matching strategies, such as matching constraints and matching propagation; (iii) dense pixel matching, which is mostly used in binocular stereo vision.
This study mainly aims to obtain matching results as dense as possible for different types of images on the basis of initial sparse matching and matching propagation, and this study belongs to the quasi-dense matching domain. The use of matching constraints is key in this method, and the widely used epipolar constraint can be used to reduce the matching search space from two dimensions (2D) to one dimension (1D). However, a large search space may exist, and this space is generally combined with other various geometric constraints, such as parallax constraints [11], Voronoi diagram [12,13,14], and Delaunay triangulation [15,16]. The search range of the epipolar line can be limited to the internal corresponding triangle when combined with the triangulation constraint [17]. Li et al. [18] proposed a Voronoi diagram-based propagation stereo matching algorithm that firstly subdivides the whole image into N cells by using the Voronoi diagram of the seed feature points; then, the corresponding relations are propagated from the seed in each cell until all pixels within this cell are processed. Similarly, Delaunay triangulation is widely used in matching [19,20,21]. Delaunay triangulation is a form of a vector-based digital surface model constructed by a set of unorganized points, whose main contributions in matching are constraint matching and matching propagation through region segmentation of triangle meshes [22]. Zhao et al. [23] used Delaunay triangulation to remove outliers in iterative matching by comparing Delaunay graph structures and then used the dual graph of Delaunay to check the above-mentioned results, recover inliers, and remove outliers. Ahuja et al. [24] proposed that a point within a triangle in a single image be matched to a point within the corresponding triangle in another image, and this approach is called local position constraint. Guo et al. [25] utilized the relationship between the matching point and the vertices of a triangle for constraint matching candidates. Zhu et al. [26,27] added a parallax gradient to the constraint matching candidates on the basis of the corresponding triangles established by the existing corresponding points on the epipolar stereo pair. The method requires a vertex of a triangle with the highest matching reliability to be selected as the reference point, and the parallax gradient is assumed to be related to the ratio of the distance between the reference and potential matching points to their difference in the parallax. The aforementioned methods can produce reliable matching results, but they only consider the point matching. Wu et al. [28] proposed a matching method that integrates point and edge matching based on self-adaptive edge-constrained triangulations to incorporate edge features in the matching and improve the image matching performance of poor textural images. The method takes advantage of edge information and updates triangulations dynamically. It is used for space-borne, airborne, and terrestrial image matching and can produce reliable matching results for poor textural images. However, it only uses image feature information. Jia et al. [29] proved the mathematical property of equal proportion of triangulation affine transformation as a means of obtaining much denser matching results and proposed a dense matching method for wide baseline images based on the theory. In their method, the equal proportion points on the corresponding edges of triangles in two views that satisfy the similarity constraints are regarded as corresponding points. The method can achieve dense matching of images with continuous scenes but lacks robustness for different types of images, especially large viewpoint change images with discontinuous surfaces.
This study proposes a new quasi-dense matching method based on triangulation constraint and propagation. The proposed method can achieve dense matching and consider the edge features of the image. Compared with existing approaches, our method has several advantages. Firstly, we introduce the intersections between edge line segments and triangulation as matching primitives. In this manner, more matches can be generated, and the image edge feature is considered at the same time. Secondly, we propose a new descriptor for point dense matching that adopts overlapped subregions to increase the correlation between description vectors and contributes to the generation of dense matching results. Thirdly, we adopt the hierarchical matching strategy by integrating additional matching into the second stage to reach subpixel accuracy. An overall similarity score is defined to integrate the descriptor similarity measure, the distance of the point from a triangle’s edge, Mahalanobis distance, and the distance of the point from the epipolar line as the four similarity measures with different weights to determine the best matching result. This stage helps generate more corresponding points and improves matching results.
The remainder of the paper is organized as follows: Section 2 introduces our hierarchical point matching method; Section 3 presents the experiments conducted on twelve image pairs and demonstrates the performance of our method by integrating additional matching into the hierarchical matching strategy; Section 4 offers the concluding remarks and suggests future work.

2. Proposed Method

The flowchart of the proposed dense matching algorithm is illustrated in Figure 1. The input data of our method consist of two images. An image is selected as the reference image while the other image acts as the search image. The scale-invariant feature transform (SIFT) algorithm [30] is used to obtain the initial corresponding points. Incorrect matches are removed by the random sample consensus algorithm (RANSAC) [31], and correct point-to-point correspondences are obtained. Next, the corresponding Delaunay triangulation is constructed in both images. The line segments extracted from the reference image are used to generate intersection points with triangles in the reference image, which are used as matching primitives for the succeeding matching. For each triangle with an area greater than the given threshold T s , the midpoints of the three sides and the intersection points mentioned above are matched in turn. Whether new corresponding points are generated after all the triangles are processed should be assessed. If new corresponding points are generated, then Delaunay triangulation is updated with these new points, and the aforementioned steps are repeated iteratively until no new corresponding points are generated. Otherwise, the matching is stopped. Thereafter, the final matching results are outputted.
Two types of matching primitives are applied in our study. One includes the midpoints of three edges of the triangle, and the other is the intersection point of the extracted line segments and edges of the triangles. In the following sections, both primitives are referred to as the midpoint and the intersection point, respectively. The proposed method is an iterative matching process that includes two stages in each iteration, with each stage corresponding to a matching strategy. In the first matching stage, we match points only by using the proposed SIFT-like local gradient descriptors and Euclidean distances. In the case of unmatched points, we further match them in the second stage by utilizing multiple constraints, which is described in Section 2.2.

2.1. Point Matching Based on Descriptors with Overlapping Subregions

For any triangle with an area greater than T s in the reference image, the midpoints of the three sides and the intersections between the triangle and line segments are calculated. We suppose that the triangles Δ a b c and Δ a b c are a pair of corresponding triangles (Figure 2), and the corresponding vertices ( A , A ) , ( B , B ) , and ( C , C ) are three pairs of corresponding points. For any midpoint a , its candidate point a is the midpoint of the corresponding side of the corresponding triangle in the search image. For any intersection point d , its candidate point d is the intersection point between the epipolar line of point d in the search image and the corresponding side of the corresponding triangle in the search image. For each correspondence hypothesis ( a , a ) or ( d , d ) , we propose a new descriptor for similarity measure calculation and determine whether the hypothesis offers the correct match.

2.1.1. Descriptor Construction

The feature descriptor in the feature matching process is an important algorithm because it extracts pertinent information and encodes it into feature vectors, which are then used to differentiate a feature from another or find the same features. With the popularization of the SIFT descriptor [32,33,34], many descriptors based on histograms of gradient orientations, such as ASIFT [35] and Daisy [36], were proposed. These descriptors have high robustness for different types of images during matching, but they also have high computational complexity. As such, we propose a new SIFT-like descriptor for the local neighborhood of a point that not only offers the advantages of the SIFT descriptor but also reduces the computation cost. The descriptor construction is processed as described below.
Firstly, Gaussian filtering is used to reduce the image noise. Then, the neighborhood window with a center at the target point is built for the descriptor construction. The size of the window is ( 2 w + 1 ) × ( 2 w + 1 ) ,   ( w = 3 , 4 , n ) . In this study, we set w = 4 . We assume that the coordinates of the target point are ( x 0 , y 0 ) and the coordinates of the four corners of the neighborhood window are ( x 0 w , y 0 w ) , ( x 0 + w , y 0 w ) , ( x 0 w , y 0 + w ) , and ( x 0 + w , y 0 + w ) . The neighborhood window is then divided into four subregions, in which the square areas with the target point and the four corners of the neighborhood window are used to set the diagonals. The size of each subregion is ( w + 1 ) × ( w + 1 ) . The schematic is shown in Figure 3. The four subregions with overlapping regions, which all include the row and column where the target point is located, are obtained. This strategy increases the correlation between regions and avoids the influence of region inconsistency caused by large viewpoint changes.
Secondly, each subregion is again divided into four square regions with the same sizes, and 16 small patches are obtained around the target point. A gradient magnitude histogram with four orientations is then calculated in each small patch. As the impact of the nearby points on the descriptor increases, the distance d i from each point in the patch to the target point is calculated, and each gradient magnitude is weighted by the corresponding distance d i with the function g r a d i e n t t r u e = g r a d i e n t × e d i in the statistical process. Then, a statistical description vector about the gradient magnitudes in the four directions for each small patch is obtained, the description vectors of four small patches in the subregion are stacked, and a 4 × 4 description matrix for this subregion is obtained. Thereafter, inspired by the mean-standard deviation line descriptor (MSLD) descriptor [37], the mean and standard deviation vectors of matrix are calculated and normalized, respectively, to unit norm to make the descriptor invariant to linear changes of illumination. Finally, the descriptor vector with a length of 32 is obtained by concatenating the normalized mean vector of four subregions with the normalized standard deviation vector of four subregions into a single vector.
Our design of the descriptor adopts the subregions overlapping one another in the feature description, which increases the correlation between description vectors and contributes to the generation of dense matching results.

2.1.2. Dissimilarity Measure

For two points in any set of matching hypotheses in the first matching stage, the two descriptors d e s c r i p t o r 1 and d e s c r i p t o r 2 are obtained for the reference and search image points, and Euclidean distance is used to calculate the dissimilarities between the descriptors. If the Euclidean distance d i s t a n c e 1 2 is smaller than the given threshold T 1 , then this set of matching hypotheses is considered correct, that is, the two points in the matching hypotheses are a pair of corresponding points.

2.2. Point Matching with Multiple Constraints

Points that could not be matched in the first stage are further matched using a strategy with multiple constraints as described below.
Firstly, for reference point a with an initial candidate point denoted as a , we define the searching region as a square with a center at point a and a length of ( 2 m + 1 ) , where m is a user-defined parameter. In this study, we set m = 1 . All pixels in the searching window are regarded as matching candidates. For each candidate, we build its descriptor and calculate the Euclidean distance d i s t a n c e 1 2 between two descriptors of reference and candidate point. If at least one candidate point satisfies the condition in which the d i s t a n c e 1 2 value is smaller than the threshold T 2 , then the optimal matching with subpixel accuracy should be determined further by utilizing multiple constraints.
Secondly, we divide each candidate point satisfied by the Euclidean distance constraint mentioned above into grids and obtain a set of candidate subpixels. For each candidate subpixel, three dissimilarity measures are defined as described below.
The first dissimilarity measure d i s t a n c e p l is the distance of the candidate subpixel from the epipolar line corresponding to the reference point (Figure 4). In this study, the epipolar line is calculated using the fundamental matrix obtained by corresponding points [38]. Its formula is l e = F x   ;   l e = F x , where x and x are the corresponding points on both images, F is the fundamental matrix, and l e and l e are the corresponding epipolar lines on both images. In theory, the corresponding points are located on the corresponding epipolar lines, but small deviations may exist due to the inaccuracy of the fundamental matrix.
The second dissimilarity measure d i s t a n c e p s is the distance of the candidate subpixel from the corresponding side in the corresponding triangle, as shown in Figure 4.
The third dissimilarity measure is related to the Mahalanobis distance, which was introduced by P.C. Mahalanobis in 1936. This distance is a measure of the distance between a point and a distribution. The formula is as follows:
M a h a l a n o b i s   d i s t a n c e = ( x μ ) T 1 ( x μ )
where x is a sample point, μ is the mean value of a set of sample points, and is the covariance matrix. In this study, the Mahalanobis distance can also be defined as a dissimilarity measure between two random vectors x and y of the same distribution with the covariance matrix ,
M a h a l a n o b i s   d i s t a n c e = ( x y ) T 1 ( x y )
where x and y are two matched points in a single image. We suppose the corresponding points on both images are denoted by the two sets s and s . For any new point-to-point correspondence ( a , a ) on both images, the Mahalanobis distance between a and the corresponding points in the set s is similar to that between a and the corresponding points in the set s . The dissimilarity measure d i s t a n c e m is defined on the basis of this principle. To reduce the computational complexity, we only calculate the Mahalanobis distance between the point and the three vertices of the triangle where the point is located. We assume that d m 1 and d m 2 are the Mahalanobis distance vector of 3 × 1 corresponding to the reference and candidate points, respectively. We define the third dissimilarity measure as d i s t a n c e m = m e a n ( | d m 1 d m 2 | ) , where m e a n ( ) denotes the ratio of the sum of all elements in the vector and the total number of elements in the vector.
On the basis of the third dissimilarity measure mentioned above, we define two thresholds T k , where k { 3 , 4 } , and eliminate the candidate subpixels with any element in | d m 1 d m 2 | greater than T 3 or d i s t a n c e m > T 4 . Thus, only the subpixels fulfilling the two constraints above are retained. Then, an overall matching similarity score Θ s related to four dissimilarity measures is determined to achieve the optimal matching subpixel for the reference point.
Θ s = w 1 e d i s t a n c e 1 2 + w 2 e d i s t a n c e m + w 3 e d i s t a n c e p l + w 4 e d i s t a n c e p s
where w k represents the different weights corresponding to four dissimilarity measures, in which k { 1 , 2 , 3 , 4 } . In this study, we set the weights to 0.45, 0.25, 0.15, and 0.15, respectively. As for all the remaining potential matching subpixels, we calculate Θ s and find the highest overall similarity score if the highest value is larger than the given threshold T 5 , which corresponds to the subpixel selected as the final matching point.

2.3. Matching Propagation

In the existing graph-based constraint matching, the images are divided into different regions by Delaunay triangulation or Voronoi, and many corresponding regions are usually obtained from two images. Matching propagation is performed within these regions (i.e., graph constraint matching propagation). On the basis of the principle in which the corresponding points are located within the corresponding regions, the matching search is conducted in the corresponding regions in different views. Most methods use the feature points extracted in the region as the matching primitives [16,26,28]. This propagation method belongs to the patch-to-patch propagation strategy or the triangle-to-triangle propagation strategy. Different from the region propagation strategies, a new edge-to-edge matching propagation strategy is adopted in this study. This new strategy uses the edge of the triangle as the carrier of matching propagation and the two types of points (midpoints and intersections) on the edge as the matching primitive. As the main principle of this propagation method, triangulation is updated by inserting newly matched points, new edges are generated as a means of obtaining new matching points, and matching is performed on the corresponding edges of the corresponding triangles in different views. Compared with the region-to-region propagation strategy, our proposed strategy offers the advantage of having an edge that is more stable than the region in theory. The principle of matching propagation is described in Algorithm 1.
Algorithm 1 Matching propagation strategy to obtain quasi-dense matching
Input: Reference image I 1 and search image I 2 ; Initial reliable corresponding points Pt(Pt1,Pt2) ; Line segments Le extracted from the reference image; T s ;   T k ,   k = 1 , 2 , 3 , 4 , 5 ;   w k ,   k = 1 , 2 , 3 , 4 ;

Output: Corresponding points P t
n u m 1 =   0   ,   n u m 2   =   1 W h i l e   ( n u m 2   >   n u m 1 )    n u m 1   = s i z e ( P t , 1 )   ;   T r i 1 = D e l a u n a y ( P t 1 )    f o r   i   =   1   :   s i z e ( T r i 1 , 1 )     Calculate   the   area   of   triangles ,   S

     i f   S > T S       Each   correspondence   hypothesis   ( a , a )   ,   #   go   to   the   first   stage   of   mathching       Construct   the   descriptors   of   two   points ( d e s _ a , d e s _ a ) i f   d i s ( d e s _ a , d e s _ a )   T h 1   ,    P t   = [ P t ; [ a , a ] ] }   first   stage       e l s e   #   go   to   the   second   stage   of   matching

       Building   a   searching   window        f o r   j   =   1 : s i z e ( w i n d o w , 1 )   ;   f o r   k   =   1   : s i z e ( w i n d o w , 2 )         i f   d i s ( d e s _ a j , k , d e s _ a j , k )     T h 2   ,   g r i d   p i x e l          Calculate   score   for   each   subpixel   in   pixel ( j , k ) ,   e n d        e n d        i f   m a x s c o r e   T 5 ,   P t = [ P t ; [ a , s u b p i x e l m a x s c o r e ] ] ,   e n d       e n d   }   second   stage

     e n d    e n d    n u m 2 = s i z e ( P t , 1 ) e n d
Return P t
During matching propagation, we only deal with newly generating edges of triangles in the updating process of triangulation to avoid repeated matching in each iteration. The blue triangles shown in Figure 5a represent the initial triangulation, while the black and green points are the newly matched points (midpoints and intersections, respectively) generated by the first iteration. We insert the newly matched points into the initial triangulation and obtain the updated triangulation. The red edges of triangles in Figure 5b are the newly generated or split edges after the first matching, which are used in the subsequent iteration matching.
Figure 6 shows the results of the different stages of matching for the tested stereo pairs, with T s = 30 . Figure 6b presents the matching result of the second iterations, while Figure 6c shows the final matching result, that is, the result of the ninth iteration. Two types of matching primitives (the intersection points and midpoints) are matched by propagation, as shown by the green and red marks, respectively. The intersection points in Figure 7 are all located on the feature edge of the building, while the midpoints have an approximately uniform distribution. The numbers of the two types of points increase in each iteration (Figure 8). As the number of iterations increases, the number of corresponding points and triangles also increases, especially in the first four iterations in which the number of corresponding points increases greatly. The results of the intersection point and the midpoint are relatively stable after five and eight iterations, respectively. These phenomena are consistent with our original assumption that the number of midpoints increases accordingly with the increase in triangles in the iterations, whereas the number of extracted line segments in the image is fixed.

3. Further Experiments and Analysis

The performance of our method is evaluated by selecting twelve representative image pairs for the dense matching (Figure 9). These image pairs include various image transformations, such as viewpoint changes, scale changes, illumination, and rotation. A computer with a 3.6-GHz central processing unit (CPU) and 16 GB of random-access memory (RAM) is used for the experiment, and the code is implemented in Matlab 2016a. Extensive experiments are conducted to select the initial proper values for certain parameters of the proposed method. Then, hierarchical matching is performed in different scene images with determined parameters.

3.1. Parameter Selection

A method for accuracy assessment based on the homography matrix is adopted to assess the matching accuracy. Firstly, the homography matrix is calculated according to the corresponding points obtained by matching. Next, the corresponding points in the reference image are mapped onto the search image by using the homography matrix. The distance histogram is obtained according to the distance between the mapping and corresponding points. Finally, the inner and outer points are determined according to the distance histogram, and the matching accuracy is calculated.
The main parameters to be tuned in our method are T k , where k { 1 , 2 , 3 , 4 , 5 } , and T s . T 1 and T 2 are the descriptors for the dissimilarity measure threshold. T 1 is used to determine whether the correspondence hypothesis is correct in the first stage, and T 2 controls the matching hypotheses to be processed in the second matching stage. T 3 and T 4 are the parameters for the Mahalanobis distance constraint, and T 5 is the overall similarity score threshold used to determine the final matching points in the second stage. We firstly determine the approximate ranges of the parameters and steps according to the histogram distribution of the operation results of all image pairs. Then, initial parameter values are selected (Table 1) after the experimental analysis of representative image pairs. Thereafter, all the image pairs are used to for parametric analysis. Owing to the relatively complex interaction between the parameters in the iterative process and the parameter analysis of the unsupervised results, we adopt the following strategy to determine the parameters. We firstly set the initial parameter values and then update each of the parameters in turn according to the matching results. After parametric analysis of the first eight image pairs (Figure 9a–h) according to the above-mentioned strategy, we find that all test image pairs satisfy a set of parameter values optimally or have only small deviations from the initial parameter values; thus, this set of parameters is chosen as the final parameters for matching (Table 2).
The image pair (a) in Figure 9, which is one of the image pairs corresponding to the optimal parameter, is taken as an example. We update parameters T k , where k { 1 , 2 , 3 , 4 , 5 } , and T s successively. The resulting curves with different parameter values are shown in Figure 10 and Figure 11. The graphs on the left show the number and accuracy of the matching, while the graphs on the right show the number of iterations and running time. The larger the two values on the left figures are and the smaller the two values in the right figures are, the better the matching result will be. As such, the optimal settings can be selected for the parameters or thresholds by combining the four factors mentioned above.
We firstly determine T 1 . The matching results with different values are shown in Figure 10a,b. The number and accuracy can reach their maximum values at the same time at   T 1 = 0.8 , and the numbers of iterations are all 10. Clearly, the running time decreases with the increase in T 1 . This outcome is attributed to the number of matching hypotheses decreasing in the second matching stage during each iteration with the increase in T 1 , while the running time of the second stage is more time-consuming than the first stage. Thus, the running time is reduced.
We fix the value to T 1 = 0.8 and conduct the experiment with different T 2 values. The result is shown in Figure 10c,d. The accuracy yields the local peak at T 2 = 1.8 , and the number of iterations and run time reach the local minimum peak at the same time. The findings indicate that the initial T 2 value can be retained at T 2 = 1.8 .
Figure 10e,f show the results with different T 3 values. The number of matching result shows a rising trend with the increase in T 3 , and the accuracy yields a peak at T 3 = 0.011 . The time and the number of iterations are relatively stable even though the number of iterations is nine when T 3 is 0.007 or 0.008, and the corresponding accuracy is low. Therefore, T 3 must be selected in a manner wherein the above-mentioned aspects are balanced and T 3 is updated to 0.011.
Similar to T 3 , parameter T 4 constrains the matching candidates in the second stage. The larger both values are, the more candidate matches will be generated in each iteration. As shown in Figure 11a,b, the accuracy of the result can reach the local peak at T 4 = 0.005 , and the number, run time, and iteration times begin to gradually flatten from T 4 = 0.005 . This finding suggests that the initial T 4 value can be retained.
For matching candidates obtained by the above-mentioned T 3 and T 4 constraints, we calculate the overall similarity score for each candidate and determine the highest score. If the highest score is larger than T 5 , then the candidate corresponding to the highest score will be selected as the final corresponding point. As shown in Figure 11c,d, the number of matching presents a downtrend with the increase in T 5 , and the accuracy yields a peak at T 5 = 0.55 . This result indicates that the initial T 5 value can be retained.
T s is the area of the triangle threshold that controls the size of the propagation triangle. The smaller T s is, the denser the matching results will be in general. This scenario can also be observed in Figure 11e,f, in which the number of corresponding points increases with the decrease in T s , and the running time is significantly increased at the same time. Although the number of matched points can reach the local maximum at T s = 20 , T s = 30 , which has the highest accuracy, is eventually selected in this study when running time is considered.
Notably, our method has low sensitivity toward different parameters within a certain range, in which the numbers and accuracies of matching points only have slight differences, and the matching results are relatively stable. Given that the proposed method is an iteration matching method, the last matching result will directly affect the matching result of the next iteration, and the effect of the different parameters on the result is relatively complex. Therefore, some signs can still be followed despite the difficulty in achieving consistent regularity for the whole scheme.

3.2. Different Matching Stages

Table 2 presents the required parameters that are finally determined according to the parametric analysis in Section 3.1. The proposed method is used for the dense matching of the 12 sets of images previously shown in Figure 9. The experimental results are listed in Table 3 and depicted in Figure 12. The columns in Table 3 represent the number of initial corresponding points, the stage of finished matching, the number of matching iterations, the number of matched midpoints, the number of matched intersection points, the number of all corresponding points, the number of incorrect matching, the matching accuracy, and the running time, respectively.
The results in Table 3 show several findings. Firstly, the proposed method utilizes the two-stage hierarchical matching strategy. The number of corresponding points can be increased by adding an implementation of the second matching stage, which will improve the accuracy of some results. However, with the increase in the number of corresponding points and the computational complexity in the second matching stage, the method will be a time-consuming process. Secondly, the number of matched midpoints is higher than the number of matched intersections for the two matching primitives. This phenomenon can be explained by the number of midpoints, which increases dynamically in the iterative matching process. By contrast, the number of extracted line segments is fixed, while that of the matched intersections tends to be stable after a few iterations. Thirdly, as shown by the corresponding points and their distribution in different colors in Figure 11, the corresponding points manifest a consistent distribution. Lastly, as for the experimental images selected in this study, the matching accuracies of the method all exceed 98%, which indicates the method’s high accuracy. This result can be attributed to the strict constraints of the algorithm and the reliability of the descriptors. The robustness of the method is also verified by the different types of image matching experiments.

3.3. Comparison with Jia’s Method

We compare the results of the proposed method with the results of the quasi-dense matching method presented in Jia (2019). In order to perform a comparative analysis of the two methods under consistent conditions, we utilize the same initial corresponding points to construct initial triangulations during the implementation of Jia’s method. Similar to our method, only triangles with an area greater than 30 are used for matching propagation in the matching process. The comparison results of the number of corresponding points, accuracy, the number of iterations, and running times of the two methods for 12 sets of image pairs are provided in Figure 13. As observed, the number of corresponding points of the proposed method is greater than that of Jia’s method for all test images, where Jia’s method fails in image pairs (e) and (f). In terms of accuracy, the accuracy of both methods is above 97%. However, the accuracy of the proposed method is higher than that of Jia’s method for most of the image pairs. Notably, Jia’s method has an accuracy of 100% at image pair (f), but this accuracy belongs to the accuracy of the initial seed point because Jia’s method fails in image pair (f) and does not obtain the corresponding point. However, no regularity is observed in the number of iterations between the two methods due to the complexity of the iterative process. We also compare the computation times required by the two methods. The results show that our method runs 0.1–4.1 times slower than Jia’s method, except for the image pairs (e) and (f). These results are due to three reasons. Firstly, the proposed method is designed using two matching stages; points that do not meet the descriptor similarity constraint in the first stage are further matched by extending the matching candidates and using multiple constraints, which increases the number of matches while ensuring match reliability. Secondly, compared with similarity constraints of single-pixel color information in Jia’s method, our constructed gradient descriptors are more robust for different images, especially for images with illumination change. The proposed method produces higher dense matching results, whereas Jia’s method fails to match. Thirdly, although the above-mentioned aspects contribute to the dense matching results of the proposed method, they require heavy workload. As a result, Jia’s method achieves better computation times than our method, but the proposed method outperforms Jia’s method in terms of robustness and density.

4. Conclusions

A hierarchical method based on triangulation constraint and propagation for the dense matching of points from a close-range image is proposed. The matching begins from the initial triangulation constructed by the initial corresponding points generated using the SIFT method. The correspondences between the image pairs are established with the midpoints and intersection points in the edges of the triangles as matching primitives. The characteristics of the proposed method are manifold. Firstly, we utilize points in the edges of triangles as the matching primitives and then conduct edge-to-edge matching propagation in the process of matching iteration. Our approach is more stable in theory than the region-to-region propagation. Secondly, we use two types of matching strategies for the different matching stages. We further match points that cannot be matched by the descriptor similarity in the first stage by utilizing multiple constraints. An overall matching similarity score is defined for the multiple constraints with different weights. In this manner, the number of matching results can be increased, and the reliability of the matching can be enhanced. Thirdly, two aspects need to be considered in improving the efficiency of the matching. The first aspect focuses on the multithreading technique for improving processing performance. The second aspect is only for the newly generated triangle edges, which are processed in each iteration matching to avoid repeated matching. Fourthly, the descriptor construction with overlapping subregions can improve the reliability of the descriptors and effectively restrain the propagation of mismatches. Lastly, intersections between feature line segments and triangle edges are introduced as the matching primitive. This way enriches the edge information of the result.
The results of the 12 typical close-range image pairs show that matching accuracy can exceed 98%. Thus, the proposed method has high robustness for different images and can obtain reliable matching results. In the future, we plan to change the range of matching results in images, from the coverage of corresponding points to the maximum overlap of image pairs, by performing an outward growth of a region with seed points. We plan to conduct tests to validate if the method can also obtain reliable results for aerial images. We also intend to transform the matching results into discrete point clouds for further 3D reconstruction.

Author Contributions

Writing—review and editing, Jingxue Wang; Methodology, Writing—original draft, Ning Zhang; Resources, Xiangqian Wu; Funding acquisition, Jingxue Wang and Weixi Wang. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China grant number 41871379 and 41971354.

Acknowledgments

We would like to thank all the anonymous reviewers for their valuable comments, which helped us to improve this manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, D.; Liu, H.; Cheng, X. A miniature binocular endoscope with local feature matching and stereo matching for 3d measurement and 3d reconstruction. Sensors 2018, 18, 2243. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Yang, W.; Li, X.; Yang, B.; Fu, Y. A novel stereo matching algorithm for digital surface model (DSM) generation in water areas. Remote Sens. 2020, 12, 870. [Google Scholar] [CrossRef] [Green Version]
  3. Su, N.; Yan, Y.; Qiu, M.; Zhao, C.; Wang, L. Object-based dense matching method for maintaining structure characteristics of linear buildings. Sensors 2018, 18, 1035. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Lhuillier, M.; Quan, L. Match propagation for image-based modeling and rendering. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 1140–1146. [Google Scholar] [CrossRef]
  5. Tola, E.; Lepetit, V.; Fua, P. A fast local descriptor for dense matching. In Proceedings of the 2008 26th IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–15. [Google Scholar] [CrossRef] [Green Version]
  6. Gruen, A. Development and Status of Image Matching in Photogrammetry. Photogramm. Rec. 2012, 27, 36–57. [Google Scholar] [CrossRef]
  7. Yu, Z.; Guo, X.; Ling, H.; Lumsdaine, A.; Yu, J. Line assisted light field triangulation and stereo matching. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2792–2799. [Google Scholar] [CrossRef] [Green Version]
  8. Revaud, J.; Weinzaepfel, P.; Harchaoui, Z.; Schmid, C. DeepMatching: Hierarchical Deformable Dense Matching. Int. J. Comput. Vis. 2016, 120, 300–323. [Google Scholar] [CrossRef] [Green Version]
  9. Dong, X.; Shen, J.; Shao, L. Hierarchical Superpixel-to-Pixel Dense Matching. IEEE Trans. Circuits Syst. Video Technol. 2017, 27, 2518–2526. [Google Scholar] [CrossRef] [Green Version]
  10. Jia, D.; Zhao, M.; Cao, J. FDM: Fast dense matching based on sparse matching. Signal Image Video Process. 2020, 14, 295–303. [Google Scholar] [CrossRef]
  11. Zhang, K.; Sheng, Y.; Ye, C. Stereo image matching for vehicle-borne mobile mapping system based on digital parallax model. Int. J. Veh. Technol. 2011, 2011, 1–11. [Google Scholar] [CrossRef] [Green Version]
  12. Laraqui, M.; Saaidi, A.; Mouhib, A.; Abarkan, M. Images Matching Using Voronoï Regions Propagation. 3D Res. 2015, 6, 1–16. [Google Scholar] [CrossRef]
  13. Khazaei, H.; Mohades, A. Fingerprint matching algorithm based on voronoi diagram. In Proceedings of the 2008 International Conference on Computational Sciences and Its Applications, Perugia, Italy, 30 June–3 July 2008; pp. 433–440. [Google Scholar] [CrossRef]
  14. Wu, J.; Wan, Y.; Chiang, Y.Y.; Fu, Z.; Deng, M. A Matching Algorithm Based on Voronoi Diagram for Multi-Scale Polygonal Residential Areas. IEEE Access 2018, 6, 4904–4915. [Google Scholar] [CrossRef]
  15. Soleymani, R.; Chehel Amirani, M. A hybrid fingerprint matching algorithm using Delaunay triangulation and Voronoi diagram. In Proceedings of the 20th Iranian Conference on Electrical Engineering (ICEE2012), Tehran, Iran, 15–17 May 2012; pp. 752–757. [Google Scholar] [CrossRef]
  16. Jiang, S.; Jiang, W. Reliable image matching via photometric and geometric constraints structured by Delaunay triangulation. ISPRS J. Photogramm. Remote Sens. 2019, 153, 1–20. [Google Scholar] [CrossRef]
  17. Zhu, Q.; Wu, B.; Tian, Y. Propagation strategies for stereo image matching based on the dynamic triangle constraint. ISPRS J. Photogramm. Remote Sens. 2007, 62, 295–308. [Google Scholar] [CrossRef]
  18. Li, T.; Sui, H.T.; Wu, C.K. Dense Stereo Matching Based on Propagation with a Voronoi Diagram. In Proceedings of the Third Indian Conference on Computer Vision, Graphics & Image Processing, Ahmadabad, India, 16–18 December 2002. [Google Scholar]
  19. Finch, A.M.; Wilson, R.C.; Hancock, E.R. Matching delaunay triangulations by probabilistic relaxation. In International Conference on Computer Analysis of Images and Patterns; Springer: Berlin/Heidelberg, Germany, 1995; Volume 970, pp. 350–358. [Google Scholar] [CrossRef]
  20. Lhuillier, M.; Quan, L. Image interpolation by joint view triangulation. In Proceedings of the 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Fort Collins, CO, USA, 23–25 June 1999; Volume 2, pp. 139–145. [Google Scholar] [CrossRef] [Green Version]
  21. Sedaghat, A.; Ebadi, H.; Mokhtarzade, M. Image Matching of Satellite Data Based on Quadrilateral Control Networks. Photogramm. Rec. 2012, 27, 423–442. [Google Scholar] [CrossRef]
  22. Liu, Z.; An, J.; Jing, Y. A simple and robust feature point matching algorithm based on restricted spatial order constraints for aerial image registration. IEEE Trans. Geosci. Remote Sens. 2012, 50, 514–527. [Google Scholar] [CrossRef]
  23. Zhao, M.; An, B.; Wu, Y.; Chen, B.; Sun, S. A Robust delaunay triangulation matching for multispectral/multidate remote sensing image registration. IEEE Geosci. Remote Sens. Lett. 2015, 12, 711–715. [Google Scholar] [CrossRef]
  24. Ma, J.; Ahuja, N. Region Correspondence by Global Con guration Matching and Progressive Delaunay Triangulation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Hilton Head Island, SC, USA, 13–15 June 2000; Volume 2, pp. 637–642. [Google Scholar] [CrossRef]
  25. Guo, X.; Cao, X. Good match exploration using triangle constraint. Pattern Recognit. Lett. 2012, 33, 872–881. [Google Scholar] [CrossRef]
  26. Zhu, Q.; Zhao, J.; Lin, H.; Gong, J. Triangulation of well-defined points as a constraint for reliable image matching. Photogramm. Eng. Remote Sens. 2005, 71, 1063–1069. [Google Scholar] [CrossRef]
  27. Zhu, Q.; Zhang, Y.; Wu, B.; Zhang, Y. Multiple close-range image matching based on a self-adaptive triangle constraint. Photogramm. Rec. 2010, 25, 437–453. [Google Scholar] [CrossRef]
  28. Wu, B.; Zhang, Y.; Zhu, Q. Integrated point and edge matching on poor textural images constrained by self-adaptive triangulations. ISPRS J. Photogramm. Remote Sens. 2012, 68, 40–55. [Google Scholar] [CrossRef]
  29. Jia, D.; Wu, S.; Zhao, M. Dense matching for wide baseline images based on equal proportion of triangulation. Electron. Lett. 2019, 55, 380–382. [Google Scholar] [CrossRef]
  30. Low, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  31. Fischler, M.A.; Bolles, R.C. Paradigm for Model. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  32. Lowe, D.G. Object recognition from local scale-invariant features. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; Volume 2, pp. 1150–1157. [Google Scholar] [CrossRef]
  33. Sun, Y.; Zhao, L.; Huang, S.; Yan, L.; Dissanayake, G. L2-SIFT: SIFT feature extraction and matching for large images in large-scale aerial photogrammetry. ISPRS J. Photogramm. Remote Sens. 2014, 91, 1–16. [Google Scholar] [CrossRef]
  34. Wang, E.; Jiao, J.; Yang, J.; Liang, D.; Tian, J. Tri-SIFT: A triangulation-based detection and matching algorithm for fish-eye images. Information 2018, 9, 299. [Google Scholar] [CrossRef] [Green Version]
  35. Morel, J.M.; Yu, G. ASIFT: A new framework for fully affine invariant image comparison. SIAM J. Imaging Sci. 2009, 2, 438–469. [Google Scholar] [CrossRef]
  36. Tola, E.; Lepetit, V.; Fua, P. DAISY: An efficient dense descriptor applied to wide-baseline stereo. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 815–830. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Wang, Z.; Wu, F.; Hu, Z. MSLD: A robust descriptor for line matching. Pattern Recognit. 2009, 42, 941–953. [Google Scholar] [CrossRef]
  38. Moisan, L.; Stival, B. A Probabilistic Criterion to Detect Rigid Point Matches Between Two. Int. J. Comput. Vis. 2004, 57, 201–218. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the proposed method.
Figure 1. Flowchart of the proposed method.
Ijgi 09 00347 g001
Figure 2. Two types of matching primitives and their corresponding matching hypotheses. Three sets of midpoint correspondence hypotheses are ( a , a ) , ( b , b ) , and ( c , c ) ; ( d , d ) is an intersection point correspondence hypothesis.
Figure 2. Two types of matching primitives and their corresponding matching hypotheses. Three sets of midpoint correspondence hypotheses are ( a , a ) , ( b , b ) , and ( c , c ) ; ( d , d ) is an intersection point correspondence hypothesis.
Ijgi 09 00347 g002
Figure 3. Schematic of the point descriptor construction.
Figure 3. Schematic of the point descriptor construction.
Ijgi 09 00347 g003
Figure 4. Multiple constraints.
Figure 4. Multiple constraints.
Ijgi 09 00347 g004
Figure 5. Propagating process of matching: (a) initial triangulation and two types of newly matched points; (b) updated triangulation using newly matched points.
Figure 5. Propagating process of matching: (a) initial triangulation and two types of newly matched points; (b) updated triangulation using newly matched points.
Ijgi 09 00347 g005
Figure 6. Propagating process of matching: (a) initial matching result; (b) further matching result; (c) final matching result.
Figure 6. Propagating process of matching: (a) initial matching result; (b) further matching result; (c) final matching result.
Ijgi 09 00347 g006
Figure 7. Zoomed view of the final triangulation from the two types of matching primitives.
Figure 7. Zoomed view of the final triangulation from the two types of matching primitives.
Ijgi 09 00347 g007
Figure 8. Numbers of different types of corresponding points and triangles in varying stages of matching iterations.
Figure 8. Numbers of different types of corresponding points and triangles in varying stages of matching iterations.
Ijgi 09 00347 g008
Figure 9. Image dataset. (a)Scale, (b) Rotation, (c) Viewpoint, (d) Viewpoint, (e) Illumination, (f) Illumination, (g) Rotation, (h) Rotation + Scale, (i) Rotation, (j) Blur, (k) Viewpoint, (l) JPEG compression.
Figure 9. Image dataset. (a)Scale, (b) Rotation, (c) Viewpoint, (d) Viewpoint, (e) Illumination, (f) Illumination, (g) Rotation, (h) Rotation + Scale, (i) Rotation, (j) Blur, (k) Viewpoint, (l) JPEG compression.
Ijgi 09 00347 g009
Figure 10. Matching results of different parameters: (a,c,e) represent the number of corresponding points and accuracy of matching; (b,d,f) represent the number of matching iterations and running time of the proposed method.
Figure 10. Matching results of different parameters: (a,c,e) represent the number of corresponding points and accuracy of matching; (b,d,f) represent the number of matching iterations and running time of the proposed method.
Ijgi 09 00347 g010
Figure 11. Matching results of different parameters: (a,c,e) represent the number of corresponding points and accuracy of matching; (b,d,f) represent the number of matching iterations and running time of the proposed method.
Figure 11. Matching results of different parameters: (a,c,e) represent the number of corresponding points and accuracy of matching; (b,d,f) represent the number of matching iterations and running time of the proposed method.
Ijgi 09 00347 g011
Figure 12. Matching result of the proposed method for image pairs, with corresponding points shown in green and blue color gradients. (a,c,e,g,i,k,m,o,q,s,u,w) are the reference image in each image pair, respectively; (b,d,f,h,j,l,n,p,r,t,v,x) are the search image in each image pair, respectively.
Figure 12. Matching result of the proposed method for image pairs, with corresponding points shown in green and blue color gradients. (a,c,e,g,i,k,m,o,q,s,u,w) are the reference image in each image pair, respectively; (b,d,f,h,j,l,n,p,r,t,v,x) are the search image in each image pair, respectively.
Ijgi 09 00347 g012
Figure 13. Comparison of the proposed method and Jia’s method for 12 sets of image pairs: (a) number of matched points obtained by both methods; (b) accuracy of the two methods; (c) number of iterations of the two methods; (d) running times of the two methods.
Figure 13. Comparison of the proposed method and Jia’s method for 12 sets of image pairs: (a) number of matched points obtained by both methods; (b) accuracy of the two methods; (c) number of iterations of the two methods; (d) running times of the two methods.
Ijgi 09 00347 g013
Table 1. Initial threshold values for matching.
Table 1. Initial threshold values for matching.
Parameters T 1 T 2 T 3 T 4 T 5 T s
Value 1.80.0090.0050.5550
Range0.5–1.51.2–20.006–0.0150.002–0.0150.45–0.7550–20
Step0.10.10.0010.0010.05−5
Table 2. Final selected thresholds and parameters for matching.
Table 2. Final selected thresholds and parameters for matching.
Parameters T 1 T 2 T 3 T 4 T 5 T s w 1 w 2 w 3 w 4
Value0.81.80.0110.0050.55300.450.250.150.15
Table 3. Comparison of matching results of the proposed method in different stages for twelve image pairs.
Table 3. Comparison of matching results of the proposed method in different stages for twelve image pairs.
ImagesSeed PointsStageNumber of IterationsNumber of MidpointsNumber of IntersectionsTotalWrongAccuracy (%)Time (s)
(a)9111132760176754362399.5883
2126016218590912399.75623
(b)2131144749267476342499.6964
2128014290511,1166799.4591
(c)63512410,592286314,0886099.57852
22719,069382623,48811999.493916
(d)4631136794461311,8685699.53211
2108530486513,8406399.54655
(e)30211810,028395214,2804099.72336
21213,838406018,1791799.911134
(f)391213415185553072499.5527
2124581193765501899.73125
(g)7501372351211952185898.89117
21516,468569022,802899.96904
(h)1901142988192050964699.0980
21910,139451614,7726799.54536
(i)891181047122623601599.36108
2129090219511,3436699.421058
(j)19212112,994562018,80413399.291232
21419,509568425,34412799.502392
(k)2331146265372910,2255399.48404
22311,533416215,8988199.501479
(l)72411810,958366715,3474299.79649
21515,591364019,9392399.891683

Share and Cite

MDPI and ACS Style

Wang, J.; Zhang, N.; Wu, X.; Wang, W. Hierarchical Point Matching Method Based on Triangulation Constraint and Propagation. ISPRS Int. J. Geo-Inf. 2020, 9, 347. https://0-doi-org.brum.beds.ac.uk/10.3390/ijgi9060347

AMA Style

Wang J, Zhang N, Wu X, Wang W. Hierarchical Point Matching Method Based on Triangulation Constraint and Propagation. ISPRS International Journal of Geo-Information. 2020; 9(6):347. https://0-doi-org.brum.beds.ac.uk/10.3390/ijgi9060347

Chicago/Turabian Style

Wang, Jingxue, Ning Zhang, Xiangqian Wu, and Weixi Wang. 2020. "Hierarchical Point Matching Method Based on Triangulation Constraint and Propagation" ISPRS International Journal of Geo-Information 9, no. 6: 347. https://0-doi-org.brum.beds.ac.uk/10.3390/ijgi9060347

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop