Next Article in Journal
Motion Control of a Gecko-like Robot Based on a Central Pattern Generator
Next Article in Special Issue
Correction of Biogeochemical-Argo Radiometry for Sensor Temperature-Dependence and Drift: Protocols for a Delayed-Mode Quality Control
Previous Article in Journal
Surface Modification Enabling Reproducible Cantilever Functionalization for Industrial Gas Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Mosaic Method for Side-Scan Sonar Strip Images Based on Curvelet Transform and Resolution Constraints

Department of Military Oceanography and Hydrography and Cartography, Dalian Naval Academy, Dalian 116018, China
*
Author to whom correspondence should be addressed.
Submission received: 11 July 2021 / Revised: 1 September 2021 / Accepted: 8 September 2021 / Published: 9 September 2021
(This article belongs to the Special Issue Advanced Sensing Technology for Ocean Observation)

Abstract

:
Due to the complex marine environment, side-scan sonar signals are unstable, resulting in random non-rigid distortion in side-scan sonar strip images. To reduce the influence of resolution difference of common areas on strip image mosaicking, we proposed a mosaic method for side-scan sonar strip images based on curvelet transform and resolution constraints. First, image registration was carried out to eliminate dislocation and distortion of the strip images. Then, the resolution vector of the common area in two strip images were calculated, and a resolution model was created. Curvelet transform was then performed for the images, the resolution fusion rules were used for Coarse layer coefficients, and the maximum coefficient integration was applied to the Detail layer and Fine layer to calculate the fusion coefficients. Last, inverse Curvelet transform was carried out on the fusion coefficients to obtain images in the fusion area. The fusion images in multiple areas were then combined in the registered images to obtain the final image. The experiment results showed that the proposed method had better mosaicking performance than some conventional fusion algorithms.

1. Introduction

As the depth of global ocean exploration continues to increase, understanding the seafloor surface and near-surface is of great significance in the “digital ocean” and “transparent ocean” era. Currently, side-scan sonar is an important means to explore seafloor geomorphology [1], and side-scan sonar images provide important data for seafloor object identification, classification of seafloor sediments, and exploration of marine resources [2,3]. In order to obtain side-scan sonar images in the entire testing zone, the most important task is to mosaic the strip images, in addition to seafloor tracking, slant-range correction, gain correction, and geocoding [4]. Side-scan sonar is generally operated using towing cables, which leads to inaccurate location information. If the coordinate information is directly used to mosaic the images, there will be distortion in the images [5,6,7,8]. Currently, a large number of studies have been carried out to achieve a mosaic of object-level strip images that produce images with complete information and of high quality.
By dividing sonar strip images into paired objects and shadows, Daniel et al. [9] realized rigid registration of side-scan sonar images using a decision tree. Through region segmentation, Thisen et al. [10] extracted shadow areas from side-scan sonar images and calculated the displacement between two images using the cross-correlative method, thereby achieving rigid registration. Vandrish et al. [11] showed that the scale invariant feature transform (SIFT) algorithm can be used for registration of sonar images, although the accuracy was not ideal. Using correlation coefficients and mutual information as similarity parameters, Chailloux et al. [12] extracted a series of significantly correlated areas on adjacent strip images and calculated the global rigid transformation parameters and local elastic transformation parameters, thereby eventually realizing mosaic of adjacent strip images. Wang et al. [13] improved the pre-processing method of side-scan sonar images to extract feature points more accurately and effectively after preprocessing; they also proposed a sped up robust feature (SURF)-based elastic mosaic algorithm to achieve feature-level conformal mosaic of the images. Moreover, Cao et al. [14] used wavelet transform in a strip image mosaic, yet it required the 3D posture information of the side-scan sonar. Zhao et al. [15] extracted SURF features of the pre-processed strip images and then performed block registration, which achieved good mosaic results. To obtain sonar images of large-area seafloor surface, Zhao et al. [16] also proposed a side-scan image mosaicking method based on the coupling feature points of position constraints. In addition, He et al. [17] used the unsharp masking (USM) algorithm to enhance the side-scan images and the SURF algorithm for image mosaicking; experiments showed that their method effectively enhanced image features and increased the amount of image information, but the average gray values of the images were affected.
The above image mosaic algorithms primarily focused on the extraction and registration of features points of adjacent strip images, and most adopted the wavelet fusion algorithm after image registration, without further exploration for alternative image fusion algorithms. Due to the complex marine environment during ocean exploration, it is nearly impossible to ensure that the sonar images on one survey line are always better than those of an adjacent strip image. Therefore, it is necessary to take into account the differences in image resolution during strip image mosaicking and retain clear image information while screening necessary information in blurred images. To address this problem, we performed image fusion using curvelet transform, which can reveal more detailed information of strip images than wavelet transform. Then, the resolution of strip images was evaluated using a resolution weight model to constrain the curvelet transform, thereby achieving mosaicked strip images with better quality. The contents of this paper were arranged as follows: Section 2 mainly introduces seven different methods of resolution assessment, which would all be used in the calculation of resolution weight model; Section 3 mainly introduces the specific process of strip Mosaic method proposed in this paper; Section 4 uses the measured data to verify the feasibility of this method; and Section 5 contains the summary and prospects.

2. Image Resolution Assessment Methods

As an important data source of seafloor geomorphology, the resolution of side-scan sonar images directly determines the accuracy of target identification and seafloor sediments classification. The assessment of image quality can be divided into two types: subjective assessment and objective assessment [18,19]. Subjective assessment is mainly performed by trained professionals, whereas objective assessment uses mathematical models to measure the image resolution based on different indices. Thus, it is imperative to develop an objective assessment method that is in consistency with subjective assessment. Currently, common objective assessment methods can be divided into three categories according to the degree of use of reference images, i.e., full-reference quality assessment, reduced-reference quality assessment, and no-reference quality assessment [20]. Since there is no original reference image for side-scan sonar images, the no-reference quality assessment method was adopted in this study.
Image resolution is one of the most important image quality evaluation indexes and is the most important image parameter of sonar image. Therefore, the resolution of image became the main research object. A total of seven resolution assessment methods from four aspects will be introduced in this section. As the more classical parameter indexes in the assessment method, they measure the sharpness of the image from different aspects. Additionally, they will all be used in the calculation of resolution vector in Section 3, making the evaluation result more accurate and perfect.

2.1. Assessment Method Based on Image Gradient

Image gradient reflects the marginal information of images. The greater the gradient value is, the sharper the image edge and the clearer the image will be. Common gradient functions for evaluating image resolution include the following three types [21].

2.1.1. Energy Gradient Function

The energy gradient of an image is the quadratic sum of the difference in grayscale value of adjacent pixels in the horizontal and vertical direction. The summation of energy gradient values of all pixels in the image is then taken as the function value. The function is shown in Equation (1):
F E G = x y { [ f ( x + 1 , y ) f ( x , y ) ] 2 + [ f ( x , y + 1 ) f ( x , y ) ] 2 }
where x and y are pixel coordinates, and f ( x , y ) is the grayscale value of the pixel.

2.1.2. Brenner Gradient Function

Brenner gradient function is relatively the easiest gradient assessment function [22]. It calculates the quadratic sum of the grayscale difference of two adjacent pixels, meaning a small calculation amount. Yet, it is sensitive to noise. The function is shown in Equation (2):
F B r e n n e r = x y [ f ( x + 2 , y ) f ( x , y ) ] 2

2.1.3. Tenengrad Gradient Function

Krotkv et al. [23] used the Tenengrad gradient function as one of the assessment indexes of image resolution, the results of which were close to objective assessment results. In this method, the Sobel operator was first used to extract the horizontal and vertical gradient values of pixels, then the quadratic sum was compared with a threshold T. The gradient values of pixels greater than T were added to obtain the Tenengrad gradient function value. The function is shown in Equation (3):
F T e n e n g r a d = x y [ G ( x , y ) 2 ]
where G ( x , y ) is the gradient calculated by the Sobel operator, as shown in Equation (4):
G ( x , y ) = G x 2 ( x , y ) + G y 2 ( x , y )
where G x ( x , y ) and G y ( x , y ) represent the horizontal and vertical gradient values, respectively.
G x ( x , y ) = f ( x , y ) g x G y ( x , y ) = f ( x , y ) g y
where is the convolution operator, and g x and g y represent the horizontal and vertical templates of the Sobel operator, respectively:
g x = [ 1 0 1 2 0 2 1 0 1 ] g y = [ 1 2 1 0 0 0 1 2 1 ]

2.2. Assessment Method Based on Image Transform Domain

It is generally believed that a clear image contains more high-frequency components than a blurry image. Thus, some studies have attempted to transform the image to the frequency domain to perform image quality assessment [24].

2.2.1. Discrete Fourier Transform (DFT)

As the most basic time–frequency transformation methods, DFT is widely used in resolution assessment. Specifically, 2D DFT is first performed on the image, and then the zero-frequency component is shifted to the matrix center, such that the frequency diffuses from the center to the periphery and from low frequency to high frequency. The spectrum values of corresponding pixels are weighted based on the distance to the central pixel, and the resolution assessment value is the weighted average of the spectrum values of all pixels [25,26]. The function of DFT-based image resolution assessment is shown in Equation (7) [27]:
F D F T = 1 M × N μ = 0 M 1 ν = 0 N 1 μ 2 + ν 2 P ( μ , ν )
where M and N are the image dimensions, μ 2 + ν 2 represents the distance of a pixel to the central pixel, and P ( μ , ν ) is the spectrum value of a pixel after DFT.

2.2.2. Discrete Cosine Transform (DCT)

DFT-based resolution assessment methods have high sensitivity; however, they are computationally more demanding than DCT-based methods. In comparison, DCT has a general orthogonal transform property, and the base vector of DCT matrix could describe image features very well [28,29]. Therefore, by replacing DFT with DCT, the transform coefficient is changed into a real number, which reduces the computation while still obtaining the distribution of image frequency. The resolution assessment function based on DCT is shown in Equation (8):
F D C T = 1 M × N μ = 0 M 1 ν = 0 N 1 ( λ + φ ) | C ( λ , φ ) |
where C ( λ , φ ) is the spectrum value of a pixel after DCT.

2.3. Assessment Method Based on Entropy Function

The entropy of an image is an important index to measure the richness of image information. Shannon believed that the greater the entropy value, the richer information the image contains. During image resolution assessment, the clearer the image is, the more abundant grayscale distribution it has, and thus, the greater the entropy value is [30]. The definition of entropy function is shown in Equation (9):
F e n t r o p y = i = 0 255 p ( i ) log 2 p ( i )
where p ( i ) is the probability of occurrence of every grayscale value.

2.4. Assessment Method Based on Variance Function

The variance function can represent the dispersion degree of the image grayscales. The smaller the range of grayscale, the smaller the variance is and the blurrier the image is, and vice versa [31]. The definition of variance function is shown in Equation (10):
F V a r = x y { [ f ( x , y ) ε ] 2 }
where ε is the average grayscale value of the image, the definition of which is in Equation (11):
ε = 1 M × N x y f ( x , y )

3. Strip Mosaic Method Based on Curvelet Transform and Resolution Constraints

3.1. Image Fusion Algorithm Based on Curvelet Transform

To obtain a clear and continuous image that can reflect complete information of the entire testing zone, image fusion in the overlapping area of side-scan sonar strip images is required. Currently, there are three common image fusion methods, namely weighted average method, image pyramid method, and wavelet fusion method [32]. The wavelet fusion method is the most common side-scan sonar strip image mosaicking method. However, due to the limitations in algorithms, the wavelet transform can only obtain edge features in the horizontal and vertical directions, and the wavelet basis does not have the anisotropy property. Hence, it is unable to get close to the image texture features. To overcome the limitations in the wavelet transform and improve the quality of strip image mosaicking, the Curvelet transform was introduced in the current study.
The Curvetlet transform was first proposed by Candes and Donoho in 1999 [33] based on the Ridgelet transform. As a multi-resolution, band-pass, and directional multi-scale image analysis method, Curvelet transform has the three characteristics of an optimal image representation method proposed by the National Institute for Physiological Science, Japan [34]. Similar to wavelet transform, Curvelet transform calculates the correlation of spatial images using a group of base functions, thereby characterizing edges and curves at different angles. The main steps of image fusion based on Curvelet transform are as follows: Curvelet coefficients are first obtained from Curvelet decomposition of the image, the coefficients are then processed based on specific fusion rules, and lastly, inverse Curvelet transform is carried out on the fused coefficient to obtain the final fusion image [35,36].
The Curvelet coefficients are obtained using the equation below:
C ( j , θ , k 1 , k 2 ) = 0 x M , 0 y N f ( x , y ) φ j , θ , k 1 , k 2 ( x , y )
where f ( x , y ) is the input image, M × N are the image dimensions, j is the scale, θ is the direction, k 1 , k 2 is the spatial location of Curvelet, and φ ( x , y ) represents the Curvelet function, which includes a group of base functions described by parameters ( j , θ , k 1 , k 2 ) .
Different from the wavelet coefficients, the Curvelet coefficients include the low-frequency coefficient in the innermost layer (i.e., the Coarse layer), the mid-to-high frequency coefficient in the Detail layer, and the high-frequency coefficient in the outermost Fine layer. As the number of layers increases, the scale of the corresponding base function turns smaller, and there are more directions. Figure 1 shows a frequency-domain base division method. Each square in Figure 1 represents a scale, and there are five scales. The bigger the square, the higher the frequency, and the smaller the scale is; hence, more detailed information will be reflected. The radial lines represent the angles. At each scale, the angle division is different, and the higher the frequency is, the smaller the angle is.
From Jia et al. [37], the energy of coefficients is mainly concentrated in the low-frequency coefficient, and the energy gradually declines as the frequency increases. In other words, the low-frequency coefficient reflects the general trend of the image, whereas high-frequency coefficient reflects the outline and texture details of an image. By fusing the coefficients at various layers using different fusion rules, the fusion image coefficient can be obtained, and by performing inverse Curvelet transform of the fusion image coefficient, the fusion image is obtained.

3.2. Strip Image Mosaicking Based on Curvelet Transform and Resolution Constraints

Due to uncertainties in the marine environment during exploration, common areas in adjacent strip images might have large differences during actual measurement. Both strip images might have good quality, or one or both of them may not be good at all. The traditional side-scan strip image mosaicking algorithms do not take the image resolution into account. In order to ensure good mosaic results, a Curvelet coefficient fusion criterion based on the resolution weight model was proposed in the present study.
In Section 2, we have introduced seven different image resolution assessment methods, including energy gradient function, Brenner gradient function, Tenengrad gradient function, DFT, DCT, entropy function, and variance function. According to Li et al. [38] and Xie et al. [39], different resolution assessment methods may have different results for the same group of images. In other words, a single method is not able to assess the resolution of an image accurately. Hence, these seven resolution assessment methods were integrated in this study to build a resolution vector, and the image resolution was obtained based on probability and given weights.
The resolution vector Q , created based on the resolution value of the above seven methods, is shown in Equation (13):
Q = [ F E G , F B r e n n e r , F T e n e n g r a d , F D F T , F D C T , F e n t r o p y , F V a r ]
Since the resolution index in each method has a positive relationship with the image resolution, the resolution weight is obtained by comparing the resolution vectors of image 1 and image 2, Q 1 , Q 2 , respectively.
R a t i o = s u m ( Q 1 Q 2 ) 7
where the resolution weight R a t i o represents the probability of an image having better resolution than the other image. Thus, it was taken as the fusion rule in the Coarse layer of Curvelet transform, as shown in Equation (15).
C C o a r s e _ f u s i o n = R a t i o C C o a r s e _ 1 + ( 1 R a t i o ) C C o a r s e _ 2
where C C oarse _ f u s i o n , C C oarse _ 1 and C C oarse _ 2 represent the coefficient in the Coarse layer after fusion and that of image 1 and image 2, respectively.
In order to fully show the texture and details of the image, the maximum coefficient fusion approach was adopted to process the Detail layer and Fine layer coefficients, as shown in Equation (16):
C D e t a i l _ f u s i o n ( x , y ) = M a x { | C D e t a i l _ 1 ( x , y ) | , | C D e t a i l _ 2 ( x , y ) | } C F i n e _ f u s i o n ( x , y ) = M a x { | C F i n e _ 1 ( x , y ) | , | C F i n e _ 2 ( x , y ) | }
where C D e t a i l _ f u s i o n , C D e t a i l _ 1 and C D e t a i l _ 2 represent the coefficient in the Detail layer after fusion and that of image 1 and image 2, respectively. C F i n e _ f u s i o n , C F i n e _ 1 and C F i n e _ 2 represent the coefficient of the Fine layer after fusion and that of image 1 and image 2, respectively. Figure 2 shows the flowchart of the proposed mosaic method based on Curvelet transform and resolution constraints.
  • Extract and match feature points of adjacent strip images and obtain registered mosaic strips using the affine transformation.
  • Select the common area A from two strip images.
  • Perform Curvelet transform for two images to obtain the coefficients in the Coarse layer, Detail layer, and Fine layer.
  • Calculate the resolution vectors of the two images to obtain the corresponding resolution weight.
  • Fuse the Coarse layer coefficients using resolution fusion rules to obtain the low-frequency coefficients. Fuse the Detail layer and Fine layer coefficients using the maximum coefficient fusion rules to obtain the high-frequency coefficients.
  • Perform inverse Curvelet transform on the fusion coefficients to obtain the fusion image in area A, which is then mosaicked to the registered strip images.
  • Repeat steps 2–6 until the whole mosaic image is obtained.
In traditional mosaic algorithms for strip images, there are various problems, such as inconsistent resolution of adjacent strip images and image distortion. In this study, we proposed a mosaic method for strip images based on Curvelet transform and resolution constraints, which produced mosaic images with complete information and high quality.

4. Experiment and Results

To verify the effectiveness of the proposed image mosaicking method, image data collected in 2019 using the Klein4000 side-scan sonar in Jiaozhou Bay, Qingdao, Shandong Province, China was used in the experiment. The water depth of the survey area is approximately 30–40 m. The overlapping rate of adjacent strip images is 50%. After preprocessing, such as seafloor tracking, slant-range correction, gray level equalization, noise suppression, gain correction, and geocoding, a group of strip image pairs with obvious common features were selected, as shown in Figure 3.
Figure 4a shows a mosaic image calculated based on geographic coordinate information. As can be seen, there is obvious dislocation and distortion. According to the steps of our method, the feature points in the strip images were extracted and matched, as shown in Figure 4b. Figure 4c shows a registered strip image after affine transformation. Based on the results, the distortion and dislocation were eliminated after image registration, resulting in good visual effects and laying a solid foundation for image fusion in the next step.
To effectively select the fusion area and ensure the integrity of the selected features, the whole survey area was first rotated counterclockwise for a certain angle, such that the survey line was approximately along the vertical direction [40]. Another reason to rotate the strips is that a series of subsequent steps, such as Curvelet transform and image fusion, require regular rectangles. After image mosaicking, it was rotated back to the original direction. Areas 1–3 were selected, and the sonar images of two strips in these areas are shown in Figure 5.
Taking Area 1 as an example, the proposed algorithm was used to process two strips in the area. First, the coefficients in the Coarse, Detail, and Fine layers were extracted using Curvelet transform. The coefficient structure is shown in Table 1.
In both strips, Area 1 has the same dimensions of 923 × 166 . Five layer decomposition was carried out. As shown in Table 1, the dimension of the coefficient matrix increases with the increase in scale. The larger the scale in spatial domain, the smaller the scale in frequency domain, and the more detailed the description of high frequency information.
Then, the resolution vectors of two strips in Area 1 were calculated, and the results are shown in Table 2.
Then, using the proposed algorithm, the coefficients in the Coarse layer of the two images were fused based on the resolution fusion rule, and the coefficients in the Detail and Fine layers of the two images were fused using the maximum coefficient fusion approach, thereby obtaining the low-frequency and high-frequency coefficients of the fused image. Lastly, the fused image of Area 1 was obtained via inverse Curvelet transform.
In order to verify the rationality of the resolution fusion rule proposed in this paper, the resolution fusion rule, the mean fusion rule, and the maximum fusion rule are used to combine the five layer coefficients of the two images obtained by the Curvelet decomposition, respectively. As shown in Figure 6, 19 combinations of fusion coefficients were obtained.
Then, the fusion coefficients of each group were inversely transformed to obtain fusion images.
The information entropy, average gradient, and spatial frequency were used as evaluation indices of the fusion results. The information entropy reveals the amount of information contained in the image, and the greater the entropy, the better the fusion result; the average gradient reflects the image’s contrast expression of small details, and the greater the average gradient, the higher the image fusion quality; the spatial frequency represents the overall activity of the image in spatial domain, and the higher the spatial frequency, the better the fusion result. Table 3 shows the three indices of each combination, and Figure 7 shows the line chart of the analysis results.
As shown in Table 2 and Figure 7, the Curvelet coefficient fusion strategy proposed in this paper, namely the resolution fusion rule used in the Coarse layer and the maximum coefficient fusion rule used in the Detail layer and Fine layer, has the best image fusion effect.
To further demonstrate the effectiveness of the proposed algorithm, the images were fused using different algorithms, including simple average, traditional wavelet fusion and wavelet fusion with resolution constraints. The fusion results were compared with that of the proposed algorithm. The traditional wavelet fusion algorithm applies the mean fusion rule to the low-frequency information of wavelet transform and the maximum coefficient fusion rule to the high-frequency information. In the wavelet fusion with resolution constraints, the resolution fusion rule is applied to the low-frequency information of wavelet transform and the maximum coefficient fusion rule is applied to the high-frequency information.
Table 4 shows the three indices of the four fusion methods, and Figure 8 shows the fusion strip images.
As shown in Table 4, the information entropy, average gradient, and spatial frequency of the proposed algorithm are much greater than those of the other three methods, indicating that the fusion result of the proposed method is the best. By comparing the results of wavelet fusion with resolution constraints and our method, it can be seen that Curvelet fusion achieved better fusion results than wavelet fusion. In addition, based on the value of indices of traditional wavelet fusion and wavelet fusion with resolution constraints, the effectiveness of the resolution fusion rule proposed in this study was demonstrated. It can also be seen intuitively from Figure 8 that the fusion image obtained by our method has better clarity and can show more details.
To further verify the effectiveness of the proposed method, the same experiments were repeated for Areas 2 and 3. Figure 9 shows the fusion strip images in Area 2 and Area 3. The evaluation results are shown in Table 5.
As shown in Table 5, the fused images in Areas 2 and 3 of the proposed method have the highest information entropy, average gradient, and spatial frequency, suggesting the best performance in image fusion and validating the effectiveness and stability of the proposed algorithm.
Then, the fused images in the three areas were mosaicked onto the registered strip, which was then rotated clockwise to the original orientation, as shown in Figure 10. Compared with Figure 4c, it can be seen that Figure 9 better reflects the overall characteristics of the features by enhancing detail texture information while retaining the overall trend of the overlapping areas.

5. Conclusions

Current strip image mosaicking algorithms do not consider the influence of the resolution difference of common objects in adjacent images on the results of mosaicking. Moreover, a traditional wavelet fusion algorithm is not able to fully describe the image details. To address these problems, in this study, we proposed an image mosaic method based on Curvelet transform and resolution constraints. Experimental verification using actual measurement data showed that the proposed method can greatly improve the fusion results, which provides high-quality image data for subsequent submarine target recognition and sediment classification, thereby greatly benefiting ocean exploration. However, there are still a lot of improvements to be made in this method, such as human involvement in the process. In view of this, target recognition and other technologies in deep learning can be introduced in the future. Thus, it can automatically identify and extract the areas that need to be fused and achieve full automation.

Author Contributions

Conceptualization, S.J. and N.Z.; methodology, S.J. and N.Z.; software, G.B. and Y.C.; validation, L.C., Y.C. and G.B.; formal analysis, S.J.; investigation, N.Z. and L.C.; resources, S.J.; data curation, L.C.; writing—original draft preparation, N.Z.; writing—review and editing, N.Z., S.J. and G.B.; visualization, G.B.; supervision, Y.C.; project administration, S.J.; funding acquisition, S.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China, grant number: 41876103.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data is kept in_house and not available online.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Philippe, B. The Handbook of Sidescan Sonar; Springer: Berlin/Heidelberg, Germany; New York, NY, USA, 2009; ISBN 978-3-540-42641-7. [Google Scholar]
  2. Fakiris, E.; Papatheodorou, G.; Geraga, M.; Ferentinos, G. An Automatic Target Detection Algorithm for Swath Sonar Backscatter Imagery, Using Image Texture and Independent Component Analysis. Remote Sens. 2016, 8, 373. [Google Scholar] [CrossRef] [Green Version]
  3. Wang, X.; Zhao, J.; Zhu, B.; Jiang, T.; Qin, T. A Side Scan Sonar Image Target Detection Algorithm Based on a Neutrosophic Set and Diffusion Maps. Remote Sens. 2018, 10, 295. [Google Scholar] [CrossRef] [Green Version]
  4. Cervenka, P.; Moustier, C.D. Sidescan sonar image processing techniques. IEEE J. Ocean. Eng. 1993, 18, 108–122. [Google Scholar] [CrossRef] [Green Version]
  5. Chavez, P.S.C., Jr.; Isbrecht, J.A.; Galanis, P.; Gabel, G.L.; Sides, S.C.; Soltesz, D.L.; Ross, S.L.; Velasco, M.G. Processing, mosaicking and management of the monterey bay digital sidescan-sonar images. Mar. Geol. 2002, 181, 305–315. [Google Scholar] [CrossRef]
  6. Wang, A. Research on 3D Seafloor Terrian Recovery from the Side Scan Sonar Image; Wuhan University: Wuhan, China, 2014. [Google Scholar]
  7. Cervenka, P.; Moustier, C.; Lonsdale, P. Geometric corrections on sidescan sonar images based on bathymetry: Application with SeaMARC II and Sea Beam data. Mar. Geophys. Res. 1995, 17, 217–219. [Google Scholar] [CrossRef] [Green Version]
  8. Cervenka, P.; Moustier, C. Postprocessing and corrections of bathymetry derived from sidescan sonar systems: Application with SeaMARC II. IEEE J. Ocean. Eng. 1994, 19, 619–629. [Google Scholar] [CrossRef]
  9. Daniel, S.; FL, L.; Roux, C. Side-scan sonar image matching. IEEE J. Ocean. Eng. 1998, 23, 245–259. [Google Scholar] [CrossRef]
  10. Thisen, E.; Sorensen, H.B.; Stage, B. Sidescan Sonar Image Matching Using Cross Correlation; International Society for Optics and Photonics: Bellingham, WA, USA, 2003; Volume 5089. [Google Scholar]
  11. Vandrish, P.; Vardy, A.; Walker, D. Side-scan sonar image registration for AUV navigation. In Proceedings of the 2011 IEEE Symposium on Underwater Technology and Workshop on Scientific Use of Submarine Cables and Related Technologies, Tokyo, Japan, 5–8 April 2011; IEEE: Piscataway, NJ, USA, 2011. [Google Scholar]
  12. Chailloux, C.; Le Caillec, J.M.; Gueriot, D.; Zerr, B. Intensity-Based Block Matching Algorithm for Mosaicing Sonar Images. IEEE J. Ocean. Eng. 2011, 36, 627–645. [Google Scholar] [CrossRef]
  13. Wang, A.; Zhao, J.; Guo, J.; Wang, X. Elastic Mosaic Method in Block for Side-Scan Sonar Image Based on Speeded-Up Robust Features. J. Wuhan Univ. Inf. Sci. Ed. 2018, 43, 697–703. [Google Scholar]
  14. Cao, M.; Guo, J. Research on Side Scan Sonar Stripe Image Mosaic Method Based on Common Vision Target. Geomat. Spat. Inf. Technol. 2014, 46–49, 52. [Google Scholar]
  15. Zhao, J.; Wang, A.; Wang, X. Research on Segment Mosaic Method of Side Scan Sonar Stripe Image. Geomat. Inf. Sci. Wuhan Univ. 2013, 9, 29–33. [Google Scholar]
  16. Zhao, J.; Shang, X.; Zhang, H. Side-Scan Sonar Image Mosaic Using Couple Feature Points with Constraint of Track Line Positions. Remote Sens. 2018, 10, 953. [Google Scholar] [CrossRef] [Green Version]
  17. He, F.; Wu, M.; Long, R.; Chen, Z. Accurate Mosaic of side scan sonar image based on SURF feature. Chin. J. Ocean. Technol. 2020, 39, 35–41. [Google Scholar]
  18. Dharmishtha, P.; Jaliya, U.; Vasava, H. A review: No-reference/blind image quality assessment. Int. Res. J. Eng. Technol. 2017, 4, 339–343. [Google Scholar]
  19. Yang, X.; Li, F.; Liu, H. A survey of DNN methods for blind image quality assessment. IEEE Access 2019, 7, 123788–123806. [Google Scholar] [CrossRef]
  20. Wang, Z. Review of no-reference image quality assessment. Acta Autom. Sin. 2015, 41, 1062–1079. [Google Scholar]
  21. Chern, N.N.K.; Neow, P.A.; Ang, M.H. Practical issues in pixel-based autofocusing for machine vision. In Proceedings of the 2001 ICRA. IEEE International Conference on Robotics and Automation (Cat. No. 01CH37164), Seoul, Korea, 21–26 May 2001; IEEE: Piscataway, NJ, USA, 2001; pp. 2791–2796. [Google Scholar]
  22. Subbarao, M.; Tyan, J. Selecting the optimal focus measure for autofocusing and depth-from-focus. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 864–870. [Google Scholar] [CrossRef] [Green Version]
  23. Krotkov, E.P. Active Computer Vision by Cooperative Focus and Stereo; Springer: New York, NY, USA, 1989; pp. 1–17. ISBN 9781461396659. [Google Scholar]
  24. Chetouani, A.; Beghdadi, A.; Deriche, M. A new reference-free image quality index for blur estimation in the frequency domain. In Proceedings of the 2009 IEEE International Symposium on Signal Processing and Information Technology (ISSPIT), Ajman, United Arab Emirates, 14–17 December 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 155–159. [Google Scholar]
  25. Vu, C.; Phan, T.; Chandler, D. S3: A spectral and spatial measure of local perceived sharpness in natural images. IEEE Trans. Image Process. 2012, 21, 934–945. [Google Scholar] [CrossRef]
  26. Lu, Y.; Zhang, T.; Zheng, J.; LI, M.; Zhang, C. No-reference blurring image quality assessment based on local standard deviation and saliency map. J. Jilin Univ. Eng. Technol. Ed. 2016, 46, 1337–1343. [Google Scholar]
  27. Li, Z. Study on Large-Scale Image Registration Algorithm in Fourier Transform Domain; Nanjing University of Information Science & Technology: Nanjing, China, 2013. [Google Scholar]
  28. Marichal, X.; Ma, W.; Zhang, H. Blur determination in the compressed domain using DCT information. In Proceedings of the 1999 International Conference on Image Processing (Cat. 99CH36348), Kobe, Japan, 24–28 October 1999; IEEE: Piscataway, NJ, USA, 1999; pp. 386–390. [Google Scholar]
  29. Caviedes, J.; Oberti, F. A new sharpness metric based on local kurtosis, edge and energy information. Signal Process. Image Commun. 2004, 19, 147–161. [Google Scholar] [CrossRef]
  30. Zhu, Z.; Li, S.; Chen, H. Research on Automatic Focusing Function Based on Image Entropy. Opt. Precis. Eng. 2004, 5, 537–542. [Google Scholar]
  31. Wang, C.; Cui, L.; Yan, B. Research on the evaluation algorithm of microscopic image sharpness based on Variance-Brenner function. Equip. Manuf. Technol. 2020, 10, 78–82. [Google Scholar]
  32. Xu, J. Research on Key Technologies of Side-Scan Sonar Image Mosaic and Segmentation; East China University of Technology: Fuzhou, China, 2017. [Google Scholar]
  33. Candes, E. Curvelets—A Surprisingly Effective Nonadaptive Representation for Objects with Edges. 2020. Available online: http://citeseerx.ist.psu.edu/viewdoc/summary;jsessionid=0B25E5104FB2CB688F212E84E4F2BD75?doi=10.1.1.161.9294 (accessed on 1 September 2021).
  34. Chen, Z. Curving Transform and Information Extraction Application of Potential Field Data; China University of Geosciences: Beijing, China, 2012. [Google Scholar]
  35. Vijayalakshmi, G.; Shanthakumar, M.; Muthukumar, K. A Contrast Adjusted Wavelet and Curvelet Transform Based Fusion of Breast Images. IOP Conf. Ser. Mater. Sci. Eng. 2020, 994, 012027. [Google Scholar] [CrossRef]
  36. Tian, Y.; Zhang, H.; Ma, X. An Image Fusion Method Based on Curvelet Transform and Guided Filter Enhancement. Math. Probl. Eng. 2020, 2020, 9821715. [Google Scholar] [CrossRef]
  37. Jia, X. Research on Face Recognition Algorithm Based on Curvwave Transform; Harbin Engineering University: Harbin, China, 2011. [Google Scholar]
  38. Li, Z.; Li, X.; Ma, L. Research on Definition Assessment based on No-reference Digital Image Quality. Remote. Sens. Technol. Appl. 2011, 26, 239–246. [Google Scholar]
  39. Xie, X.; Zhou, J.; Wu, Q. A Nonreference Quality Evaluation Index for Image Blur. Comput. Appl. 2010, 4, 921–924. [Google Scholar]
  40. Gao, F.; Wang, X.; Yang, J.; Zhang, B.; Zhou, H.; Chen, J. Research on Multi-strip Side Scan Sonar Image Fine Mosaic Method. Sci. Technol. Innov. Appl. 2021, 5, 1–4. [Google Scholar]
Figure 1. Frequency-Domain Base of Curvelet Transform.
Figure 1. Frequency-Domain Base of Curvelet Transform.
Sensors 21 06044 g001
Figure 2. Flowchart of the proposed method.
Figure 2. Flowchart of the proposed method.
Sensors 21 06044 g002
Figure 3. Two strips used for verification. Four image pairs with obvious common features were selected.
Figure 3. Two strips used for verification. Four image pairs with obvious common features were selected.
Sensors 21 06044 g003
Figure 4. Strip image registration. A–D and E–H are four areas selected from (a) and (c) respectively. (b) shows the registration process of strips. It can be seen that there was significant dislocation in A–D. After strip registration, the dislocation effect largely disappeared in E–H.
Figure 4. Strip image registration. A–D and E–H are four areas selected from (a) and (c) respectively. (b) shows the registration process of strips. It can be seen that there was significant dislocation in A–D. After strip registration, the dislocation effect largely disappeared in E–H.
Sensors 21 06044 g004
Figure 5. Selected fusion areas.
Figure 5. Selected fusion areas.
Sensors 21 06044 g005
Figure 6. Combination diagram of fusion rules about Curvelet transform coefficients.
Figure 6. Combination diagram of fusion rules about Curvelet transform coefficients.
Sensors 21 06044 g006
Figure 7. Line chart of the fusion coefficients analysis results.
Figure 7. Line chart of the fusion coefficients analysis results.
Sensors 21 06044 g007
Figure 8. Fusion results of the four different methods.
Figure 8. Fusion results of the four different methods.
Sensors 21 06044 g008
Figure 9. (a) shows the strip images and fusion strip images in Area 2. (b) shows the strip images and fusion strip images in Area 3.
Figure 9. (a) shows the strip images and fusion strip images in Area 2. (b) shows the strip images and fusion strip images in Area 3.
Sensors 21 06044 g009aSensors 21 06044 g009b
Figure 10. Results of strip image mosaicking.
Figure 10. Results of strip image mosaicking.
Sensors 21 06044 g010
Table 1. Structure of Curvelet transform coefficients.
Table 1. Structure of Curvelet transform coefficients.
LayerScale CoefficientNumber of DirectionsMatrix Dimensions
Coarse C { 1 } 1 77 × 13
Detail C { 2 } 16 62 × 14 57 × 14 77 × 11 77 × 10
C { 3 } 32 120 × 14 115 × 15 115 × 14 115 × 15
77 × 22 78 × 21 77 × 21
C { 4 } 32 241 × 29 231 × 28 231 × 29 154 × 44
155 × 42 155 × 42 154 × 42
Fine C { 5 } 1 923 × 166
Table 2. Resolution vectors of two strips in Area 1.
Table 2. Resolution vectors of two strips in Area 1.
F E G F Brenner F Tenengrad F D F T F D C T F entropy F Var
Q 1 1.6 × 10 9 7.2 × 10 8 6.0 × 10 8 2.5 × 10 8 4.3 × 10 3 2.6 × 10 6 7.303
Q 2 2.2 × 10 9 1.4 × 10 8 1.0 × 10 8 3.4 × 10 8 6.2 × 10 3 3.3 × 10 6 6.886
Q 1 ,   Q 2 denote the resolution vectors of Strip 1 and Strip 2, respectively. Additionally, the resolution weight ratio, computed according to Equation (14), is 0.1428.
Table 3. Comparison of fusion effects in different combinations.
Table 3. Comparison of fusion effects in different combinations.
CaseInformation EntropyAverage GradientSpatial Frequency
a7.3376 9.4983 25.5944
b7.3826 11.1938 31.9010
c7.4250 12.5927 33.7882
d7.4637 13.0406 34.3080
e 17.6156 13.7586 35.4629
f7.2523 9.4976 25.5687
g7.3222 9.7393 25.8062
h7.3588 10.3818 26.4863
i7.3941 11.8541 28.6577
j7.4265 13.1232 34.4313
k7.1802 7.5718 19.2626
l7.2552 10.3865 30.7766
m7.3167 12.3194 33.4798
n7.3728 12.9310 34.1983
o7.4441 13.1217 34.4303
p7.2523 9.4976 25.5687
q7.3222 9.7393 25.8062
r7.3588 10.3818 26.4863
s7.3941 11.8541 28.6577
1 is the fusion rule combination form of our method.
Table 4. Comparison of fusion results of different methods in Area 1.
Table 4. Comparison of fusion results of different methods in Area 1.
AlgorithmsInformation EntropyAverage GradientSpatial Frequency
Our method7.615613.758635.4629
Wavelet fusion with resolution constraints7.35699.087228.6397
Traditional wavelet fusion7.22608.205026.8381
Simple average7.15847.645219.5543
Table 5. Comparison of fusion results of different methods in Areas 2 and 3.
Table 5. Comparison of fusion results of different methods in Areas 2 and 3.
RatioFusion AlgorithmsInformation EntropyAverage GradientSpatial Frequency
Area 20.1428Our method7.131811.752730.4386
Wavelet fusion with resolution constraints6.93888.256925.4472
Traditional wavelet fusion6.76147.345123.7968
Simple average6.69626.745717.0373
Area 30.2857Our method7.236711.842530.1219
Wavelet fusion with resolution constraints6.96197.815024.4447
Traditional wavelet fusion6.91747.451023.7001
Simple average6.86576.858516.9889
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, N.; Jin, S.; Bian, G.; Cui, Y.; Chi, L. A Mosaic Method for Side-Scan Sonar Strip Images Based on Curvelet Transform and Resolution Constraints. Sensors 2021, 21, 6044. https://0-doi-org.brum.beds.ac.uk/10.3390/s21186044

AMA Style

Zhang N, Jin S, Bian G, Cui Y, Chi L. A Mosaic Method for Side-Scan Sonar Strip Images Based on Curvelet Transform and Resolution Constraints. Sensors. 2021; 21(18):6044. https://0-doi-org.brum.beds.ac.uk/10.3390/s21186044

Chicago/Turabian Style

Zhang, Ning, Shaohua Jin, Gang Bian, Yang Cui, and Liang Chi. 2021. "A Mosaic Method for Side-Scan Sonar Strip Images Based on Curvelet Transform and Resolution Constraints" Sensors 21, no. 18: 6044. https://0-doi-org.brum.beds.ac.uk/10.3390/s21186044

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop