Next Article in Journal
Dissolved Gaseous Mercury (DGM) in the Gulf of Trieste, Northern Adriatic Sea
Next Article in Special Issue
Coordinated Formation Control of Discrete-Time Autonomous Underwater Vehicles under Alterable Communication Topology with Time-Varying Delay
Previous Article in Journal
A Summary of Environmental Monitoring Recommendations for Marine Energy Development That Considers Life Cycle Sustainability
Previous Article in Special Issue
Autonomous Water Sampler for Oil Spill Response
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Infrared and Visible Image Fusion Methods for Unmanned Surface Vessels with Marine Applications

1
Science and Technology on Underwater Vehicle Laboratory, Harbin Engineering University, Harbin 150001, China
2
TOEC Technology Co., Ltd., Tianjin 300210, China
3
Marine Design and Research Institue of China, Shanghai 200011, China
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2022, 10(5), 588; https://0-doi-org.brum.beds.ac.uk/10.3390/jmse10050588
Submission received: 10 March 2022 / Revised: 12 April 2022 / Accepted: 13 April 2022 / Published: 26 April 2022

Abstract

:
Infrared and visible image fusion is a very effective way to solve the degradation of sea images for unmanned surface vessels (USVs). Fused images with more clarity and information are useful for the visual system of USVs, especially in harsh marine environments. In this work, three novel fusion strategies based on adaptive weight, cross bilateral filtering, and guided filtering are proposed to fuse the feature maps that are extracted from source images. First, the infrared and visible cameras equipped on the USV are calibrated using a self-designed calibration board. Then, pairs of images containing water scenes are aligned and used as experimental data. Finally, each proposed strategy is inserted into the neural network as a fusion layer to verify the improvements in quality of water surface images. Compared to existing methods, the proposed method based on adaptive weight provides a higher spatial resolution and, in most cases, less spectral distortion. The experimental results show that the visual quality of fused images obtained based on an adaptive weight strategy is superior compared to other strategies, while also providing an acceptable computational load.

1. Introduction

Recently, USVs have received considerable attention due to their high working efficiency and strong adaptability in ocean missions, such as maritime search and rescue, port surveillance, and ocean environment monitoring [1,2,3]. To make USVs competent for these mission requirements, object detection and recognition are of the utmost significance for them to sense the surrounding environment. However, due to the complex working environment, there are still various obstacles to achieving satisfactory environment recognition performance on the sea surface, such as sea fog, sea reflections, and rainstorms [4].
To sense the environment more clearly, fusing infrared and visible images is superior in many aspects [5]. First, a USV will generally be equipped with an optoelectronic device to capture visible and infrared images at the same time, making this method easy to implement [6,7,8,9,10]. Second, infrared and visible images provide scene information from different aspects. Data from images at different frequencies are combined to enhance the knowledge obtained from expected scene information. This combination contains more information than the combination of single-modality signals [11]. Finally, infrared and visible images have complementary characteristics, which means that fused images are robust and informative. Visible images typically have a high spatial resolution and considerable detail and chiaroscuro, but they will be seriously degraded in severe weather [12,13,14], while infrared images, which depict the information of objects from a different aspect, are resistant to these disturbances at lower resolutions. Fusion technology can combine the advantages of these two kinds of images to obtain better results. Therefore, it is highly preferable to incorporate infrared and visible image fusion techniques in environmental perception to enhance the adaptability of USVs. It is worth noting that, due to its applicability in all conditions and even in high-humidity environments, medium-wave infrared (MWIR) cameras that are sensitive to a thermal energy of 3–5 μm are generally used on USVs. Thus, only medium-wave infrared was considered in this study.
Many researchers have studied the fusion of infrared and visible images and provided different methods [5]. Multi-scale transform-based methods have proven to be very effective for image fusion and other image-processing tasks [15,16,17]. A multi-scale representation of the input image is obtained by multi-scale transformation. The multi-scale coefficients of fusion are obtained according to specific fusion rules, which usually take into account the activity of the coefficients and the correlation between adjacent pixels or pixels of different scales. Finally, the fusion coefficients are inversely transformed to obtain the fused image. This image fusion framework involves two basic problems, the choice of the multi-scale decomposition method or the fusion strategy for multi-scale coefficient fusion. Sparse representation [18,19,20,21,22] has emerged as a novel signal analysis model, where the signal can be expressed as a linear combination of a few atoms that can reveal the intrinsic properties of the image. The representation of images with linear combinations of sparse bases is the key to their good performance.
In recent years, with advances in computer performance, neural network theory has been further improved [23,24]. Neural network-based methods have better real-time performance and effectiveness compared to other methods, so they are suitable for USVs. Liu et al. [25] proposed a fusion method based on convolutional sparse representation (CSR), in which two raw images are taken as input to the network, and the extracted feature details based on CSR are considered to obtain a fused image. This method is more stable than the fusion method based on sparse representation. After that, Liu et al. [26] presented a CNN-based fusion method for the task of multi-focus image fusion. This CNN-based approach encodes direct mapping from source images to weight maps so that the activity level measurement and weight assignment can be obtained directly by training the neural network. However, this method can only be used in multi-focus image fusion. Wu et al. [27] used the L1 norm and weighted-average strategy to generate several candidates for fused detail content, with features extracted by a deep learning network. The fused image is reconstructed by combining the fused base part and the detail content. However, these neural network-based methods only use the features extracted from the last layer in the neural network, which results in the loss of shallow feature information obtained in the process of feature extraction. To overcome this drawback, Wu et al. [28] presented a novel deep learning architecture named DenseFuse for infrared and visible image fusion. The encoding network is combined by convolutional networks, fusion layers, and dense blocks, in which the output of each layer is connected to every other layer. It can avoid losing the feature information in the middle layer by making more effective use of the features extracted from the original image. However, the inputs are fused according to the two fusion strategies, in which feature maps from different sources have the same importance. As a result, the complementarity of the two images cannot be effectively utilized.
Motivated by the above observations, this paper provides three fusion strategies to improve the image fusion quality for USVs under complicated sea environments. Contributions of this work are summarized as follows:
(1)
A novel calibration board was designed to calibrate infrared and visible cameras. To avoid the loss of detailed texture in infrared images that is induced by the thermal diffusion between high- and low-temperature regions, all heating elements of the calibration board are processed to be insulated, which makes the accuracy of the calibration able to be improved. Moreover, a major part of the calibration board is made of lightweight thermal insulation material. Thus, the designed calibration board not only possesses high contrast for both visible images and infrared images, but it also enjoys high portability in USV field applications.
(2)
Three novel fusion strategies, adaptive weight fusion (AWF), cross bilateral filtering fusion (CBF), and guided filtering fusion (GFF), are proposed in this paper. The AWF calculates the weight of each sub-block, rather than roughly calculating the feature maps’ weights, such that the fusion result can be more accurate. The CBF considers intensity re-semblance and geometric closeness for the computation of fusion weights. The GFF utilizes a guided filter for edge preservation in the fusion results. Compared to the previous fusion strategies [28,29], the proposed fusion strategies make more effective use of the texture features in infrared images.
(3)
The proposed algorithms are compared with two widely accepted algorithms, average weighted fusion (AVE) [29] and L1 norm weighted fusion (L1) [28], under the optoelectronic system of ‘Tianxing-1’ USV. The experiment results indicate that the proposed AWF strategy can be applied to the USV and show superior performance on water surface images.
The paper is structured as follows: Section 2 introduces the self-designed calibration board used for camera calibration which is a precondition for infrared and visible image fusion. The proposed methods are explained in detail in Section 3. Section 4 presents the experimental results and a discussion. The paper concludes with Section 5.

2. Camera Calibration

In this study, a calibration board (shown in Figure 1) was designed to calibrate infrared and visible cameras. To generate the characteristics of infrared radiation, the calibration board contained 48 holes arranged in a 6 × 8 array, with a heating element in each hole. The horizontal and vertical distance between heating elements was 40 mm. To avoid the loss of detailed texture in infrared images due to thermal diffusion between high- and low-temperature regions, we used LED lights wrapped in black tape as heating elements. The base of the calibration board was made of insulation material so that the contrast of infrared images can be improved. The image information of the calibration board could be collected by the visible and infrared cameras simultaneously when the LED lights were on. Compared to the commonly used checkerboard, the calibration board is suitable for USV field tests, as it is more portable.
We captured visible and infrared images of the calibration board simultaneously, as shown in Figure 2. Following the method in the literature [30], the intrinsic parameters of the visible and infrared cameras and the extrinsic parameters between the two cameras can be estimated. In this way, we can extract the corresponding corner from the visible and infrared images simultaneously by using the calibration board.
After the camera calibration was completed, the visible and infrared images could be registered using the estimated parameters. The result is shown in Figure 3, which shows the superposition of visible gray image and infrared image. By observing this result, it can be concluded that the scene of the visible image is consistent with the infrared image. Therefore, the objects in these two figures are well aligned.

3. Improvement of Fusion Network

As shown in Figure 4, the detail framework of the network in [28] consisted of three parts: encoder, fusion layer, and decoder. The feature maps obtained by the encoder were fused in the fusion layer then were integrated into one feature map that contains all salient features in source images. Finally, the fused image was reconstructed by a decoder network.
The input images, including infrared and visible images, are denoted as I 1 , , I k , and k 2 . Note that the input images were registered in the process of camera calibration. As shown in Figure 4, the encoder consisted of a convolutional layer (C1) and a DenseBlock that contains three convolutional layers. The architecture of the DenseBlock can preserve deep features as much as possible in the encoding network.
In the training phase, the fusion layer was discarded, and only one image was fed into the network at a time. The loss function L is a weighted combination of pixel loss L P and structural similarity (SSIM) loss L s s i m , with the weight λ as shown in Figure 5. The pixel loss L P indicates the Euclidean distance between the output O and the input I . The S S I M ( · ) in L s s i m denotes the structural similarity of two images. The λ was set as 1, 10, 100, and 1000, respectively, because there are three orders of magnitude differences between L P and L s s i m . Based on the above, the encoder and decoder network can be trained to reconstruct the input image. The detailed framework of the network in the training phase is shown in Figure 5. The architecture of the network is outlined in Table 1.
When the network has the ability to reconstruct the input images, the quality of the fusion result depends on the fusion layer (strategy). The fusion layer is used to combine the salient feature maps from different sources. The final fused image is obtained by the decoder, in which the result of the fusion layer is regarded as the input. A diagram of the fusion layer is shown in Figure 6.
Different from [28], which focused on the feature extraction and reconstruction ability of the neural network, the fusion strategy in the neural network was further studied in this study. Two fusion strategies (AVE, L1) used for the network were proposed in [28]. The AVE takes the average sum of two feature maps as the fusion result. The L1 calculated the value of the L1 norm between two feature maps as the fusion weight. It was proven that the L1 norm strategy had better performance, but it still has shortcomings that could be improved. In the L1 norm strategy, the weighted sum of feature maps is used to achieve the fusion of feature maps, in which the weight is set for the entire feature map. The interaction between different parts within the feature map is not considered. Moreover, the L1 norm strategy only calculates the values in the 3 × 3 range around the point, which leads to the loss of the remaining features. To make up for these drawbacks, we proposed three novel fusion strategies to improve the quality of the fusion result by improving the fusion layer (strategy) in the network.

3.1. Adaptive Weight Strategy

In this section, the adaptive weight-based fusion (AWF) strategy is developed. The visible and infrared feature maps are divided into blocks, with AWF utilized by the window size l × l (better result with l = 80 ). The weights are calculated for the blocks rather than the whole feature map, resulting in a more accurate fused feature map. In this way, feature loss is avoided, since the perceptual area corresponding to each point is increased.
If we define D i . j as the block located in row i , column j within the feature map, then the corresponding weight C i , j ( ϕ m ) of block can be formulated as:
C i , j ( ϕ m ) = ϕ m ( x , y ) 1 , ( x , y ) D i , j
where ϕ m ( m { 1 , , M } ) represents the feature maps extracted from one input image, with M representing the number of feature maps, and ( x , y ) denotes the coordinates of any point in block D i . j .
Then, the bilinear interpolation method was adopted to deal with the region segmentation problem induced by the blocking artifact. For any point ( x , y ) in a feature map, the weights of its nearest four neighbor blocks are denoted as C i , j , C i + 1 , j , C i , j + 1 , C i + 1 , j + 1 . For these blocks, we adopted the following method to set the weights of the point:
ω m ( x , y ) = ( 1 Δ y ) ( ( 1 Δ x ) C i , j ( ϕ m ) + Δ x C i + 1 , j ( ϕ m ) ) + Δ y ( ( 1 Δ x ) C i , j + 1 ( ϕ m ) + Δ x C i + 1 , j + 1 ( ϕ m ) )
where Δ x = | x x 0 | / l , Δ y = | y y 0 | / l , ( x 0 , y 0 ) denotes the center of the upper left block. Then, the fusion feature map was generated by weighting the visible and infrared feature map according to the following formula:
f m ( x , y ) = ω v m ( x , y ) × ϕ v m ( x , y ) + ω i r m ( x , y ) × ϕ i r m ( x , y ) ω v m ( x , y ) + ω i r m ( x , y )
where ϕ v m ( x , y ) indicates feature maps of visible image and ϕ i r m ( x , y ) for those of infrared image, as shown in Figure 6; ω v m ( x , y ) indicates the weight of ϕ v m ( x , y ) , and ω i r m ( x , y ) indicates the weight of ϕ i r m ( x , y ) . ( x , y ) denotes the coordinates of any point in the feature map.
For a position in the edge area (green area in Figure 7), the weight at the corresponding points in the feature map is obtained by linear interpolation between two adjacent blocks. For a position in the corner area (pink area in Figure 7), the weight of this block is directly used as the weight of the point corresponding to the feature map.
The experimental results of the presented adaptive weight strategy are shown in Figure 8.

3.2. Cross Bilateral Filtering

Cross bilateral filtering (CBF) is a modified weight estimation method inspired by the bilateral filter [31]. This method considers both levels of gray similarities and the geometric closeness of neighborhood pixels in image A to adjust the filter kernel and filters the image B. The weights are computed by measuring the strength of the details in the detail image obtained by CBF. This method smoothens the image by preserving the edges and applying neighborhood pixels. Therefore, this method makes better use of the texture details in the infrared image than the fusion strategy in [28], which directly takes L1 parameters as fusion weights. Note that the structural similarity between feature maps extracted from infrared and visible images meets the application conditions of CBF.
(1)
CBF
The feature maps of infrared images were used to filter the corresponding feature maps of visible images. To simplify the expression, A denotes the ϕ i r m , and B denotes the ϕ v m . The CBF output of B marked as B C B F at pixel location p is calculated as:
B C B F = 1 W q S G σ s ( p q ) × G σ r ( | A ( p ) A ( q ) | ) B ( q )
W = q S ( exp ( p q 2 2 σ s 2 ) exp ( A ( p ) A ( q ) 2 2 σ r 2 )
where G σ s ( p q ) = e p q 2 2 σ s 2 represents a geometric closeness function with the design parameter σ s , which is normally set to 1.8; p q is the Euclidean distance between p and q ; G σ r ( | A ( p ) A ( q ) | ) = e | A ( p ) A ( q ) | 2 2 σ r 2 is a gray-level similarity edge stopping function, σ r is a design parameter which is normally set to 25, A ( · ) denotes the pixel value at position · in feature map A and B ( · ) in feature map B, and S is the spatial neighborhood of p . W is a normalization constant.
The detail image of feature maps A and B can be expressed with A D and B D , respectively.
A D = A A C B F B D = B B C B F
(2)
Pixed-based fusion rule
A window of size l × l (referring to [31], l = 11 ) around a detail coefficient A D ( x , y ) or B D ( x , y ) is considered as a neighborhood to compute its weight. This neighborhood is denoted as matrix M . Each row of M is treated as an observation and each column as a variable to compute the unbiased estimate ƛ x , y h of its covariance matrix, in which ( x , y ) are the spatial coordinates of the detail coefficient A D ( x , y ) or B D ( x , y ) .
cov ( M ) = E [ ( M E [ M ] ) ( M E ( M ) ) T ]
ƛ x , y h = k = 1 l ( x k x ¯ ) ( x k x ¯ ) T l 1
where x k is the kth observation of the l -dimensional variable, and x ¯ is the mean of observation. Similarly, an unbiased covariance estimate ƛ x , y v is computed by treating each column of M as an observation and each row as a variable (opposite to that of ƛ x , y h ). The sum of these eigenvalues is directly proportional to the horizontal detail strength of the neighborhood and is denoted as S h . Similarly, the sum of eigenvalues of ƛ x , y v gives vertical detail strength S v . That is,
S h ( x , y ) = k = 1 l e i g e n k   o f   ƛ x , y h
S v ( x , y ) = k = 1 l e i g e n k   o f   ƛ x , y v
where e i g e n k is the k th eigenvalue of the unbiased estimate of the covariance matrix.
The addition operation between horizontal detail strength S h and vertical detail strength S v can lead to the weight of a particular detail factor:
w t ( x , y ) = S h ( x , y ) + S v ( x , y )
Therefore, the fused feature map is computed using the following equation:
f m ( x , y ) = A ( x , y ) w t a ( x , y ) + B ( x , y ) w t b ( x , y ) w t a ( x , y ) + w t b ( x , y )
where the weights for the detail coefficient A D and B D are denoted as w t a and w t b .
The results of the fusion strategy based on CBF are shown in Figure 9.

3.3. Guided Filtering

In this section, we improved the feature fusion layer using the guided filtering method [32], which acted as the DenseFuse network fusion rule. The L1 norm strategy in [28] calculated the L1 parametric number for both visible and infrared feature maps, resulting in them having the same importance in the algorithm. However, the texture features of infrared images were much clearer than those of visible images in actuality. Motivated by this, the fused feature map was reconstructed utilizing feature maps of infrared images ϕ i r m to guide that of visible images ϕ v m , so that the textural features in the infrared images would be fully utilized in this fusion strategy.
We assume that f m is a linear transform of ϕ i r m in a window w k centered at pixel k:
f m ( x , y ) = a k ϕ i r m ( x , y ) + b k , ( x , y ) w k
where ( a k , b k ) are some linear coefficients assumed to be constant in w k . This local linear model ensures that f m has an edge only if ϕ i r m has an edge. In other words, the output will reflect the approximate contour edge of the guide. It is supposed that the output f m consists of input ϕ v m and noise n .
f m = ϕ v m n
Then, the following cost function was used to minimize the difference between f m and ϕ v m while maintaining the linear model:
E ( a k , b k ) = ( x , y ) w k ( ( a k ϕ i r m ( x , y ) + b k ) 2 + ε a k 2 )
where ε is a regularization parameter penalizing large a k . Equation (15) is the linear ridge regression model, and its solution is given by
a k = 1 | w | ( x , y ) w k ϕ i r m ( x , y ) ϕ v m ( x , y ) μ k ϕ v m ( k ) ¯ σ k 2 + ε
b k = ϕ v m ( k ) ¯ a k μ k
where μ k and σ k 2 are the mean and variance of I in w k , | w | is the number of pixels in w k , and ϕ v m ( k ) ¯ is the mean of ϕ v m in w k . After computing ( a k , b k ) for all windows w k in the image, the filtering output was computed by
f m ( x , y ) = 1 | w | ( x , y ) w k ( a k ϕ i r m + b k ) = a ¯ i ϕ i r m ( x , y ) + b ¯ i
where a ¯ i = 1 | w | k w a k , b ¯ i = 1 | w | k w b k .
A fusion strategy based on guided filtering can take advantage of the similarities between feature maps extracted from different sources. This method can preserve the edge contours and texture feature information of source images. The results are shown in Figure 10.

4. Experimental Results and Analysis

The results of the field experiments are given in this section to show that our proposed algorithm can improve the quality of sea images and is suitable for the USV visual system. We first trained the network to reconstruct the input images (fusion layer was discarded). The MS-COCO dataset was adopted to train the encoder and decoder network. In the dataset, about 79,000 images were utilized as input images, while the remaining 1000 images were used to validate the reconstruction ability. After the training, the parameters of the encoder and decoder were no longer updated. The fusion effect of different fusion strategies was observed by the improved fusion layer in the network.
To demonstrate the effectiveness and superiority of the proposed methods under other operating scenarios, we evaluated previous fusion methods (AVE, L1) and the proposed fusion method (AWF, CBF, GFF) in both qualitative and quantitative aspects. Eight representative pairs of visible and infrared images were selected, which were specified by different illumination contained by common sea scenes, such as coast, buoy, ship, and lighthouse. The ocean surface infrared images used in the experiment were collected with an MWIR camera equipped on the USV. The ‘Tian Xing’ USV platform for the experiment is shown in Figure 11.

4.1. Qualitative Image Quality Assessment

The fusion results of the five algorithms are shown in Section 4.2. Our results seem to indicate that AWF had the best performance in clarity and contrast restoration, making the fused image look like a natural image. The fused image obtained by CBF had a complete scene structure, but the clarity was poor and the details of dense edge texture areas such as hillsides were poorly preserved. It did not have advantages over other methods. The contrast and clarity of the fused image obtained by GFF were better than the other methods. The details were clear, the texture details were retained completely, and the noise elimination effect was good. Distant ships, buoys, lighthouses, and other target objects could be clearly identified. However, the results show that there will be obvious color distortion under certain circumstances, which affects the overall image quality. In general, the proposed AWF fusion strategy had better performance under different working conditions.

4.2. Quantitative Image Quality Assessment

Quantitative image quality assessment can overcome the influence of subjective factors of the observer and make accurate and objective judgments on the effect of the image. The performance of proposed methods was evaluated in the following two aspects.
(1)
Quantitative assessment compared to baselines
To illustrate the effectiveness of the proposed three methods, they were compared with the AVE [29] and L1 in [28]. Comparison results of the three proposed methods with the AVE method and L1 method are given in Figure 12. For these results, the average gradient (AG) was adopted to measure the quality of the fusion result.
Assessments of the comparison results are summarized in Table 2. From these results, we can observe that the presented methods (i.e., AWF, CBF, GFF) possessed higher AG values than the AVE and L1 method. To be specific, compared with the AVE and L1 method, the AG value improved by 5.5% (AWF), 12.4% (CBF), and 2.7% (GFF); 14.9% (AWF), 22.6% (CBF), and 12% (GFF), respectively. These results indicate that the three proposed methods had better performance in feature fusion than the widely accepted AVE and L1.
(2)
Quantitative assessment over proposed algorithms
To further demonstrate the effectiveness and superiorities of the proposed methods, average gradient (AG), information entropy (EN), standard deviation (SD), structural similarity (SSIM), mutual information (MI), and signal-to-noise ratio (SNR) were used in a quantitative image quality assessment. There were eight pairs of input images as shown in Figure 13, the fusion results and the values of the above six metrics are shown in Figure 14 and Table A1, respectively.
Figure 15 and Table 3 show the average gradient of the output images in different sea surface scenarios with different methods. Table A1 provides the quantitative image quality assessment of eight groups of fused images.
According to the results in Figure 13 and Table A1, the AWF algorithm had a stable effect, and its overall average gradient improved by 6.7% compared to the AVE method. Other indicators also performed well. The CBF algorithm had a clear gap with the other methods in most indicators. The performance of the GFF method varied greatly in different scenarios. It performed slightly worse than the AWF method in terms of the average gradient index and showed a 5.5% improvement compared to the AVE method.
In conclusion, according to the qualitative and quantitative image quality assessment, the AWF fusion strategy is more suitable than other fusion strategies for sea images.

5. Conclusions

In this paper, three fusion strategies were proposed to combine with neural networks to improve the quality of water surface images in response to the image degradation problem existing in the marine environment. First, calibration of the infrared and visible cameras equipped on the USV was carried out using a self-designed calibration board to estimate the camera parameters. Then, visible and infrared images of the USV’s operating environment were aligned, utilizing these parameters. Finally, the pairs of aligned images were used as the input for the network to verify the superiority of the proposed methods.
The quality of images generated by the proposed methods was evaluated according to a variety of indicators. Compared to AVE, the average image fusion gradient based on the AWF increased by 6.7%. The CBF had no obvious advantage over the original algorithm (AVE and L1). The GFF performed better than other methods, but there was a color distortion problem in the images. The proposed AWF showed superior performance to other methods, in most cases, in two aspects of the qualitative and quantitative image quality assessment that was carried out.
The experiments showed that, compared to the existing methods AVE and L1 from [28], AWF is more suitable for water surface scenes and can be applied in USV visual systems. It can improve the image quality to help USVs better identify targets and complete various complex tasks.

Author Contributions

Conceptualization, R.Z.; methodology, R.Z.; software, Y.L.; validation, R.Z., Y.L. and J.F.; formal analysis, R.Z. and Y.S.; investigation, R.Z.; resources, L.Z.; data curation, Y.L.; writing—original draft preparation, R.Z.; writing—review and editing, R.Z.; visualization, J.F.; supervision, Y.S.; project administration, L.Z.; funding acquisition, L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by [Heilongjiang Provincial Excellent Youth Fund] grant number [YQ2021E013] and [Central University Fund] grant number [3072021CFT0104].

Institutional Review Board Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Objective Image Quality Assessment.
Table A1. Objective Image Quality Assessment.
Image Method Quantitative Analysis
(Percentage Change)
AG EN SD SSIM MI SNR
Image 1AVE1.6187.12859.940.8696.25621.883
L11.985
(22.7%)
7.181
(0.7%)
50.74
(−15.3%)
0.855
(−1.6%)
6.185
(−1.1%)
22.783
(4.1%)
AWF1.995
(23.3%)
7.183
(0.8%)
50.594
(−15.6%)
0.86
(−1.0%)
6.167
(−1.4%)
23.182
(5.9%)
CBF1.452
(−10.3%)
6.073
(−14.8%)
32.446
(−45.9%)
0.841
(−3.2%)
5.713
(−8.7%)
20.439
(−6.6%)
GFF1.91
(18.0%)
7.227
(1.4%)
64.736
(8.0%)
0.849
(−2.3%)
6.349
(1.5%)
20.904
(−4.5%)
Image 2AVE2.1637.19256.7040.8516.11118.096
L12.321
(7.3%)
7.232
(0.6%)
55.569
(−2.0%)
0.847
(−0.5%)
6.165
(0.9%)
21.218
(17.3%)
AWF2.349
(8.6%)
7.275
(1.2%)
57.362
(1.2%)
0.849
(−0.2%)
6.146
(0.6%)
21.177
(17.0%)
CBF2.029
(−6.2%)
6.748
(−6.2%)
43.31
(−23.6%)
0.797
(−6.3%)
6.049
(−1.0%)
20.695
(14.4%)
GFF2.514
(16.2%)
7.206
(0.2%)
57.344
(1.1%)
0.816
(−4.1%)
6.439
(5.4%)
18.801
(3.9%)
Image 3AVE2.5337.45463.0370.8426.38719.255
L12.6
(2.6%)
7.411
(−0.6%)
60.157
(−4.6%)
0.835
(−0.8%)
6.426
(0.6%)
20.015
(3.9%)
AWF2.639
(4.2%)
7.41
(−0.6%)
61.285
(−2.8%)
0.842
(0.0%)
6.361
(−0.4%)
20.288
(5.4%)
CBF2.103
(−17.0%)
6.708
(−10.0%)
53.082
(−15.8%)
0.781
(−7.2%)
6.074
(−4.9%)
20.073
(4.2%)
GFF2.386
(−5.8%)
7.485
(0.4%)
62.147
(−1.4%)
0.805
(−4.4%)
6.619
(3.6%)
18.244
(−5.3%)
Image 4AVE1.0636.57439.0430.865.41414.011
L10.975
(−8.3%)
6.569
(−0.1%)
32.592
(−16.5%)
0.858
(−0.2%)
5.857
(8.2%)
15.28
(9.1%)
AWF1.121
(5.5%)
6.859
(4.3%)
40.939
(4.9%)
0.854
(−0.7%)
5.759
(6.4%)
12.976
(−7.4%)
CBF1.195
(12.4%)
5.794
(−11.9%)
18.555
(−52.5%)
0.815
(−5.2%)
5.383
(−0.6%)
27.403
(95.6%)
GFF1.092
(2.7%)
5.631
(−14.3%)
15.599
(−60.0%)
0.819
(−4.8%)
5.016
(−7.4%)
28.412
(102.8%)
Image 5AVE1.4516.39147.170.8055.71715.075
L11.621
(11.7%)
6.538
(2.3%)
56.349
(19.5%)
0.79
(−1.9%)
5.941
(3.9%)
11.539
(−23.5%)
AWF1.624
(11.9%)
6.595
(3.2%)
58.361
(23.7%)
0.799
(−0.7%)
5.736
(0.3%)
11.368
(−24.6%)
CBF1.347
(−7.2%)
6.258
(−2.1%)
25.079
(−46.8%)
0.771
(−4.2%)
5.492
(−3.9%)
23.133
(53.5%)
GFF1.642
(13.2%)
6.729
(5.3%)
28.733
(−39.1%)
0.777
(−3.5%)
5.549
(−2.9%)
18.601
(23.4%)
Image 6AVE1.4626.47444.2040.7515.60716.286
L11.556
(6.4%)
6.44
(−0.5%)
51.687
(16.9%)
0.749
(−0.3%)
5.676
(1.2%)
14.312
(−12.1%)
AWF1.474
(0.8%)
6.401
(−1.1%)
49.625
(12.3%)
0.755
(0.5%)
5.574
(−0.6%)
14.928
(−8.3%)
CBF1.679
(14.8%)
5.964
(−7.9%)
25.7
(−41.9%)
0.745
(−0.8%)
5.409
(−3.5%)
26.396
(62.1%)
GFF1.576
(7.8%)
6.807
(5.1%)
41.737
(−5.6%)
0.707
(−5.9%)
5.963
(6.3%)
16.149
(−0.8%)
Image 7AVE1.5056.23629.3770.8545.30516.729
L11.389
(−7.7%)
6.147
(−1.4%)
34.116
(16.1%)
0.849
(−0.6%)
5.563
(4.9%)
14.587
(−12.8%)
AWF1.564
(3.9%)
6.309
(1.2%)
31.24
(6.3%)
0.863
(1.1%)
5.532
(4.3%)
15.696
(−6.2%)
CBF1.92
(27.6%)
6.638
(6.4%)
57.972
(97.3%)
0.587
(−31.3%)
5.72
(7.8%)
8.832
(−47.2%)
GFF1.656
(10.0%)
6.312
(1.2%)
38.344
(30.5%)
0.808
(−5.4%)
5.496
(3.6%)
12.878
(−23.0%)
Image 8AVE1.4916.36887.1790.8025.8768.436
L11.377
(−7.6%)
6.462
(1.5%)
78.577
(−9.9%)
0.813
(1.4%)
5.796
(−1.4%)
9.98
(18.3%)
AWF1.42
(−4.8%)
6.389
(0.3%)
82.771
(−5.1%)
0.81
(1.0%)
5.577
(−5.1%)
9.245
(9.6%)
CBF1.217
(−18.4%)
5.943
(−6.7%)
40.586
(−53.4%)
0.803
(0.1%)
5.681
(−3.3%)
12.365
(46.6%)
GFF1.222
(−18.0%)
7.091
(11.4%)
65.273
(−25.1%)
0.789
(−1.6%)
6.153
(4.7%)
11.501
(36.3%)

References

  1. Liu, Z.; Zhang, Y.; Yu, X.; Yuan, C. Unmanned surface vehicles: An overview of developments and challenges. Annu. Rev. Control. 2016, 41, 71–93. [Google Scholar] [CrossRef]
  2. Huang, B.; Zhou, B.; Zhang, S.; Zhu, C. Adaptive prescribed performance tracking control for underactuated autonomous underwater vehicles with input quantization. Ocean. Eng. 2021, 221, 108549. [Google Scholar] [CrossRef]
  3. Campbell, S.; Naeem, W.; Irwin, G.; Campbell, S.; Naeem, W.; Irwin, G. A review on improving the autonomy of unmanned surface vehicles through intelligent collision avoidance manoeuvres. Annu. Rev. Control. 2012, 36, 267–283. [Google Scholar] [CrossRef] [Green Version]
  4. Ma, Z.; Wen, J.; Zhang, C.; Liu, Q.; Yan, D. An effective fusion defogging approach for single sea fog image. Neurocomputing 2016, 173, 1257–1267. [Google Scholar] [CrossRef]
  5. Ma, J.; Ma, Y.; Li, C. Infrared and visible image fusion methods and applications: A survey. Inf. Fusion 2019, 45, 153–178. [Google Scholar] [CrossRef]
  6. Singh, R.; Vatsa, M.; Noore, A. Integrated multilevel image fusion and match score fusion of visible and infrared face images for robust face recognition. Pattern Recognit. 2008, 41, 880–893. [Google Scholar] [CrossRef] [Green Version]
  7. Zhu, C.; Zeng, J.; Huang, B.; Su, Y.; Su, Z. Saturated approximation-free prescribed performance trajectory tracking control for autonomous marine surface vehicle. Ocean. Eng. 2021, 237, 109602. [Google Scholar] [CrossRef]
  8. Zhou, B.; Huang, B.; Su, Y.; Zheng, Y.; Zheng, S. Fixed-time neural network trajectory tracking control for underactuated surface vessels. Ocean. Eng. 2021, 236, 109416. [Google Scholar] [CrossRef]
  9. Kumar, P.; Mittal, A.; Kumar, P. Fusion of Thermal Infrared and Visible Spectrum Video for Robust Surveillance. In Computer Vision, Graphics and Image Processing; Springer: Berlin/Heidelberg, Germany, 2006; pp. 528–539. [Google Scholar]
  10. Simone, G.; Farina, A.; Morabito, F.C.; Serpico, S.B.; Bruzzone, L. Image fusion techniques for remote sensing ap-plications. Inf. Fusion 2002, 3, 3–15. [Google Scholar] [CrossRef] [Green Version]
  11. Ma, Z.; Wen, J.; Liang, X. Video Image Clarity Algorithm Research of USV Visual System under the Sea Fog. In Proceedings of the International Conference in Swarm Intelligence, Harbin, China, 12–15 June 2013; Springer Science and Business Media LLC: Berlin/Heidelberg, Germany, 2013; Volume 7929, pp. 436–444. [Google Scholar]
  12. Zabolotskikh, E.V.; Mitnik, L.; Chapron, B. New approach for severe marine weather study using satellite passive microwave sensing. Geophys. Res. Lett. 2013, 40, 3347–3350. [Google Scholar]
  13. Zhu, C.; Huang, B.; Zhou, B.; Su, Y.; Zhang, E. Adaptive model-parameter-free fault-tolerant trajectory tracking control for autonomous underwater vehicles. ISA Trans. 2021, 114, 57–71. [Google Scholar] [CrossRef] [PubMed]
  14. Zhang, H.; Wang, X.; Luo, X.; Xie, S.; Zhu, S. Unmanned surface vehicle adaptive decision model for changing weather. Int. J. Comput. Sci. Eng. 2021, 24, 18–26. [Google Scholar] [CrossRef]
  15. Liu, Y.; Jin, J.; Wang, Q.; Shen, Y.; Dong, X. Region level based multi-focus image fusion using quaternion wavelet and normalized cut. Signal. Process. 2014, 97, 9–30. [Google Scholar] [CrossRef]
  16. Toet, A. Image fusion by a ratio of low-pass pyramid. Pattern Recognit. Lett. 1989, 9, 245–253. [Google Scholar] [CrossRef]
  17. Choi, M.; Kim, R.Y.; Nam, M.-R.; Kim, H.O. Fusion of Multispectral and Panchromatic Satellite Images Using the Curvelet Transform. IEEE Geosci. Remote Sens. Lett. 2005, 2, 136–140. [Google Scholar] [CrossRef]
  18. Wang, J.; Peng, J.; Feng, X.; He, G.; Fan, J. Fusion method for infrared and visible images by using non-negative sparse representation. Infrared Phys. Technol. 2014, 67, 477–489. [Google Scholar] [CrossRef]
  19. Li, S.; Yin, H.; Fang, L. Group-sparse representation with dictionary learning for medical image denoising and fusion. IEEE Trans. Biomed. Eng. 2012, 59, 3450–3459. [Google Scholar] [CrossRef]
  20. Wang, Z.; Cui, Z.; Zhu, Y. Multi-modal medical image fusion by Laplacian pyramid and adaptive sparse representation. Comput. Biol. Med. 2020, 123, 103823. [Google Scholar] [CrossRef]
  21. Zhu, Z.; Yin, H.; Chai, Y.; Li, Y.; Qi, G. A novel multi-modality image fusion method based on image decomposition and sparse representation. Inf. Sci. 2018, 432, 516–529. [Google Scholar] [CrossRef]
  22. Xing, C.; Wang, M.; Dong, C.; Duan, C.; Wang, Z. Using Taylor Expansion and Convolutional Sparse Representation for Image Fusion. Neurocomputing 2020, 402, 437–455. [Google Scholar] [CrossRef]
  23. Wu, C.; Chen, L. Infrared and visible image fusion method of dual NSCT and PCNN. PLoS ONE 2020, 15, e0239535. [Google Scholar] [CrossRef] [PubMed]
  24. Zhong, Z.; Gao, W.; Khattak, A.M.; Wang, M. A novel multi-source image fusion method for pig-body multi-feature detection in NSCT domain. Multimed. Tools Appl. 2020, 79, 26225–26244. [Google Scholar] [CrossRef]
  25. Li, H.; Wu, X.-J. Multi-focus Image Fusion Using Dictionary Learning and Low-Rank Representation. In Proceedings of the International Conference on Image and Graphics, Shanghai, China, 13–15 September 2017; Springer Science and Business Media LLC: Berlin/Heidelberg, Germany, 2017; pp. 675–686. [Google Scholar]
  26. Liu, Y.; Chen, X.; Peng, H.; Wang, Z. Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion 2017, 36, 191–207. [Google Scholar] [CrossRef]
  27. Liu, Y.; Chen, X.; Ward, R.K.; Wang, Z.J. Image Fusion With Convolutional Sparse Representation. IEEE Signal. Process. Lett. 2016, 23, 1882–1886. [Google Scholar] [CrossRef]
  28. Li, H.; Wu, X.-J. DenseFuse: A Fusion Approach to Infrared and Visible Images. IEEE Trans. Image Process. 2019, 28, 2614–2623. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Ram Prabhakar, K.; Sai Srikar, V.; Venkatesh Babu, R. Deepfuse: A deep unsupervised approach for exposure fusion with extreme exposure image pairs. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4714–4722. [Google Scholar]
  30. An, G.H.; Lee, S.; Seo, M.-W.; Yun, K.; Cheong, W.-S.; Kang, S.-J. Charuco Board-Based Omnidirectional Camera Calibration Method. Electronics 2018, 7, 421. [Google Scholar] [CrossRef] [Green Version]
  31. Ch, M.M.I.; Riaz, M.M.; Iltaf, N.; Ghafoor, A.; Ahmad, A. Weighted image fusion using cross bilateral filter and non-subsampled contourlet transform. Multidimens. Syst. Signal. Process. 2019, 30, 2199–2210. [Google Scholar] [CrossRef]
  32. Hayat, N.; Imran, M. Ghost-free multi exposure image fusion technique using dense SIFT descriptor and guided filter. J. Vis. Commun. Image Represent. 2019, 62, 295–308. [Google Scholar] [CrossRef]
Figure 1. Calibration board.
Figure 1. Calibration board.
Jmse 10 00588 g001
Figure 2. (a) Visible image and (b) infrared image of the calibration board.
Figure 2. (a) Visible image and (b) infrared image of the calibration board.
Jmse 10 00588 g002
Figure 3. Image after registration.
Figure 3. Image after registration.
Jmse 10 00588 g003
Figure 4. Architecture of the fusion network [28].
Figure 4. Architecture of the fusion network [28].
Jmse 10 00588 g004
Figure 5. Framework of the training process [28].
Figure 5. Framework of the training process [28].
Jmse 10 00588 g005
Figure 6. Diagram of the fusion layer.
Figure 6. Diagram of the fusion layer.
Jmse 10 00588 g006
Figure 7. Bilinear interpolation representation.
Figure 7. Bilinear interpolation representation.
Jmse 10 00588 g007
Figure 8. Fusion result of adaptive weight strategy: (a) visible image, (b) infrared image, and (c) fused image.
Figure 8. Fusion result of adaptive weight strategy: (a) visible image, (b) infrared image, and (c) fused image.
Jmse 10 00588 g008
Figure 9. Fusion result of cross bilateral filtering: (a) visible image, (b) infrared image, and (c) fused image.
Figure 9. Fusion result of cross bilateral filtering: (a) visible image, (b) infrared image, and (c) fused image.
Jmse 10 00588 g009aJmse 10 00588 g009b
Figure 10. Fusion result of guided filtering: (a) visible image, (b) infrared image, and (c) fusion image.
Figure 10. Fusion result of guided filtering: (a) visible image, (b) infrared image, and (c) fusion image.
Jmse 10 00588 g010
Figure 11. ‘Tian Xing’ USV.
Figure 11. ‘Tian Xing’ USV.
Jmse 10 00588 g011
Figure 12. Fusion results of different methods: (a) AVE; (b) L1; (c) AWF; (d) CBF; (e) GFF.
Figure 12. Fusion results of different methods: (a) AVE; (b) L1; (c) AWF; (d) CBF; (e) GFF.
Jmse 10 00588 g012
Figure 13. Eight pairs of source images. (a) buoy; (b) lighthouse; (cg) ship; (h) coast. The first and third rows contain visible images, and the second and fourth rows contain infrared images.
Figure 13. Eight pairs of source images. (a) buoy; (b) lighthouse; (cg) ship; (h) coast. The first and third rows contain visible images, and the second and fourth rows contain infrared images.
Jmse 10 00588 g013
Figure 14. Experiment on sea images containing typical objects: (a) AVE; (b) L1; (c) AWF; (d) CBF; (e) GFF.
Figure 14. Experiment on sea images containing typical objects: (a) AVE; (b) L1; (c) AWF; (d) CBF; (e) GFF.
Jmse 10 00588 g014aJmse 10 00588 g014b
Figure 15. Average gradient line chart of fused images in each group.
Figure 15. Average gradient line chart of fused images in each group.
Jmse 10 00588 g015
Table 1. Architecture of the training process. Conv, convolutional block consisting of a convolutional layer and an activation layer; Dense, DenseBlock.
Table 1. Architecture of the training process. Conv, convolutional block consisting of a convolutional layer and an activation layer; Dense, DenseBlock.
LayerSizeStrideChannel
(Input)
Channel
(Output)
Activation
EncoderConv (C1)31116ReLU
Dense
DecoderConv (C2)316464ReLU
Conv (C3)316432ReLU
Conv (C4)313216ReLU
Conv (C5)31161ReLU
Dense
(DenseBlock)
Conv (DC1)311616ReLU
Conv (DC1)313216ReLU
Conv (DC1)314816ReLU
Table 2. Average gradient of fusion results.
Table 2. Average gradient of fusion results.
AVEL1AWFCBFGFF
Average gradient 1.0630.9571.1211.1951.092
Percentage change for AVE--+5.5%+12.4%+2.7%
Percentage change for L1--+14.9%+22.6%+12%
Table 3. Average gradient of results based on different fusion strategies.
Table 3. Average gradient of results based on different fusion strategies.
ImageAverage Gradient of Results
(Percentage Change)
AVEL1AWFCBFGFF
Figure 13a1.6181.985
(22.7%)
1.995
(23.3%)
1.452
(−10.3%)
1.91
(18.0%)
Figure 13b2.1632.321
(7.3%)
2.349
(8.6%)
2.029
(−6.2%)
2.514
(16.2%)
Figure 13c2.5332.6
(2.6%)
2.639
(4.2%)
2.103
(−17.0%)
2.386
(−5.8%)
Figure 13d1.4511.621
(11.7%)
1.624
(11.9%)
1.347
(−7.2%)
1.642
(13.2%)
Figure 13e1.0630.975
(−8.3%)
1.121
(5.5%)
1.195
(12.4%)
1.092
(2.7%)
Figure 13f1.4621.556
(6.4%)
1.474
(0.8%)
1.679
(14.8%)
1.576
(7.8%)
Figure 13g1.5051.389
(−7.7%)
1.564
(3.9%)
1.92
(27.6%)
1.656
(10.0%)
Figure 13h1.4911.377
(−7.6%)
1.42
(−4.8%)
1.217
(−18.4%)
1.222
(−18.0%)
Average--
(3.4%)
-
(6.7%)
-
(−0.5%)
-
(5.5%)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, R.; Su, Y.; Li, Y.; Zhang, L.; Feng, J. Infrared and Visible Image Fusion Methods for Unmanned Surface Vessels with Marine Applications. J. Mar. Sci. Eng. 2022, 10, 588. https://0-doi-org.brum.beds.ac.uk/10.3390/jmse10050588

AMA Style

Zhang R, Su Y, Li Y, Zhang L, Feng J. Infrared and Visible Image Fusion Methods for Unmanned Surface Vessels with Marine Applications. Journal of Marine Science and Engineering. 2022; 10(5):588. https://0-doi-org.brum.beds.ac.uk/10.3390/jmse10050588

Chicago/Turabian Style

Zhang, Renran, Yumin Su, Yifan Li, Lei Zhang, and Jiaxiang Feng. 2022. "Infrared and Visible Image Fusion Methods for Unmanned Surface Vessels with Marine Applications" Journal of Marine Science and Engineering 10, no. 5: 588. https://0-doi-org.brum.beds.ac.uk/10.3390/jmse10050588

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop