Next Article in Journal
Discrimination of Biological Scatterers in Polarimetric Weather Radar Data: Opportunities and Challenges
Previous Article in Journal
GNSS/INS Integration with Integrity Monitoring for UAV No-fly Zone Management
Previous Article in Special Issue
Correction: Wang, L., et al. Assessment of the Dual Polarimetric Sentinel-1A Data for Forest Fuel Moisture Content Estimation. Remote Sensing 2019, 11(13), 1568
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image Similarity Metrics Suitable for Infrared Video Stabilization during Active Wildfire Monitoring: A Comparative Analysis

1
Centre for Technological Risk Studies, Universitat Politècnica de Catalunya, 08034 Barcelona, Spain
2
IDLab, Ghent University – imec, 9502 Ghent, Belgium
3
Missoula Fire Sciences Lab, US Forest Service Rocky Mountain Research Station, Missoula, MT 59808, USA
4
National Center for Landscape Fire Analysis, University of Montana, Missoula, MT 59812, USA
*
Author to whom correspondence should be addressed.
Submission received: 17 December 2019 / Revised: 13 January 2020 / Accepted: 21 January 2020 / Published: 6 February 2020
(This article belongs to the Special Issue Remote Sensing and Image Processing for Fire Science and Management)

Abstract

:
Aerial Thermal Infrared (TIR) imagery has demonstrated tremendous potential to monitor active forest fires and acquire detailed information about fire behavior. However, aerial video is usually unstable and requires inter-frame registration before further processing. Measurement of image misalignment is an essential operation for video stabilization. Misalignment can usually be estimated through image similarity, although image similarity metrics are also sensitive to other factors such as changes in the scene and lighting conditions. Therefore, this article presents a thorough analysis of image similarity measurement techniques useful for inter-frame registration in wildfire thermal video. Image similarity metrics most commonly and successfully employed in other fields were surveyed, adapted, benchmarked and compared. We investigated their response to different camera movement components as well as recording frequency and natural variations in fire, background and ambient conditions. The study was conducted in real video from six fire experimental scenarios, ranging from laboratory tests to large-scale controlled burns. Both Global and Local Sensitivity Analyses (GSA and LSA, respectively) were performed using state-of-the-art techniques. Based on the obtained results, two different similarity metrics are proposed to satisfy two different needs. A normalized version of Mutual Information is recommended as cost function during registration, whereas 2D correlation performed the best as quality control metric after registration. These results provide a sound basis for image alignment measurement and open the door to further developments in image registration, motion estimation and video stabilization for aerial monitoring of active wildland fires.

1. Introduction

Forest fires have been studied through remote sensing techniques for decades. A number of spaceborne sensors have successfully been used to analyze various fire aspects and post-fire effects [1]. Existing applications include the detection of active fires [2,3,4], burned area measurement [5,6,7,8,9], sensing of radiated energy [10,11] and the estimation of pyrogenic gas emissions [12,13], among others. Similarly, airborne imaging systems are being increasingly employed to gain detailed insight into fire behavior variables such as fire rate of spread, fire line intensity and fire radiative power [14,15,16,17,18,19,20,21,22]. Unmanned and remotely piloted aircraft further simplify sensor deployment while significantly reducing operation costs and risk [23,24,25].
Although a few successful experiences have been reported that use airborne monitoring systems in large-scale wildfires [17,26], the majority of developments in fire detection and monitoring occur via sensing prescribed fires, which are often restricted in areal extent as well as fire line radiative intensity [27]. In these cases, the remote sensor is usually placed in a fixed position or a hovering aircraft and it is deployed to collect high spatial resolution images with a moderate temporal resolution for the full duration of flaming combustion [28,29,30]. This type of deployment rarely follows a large-area mapping mission profile where parallel and overlapping flight lines are required. Nevertheless, turbulence from the fire often results in significant roll, pitch and yaw variations that are hard to cancel with mechanical stabilization systems only. Given camera motion during the acquisition, image registration and rectification are required before spatial inference can be completed (for example, to measure fire residence time per pixel or rate of spread).
Within sensor types suitable for wildfire monitoring, optical cameras working in the thermal infrared (TIR) range are widely applied to characterize active fire behavior due to their high availability and versatility [11,15,17,31,32]. Airborne TIR cameras allow measuring fire geometry and radiated energy with high spatial and moderate temporal resolution even in the presence of smoke. Due to these advantages, several TIR image processing algorithms have been developed for automated computation of fire behavior metrics [15,20,21]. Automated methodologies allow not only faster but also more rigorous quantitative studies by removing bias and ensuring a systematic analysis framework. In order to draw meaningful conclusions, fire behavior metrics must be measured explicitly in time and space, avoiding long-term and wide-field average values whenever possible.
However, a number of limitations remain in the automated processing of fire thermal infrared imagery. Among existing needs, image registration tools that allow camera motion estimation and cancellation are in high demand [24,33]. The current approach used to georeference aerial TIR fire imagery is based on the manual annotation of ground control points in every video frame [15,18,21,34]. This methodology is not only very time consuming but also prone to errors and hard to implement operationally. Important difficulties with image georeferencing have been reported in previous studies, sometimes resulting in loss of data [18,24]. Because of the highly variable energy emitted by the fire and the dynamic radiance range used by many cameras, background objects are sometimes not well resolved. This fact prevents the identification of GPSed ground control points in the video. Even when it is successfully performed, manual georeferencing has been identified as one of the most significant sources of uncertainty in the study of wildfire behavior from aerial TIR imagery [22]. These limitations seriously restrict the amount of quantitative information obtainable through remote sensing as well as its quality. Consequently, the achievement of accurate automated image georeferencing is a high priority necessity for wildfire science.
A fundamental operation during image registration is the measurement of image similarity, usually with the ultimate goal of maximizing such measure. There are three major types of dissimilarities that can be observed when comparing two or more images [35]. The first type is misalignment, which appears due to variations in the position of the acquisition sensor. These variations are relatively easy to model as geometric transformations. Frequently, prior knowledge of the scene determines the class of transformations to be explored, and this selection in turn determines the most suitable registration method. The second type of image dissimilarities are generated by variations in external conditions during image acquisition. Contrary to first-type image differences, second-type dissimilarities are frequently not easy to model. Lighting and atmospheric conditions, among others, fall within this category. Finally, the third type of image differences are caused by changes in the scene itself. The observer is usually interested in these changes, which include movement of the objects under study, among others.
The main objective of image registration techniques is to find the correct spatial transformation that cancels the first type of image dissimilarities, without being affected by variations of the second type. By doing this, the third type of image differences are then easier to analyze. Differences of types two and three are not cancelled by registration methods but constitute a challenge for them because they prevent an exact match between images that must be compared. In the wildfire context, the movement of a drone operating the camera constitutes an example of type-one variations. Conversely, differences in lighting conditions and smoke concentration between camera and fire fall within the second category because they produce variations in the imaging scenario and they can affect image processing algorithms. Finally, changes in the fire itself, which is dynamic, are the best example of type-three image dissimilarities.
In contrast with visible imagery, fire thermal infrared video entails several challenges that have so far prevented the achievement of automated image registration. Fire monitoring requires high measurement ranges for brightness temperature, usually starting over 200 °C. This fact diminishes the amount of detail distinguishable in the cold background. Moreover, fire usually occupies a large portion of the camera field of view. Because fire is dynamic, this fact significantly hinders the identification of persistent features between images acquired at different times.
This article analyzes the problem of image similarity measurement in the context of forest fire aerial remote sensing, specifically focusing on TIR imagery of active fires acquired from a vantage point and an oblique perspective. The ultimate goal of this study is the identification of image similarity metrics suitable for inter-frame registration and video stabilization. State-of-the-art methodologies used to measure image similarity in other fields were surveyed and benchmarked. Tested methods include metrics based on gray value difference, gray value correlation and information theory. Metrics were assessed based on their ability to meet two specific needs: on the one hand, a well-behaving cost function is needed during registration; on the other, a robust estimator of absolute image alignment is required after registration for quality control.

2. Background: Image Similarity Metrics

The most popular approach to measure image similarity has historically been based on gray difference statistical metrics such as intensity mean squared difference and two-dimensional correlation [35,36,37]. Cross-correlation has been used for image registration during decades and it is still in use in several applications, including remote sensing [38,39,40]. Recently the use of direct gray difference measurements has decayed in favor of more powerful metrics based on information theory, such as Mutual Information (MI) [39,41,42]. Translating gray level values into the more general measure of information content provides enormous flexibility. For this reason, MI is vastly employed not only for image similarity measurement but also to fuse multispectral [39,43,44] and multi-modal [45,46,47] information. However, this flexibility entails an elevated computational cost that may become prohibitive under certain circumstances. Speed requirements (e.g., for real-time processing) and hardware limitations (e.g., for deployment aboard satellites or unmanned aircraft) usually motivate the use of low-complexity algorithms [48,49,50]. In addition to this wide variety of advantages and drawbacks, the performance of each methodology varies significantly with the field of application. Fire IR imagery presents important singularities with respect to other remote sensing and computer vision scenarios. In order to address this issue, we analyzed the suitability for fire monitoring of some of the most widely employed image similarity metrics.

2.1. Intensity 2D Correlation

Cross-correlation and the two-dimensional correlation coefficient are statistical indices widely used in image registration [35]. The 2D correlation coefficient between two images c o r r 2 D ( I 1 , I 2 ) (Equation (1)) provides a scalar measurement of their global similarity, whereas cross-correlation C ( u , v ) measures the degree of similarity between a reference image I and a template T shifted u and v pixels in the x and y direction, respectively (Equation (2)). Cross-correlation is frequently used for template matching and pattern recognition.
c o r r 2 D ( I 1 , I 2 ) = c o v a r i a n c e ( I 1 , I 2 ) σ 1 σ 2 = i j ( I 1 ( i , j ) μ 1 ) ( I 2 ( i , j ) μ 2 ) i j I 1 ( i , j ) μ 1 ) 2 I 2 ( i , j ) μ 2 ) 2
C ( u , v ) = x y T ( x , y ) I ( x u , y v ) x y I 2 ( x u , y v )
In Equation (1), μ i and σ i represent, respectively, the gray value average and standard deviation within each image.
The 2D correlation coefficient has two important advantages. First, it provides a similarity measurement in the fixed range [−1, 1]. Secondly, it shows a linear relationship with image similarity under certain statistical assumptions [35]. Both properties are particularly useful in an image registration scheme because they allow an absolute assessment of the achieved registration quality. This way, the estimated registration transformation can be accompanied by a confidence assessment. A correlation coefficient of 1 represents a perfect match, achieved when identical images are perfectly aligned.

2.2. Intensity Mean Squared Difference (IMSD)

Intensity mean squared difference (Equation (3)) falls within the group of the simplest metrics useful to measure dissimilarity between two variables. It is simple and computationally efficient, and it also provides an absolute similarity estimation. In this case, perfect match is denoted by an IMSD value of 0.
IMSD ( I 1 , I 2 ) = 1 N · M i = 1 N j = 1 M ( I 1 ( i , j ) I 2 ( i , j ) ) 2
In Equation (3), N and M indicate the number of rows and columns in images I 1 and I 2 , which must both be the same size.

2.3. Mutual Information

Mutual Information (MI) quantifies the dependence between two variables, more specifically the amount of information that one variable contains about the other [51]. This definition allows for a criterion frequently used in image registration problems, which states that two images are geometrically aligned when MI between the intensity values of corresponding pixels -or voxels- is maximal [52].
Image similarity metrics based on Mutual Information were first proposed by Viola [53] and Collignon et al. [54]. Since then, they have been extensively used in the field of medical imaging [41,46,52,55] and, more recently, in remote sensing [39,42,44,56].
If A and B are two random variables with marginal probability distributions p A ( a ) and p B ( b ) , the Mutual Information between them I ( A , B ) measures the distance between their joint distribution p A B ( a , b ) and the joint distribution they would have if they were completely independent, p A ( a ) · p B ( b ) . This distance represents the degree of dependence between A and B and it is usually computed using the Kullback–Leibler measure (Equation (4)).
I ( A , B ) = a , b p A B ( a , b ) l o g p A B ( a , b ) p A ( a ) · p B ( b )
Alternatively, MI can be defined using the concept of image entropy. Entropy (H) measures the uncertainty of a random variable. It is widely used in information theory and its most common mathematical definition was proposed by Shannon [57] (Equation (5)).
H = i p i l o g p i
It can be demonstrated [41] that Equation (4) can be rewritten as Equations (6)–(8) introducing Shannon entropy:
I ( A , B ) = H ( A ) + H ( B ) H ( A , B )
= H ( A ) H ( A B )
= H ( B ) H ( B A )
where H ( A ) and H ( B ) are the entropy of images A and B, respectively, H ( A , B ) is their joint entropy, H ( A B ) is the conditional entropy of A given B and H ( B A ) is the conditional entropy of B given A. H ( A ) measures the uncertainty of A, whereas H ( A B ) represents the amount of uncertainty left in A when knowing B. Consequently, I ( A , B ) can be understood as the reduction in uncertainty of A caused by the knowledge of B. In other words, I ( A , B ) represents the amount of information that B contains about A.
Mutual Information is very powerful because it allows a general comparison of two images without assumptions about their nature or the nature of their relation, and with no need for prior segmentation. This provides the additional capability of comparing images of a different nature, a property that has been exploited for image fusion purposes [58,59]. Furthermore, MI presents some significant advantages over metrics based on cross-correlation, which are affected by changes in lighting conditions and reflectance dependence on wavelength [39,42].
Despite these good properties, there are also some drawbacks related to Mutual Information. The high computational cost of MI computation, together with interpolation artefacts and the relatively high amount of noise present in the MI surface and its derivatives when it is undersampled [39], hinder convergence of MI-based registration algorithms. Additionally, the MI registration function may contain local maxima, which can result in misregistration [60,61].
Furthermore, the original MI formulation presents two significant limitations when used as image similarity metric alone, decoupled from the image registration framework. On the one hand, MI value is sensitive to the amount of overlap between compared images [41,62]. On the other hand, it does not provide an absolute measurement of how well two images are aligned. MI can estimate relative agreement between two images: the more similar two images are, the greater their MI value. However, this MI value cannot be directly compared against an absolute similarity scale and identical images do not always reach the same MI value. The main consequence of this is that the quality of a certain registration algorithm cannot be absolutely measured using MI, and therefore there is no means of verifying whether the achieved optimum MI is acceptable.

2.4. Normalized Mutual Information (NMI)

Several improvements have been proposed to overcome MI limitations. Studholme et al. [62] suggested a revised version of MI, invariant to image overlap (Equation (9)). Although the original authors called this revision normalized, its value is in fact not comprised in the interval 0 , 1 [59]. Therefore, we refer to it as Studholme’s Mutual Information (SMI).
SMI ( A , B ) = H ( A ) + H ( B ) H ( A , B )
Similar to SMI is the so-called entropy correlation coefficient (ECC, Equation (10)), first proposed by Astola and Virtanen [63] and tested by Maes et al. [52]. The behavior of ECC is similar to SMI by definition [41,61] and Maes et al. [52] did not find a clear difference in performance when compared with original MI.
ECC ( A , B ) = 2 I ( A , B ) H ( A ) + H ( B )
In this paper, we propose an alternative MI formulation that provides actual normalization and, consequently, an absolute measurement of image similarity. We assessed the performance of the Normalized Mutual Information (NMI) as defined in Equation (11). This formulation had previously been used in machine learning algorithms [64] but it had barely received attention in image analysis problems, being Bai et al. [65] and Pillai and Vatsavai [66] the only two exceptions we found. This formulation is based on the fact that H ( X ) = I ( X , X ) and constitutes an analogy with a normalized inner product in Hilbert space. NMI is symmetric and restricted to the range [0,1] by definition.
NMI ( A , B ) = I ( A , B ) H ( A ) H ( B )
Estévez et al. [67] took a similar approach when comparing a feature set F with a subset of itself S. They defined the normalized mutual information between f i F and f s S by dividing their MI by the minimum entropy of both sets (Equation (12)):
NMI Estevez ( f i ; f s ) = I ( f i ; f s ) m i n H ( f i ) , H ( f s )
NMI Estevez is also symmetric and takes values in [0,1]. However, its formulation is less convenient for our image registration problem. NMI Estevez similarity values may change abruptly when modifying the reference frame, and consequently the NMI Estevez distribution is subject to discontinuities along a video sequence. Updating the reference frame is especially important during wildfire video stabilization to account for fire evolution. Therefore, we suggest using the NMI formulation shown in Equation (11).

3. Methodology

Similarity metrics described in Section 2 were subject to Global and Local Sensitivity Analyses (GSA and LSA, respectively) in order to assess their response to different variables of interest such as camera movement, video temporal resolution and natural variations in fire, background and ambient conditions. GSA was conducted first for general screening of significant variable relationships. Subsequently, metric sensitivity was studied locally in the region of interest where similarity must be measured. This local region consists of a limited range of camera translations, rotations and scaling around the nominal recording point of zero perturbation. Details about the GSA and LSA workflows are provided in Section 3.3 and Section 3.4, respectively.
These analyses were performed on six TIR video sequences of active spreading fire recorded during laboratory and field experiments. All cameras were installed at fixed elevated vantage points so that the obtained video was stable. Recorded TIR video was used as reference and frames were perturbed through synthetic image translations, rotations and changes in scale that produced virtual camera movement. Individual experiment, camera and footage details are described in Section 3.1. Due to its higher computational cost, GSA was limited to four video sequences representative of different experimental setups, whereas LSA was performed for all available footage. Figure 1 summarizes the followed methodology.

3.1. Test Data

GSA and LSA studies were applied to six experimental scenarios ranging from small laboratory tests to large-scale field experiments. In all cases, fire evolution was recorded from fixed vantage points using thermal infrared cameras. Employed cameras and setups varied, but all video sequences were stable. The resulting dataset allowed a systematic study under controlled, yet dissimilar, conditions.
Scenario 1 was recorded in the Centre for Technological Risk Studies at Universitat Politècnica de Catalunya. A homogeneous bed of straw was burned on a 1.5 m × 3 m combustion table to reproduce fire spread on a flat horizontal surface with no wind. Scenarios 2 and 3 were recorded at the Tall Timbers Research Station in Tallahassee, FL, USA, in April 2017. These video sequences were acquired during a set of small-scale experimental burns on mixed rough/long leaf pine fuels. Scenarios 4–6 were monitored during one of the most complete large-scale experimental campaigns conducted so far, RxCADRE [29]. Video sequences 4, 5 and 6 correspond to plots S3, S4 and S5 of this experiment, respectively. These three plots were recorded with a high-resolution IR camera mounted on a boom lift [68]. Burned vegetation was a mix of grass and shrubs, predominantly turkey oak. Figure 2 shows sample frames from all scenarios, while Table 1 summarizes the technical details of the employed thermal cameras. Figure 3 provides additional information about the experimental setups.
Crown fire did not occur in any of the presented scenarios. Still, this dataset can be considered representative of the typical use of thermal IR cameras in forest fire research.

3.2. Approach Overview

Sensitivity analysis is defined as the study of how uncertainty in the output of a model can be apportioned to different sources of uncertainty in the model input [69]. In other words, it attempts to quantify the effect that deviations in input parameters have on model outputs. In modeling literature, this has traditionally been achieved in practice through the estimation of partial derivatives of a particular model output versus a particular input. Because such derivatives must be estimated locally at a point of interest, this approach is usually referred to as local sensitivity analysis (LSA).
While LSA may provide valuable insight into model behavior, the scientific discipline of sensitivity analysis has evolved towards more general terms. Statistical studies, risk analysis and reliability assessments—just to name a few—require a broader approach in which the influence of factors on outputs is studied along the entire input space. Consequently, modern global sensitivity analyses are often based on Monte Carlo space sampling and the computation of statistical measures [70].
To apply sensitivity analysis techniques to the study of image similarity measurement, one can understand metrics as models that allow computing a certain output of interest—i.e., similarity—from a set of inputs—i.e., two images. Furthermore, one can disaggregate the actual input parameters into simpler components. In the specific case of active fire video stabilization, the two images to be compared will typically be extracted from the same video sequence by sampling the video at a given frequency. Time elapsed between both frames, which is related to sampling frequency, can significantly affect image similarity, mostly due to fire dynamics. Moreover, we can assume the second image misaligned with respect to the first one if the camera moved between both acquisitions. Such misalignment can be expressed in terms of a two-dimensional relative translation, a rotation angle and a scale coefficient—if only affine transformations are considered. Beyond sampling frequency and misalignment, similarity values computed between the first frame and the second may vary greatly with the state of the scene. Among others, the portion of the field of view that is covered by fire and the flaming intensity can importantly affect similarity metric behavior. In this study, we grouped all factors contributing to absolute changes in the state of the scene under the variable time. time represents the time at which the first frame was acquired, measured with respect to whichever time reference is selected—typically, the start of the video sequence.
Therefore, the similarity value provided by any of the investigated metrics can be understood as a function of sampling frequency (f), time (t) and geometric misalignments—relative translations ( T x , T y ), rotation ( θ ) and scaling ( s c ) (Equation (13)). This function is our model, and we studied the sensitivity of its only output—similarity—to each of its inputs, which were assumed independent of each other. The explored input space is summarized in Table 2.
Similarity ( I 1 , I 2 ) = F ( f , t , T x , T y , θ , s c )

3.3. Global Sensitivity Analysis

Among existing GSA methods, the most powerful approach consists of estimating the conditional variance of model outputs with respect to each of its inputs. This strategy, named Variance-Based Sensitivity Analysis (VBSA), allows not only understanding model sensitivity but also quantifying it. Saltelli et al. [70] defined conditional variance of a model output Y as its variance when one of the inputs X i is fixed to a specific value x i . This can be represented as V X i Y | X i = x i , where V X i indicates that the resulting variance is taken over all factors but X i .
This definition of conditional variance leads to the formulation of two types of sensitivity indices widely used in GSA: first-order and total sensitivity indices. The first-order sensitivity index of X i on Y, S i (Equation (14)), measures the variance produced in Y when only X i is modified. This effect is averaged over all possible values of X i to provide a general measure not limited to a specific point in the input space. The reader is referred to Saltelli et al. [70] for the complete mathematical derivation of this metric.
S i = V X i E X i Y | X i V ( Y )
First-order sensitivity indices do not take into consideration potential input interactions, which may be relevant in non-linear models. In order to account for them, higher-order indices can be defined analogously to Equation (14). In practice, although higher-order indices can be estimated, their detailed computation has an important drawback. The amount of sensitivity indices increases exponentially with the number of inputs. Specifically, a system with k parameters will have 2 k 1 indices including first-order and higher-order terms. The computation of all terms is usually impractical, especially because this detailed information can be replaced with an indirect measurement of higher-order effects through the so-called total effects.
The total effect of factor X i is defined as the sum of all terms of any order that include X i . In other words, S T i encompasses all possible contributions of X i —both direct and indirect—to the output variance. This can be expressed through Equation (15):
S T i = 1 V E Y | X X i V ( Y ) = E V Y | X X i V ( Y )
where V E Y | X X i V ( Y ) combines all terms of any order that do not include factor X i .
According to Saltelli et al. [70] (also stated in Saltelli et al. [69] and references therein), the computation of all first-order indices and total effects of a model provides sufficient characterization of its sensitivity pattern while keeping computational cost acceptable in most cases.
Still, estimation of variances and expected values requires a high amount of model runs for a meaningful sample of the input space. Such computation may become unfeasible for complex models, which has motivated the development of alternative methods to gain insight into model sensitivity. Example algorithms developed to find approximate sensitivity information include the Elementary Effect Test, Monte Carlo Filtering and the Fourier Amplitude Sensitivity Test [71].
Due to the manageable cost of computing similarity between two images, VBSA could be applied in this study, although it was limited to four TIR sequences, namely scenarios 1, 2, 4 and 5. These 4 scenarios were considered representative of different fire scales and conditions.
Our implementation followed recommendations given by Saltelli et al. [70]. These authors claim to provide the best algorithm available today to compute first-order and total-effect indices purely from model evaluations. Their method builds on the original approach proposed by Sobol [72] and is based on the following steps:
  • Generate a sample of the model input space of size 2 N . This can be accomplished through random sampling or using sequences of quasi-random numbers. The latter approach allows a significant reduction on the sample size necessary to achieve convergence in estimated statistics.
  • Split the input sample into two groups. The result will be two matrices of size N × M , where M is the number of model inputs. We call these matrices A and B.
  • Create a third matrix C by combining columns from A and B. Specifically, C will be a vertical concatenation of M submatrices C i , where each C i is composed of all columns of B except the ith column, which is taken from A.
  • Run the model for each sample in matrices A, B and C, thus obtaining output vectors Y A , Y B and Y C .
  • Compute first-order ( S i ) and total-effect ( S T i ) sensitivity indices defined in Equations (14) and (15). S i and S T i can be computed from vectors Y A , Y B and Y C using Equations (16) and (17), respectively.
S i = V E Y | X i V ( Y ) = Y A · Y C i f 0 2 Y A · Y A f 0 2 = 1 N j = 1 N y A j y C i j f 0 2 1 N j = 1 N y A j 2 f 0 2
S T i = 1 V E Y | X i V ( Y ) = 1 Y B · Y C i f 0 2 Y A · Y A f 0 2 = 1 1 N j = 1 N y B j y C i j f 0 2 1 N j = 1 N y A j 2 f 0 2
In Equations (16) and (17), y A j , y B j and y C i j represent the jth element of vectors Y A , Y B and Y C i , respectively, and f 0 is the mean of Y A elements (Equation (18)).
f 0 = 1 N j = 1 N y A j
Input space samples were generated using Latin Hypercube Sampling (LHS) [73] in the parameter ranges indicated in Table 2. Sample size was 10 6 in all scenarios and the probability distribution was considered uniform for all inputs.In addition to computing main and total effects for each similarity index, bootstrapping was used to estimate confidence intervals for these sensitivity indices. 500 subsamples were used in all scenarios for bootstrapping. Finally, index convergence was assessed by sequentially increasing the number of model runs used to compute S i and S T i . The practical implementation of this method was conducted with help of the MATLAB toolbox provided by Pianosi et al. [74].

3.4. Local Sensitivity Analysis

In addition to GSA, local sensitivity analysis was conducted to gain further insight about metric performance around their nominal point of operation. Whereas GSA allowed the general assessment of metric behavior throughout the complete parameter input space, LSA facilitated a more detailed analysis of their application in practice. Requirements for image similarity measurement are not identical during and after registration of consecutive video frames. For the former, the metric must be robust and allow finding the point of maximum similarity. For the latter, the metric must provide a reliable absolute similarity estimation. Both applications were studied by means of local sensitivity analysis around the point of perfect alignment.
To achieve this, stable video sequences were sampled at an approximate frequency of 1 Hz, which was considered representative of typical operation conditions in a wildfire scenario. Each sampled frame was perturbed systematically through affine geometric transformations. Horizontal translations, vertical translations, rotations and scaling were applied separately. Their intensity varied sequentially within the ranges indicated in Table 2 in steps of 1% for translations, 1deg for rotations and 1% for scaling.
Once perturbed, each frame was compared to its original (i.e., stable) version to assess metric response to each movement component when the effect of all other factors—including scene variations and sampling frequency—was blocked. This approach was named idealized operation conditions. Additionally, scene variations and time were considered by comparing each perturbed frame with the previous sampled frame. We refer to this approach as realistic operation conditions.

4. Results

This section summarizes and discusses the main results of this study. For the sake of readability, only a reduced subset with the most important results is included here. The interested reader can find a comprehensive compilation of all produced data for each considered scenario in Supplementary Materials.
Variance-based global sensitivity analysis techniques described in Section 3.3 were used to assess the general response of various image similarity metrics to six variables of interest, namely: horizontal translation, vertical translation, rotation, scale, time and sampling frequency. The first four parameters represent the camera movement to be detected, whereas time and frequency account for additional sources of image differences which may affect metric performance. An ideal image similarity measure should be highly sensitive to camera movement and robust in the presence of recording frequency variations and image content differences appearing over time.

4.1. GSA Convergence Considerations

Metric sensitivity was assessed using Main Effect (ME) and Total Effect (TE) indices as defined in Equations (14) and (15), respectively. ME and TE were estimated through Equations (16) and (17), with N = 10 6 LHS samples of the input space. This sample size was enough to achieve index convergence in all studied cases (see convergence results in Supplementary Materials).
However, ME and TE approximations converged to meaningless values in some cases (see Figure 4 for an example). This occurred when output similarity values did not follow a standard normal distribution, because y values used in Equations (16) and (17) are expected to follow a standard normal distribution. Although output probability was approximately normally distributed in all cases, each similarity metric uses a different optimum value for perfect image match. As a result, output distributions provided by some similarity metrics were displaced with respect to the standard normal distribution.
This limitation was solved by centering the y distributions provided by similarity metrics. Centering was achieved using Equation (19),
y c e n t r e d = ( y y ¯ ) / σ y
where y ¯ represents the average of y and σ y , its standard deviation. The application of Equations (16) and (17) to the centered distributions allowed the correct estimation of ME and TE. Figure 4 demonstrates the effect of centering y distributions in a sample case. Verifying that model outputs approximately follow a standard normal distribution is essential to ensure correct estimation of variance-based sensitivity indices. However, common GSA libraries, including the code provided by Pianosi et al. [74], do not usually include the centering step by default. We therefore suggest paying special attention to this aspect and double check model output distributions obtained through Multi-Carlo sampling before proceeding further.

4.2. GSA Results

Average GSA results—obtained after solving issues with y distributions—are displayed in Figure 5. According to them, all metrics had a similar response to frequency, whereas the strongest variation with time corresponds to IMSD and MI. On the contrary, 2D correlation showed the strongest response to all movement components while being less affected by time than IMSD and MI. NMI showed a better performance than MI, closely following 2D correlation.
One important conclusion that can be drawn from Figure 5 is the existence of important interactions among the six studied variables. While main effects account for the variance increase observed in model outputs when a single input parameter is varied, total effects include the effect produced by one parameter when the other input parameters are also allowed to vary. The fact that TE are significantly higher than ME demonstrates that individual parameter contributions to output variance are boosted when several input variables vary simultaneously.
The phenomenon of coupled response is especially important for 2D correlation, which could otherwise have been awarded the first place in this comparative study. High sensitivity to camera movement is ideal only if the source of image dissimilarity can be identified correctly. Conversely, there is little use in knowing that two images are different if the cause of this difference cannot be attributed to a single parameter. This fact reinforced the need for a local sensitivity analysis to gain further insight into model response to each individual parameter around the nominal point of operation.

4.3. LSA Results

LSA was first used to assess existing differences in behavior between MI, SMI and NMI. Although their absolute values vary, all three metrics behaved similarly with camera movement (Figure 6). On the contrary, there was an important difference in their response to image content. Figure 7 displays metric behavior under idealized conditions in 3D. This representation highlights a significant variation of MI values with time. Considering the nature of fire TIR images, these differences are presumably due to the fact that the majority of the image entropy is provided by the fire. Consequently, image entropy increases as the image portion filled with fire grows over time. In its original formulation, Mutual Information between two images increases with the individual entropy of each of these images (see Equations (6)–(8)). Conversely, entropy normalization introduced in SMI and NMI cancels this effect (see Equations (9) and (11)).
A further difference to be noted in Figure 7 is the maximum value achieved by each metric. Whereas SMI proved insensitive to time, Figure 7 demonstrates that it is indeed not normalized as its maximum value is not equal to one. NMI, while having a similar behavior, takes values in the restricted range [ 0 , 1 ] , where 1 designates a perfect image match. Based on these results, NMI was deemed the best MI alternative for fire thermal image similarity analysis.
After selecting normalized mutual information among MI-based candidates, its performance was compared to 2D correlation and intensity mean squared difference. Figure 8 compares the average response of these metrics to synthetic camera movement. These results corroborate that IMSD is significantly less sensitive to image misalignments than NMI and 2Dcorr, as previously concluded from the global sensitivity analysis. Conversely, both NMI and 2Dcorr present an important peak at the position of perfect alignment, which makes them useful for image registration algorithms.
Nevertheless, results displayed in Figure 8 were obtained under idealized conditions in which each image was compared with a displaced version of itself. In a real scenario, the two IR images to be registered will not be identical. Typically, they will have been acquired from different perspectives or using dissimilar cameras. In a video stabilization problem, each video frame is to be compared to a previous frame of the same sequence. The amount of time elapsed between the acquisition of both frames will prevent an exact match even in the position of perfect alignment. This limitation affects similarity metrics differently, as demonstrated in Figure 9. Whereas 2D correlation can maintain optimum values close to 1 under real working conditions, NMI optimum values drop significantly due to the fact that a perfect image match is impossible. This behavior becomes more accentuated as sampling frequency diminishes, as shown in Figure 10.
Results displayed in Figure 9 highlight the first important difference in performance between 2D correlation and NMI. These results suggest that whereas both metrics can search for the perfect alignment position through an optimization strategy, they are not equally capable of assessing the quality of the achieved registration. In the specific case of IR fire video stabilization, 2D correlation values provide a reliable estimation of the alignment between consecutive frames. Two correctly registered frames will have a 2D correlation coefficient close to 1, whereas lower values can be attributed to misalignment. On the contrary, NMI may reach different maximum values depending on the video recording frequency, which prevents an absolute quality estimation. Therefore, we recommend the use of 2D correlation as a quality control metric after image registration.
To select the best-behaving metric during registration, one more property was analyzed: confidence of similarity values provided under real working conditions. Metric confidence was assessed first through their standard deviation when only one movement component was varied at a time (Figure 11). Additionally, metric robustness was analyzed in a general case in which the camera was allowed to move freely through a combination of translations, rotation and scaling (Figure 12).
Figure 11 shows significantly higher standard deviation for 2D correlation than NMI, the only exception occurring at the point of perfect alignment. According to this, 2Dcorr values computed for a certain misalignment are subject to greater differences due to changes in image content and recording conditions. Therefore, confidence on image similarity estimation provided by 2D correlation can be considered lower in general. Interestingly, this does not hold for a small region around the point of perfect alignment, where 2Dcorr values became more precise. Based on these results, 2Dcorr can still be considered suitable for robust estimation of achieved registration quality, whereas NMI outperformed 2Dcorr in situations far from perfect alignment, i.e., during registration.
These results are supported by Figure 12, which shows the statistical difference in similarity values computed under real and idealized conditions. Such difference was assessed through Bland-Altman plots, a common tool widely used to compare results provided by two methods designed to measure the same property. Bland-Altman plots are built by graphically displaying measurement differences along the complete range of measured values. Bias and limits of agreement are superimposed to the scatter plot. Bias is computed as the average difference, whereas limits of agreement are estimated as bias plus and minus 1.96 times the difference standard deviation [75]. Finally, both bias and limits of agreement are accompanied by their respective 95% confidence intervals, which were computed here using the approximated estimations proposed by Bland and Altman [76]. Confidence intervals are not always computed in the literature when using Bland-Altman plots, although they have been considered essential by some authors [77].
Figure 12 supports the hypothesis that NMI is more robust than 2D correlation under real conditions, especially when various movement components are combined. Although no significant bias was appreciated in any method, narrower limits of agreement mean that similarity estimations provided by NMI under real conditions are in general closely related to estimations provided under idealized conditions. This implies that NMI sensitiveness to changes in the reference frame (Figure 9) is limited to a small region around the point of perfect alignment. On average, when considering the complete camera movement space, NMI similarity estimations were more robust than those provided by 2D correlation and MI.
For this reason, we propose the use of NMI as the best performing image similarity metric for inter-frame registration in video stabilization frameworks. Maximization of normalized mutual information is likely to cancel misalignments for a wide range of video recording frequencies. Therefore, we encourage the application of optimization algorithms such as the ones proposed by Chen et al. [42] or Kern and Pattichis [39] for MI. However, optimum NMI achieved close to the point of perfect alignment cannot be used as a reliable estimation of absolute registration quality. Instead, we recommend using the 2D correlation coefficient for absolute alignment measurement.

5. Discussion

Our results are aligned with previous findings published in related fields. Due to its higher robustness to outliers and noise, Mutual Information has been selected as the primary similarity metric for image registration in multiple applications including medical imaging [46,78], stereo processing [79] and object tracking [80,81], among others. Previous comparative studies also found that MI produces consistently sharper optimum peaks at the correct registration values than correlation [82]. We observed a similar behavior in fire TIR imagery.
The use of the 2D correlation coefficient as a quality control metric is not so common in previous literature. When new algorithms are developed, registration accuracy is usually measured in a controlled environment where ground-truth camera movement is known or synthetically generated. In these cases, algorithm performance is assessed by comparing predicted and ground-truth registration transformations [39,44,83,84,85,86]. However, this approach cannot be used for quality control in a general operational scenario where the ground-truth registration transformation is unknown.
Other authors used pixel gray value mean absolute difference or root mean squared difference to assess registration quality [87]. However, similarity metrics based on gray differences are highly sensitive not only to the image relative position but also to the image content, as demonstrated here. Given two pairs of images with the same relative position, gray difference metrics are likely to give a higher registration rating to the image pair with lower contrast. We therefore recommend the use of 2D correlation for this purpose instead.
Given these results, the most immediate follow-up work for this paper consists in using the selected methods to build an IR video stabilization system suitable for aerial wildfire monitoring. In addition, this study could be extended to other image similarity metrics not considered here. We assessed and discussed some of the most popular methods in their basic formulation. However, there exist a wealth of variants that were derived from the basic algorithms to improve their performance. Among other adaptations, several authors have proposed the use of multiresolution schemes to measure similarity in image registration problems [80,88]. These methodologies provide additional improvements and their behavior should be analyzed.

6. Conclusions

This article analyzed alternative approaches to measure image similarity within a TIR fire image registration framework. Image registration is an important pre-processing step for the study of wildfire behavior through remote sensing. Within registration, image similarity measurement requires special attention because the estimation of image misalignment is essential to accomplish image registration as well as to control its quality.
Performance of any image similarity metric is highly dependent on the specific application for which it is used. For this reason, the primary goal of this study was the assessment of general similarity measurement approaches for the specific problem of fire thermal image registration. Image registration requires image similarity measurement for two different purposes. First, image similarity is treated as a cost function during the optimization problem linked to registration. Secondly, a robust estimation of absolute image similarity is essential for quality control. The highly dynamic nature of any wildfire scenario adds important difficulties when comparing images acquired at different times. This can result in mismatches at any time, even with the most accurate registration method. Such outliers must be automatically detected if the algorithm is meant to work unsupervisedly.
Without an in-depth analysis, it may seem that the methodology used to estimate image similarity does not have further implications as long as metrics provide higher similarity values for better aligned images. Results presented here show that this is not the case and careful attention should be paid to the method used to measure quality of alignment. Our results demonstrate that different image similarity metrics are affected differently by camera translation, rotation, distance to fire, size of fire, recording frequency and temporal changes in the scene. Such distinct behavior may motivate the selection of one metric or another depending on the specific goal to be achieved. In the case of video stabilization, we suggest using Normalized Mutual Information (NMI) for similarity maximization during inter-frame registration, whereas the 2D correlation coefficient is recommended for absolute alignment assessment and quality control.
These results constitute a key departing point for further studies into remote sensing of active wildfires through aerial TIR imagery. Furthermore, we described a generic and systematic methodology that can be replicated for analogous studies. GSA and LSA are usually complementary, and so were they in this study. While LSA provided detailed insight into metric behavior around the nominal operation point, GSA allowed measurement of the general sensitivity of metric candidates to factors that can vary widely and do not have fixed nominal values, such as scene content and video sampling frequency.
Finally, our analysis of MI-based metrics may help in other image registration problems. NMI inherits MI strengths and solves MI limitations regarding sensitiveness to image overlap and dependence on absolute image entropy. Because MI-based metrics allow general comparison between multi-modal images without making assumptions about their nature, NMI may be a powerful metric for multi-modal image registration or image fusion. As with thermal cameras, multi- and hyper-spectral sensors are becoming more compact and lighter, which will boost their potential for near-range remote sensing in wildfire operations. In a likely future scenario, several remote sensing platforms may be flying simultaneously over an active fire, each one carrying different sensors and acquiring complementary views at different times from different perspectives. How to fuse these data will undoubtedly be a topic of interest in the near future, and studies such as the one we present here contribute towards a better understanding of image processing alternatives.

Supplementary Materials

The following are available online at https://0-www-mdpi-com.brum.beds.ac.uk/2072-4292/12/3/540/s1, Figures S1–S4: GSA index convergence in each scenario, Figure S5: Global sensitivity indices in each scenario, Figures S6–S10: LSA of MI-based metrics under idealized conditions in each scenario, Figure S11: LSA results under idealized conditions in each scenario, Figure S12: LSA results under real conditions in each scenario, Figure S13: Metric value dispersion under real conditions in each scenario, Figure S14: Bland–Altman plots comparing metric behavior under real and idealized conditions in each scenario.

Author Contributions

Conceptualization, M.M.V. and S.V.; methodology, M.M.V. and S.V.; software, M.M.V.; validation, C.M., E.P. (Elsa Pastor), E.P. (Eulàlia Planas), M.M.V., O.R. and S.V.; resources, D.J., E.P. (Elsa Pastor), E.P. (Eulàlia Planas), L.Q. and S.V.; data curation, C.M., D.J., L.Q. and M.M.V.; writing—original draft preparation, M.M.V.; writing—review and editing, all authors; visualization, C.M., M.M.V. and O.R.; supervision, E.P. (Elsa Pastor), E.P. (Eulàlia Planas) and S.V.; project administration, E.P. (Elsa Pastor) and E.P. (Eulàlia Planas); funding acquisition, M.M.V., E.P. (Elsa Pastor) and E.P. (Eulàlia Planas). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Spanish Ministry of Education, Culture and Sport (Grant FPU13/05876), the Spanish Ministry of Economy and Competitiveness (projects CTM2014- 57448-R and CTQ2017-85990-R, co-financed with FEDER funds), the Erasmus+ Traineeship Program and Obra Social La Caixa research mobility grants.

Acknowledgments

The authors thank Valentijn Hoff for his help with data collection and Bret Butler for his support during this research and his comments on the manuscript draft.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Lentile, L.B.; Holden, Z.A.; Smith, A.M.S.; Falkowski, M.J.; Hudak, A.T.; Morgan, P.; Lewis, S.A.; Gessler, P.E.; Benson, N.C. Remote sensing techniques to assess active fire characteristics and post-fire effects. Int. J. Wildland Fire 2006, 15, 319. [Google Scholar] [CrossRef]
  2. Giglio, L.; Descloitres, J.; Justice, C.O.; Kaufman, Y.J. An Enhanced Contextual Fire Detection Algorithm for MODIS. Remote Sens. Environ. 2003, 87, 273–282. [Google Scholar] [CrossRef]
  3. Ichoku, C.; Kaufman, Y.J.; Giglio, L.; Li, Z.; Fraser, R.H.; Jin, J.Z.; Park, W.M. Comparative analysis of daytime fire detection algorithms using AVHRR data for the 1995 fire season in Canada: Perspective for MODIS. Int. J. Remote Sens. 2003, 24, 1669–1690. [Google Scholar] [CrossRef]
  4. Dennison, P.E. Fire detection in imaging spectrometer data using atmospheric carbon dioxide absorption. Int. J. Remote Sens. 2006, 27, 3049–3055. [Google Scholar] [CrossRef]
  5. Justice, C.; Giglio, L.; Korontzi, S.; Owens, J.; Morisette, J.; Roy, D.; Descloitres, J.; Alleaume, S.; Petitcolin, F.; Kaufman, Y. The MODIS fire products. Remote Sens. Environ. 2002, 83, 244–262. [Google Scholar] [CrossRef]
  6. Smith, A.M.S.; Wooster, M.J.; Powell, A.K.; Usher, D. Texture based feature extraction: Application to burn scar detection in Earth observation satellite sensor imagery. Int. J. Remote Sens. 2002, 23, 1733–1739. [Google Scholar] [CrossRef]
  7. Holden, Z.A.; Smith, A.M.S.; Morgan, P.; Rollins, M.G.; Gessler, P.E. Evaluation of novel thermally enhanced spectral indices for mapping fire perimeters and comparisons with fire atlas data. Int. J. Remote Sens. 2005, 26, 4801–4808. [Google Scholar] [CrossRef]
  8. Roy, D.; Jin, Y.; Lewis, P.; Justice, C. Prototyping a global algorithm for systematic fire-affected area mapping using MODIS time series data. Remote Sens. Environ. 2005, 97, 137–162. [Google Scholar] [CrossRef]
  9. Giglio, L.; Loboda, T.; Roy, D.P.; Quayle, B.; Justice, C.O. An active-fire based burned area mapping algorithm for the MODIS sensor. Remote Sens. Environ. 2009, 113, 408–420. [Google Scholar] [CrossRef]
  10. Wooster, M.J.; Zhukov, B.; Oertel, D. Fire radiative energy for quantitative study of biomass burning: Derivation from the BIRD experimental satellite and comparison to MODIS fire products. Remote Sens. Environ. 2003, 86, 83–107. [Google Scholar] [CrossRef]
  11. Zhukov, B.; Lorenz, E.; Oertel, D.; Wooster, M.; Roberts, G. Spaceborne detection and characterization of fires during the bi-spectral infrared detection (BIRD) experimental small satellite mission (2001–2004). Remote Sens. Environ. 2006, 100, 29–51. [Google Scholar] [CrossRef]
  12. Roberts, G.; Wooster, M.J.; Perry, G.L.W.; Drake, N.; Rebelo, L.M.; Dipotso, F. Retrieval of biomass combustion rates and totals from fire radiative power observations: Application to southern Africa using geostationary SEVIRI imagery. J. Geophys. Res. Atmos. 2005, 110, 1–19. [Google Scholar] [CrossRef] [Green Version]
  13. Wooster, M.J.; Roberts, G.; Perry, G.L.W.; Kaufman, Y.J. Retrieval of biomass combustion rates and totals from fire radiative power observations: FRP derivation and calibration relationships between biomass consumption and fire radiative energy release. J. Geophys. Res. Atmos. 2005, 110, 1–24. [Google Scholar] [CrossRef]
  14. Riggan, P.J.; Tissell, R.G. Chapter 6 Airborne Remote Sensing of Wildland Fires. Dev. Environ. Sci. 2008, 8, 139–168. [Google Scholar] [CrossRef]
  15. Paugam, R.; Wooster, M.J.; Roberts, G. Use of Handheld Thermal Imager Data for Airborne Mapping of Fire Radiative Power and Energy and Flame Front Rate of Spread. IEEE Trans. Geosci. Remote Sens. 2013, 51, 3385–3399. [Google Scholar] [CrossRef]
  16. Plucinski, M.; Pastor, E. Criteria and methodology for evaluating aerial wildfire suppression. Int. J. Wildland Fire 2013, 22, 1144–1154. [Google Scholar] [CrossRef]
  17. Stow, D.A.; Riggan, P.J.; Storey, E.J.; Coulter, L.L. Measuring fire spread rates from repeat pass airborne thermal infrared imagery. Remote Sens. Lett. 2014, 5, 803–812. [Google Scholar] [CrossRef]
  18. Dickinson, M.B.; Hudak, A.T.; Zajkowski, T.; Loudermilk, E.L.; Schroeder, W.; Ellison, L.; Kremens, R.L.; Holley, W.; Martinez, O.; Paxton, A.; et al. Measuring radiant emissions from entire prescribed fires with ground, airborne and satellite sensors—RxCADRE 2012. Int. J. Wildland Fire 2016, 25, 48–61. [Google Scholar] [CrossRef]
  19. Mueller, E.V.; Skowronski, N.; Clark, K.; Gallagher, M.; Kremens, R.; Thomas, J.C.; El Houssami, M.; Filkov, A.; Hadden, R.M.; Mell, W.; et al. Utilization of remote sensing techniques for the quantification of fire behavior in two pine stands. Fire Saf. J. 2017, 91, 845–854. [Google Scholar] [CrossRef] [Green Version]
  20. Johnston, J.M.; Wooster, M.J.; Paugam, R.; Wang, X.; Lynham, T.J.; Johnston, L.M. Direct estimation of Byram’s fire intensity from infrared remote sensing imagery. Int. J. Wildland Fire 2017, 26, 668–684. [Google Scholar] [CrossRef] [Green Version]
  21. Valero, M.; Rios, O.; Pastor, E.; Planas, E. Automated location of active fire perimeters in aerial infrared imaging using unsupervised edge detectors. Int. J. Wildland Fire 2018, 27, 241–256. [Google Scholar] [CrossRef] [Green Version]
  22. Stow, D.; Riggan, P.; Schag, G.; Brewer, W.; Tissell, R.; Coen, J.; Storey, E. Assessing uncertainty and demonstrating potential for estimating fire rate of spread at landscape scales based on time sequential airborne thermal infrared imaging. Int. J. Remote Sens. 2019, 40, 4876–4897. [Google Scholar] [CrossRef]
  23. Pastor, E.; Barrado, C.; Royo, P. Architecture for a helicopter-based unmanned aerial systems wildfire surveillance system. Geocarto Int. 2011, 26, 113–131. [Google Scholar] [CrossRef]
  24. Zajkowski, T.J.; Dickinson, M.B.; Hiers, J.K.; Holley, W.; Williams, B.W.; Paxton, A.; Martinez, O.; Walker, G.W. Evaluation and use of remotely piloted aircraft systems for operations and research—RxCADRE 2012. Int. J. Wildland Fire 2016, 25, 114–128. [Google Scholar] [CrossRef]
  25. Moran, C.J.; Seielstad, C.A.; Cunningham, M.R.; Hoff, V.; Parsons, R.A.; Queen, L.; Sauerbrey, K.; Wallace, T. Deriving Fire Behavior Metrics from UAS Imagery. Fire 2019, 2, 36. [Google Scholar] [CrossRef] [Green Version]
  26. Ambrosia, V.; Wegener, S.; Zajkowski, T.; Sullivan, D.; Buechel, S.; Enomoto, F.; Lobitz, B.; Johan, S.; Brass, J.; Hinkley, E. The Ikhana unmanned airborne system (UAS) western states fire imaging missions: From concept to reality (2006–2010). Geocarto Int. 2011, 26, 85–101. [Google Scholar] [CrossRef]
  27. Hudak, A.T.; Dickinson, M.B.; Bright, B.C.; Kremens, R.L.; Loudermilk, E.L.; O’Brien, J.J.; Hornsby, B.S.; Ottmar, R.D. Measurements relating fire radiative energy density and surface fuel consumption—RxCADRE 2011 and 2012. Int. J. Wildland Fire 2016, 25, 25–37. [Google Scholar] [CrossRef] [Green Version]
  28. Clements, C.B.; Davis, B.; Seto, D.; Contezac, J.; Kochanski, A.; Fillipi, J.B.; Lareau, N.; Barboni, B.; Butler, B.; Krueger, S.; et al. Overview of the 2013 FireFlux II grass fire field experiment. In Advances in Forest Fire Research—Proceedings of the 7th International Conference on Forest Fire Research; Coimbra University Press: Coimbra, Portugal, 2014; pp. 392–400. [Google Scholar] [CrossRef]
  29. Ottmar, R.D.; Hiers, J.K.; Butler, B.W.; Clements, C.B.; Dickinson, M.B.; Hudak, A.T.; O’Brien, J.J.; Potter, B.E.; Rowell, E.M.; Strand, T.M. Measurements, datasets and preliminary results from the RxCADRE project–2008, 2011 and 2012. Int. J. Wildland Fire 2016, 25, 1–9. [Google Scholar] [CrossRef]
  30. Hudak, A.; Freeborn, P.; Lewis, S.; Hood, S.; Smith, H.; Hardy, C.; Kremens, R.; Butler, B.; Teske, C.; Tissell, R.; et al. The Cooney Ridge Fire Experiment: An Early Operation to Relate Pre-, Active, and Post-Fire Field and Remotely Sensed Measurements. Fire 2018, 1, 10. [Google Scholar] [CrossRef] [Green Version]
  31. Riggan, P.J.; Hoffman, J.W. Firemappertm: A thermal-imaging radiometer for wildfire research and operations. IEEE Aerosp. Conf. Proc. 2003, 4, 1843–1854. [Google Scholar] [CrossRef] [Green Version]
  32. Valero, M.M.; Jimenez, D.; Butler, B.W.; Mata, C.; Rios, O.; Pastor, E.; Planas, E. On the use of compact thermal cameras for quantitative wildfire monitoring. In Advances in Forest Fire Research 2018; Viegas, D.X., Ed.; University of Coimbra Press: Coimbra, Portugal, 2018; Chapter 5; pp. 1077–1086. [Google Scholar] [CrossRef] [Green Version]
  33. Yuan, C.; Zhang, Y.; Liu, Z. A survey on technologies for automatic forest fire monitoring, detection, and fighting using unmanned aerial vehicles and remote sensing techniques. Can. J. For. Res. 2015, 45, 783–792. [Google Scholar] [CrossRef]
  34. Pérez, Y.; Pastor, E.; Planas, E.; Plucinski, M.; Gould, J. Computing forest fires aerial suppression effectiveness by IR monitoring. Fire Saf. J. 2011, 46, 2–8. [Google Scholar] [CrossRef]
  35. Brown, L.G. A survey of image registration techniques. ACM Comput. Surv. 1992, 24, 325–376. [Google Scholar] [CrossRef]
  36. Kaneko, S.; Murase, I.; Igarashi, S. Robust image registration by increment sign correlation. Pattern Recognit. 2002, 35, 2223–2234. [Google Scholar] [CrossRef]
  37. Yang, Q.; Ma, Z.; Xu, Y.; Yang, L.; Zhang, W. Modeling the Screen Content Image Quality via Multiscale Edge Attention Similarity. IEEE Trans. Broadcast. 2020. [Google Scholar] [CrossRef]
  38. Zitová, B.; Flusser, J. Image registration methods: A survey. Image Vis. Comput. 2003, 21, 977–1000. [Google Scholar] [CrossRef] [Green Version]
  39. Kern, J.P.; Pattichis, M.S. Robust Multispectral Image Registration Using Mutual-Information Models. IEEE Trans. Geosci. Remote Sens. 2007, 45, 1494–1505. [Google Scholar] [CrossRef]
  40. Wu, Y.; Ma, W.; Su, Q.; Liu, S.; Ge, Y. Remote sensing image registration based on local structural information and global constraint. J. Appl. Remote Sens. 2019, 13. [Google Scholar] [CrossRef]
  41. Pluim, J.P.; Maintz, J.B.; Viergever, M.A. Mutual-information-based registration of medical images: A survey. IEEE Trans. Med. Imaging 2003, 22, 986–1004. [Google Scholar] [CrossRef]
  42. Chen, H.M.; Varshney, P.K.; Arora, M.K. Performance of Mutual Information Similarity Measure for Registration of Multitemporal Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2003, 41, 2445–2454. [Google Scholar] [CrossRef]
  43. Jones, C.; Christens-Barry, W.A.; Terras, M.; Toth, M.B.; Gibson, A. Affine registration of multispectral images of historical documents for optimized feature recovery. Digit. Scholarsh. Humanit. 2019. [Google Scholar] [CrossRef]
  44. Liu, D.; Mansour, H.; Boufounos, P.T. Robust mutual information-based multi-image registration. In Proceedings of the IGARSS 2019—2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 915–918. [Google Scholar] [CrossRef]
  45. Baillet, S.; Garnero, L.; Marin, G.; Hugonin, J.P. Combined MEG and EEG Source Imaging by Minimization of Mutual Information. IEEE Trans. Biomed. Eng. 1999, 46, 522–534. [Google Scholar] [CrossRef] [PubMed]
  46. Panin, G. Mutual information for multi-modal, discontinuity-preserving image registration. In International Symposium on Visual Computing (ISVC); Springer: Berlin/Heidelberg, Germany, 2012. [Google Scholar] [CrossRef]
  47. Eikhosravi, A.; Li, B.; Liu, Y.; Eliceiri, K.W. Intensity-based registration of bright-field and second-harmonic generation images of histopathology tissue sections. Biomed. Opt. Express 2020, 11, 160–173. [Google Scholar] [CrossRef] [PubMed]
  48. Barnea, D.I.; Silverman, H.F. A class of algorithm for fast digital image rectification. IEEE Trans. Comput. 1972, C-21, 179–186. [Google Scholar] [CrossRef]
  49. Ertürk, S. Digital image stabilization with sub-image phase correlation based global motion estimation. IEEE Trans. Consum. Electron. 2003, 49, 1320–1325. [Google Scholar] [CrossRef]
  50. Lüdemann, J.; Barnard, A.; Malan, D.F. Sub-pixel image registration on an embedded Nanosatellite Platform. Acta Astronaut. 2019, 161, 293–303. [Google Scholar] [CrossRef]
  51. Cover, T.M.; Thomas, J.A. Elements of Information Theory; John Wiley & Sons Inc.: New York, USA, 1991. [Google Scholar]
  52. Maes, F.; Collignon, A.; Vandermeulen, D.; Marchal, G.; Suetens, P. Multimodality Image Registration by Maximization of Mutual Information. IEEE Trans. Med. Imaging 1997, 16, 187–198. [Google Scholar] [CrossRef] [Green Version]
  53. Viola, P.A. Alignment by Maximization of Mutual Information. Ph.D. Thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 1995. [Google Scholar]
  54. Collignon, A.; Maes, F.; Delaere, D.; Vandermeulen, D.; Suetens, P.; Marchal, G. Automated multi-modality image registration based on information theory. Inf. Process. Med. Imaging 1995, 3, 263–274. [Google Scholar] [CrossRef]
  55. Xu, R.; Chen, Y.w.; Tang, S.Y.; Morikawa, S.; Kurumi, Y. Parzen-Window Based Normalized Mutual Information for Medical Image Registration. IEICE Trans. Inf. Syst. 2008, E91-D, 132–144. [Google Scholar] [CrossRef]
  56. Zhuang, Y.; Gao, K.; Miu, X.; Han, L.; Gong, X. Infrared and visual image registration based on mutual information with a combined particle swarm optimization—Powell search algorithm. Optik 2016, 127, 188–191. [Google Scholar] [CrossRef]
  57. Shannon, C. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
  58. Wang, Q.; Shen, Y.; Zhang, Y.; Zhang, J.Q. A quantitative method for evaluating the performances of hyperspectral image fusion. IEEE Trans. Instrum. Meas. 2003, 52, 1041–1047. [Google Scholar] [CrossRef]
  59. Yan, L.; Liu, Y.; Xiao, B.; Xia, Y.; Fu, M. A Quantitative Performance Evaluation Index for Image Fusion: Normalized Perception Mutual Information. In Proceedings of the 31st Chinese Control Conference, Hefei, China, 25–27 July 2012; pp. 3783–3788. [Google Scholar]
  60. Penney, G.P.; Weese, J.; Little, J.A.; Desmedt, P.; Hill, D.L.; Hawkes, D.J. A comparison of similarity measures for use in 2-D-3-D medical image registration. IEEE Trans. Med. Imaging 1998, 17, 586–595, [0912.0405]. [Google Scholar] [CrossRef]
  61. Pluim, J.P.W.; Maintz, J.B.A.; Viergever, M.A. Image registration by maximization of combined mutual information and gradiant information. IEEE Trans. Med. Imaging 2000, 19, 809–814. [Google Scholar] [CrossRef]
  62. Studholme, C.; Hill, D.; Hawkes, D. An overlap invariant entropy measure of 3D medical image alignment. Pattern Recognit. 1999, 32, 71–86. [Google Scholar] [CrossRef]
  63. Astola, J.; Virtanen, I. Entropy Correlation Coefficient, a Measure of Statistical Dependence for Categorized Data; Discussion Papers, 44; University of Vaasa: Vaasa, Finland, 1982. [Google Scholar]
  64. Strehl, A.; Ghosh, J. Cluster ensembles—A knowledge reuse framework for combining multiple partitions. J. Mach. Learn. Res. 2002, 3, 583–617. [Google Scholar] [CrossRef]
  65. Bai, X.; Zhao, Y.; Huang, Y.; Luo, S. Normalized joint mutual information measure for image segmentation evaluation with multiple ground-truth images. In Proceedings of the 14th International Conference on Computer Analysis of Images and Patterns, Seville, Spain, 29–31 August 2011; Volume Part 1, pp. 110–117. [Google Scholar] [CrossRef]
  66. Pillai, K.G.; Vatsavai, R.R. Multi-sensor remote sensing image change detection: An evaluation of similarity measures. In Proceedings of the IEEE 13th International Conference on Data Mining Workshops, Dallas, TX, USA, 7–10 December 2013; pp. 1053–1060. [Google Scholar] [CrossRef]
  67. Estévez, P.A.; Member, S.; Tesmer, M.; Perez, C.A.; Member, S.; Zurada, J.M. Normalized Mutual Information Feature Selection. IEEE Trans. Neural Netw. 2009, 20, 189–201. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  68. O’Brien, J.J.; Loudermilk, E.L.; Hornsby, B.; Hudak, A.T.; Bright, B.C.; Dickinson, M.B.; Hiers, J.K.; Teske, C.; Ottmar, R.D. High-resolution infrared thermography for capturing wildland fire behaviour: RxCADRE 2012. Int. J. Wildland Fire 2016, 25, 62–75. [Google Scholar] [CrossRef]
  69. Saltelli, A.; Tarantola, S.; Campolongo, F.; Ratto, M. Sensistivity Analysis in Practice: A Guide to Assessing Scientific Models; John Wiley & Sons Ltd.: Chichester, UK, 2004. [Google Scholar]
  70. Saltelli, A.; Ratto, M.; Andres, T.; Campolongo, F.; Cariboni, J.; Gatelli, D.; Saisana, M.; Tarantola, S. Global Sensitivity Analysis—The Primer; John Wiley & Sons Ltd.: Hoboken, NJ, USA, 2008; pp. 1–305. [Google Scholar] [CrossRef] [Green Version]
  71. Cukier, R.; Fortuin, C.; Shuler, K. Study of the sensitivity of coupled reaction systems to uncertainties in rate coefficients. I Theory. J. Chem. Phys. 1973, 59. [Google Scholar] [CrossRef]
  72. Sobol, I. Sensitivity analysis for nonlinear mathematical models. Math. Model. Comput. Exp. 1993, 1, 407–414. [Google Scholar] [CrossRef]
  73. McKay, M.; Beckman, R.; Conover, W. A Comparison of Three Methods for Selecting Values of Input Variables in the Analysis of Output from a Computer Code. Technometrics 1979, 21, 239–245. [Google Scholar]
  74. Pianosi, F.; Sarrazin, F.; Wagener, T. A Matlab toolbox for Global Sensitivity Analysis. Environ. Model. Softw. 2015, 70, 80–85. [Google Scholar] [CrossRef] [Green Version]
  75. Bland, J.M.; Altman, D.G. Statistical methods for assessing agreement between two methods of clinical measurement. Lancet 1986, 327, 307–310. [Google Scholar] [CrossRef]
  76. Bland, J.M.; Altman, D.G. Measuring agreement in method comparison studies. Stat. Methods Med. Res. 1999, 8, 135–160. [Google Scholar] [CrossRef] [PubMed]
  77. Carkeet, A. Exact Parametric Confidence Intervals for Bland-Altman Limits of Agreement. Optom. Vis. Sci. 2015, 92, 71–80. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  78. Yaegashi, Y.; Tateoka, K.; Fujimoto, K.; Nakazawa, T.; Nakata, A.; Saito, Y.; Abe, T.; Yano, M.; Sakata, K. Assessment of Similarity Measures for Accurate Deformable Image Registration. J. Nucl. Med. Radiat. Ther. 2012, 42. [Google Scholar] [CrossRef] [Green Version]
  79. Hirschmüller, H. Stereo Processing by Semiglobal Matching and Mutual Information. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 328–341. [Google Scholar] [CrossRef]
  80. Panin, G.; Knoll, A. Mutual information-based 3D object tracking. Int. J. Comput. Vis. 2008, 78, 107–118. [Google Scholar] [CrossRef]
  81. Dame, A.; Marchand, E. Accurate real-time tracking using mutual information. In Proceedings of the 9th IEEE International Symposium on Mixed and Augmented Reality 2010: Science and Technology (ISMAR 2010), Seoul, Korea, 13–16 October 2010; pp. 47–56. [Google Scholar] [CrossRef] [Green Version]
  82. Cole-Rhodes, A.A.; Johnson, K.L.; LeMoigne, J.; Zavorin, L. Multiresolution registration of remote sensing imagery by optimization of mutual information using a stochastic gradient. IEEE Trans. Image Process. 2003, 12, 1495–1510. [Google Scholar] [CrossRef]
  83. Bentoutou, Y.; Taleb, N.; Kpalma, K.; Ronsin, J. An automatic image registration for applications in remote sensing. IEEE Trans. Geosci. Remote Sens. 2005, 43, 2127–2137. [Google Scholar] [CrossRef]
  84. Sakai, T.; Sugiyama, M.; Kitagawa, K.; Suzuki, K. Registration of infrared transmission images using squared-loss mutual information. Precis. Eng. 2015, 39, 187–193. [Google Scholar] [CrossRef] [Green Version]
  85. Li, H.; Ding, W.; Cao, X.; Liu, C. Image registration and fusion of visible and infrared integrated camera for medium-altitude unmanned aerial vehicle remote sensing. Remote Sens. 2017, 9, 441. [Google Scholar] [CrossRef] [Green Version]
  86. Ma, W.; Wen, Z.; Wu, Y.; Jiao, L.; Gong, M.; Zheng, Y.; Liu, L. Remote sensing image registration with modified sift and enhanced feature matching. IEEE Geosci. Remote Sens. Lett. 2017, 14, 3–7. [Google Scholar] [CrossRef]
  87. Wang, X.; Li, Y.; Wei, H.; Liu, F. An ASIFT-Based Local Registration Method for Satellite Imagery. Remote Sens. 2015, 7, 7044–7061. [Google Scholar] [CrossRef] [Green Version]
  88. Thévenaz, P.; Unser, M. Optimization of mutual information for multiresolution image registration. IEEE Trans. Image Process. 2000, 9, 2083–2099. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Design of the comparative analysis conducted to assess the behavior of various image similarity metrics. Desired properties were high sensitivity to camera movement and low sensitivity to scene variations and sampling frequency. LSA was applied to the 6 available video sequences. GSA was applied to sequences 1, 2, 4 and 5 due to computational restrictions.
Figure 1. Design of the comparative analysis conducted to assess the behavior of various image similarity metrics. Desired properties were high sensitivity to camera movement and low sensitivity to scene variations and sampling frequency. LSA was applied to the 6 available video sequences. GSA was applied to sequences 1, 2, 4 and 5 due to computational restrictions.
Remotesensing 12 00540 g001
Figure 2. Sample frames of the six video sequences (af) employed in this study.
Figure 2. Sample frames of the six video sequences (af) employed in this study.
Remotesensing 12 00540 g002aRemotesensing 12 00540 g002b
Figure 3. Experimental setups used to record the six video sequences (af) employed in this study.
Figure 3. Experimental setups used to record the six video sequences (af) employed in this study.
Remotesensing 12 00540 g003
Figure 4. Effect of centering model output distributions computed during GSA. The top row shows sensitivity indices and their convergence obtained with the original distribution. The bottom row shows results achieved after centering the model output distribution. Mean values (solid lines) and confidence bounds (dashed lines) were estimated using bootstrapping with 500 resamples. Example results for Studholme’s Mutual Information (SMI) in Scenario 1.
Figure 4. Effect of centering model output distributions computed during GSA. The top row shows sensitivity indices and their convergence obtained with the original distribution. The bottom row shows results achieved after centering the model output distribution. Mean values (solid lines) and confidence bounds (dashed lines) were estimated using bootstrapping with 500 resamples. Example results for Studholme’s Mutual Information (SMI) in Scenario 1.
Remotesensing 12 00540 g004
Figure 5. Global sensitivity indices of image similarity metrics to the six considered parameters, averaged over scenarios 1, 2, 4 and 5. Individual results for each scenario can be consulted in Supplementary Materials.
Figure 5. Global sensitivity indices of image similarity metrics to the six considered parameters, averaged over scenarios 1, 2, 4 and 5. Individual results for each scenario can be consulted in Supplementary Materials.
Remotesensing 12 00540 g005
Figure 6. Local response of MI-based metrics to independent camera movement components under idealized conditions, i.e., when each video frame is compared to the stable version of itself. Averaged values along all studied video sequences. Camera movement components are: translation in the X direction ( T x ), translation in the Y direction ( T y ), rotation ( θ ) and scaling.
Figure 6. Local response of MI-based metrics to independent camera movement components under idealized conditions, i.e., when each video frame is compared to the stable version of itself. Averaged values along all studied video sequences. Camera movement components are: translation in the X direction ( T x ), translation in the Y direction ( T y ), rotation ( θ ) and scaling.
Remotesensing 12 00540 g006
Figure 7. 3D representation of MI-based metrics response to time and image rotation under idealized conditions. Analogous results for translations and scaling are included in Supplementary Materials.
Figure 7. 3D representation of MI-based metrics response to time and image rotation under idealized conditions. Analogous results for translations and scaling are included in Supplementary Materials.
Remotesensing 12 00540 g007
Figure 8. Metric response to independent camera movement components under idealized conditions, i.e., when each video frame is compared to the stable version of itself. Averaged values along all studied video sequences. Camera movement components are: translation in X direction ( T x ), translation in Y direction ( T y ), rotation ( θ ) and scaling. 1-IMSD is displayed for consistency with the rest of metrics.
Figure 8. Metric response to independent camera movement components under idealized conditions, i.e., when each video frame is compared to the stable version of itself. Averaged values along all studied video sequences. Camera movement components are: translation in X direction ( T x ), translation in Y direction ( T y ), rotation ( θ ) and scaling. 1-IMSD is displayed for consistency with the rest of metrics.
Remotesensing 12 00540 g008
Figure 9. Metric response to independent camera movement components under realistic operation conditions, i.e., when each video frame is compared to the stable version of the previous frame. Averaged values along all studied video sequences. Camera movement components are: translation in the X direction ( T x ), translation in the Y direction ( T y ), rotation ( θ ) and scaling. 1-IMSD is displayed for consistency with the rest of metrics.
Figure 9. Metric response to independent camera movement components under realistic operation conditions, i.e., when each video frame is compared to the stable version of the previous frame. Averaged values along all studied video sequences. Camera movement components are: translation in the X direction ( T x ), translation in the Y direction ( T y ), rotation ( θ ) and scaling. 1-IMSD is displayed for consistency with the rest of metrics.
Remotesensing 12 00540 g009
Figure 10. Metric response to video recording frequency. Displayed values show similarity between two stable consecutive frames, time-averaged along each video sequence. 1-IMSD is displayed for consistency with the rest of metrics.
Figure 10. Metric response to video recording frequency. Displayed values show similarity between two stable consecutive frames, time-averaged along each video sequence. 1-IMSD is displayed for consistency with the rest of metrics.
Remotesensing 12 00540 g010
Figure 11. Metric value dispersion under real operation conditions, i.e., when each video frame is compared to the previous frame. Output standard deviation, computed along all studied video sequences, provides a quantitative assessment of how robust each metric is in front of natural image dissimilarities and recording conditions.
Figure 11. Metric value dispersion under real operation conditions, i.e., when each video frame is compared to the previous frame. Output standard deviation, computed along all studied video sequences, provides a quantitative assessment of how robust each metric is in front of natural image dissimilarities and recording conditions.
Remotesensing 12 00540 g011
Figure 12. Bland-Altman plots comparing metric behavior under real and idealized conditions. S i m i d e a l : similarity measured between each perturbed frame and the stable version of itself; S i m r e a l : similarity measured between each perturbed frame and the previous stable frame. Black dots are individual random samples along all studied scenarios; red solid lines indicate mean bias; red dashed lines indicate Limits of Agreement (LoA); red dotted lines represent 95% confidence intervals for estimated bias and LoA. Wide LoA are representative of significant sensitivity to changes in the reference frame used for registration.
Figure 12. Bland-Altman plots comparing metric behavior under real and idealized conditions. S i m i d e a l : similarity measured between each perturbed frame and the stable version of itself; S i m r e a l : similarity measured between each perturbed frame and the previous stable frame. Black dots are individual random samples along all studied scenarios; red solid lines indicate mean bias; red dashed lines indicate Limits of Agreement (LoA); red dotted lines represent 95% confidence intervals for estimated bias and LoA. Wide LoA are representative of significant sensitivity to changes in the reference frame used for registration.
Remotesensing 12 00540 g012
Table 1. Camera properties and parameters used to record analyzed footage.
Table 1. Camera properties and parameters used to record analyzed footage.
ScenarioCamera Commercial NameSpectral Range Wavelength ( μ m)Brightness Temperature Range (°C)Image Resolution (Pixels)Field of View (°)Thermal Sensitivity (mK)Recording Frequency (Hz)
1Optris PI 640[7.5, 13][20, 900]640 × 48060 × 457532
2Optris PI 400[7.5, 13][200, 1500]382 × 28860 × 457527
3Optris PI 400[7.5, 13][200, 1500]382 × 28860 × 457527
4FLIR SC660[7.5, 13][300, 1500]640 × 48045 × 30301
5FLIR SC660[7.5, 13][300, 1500]640 × 48045 × 30301
6FLIR SC660[7.5, 13][300, 1500]640 × 48045 × 30301
Table 2. Input parameter ranges considered for sensitivity analysis. Time ranges were set as wide as possible, provided that fire was present in the scene. Maximum sampling frequency corresponds to video recording frequency.
Table 2. Input parameter ranges considered for sensitivity analysis. Time ranges were set as wide as possible, provided that fire was present in the scene. Maximum sampling frequency corresponds to video recording frequency.
Video SequenceTranslation Range (% of Width/Height)Rotation Range (deg)Scaling RangeFrequency Range (Hz)Time Range (s)
1[−20, 20][−25, 25][0.8, 1.2][0.1, 32][60, 240]
2[−20, 20][−25, 25][0.8, 1.2][0.1, 27][8, 660]
3[−20, 20][−25, 25][0.8, 1.2][0.1, 27][23, 700]
4[−20, 20][−25, 25][0.8, 1.2][0.1, 0.86][90, 1560]
5[−20, 20][−25, 25][0.8, 1.2][0.1, 0.88][45, 700]
6[−20, 20][−25, 25][0.8, 1.2][0.1, 0.87][90, 770]

Share and Cite

MDPI and ACS Style

Valero, M.M.; Verstockt, S.; Mata, C.; Jimenez, D.; Queen, L.; Rios, O.; Pastor, E.; Planas, E. Image Similarity Metrics Suitable for Infrared Video Stabilization during Active Wildfire Monitoring: A Comparative Analysis. Remote Sens. 2020, 12, 540. https://0-doi-org.brum.beds.ac.uk/10.3390/rs12030540

AMA Style

Valero MM, Verstockt S, Mata C, Jimenez D, Queen L, Rios O, Pastor E, Planas E. Image Similarity Metrics Suitable for Infrared Video Stabilization during Active Wildfire Monitoring: A Comparative Analysis. Remote Sensing. 2020; 12(3):540. https://0-doi-org.brum.beds.ac.uk/10.3390/rs12030540

Chicago/Turabian Style

Valero, Mario M., Steven Verstockt, Christian Mata, Dan Jimenez, Lloyd Queen, Oriol Rios, Elsa Pastor, and Eulàlia Planas. 2020. "Image Similarity Metrics Suitable for Infrared Video Stabilization during Active Wildfire Monitoring: A Comparative Analysis" Remote Sensing 12, no. 3: 540. https://0-doi-org.brum.beds.ac.uk/10.3390/rs12030540

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop