Next Article in Journal
Analytical Model of Induction Machines with Multiple Cage Faults Using the Winding Tensor Approach
Previous Article in Journal
Can Intraoperative Intra-Articular Loads Predict Postoperative Knee Joint Laxity Following Total Knee Arthroplasty? A Cadaver Study with Smart Tibial Trays
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Three-Dimensional Interferometric ISAR Imaging Algorithm Based on Cross Coherence Processing

School of Physics and Information Technology, Shaanxi Normal University, West Chang’an Avenue, No. 620, Xi’an 710119, China
*
Author to whom correspondence should be addressed.
Submission received: 9 June 2021 / Revised: 9 July 2021 / Accepted: 20 July 2021 / Published: 27 July 2021

Abstract

:
Interferometric inverse synthetic aperture radar (InISAR) has received significant attention in three-dimensional (3D) imaging due to its applications in target classification and recognition. The traditional two-dimensional (2D) ISAR image can be interpreted as a filtered projection of a 3D target’s reflectivity function onto an image plane. Such a plane usually depends on unknown radar-target geometry and dynamics, which results in difficulty interpreting an ISAR image. Using the L-shape InISAR imaging system, this paper proposes a novel 3D target reconstruction algorithm based on Dechirp processing and 2D interferometric ISAR imaging, which can jointly estimate the effective rotation vector and the height of scattering center. In order to consider only the areas of the target with meaningful interferometric phase and mitigate the effects of noise and sidelobes, a special cross-channel coherence-based detector (C3D) is introduced. Compared to the multichannel CLEAN technique, advantages of the C3D include the following: (1) the computational cost is lower without complex iteration and (2) the proposed method, which can avoid propagating errors, is more suitable for a target with multi-scattering points. Moreover, misregistration and its influence on target reconstruction are quantitatively discussed. Theoretical analysis and numerical simulations confirm the suitability of the algorithm for 3D imaging of multi-scattering point targets with high efficiency and demonstrate the reliability and effectiveness of the proposed method in the presence of noise.

1. Introduction

Due to its superiority in all-day, all-weather and high-resolution applications, inverse synthetic aperture radar (ISAR) has attracted much attention in both civil and military fields [1,2,3,4,5,6]. Typically, the conventional ISAR imaging system can produce a two-dimensional (2D) image of the moving target, which can be interpreted as a filtered projection of the three-dimensional (3D) target’s reflectivity function onto a so-called image projection plane (IPP) [7,8,9,10]. Due to the dependence of the IPP on the target’s own motions and radar-target geometry, such projection cannot be predicted beforehand, which makes it more difficult to interpret ISAR images. A previous attempt to estimate IPP information in advance has been conducted, but its applicability and effectiveness are not confirmed [11]. To radically solve the problem of interpreting 2D ISAR images and support automatic target classification and recognition (ATC/ATR), 3D imaging techniques have emerged and motivated the development of various configurations of 3D ISAR imaging systems [12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33]. These methods for creating 3D ISAR images can be mainly divided into three different categories.
The first class of method aims at producing 3D target reconstruction by exploiting a single sensor ISAR image sequence [27]. The 3D position of each target’s scattering center is determined by processing the tracks on the ISAR image plane. In this case, knowledge of the relative dynamics corresponding to each ISAR image must be obtained precisely. In addition, this method requires a long target observation time to generate the ISAR image sequence, which is usually uncontrolled and unsatisfactory in a complex imaging environment (such as highly maneuvering airplanes and ships fluctuating with oceanic waves). The second type of approach to create 3D images is based on a 2D antenna array [31,32,33]. Although such an approach circumvents complex motion compensation, it seriously restricts the observation area [28]. In order to acquire fine spatial resolution, a very wide antenna aperture and a large number of antenna elements are necessary, which undoubtedly increases the complexity of the radar system and causes difficulties in signal processing. The third variety of method is interferometric imaging techniques that achieve 3D target reconstruction by combining interferometric processing with ISAR processing [13,14,15,16,17,18,19,20,21,22,23,24,25,26,29,30]. In comparison with the aforementioned methods, both the InISAR system and signal processing are easier to realize. Reported applications of InISAR include target scattering diagnosis, motion modeling, 3D imaging of moving targets, automatic aircraft landing and target recognition [34].
For InISAR imaging, the target geometry is estimated using the interferometric phase of ISAR images acquired from different aperture centers. For this purpose, many 3D InISAR imaging algorithms based on continuous wave (CW) radar have been proposed and applied in practice. In [18], a dual interferometric ISAR system with multichannel CLEAN technique (MC-CLEAN) was utilized to reconstruct 3D images of moving targets. The real data from a suitably designed Multi-Channel ground-based radar was used to verify the reliability of a 3D InISAR technique in [21]. Meanwhile, several successful 3D InISAR algorithms based on a linear frequency modulation (LFM) signal model have also been proposed. In [24], a practical maneuvering target 3D imaging algorithm based on an arbitrary three-antenna InISAR configuration was investigated. To produce high-resolution InISAR images of a maneuvering target, a novel algorithm using the peak extraction technique was proposed in [25]. In the above two references, emphasis was also placed on target motion compensation and motion error, which is of great value for practical applications. In [29,30], time-frequency transform (TFT) and clean technique were used to achieve 3D reconstruction of targets with nonsevere maneuverability. In [23], 3D InISAR imaging combining bistatic configuration with compressive sensing (CS) was proposed, which was especially useful for ship targets, whereas a bistatic radar system on its own was subject to synchronization error.
However, there are several aspects that restrict the applications of the InISAR algorithms above. First, for CW radar, the configurations generally are of one of the following two kinds: (1) three antennas that both transmit and receive, which must utilize orthogonal codes to separate the three channels or (2) one transmitter and a number of collocated receivers. Unfortunately, in either case, such an approach suffers from complex system processing and hardware consumption. Second, in the InISAR imaging system, the multichannel 2D ISAR images are usually utilized to perform interferometric processing. Thus, extraction of the scattering center is carried out using the MC-CLEAN technique, of which the computational cost is at least three times more than that of traditional 2D ISAR imaging [23]. The methods in [18,21,24,25,33,34] all use the CLEAN technique, which is computationally expensive, especially for multi-scattering point targets. Third, the effective interferometric phase provides useful information only in areas with high coherence among three ISAR images [21]. In order to guarantee the precision of parameter estimation, it is crucial to obtain well-focused 2D ISAR images and accurate scattering center extraction. However, some areas without benefit to 3D reconstruction, such as areas with noise and low power areas (in particular secondary lobes from prominent scatters), may also have high coherence, which degrade the quality of the 3D InISAR reconstruction.
To overcome the above drawbacks, this paper presents an InISAR imaging system consisting of two orthogonal baselines based on an LFM signal model, which is more suitable for high resolution imaging purposes [35,36,37,38]. Dechirp processing and Fourier Transformation (FT) are applied to obtain the analytical expression of the complex ISAR image in range frequency-Doppler domain. As Dechirp processing is simple and flexible for hardware, it is now widely applied in ISAR imaging and the radar signal processing field [39]. Finally, the 3D coordinates of the target can be acquired by interferometry and mathematical processing. In order to restrain noise and sidelobes effectively, this paper introduces a detector combining cross-channel coherence with power using three ISAR images. With this method, the interferometric phase of interest can be easily selected and computational cost can be saved by avoiding use of the MC-CLEAN technique. Quantitative analysis of coregistration between two ISAR images from different interferometric antennas for the same baseline is conducted. The simulations are given to verify the proposed algorithm, illustrating that the algorithm is suitable for 3D imaging of multiple scattering points at a low computational cost.
The organization of this paper is as follows. In Section 2, the received signal model is established. The proposed 3D InISAR imaging algorithm is discussed in detail in Section 3. Section 4 gives detailed performance analysis of the proposed algorithm. The simulations of the 3D InISAR imaging algorithm and the conclusions are given in Section 5 and Section 6, respectively.

2. Received Signal Model

In this section, to make the proposed 3D reconstruction procedure straightforward, the overall flowchart of the 3D InISAR reconstruction processing is first depicted in Figure 1. Then, the geometry of the system is described, followed by the derivation of the received signal model.

2.1. System Geometry

In accordance with [18,29], the 3D InISAR imaging system geometry shown in Figure 2a is composed of three receiver antennas labeled AV, AC, and AH, where antenna AC acting as the transmitter is located at the origin of the reference system. In Figure 2b, receivers AV and AH are located on a horizontal and a vertical baseline with coordinates ( 0 , 0 , d v ) and ( d h , 0 , 0 ) respectively. The Cartesian reference system T z ( z 1 , z 2 , z 3 ) is embedded in the radar system with its origin at the array phase center. The axis z 2 is oriented along the line of sight (LOS) direction from the radar to the center of the target, while z 1 and z 3 correspond to the horizontal and vertical baseline respectively. The projection of the total target rotation vector Ω T onto the plane orthogonal to the LOS defines the effective rotation vector Ω e f f . The Cartesian reference system T x ( x 1 , x 2 , x 3 ) is embedded in the target and centered at the focusing point O . Axis x 2 is oriented along z 2 , whereas x 3 is oriented along Ω e f f . Since the effective rotation vector changes its orientation in time, the reference system T x is time varying. Another reference system T y is defined to be centered on the target with fixed orientation and coincident with T x for slow time t m = 0 . R 0 is the distance of the radar to the focus center of the target at t m = 0 .
Interestingly, the matrix M z x , which describes the rotation of T x with respect to T z , is defined as follows:
M z x = [ cos ϕ 0 sin ϕ 0 1 0 sin ϕ 0 cos ϕ ]
Thus, the LOS unit vectors i L O S x can be written as Equation (2) by means of the rotation matrix M z x :
i L O S x A V = [ d v sin ϕ R 0 ( t m ) + d v 2 R 0 ( t m ) R 0 ( t m ) + d v 2 d v cos ϕ R 0 ( t m ) + d v 2 ] [ d v sin ϕ R 0 1 d v cos ϕ R 0 ] i L O S x A C = [ 0 1 0 ] i L O S x A H = [ d h cos ϕ R 0 ( t m ) + d h 2 R 0 ( t m ) R 0 ( t m ) + d h 2 d h sin ϕ R 0 ( t m ) + d h 2 ] [ d h cos ϕ R 0 1 d h sin ϕ R 0 ]
where R 0 ( t m ) R 0 is valid in a far field condition, so the terms R 0 + d v 2 and R 0 + d h 2 are approximated as R 0 for a small observation time. y = [ y 1 , y 2 , y 3 ] denotes the position for a generic scatterer at t m = 0 in T y .
Under the assumption of a small observation time value T a , the total rotation vector Ω T can be considered as constant and the image plane as fixed with respect to T z [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19]. Therefore, x ( t m ) , the location of an arbitrary scatterer at a time instant t m in T x , can be expressed in a closed form by solving the differential equation system x ˙ ( t m ) = Ω T × x ( t m ) with initial condition x ( 0 ) = y . The result is as follows:
x ( t m ) = a + b cos ( Ω T t m ) + c Ω T sin ( Ω T t m )
where Ω T = | Ω T | , a = Ω T y Ω T 2 Ω T , b = y Ω T y Ω T 2 Ω T , c = Ω T × y .
For values of the observation time lower than 1s, the term x ( t m ) in Equation (3) can be reasonably approximated by its first-order Taylor series at t m = 0 [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18]:
x ( t m ) = y + c t m
where c = [ c 1 , c 2 , c 3 ] .
As a consequence, the distance sum Δ R p ( t m ) = [ Δ R p A V ( t m ) , Δ R p A C ( t m ) , Δ R p A H ( t m ) ] T between the focusing point O and the projection of the scatterer on the LOS in the different channels (AV, AC, and AH) can be written as follows:
Δ R p ( t m ) = ( i L O S x A C + i L O S x A Γ ) x ( t m ) T = K 0 + K 1 t m
where K 0 = [ K 0 A V , K 0 A C , K 0 A H ] T , K 1 = [ K 1 A V , K 1 A C , K 1 A H ] T and
K 0 A V = 2 y 2 d v R 0 ( y 1 sin ϕ + y 3 cos ϕ ) ,   K 1 A V = 2 c 2 d v R 0 ( c 1 sin ϕ + c 3 cos ϕ ) K 0 A C = 2 y 2 ,   K 1 A C = 2 c 2 K 0 A H = 2 y 2 d h R 0 ( y 1 cos ϕ y 3 sin ϕ ) ,   K 1 A H = 2 c 2 d h R 0 ( c 1 cos ϕ c 3 sin ϕ )

2.2. Received Signal Modeling

Assume that the radar transmits a normalized LFM signal, which takes the form of:
s t ( t ^ , t m ) = r e c t ( t ^ T s ) exp ( j 2 π ( f c t ^ + 1 2 γ t ^ 2 ) ) ,
where t ^ is the fast time, T s is the pulse duration, f c is the carrier frequency, γ is the frequency modulation rate, and r e c t ( ) denotes a unit rectangle function. t m = m T r ( m = 0 , 1 , 2 ) is the slow time, where T r is the pulse repetition period. The complex envelope of the received signals from the p t h scattering center on the target is given as follows:
s r ( t ^ , t m ) = δ p r e c t ( t ^ R p ( t m ) c T s ) e x p [ j 2 π f c ( t ^ R p ( t m ) c ) ] e x p [ j π γ ( t ^ R p ( t m ) c ) 2 ]
where s r ( t ^ , t m ) = [ s r A V ( t ^ , t m ) , s r A C ( t ^ , t m ) , s r A H ( t ^ , t m ) ] T is expressed by means of a multichannel vector. When the baselines are short compared with the radar–target distance, δ p can be considered as the same for channels AV, AC, and AH. c is the speed of light, and R p ( t m ) = [ R p A V ( t m ) , R p A C ( t m ) , R p A H ( t m ) ] T denotes the sum of the distance from the p t h scattering center to the transmitter (AC) and the receivers (AV, AC, AH).
As demonstrated in [40], the signal from the focus center of the ISAR image is selected as the reference signal to Dechirp the received signal at the receivers AV, AC, and AH. The reference signal from the reference point can be expressed as:
s r e f ( t ^ , t m ) = r e c t ( t ^ R r e f / c T r e f ) e x p [ j 2 π f c ( t ^ R r e f c ) ] e x p [ j π γ ( t ^ R r e f c ) 2 ]
where T r e f is the duration of the reference signal and T r e f > T s , R r e f 2 R 0 . The received signal after dechirping can be expressed as in [40]:
s i f ( t ^ , t m ) = s r ( t ^ , t m ) × [ s r e f ( t ^ , t m ) ] *
where * denotes the conjugation operation, and s r e f ( t ^ , t m ) = [ s r e f A V ( t ^ , t m ) , s r e f A C ( t ^ , t m ) , s r e f A H ( t ^ , t m ) ] T . In order to obtain well-focused images for moving targets, motion compensation is required for 2D ISAR imaging. However, if the translational motion compensation is made for three channels independently, the ISAR images obtained from three channels will have different focus centers, which may result in pixel misregistration and great difficulties in scattering center extraction. Discussion of coherent translational motion compensation is outside the scope of this paper, and the reader can refer to [41,42] for more details. Assuming that coherent translational motion compensations have been completed, and thus only focusing on the contribution of the rotational motion to ISAR imaging, the signal in Equation (9) can be represented as:
s i f ( t ^ , t m ) = δ p r e c t ( t ^ R r e f / c T r e f ) e x p [ j 2 π γ c ( t ^ R r e f c ) Δ R p ( t m ) ]   × e x p [ j 2 π f c c Δ R p ( t m ) ] e x p [ j π γ c 2 ( Δ R p ( t m ) ) 2 ]
Taking the time of the reference point as a basis, fast Fourier transform (FFT) is performed on Equation (10) with respect to t ^ , obtaining:
S I F ( f r , t m ) = δ p T s sinc [ T s ( f r + γ c Δ R p ( t m ) ) ] exp ( j 2 π f c c Δ R p ( t m ) )   × exp ( j 2 π f r c Δ R p ( t m ) ) exp ( j π γ c 2 ( Δ R p ( t m ) ) 2 )
where S I F ( f r , t m ) = [ S I F A V ( f r , t m ) , S I F A C ( f r , t m ) , S I F A H ( f r , t m ) ] T , f r denotes the range frequency. As Equation (11) demonstrates, the energy of the received signal peaks along the line f r = γ c Δ R p ( t m ) in the frequency domain. After envelope alignment and removing the residual video phase (RVP) by using the compensation function S C ( f r , t m ) = exp ( j π f r 2 / γ ) , the multichannel received signal can be rewritten as follows [40]:
S ( f r , t m ) = S I F ( f r , t m ) S C ( f r , t m ) = δ p T s sin c [ T s ( f r + γ c K 0 ) ] exp [ j 2 π f c c ( K 0 + K 1 t m ) ]
It can be seen from Equation (12) that the radial distance of different scattering centers on the target corresponds with the different frequency range of the received signal after Dechirping. Therefore, through frequency analysis, accurate resolution in the range dimension can be achieved. Similarly, by performing FFT on Equation (12) with respect to slow time t m , the received signal in the range frequency–Doppler domain can be written as follows:
S ( f r , f m ) = δ p T s T a exp [ j 2 π λ K 0 ] × sin c [ T s ( f r + γ c K 0 ) ] × sin c [ T a ( f m + K 1 λ ) ]
where f m denotes the Doppler frequency. In Equation (13), if the same scatterer appears in the same pixel in all three images, the interferometric phase can be calculated directly without need for the coregistration algorithm, which will be analyzed in Section 4.

3. 3D Target Reconstruction Based on the Proposed Algorithm

In Section 2, three 2D ISAR images are obtained by Dechirp processing and a FFT operation. In this section, a novel ISAR imaging algorithm is proposed to reconstruct the 3D target image, which significantly reduces computational cost and avoids image misregistration. The main steps of the proposed algorithm include: (i) obtaining three channel ISAR images with basic signal processing, such as motion compensation and pulse compression by Dechirping and a FFT operation; (ii) applying a cross-channel coherence technique to extract effective scattering center among three ISAR images; (iii) estimating the effective rotation vector Ω e f f and reconstructing the 3D ISAR image.

3.1. Scattering Center Extraction via Cross-Channel Coherence-Based Detector

By considering the received signal model in Section 2, in order to make the following analysis clearer, the ISAR images from the three channels in Equation (13) can be rewritten as:
S ( f r , f m ) = { S A V ( f r , f m ) = δ p T s T a exp [ j 2 π λ K 0 A V ] × sin c [ T s ( f r + γ c K 0 A V ) ] × sin c [ T a ( f m + K 1 A V λ ) ] S A C ( f r , f m ) = δ p T s T a exp [ j 2 π λ K 0 A C ] × sin c [ T s ( f r + γ c K 0 A C ) ] × sin c [ T a ( f m + K 1 A C λ ) ] S A H ( f r , f m ) = δ p T s T a exp [ j 2 π λ K 0 A H ] × sin c [ T s ( f r + γ c K 0 A H ) ] × sin c [ T a ( f m + K 1 A H λ ) ]
Therefore, the phase differences at the peaks of the sin c function can be computed by exploiting the received signal in Equation (14) as follows:
Δ ϑ A V = ( S A V ( f r , f m ) S A C * ( f r , f m ) ) Δ ϑ A H = ( S A H ( f r , f m ) S A C * ( f r , f m ) )
It is known that noise and secondary lobes from prominent scatters influence the interferometric phase of interest. In order to remove these effects, thresholds in power and coherence are jointly used to detect the areas of interest. Normally, only when image pixels exceed both the coherence threshold and the power threshold will the areas be extracted to form the final InISAR image.
The cross-channel coherence among three ISAR images for each resolution cell is computed as follows:
C i j ( f r , f m ) = | S i ( f r , f m ) S j ( f r , f m ) * | | S i ( f r , f m ) | 2 | S j ( f r , f m ) | 2 ,   i j
where i = 1 , 2 , 3 denotes channels AV, AC, and AH respectively. In addition, it can be seen that:
P i ( f r , f m ) = | S i ( f r , f m ) |
Therefore, an effective detection algorithm, known as a cross-channel coherence-based detector (C3D), is defined by:
T ( f r , f m ) = [ i = 1 , i j 3 C i j ( f r , f m ) > C t h ] & [ i = 1 3 P i ( f r , f m ) > P t h ]
where & denotes Boolean product. Thus, extraction of the scattering center from an ISAR image can be performed with such a detection mask. Only image pixels exceeding the coherence threshold C t h and power threshold P t h will contribute to the 3D target reconstruction. Therefore, for each ISAR image, a noise area where the target of interest does not appear can be selected to estimate the noise level P noise . The detection threshold can be obtained by estimating the noise level of each ISAR image. For single channel ISAR image detection, we can set P sin gle = P noise + 15   dB , and the coherence threshold would be prepared in a similar way to the power threshold [21].

3.2. 3D Target Reconstruction via Joint Estimation of Ω e f f

It is observed from Equations (14) and (15) that the phase differences at the peak of the sinc function for horizontal and vertical configurations can be computed as:
Δ ϑ A V = ( S A V ( f r , f m ) S A C * ( f r , f m ) ) = 2 π λ ( K 0 A C K 0 A V ) = 2 π λ d v R 0 ( y 1 sin ϕ + y 3 cos ϕ ) Δ ϑ A H = ( S A H ( f r , f m ) S A C * ( f r , f m ) ) = 2 π λ ( K 0 A C K 0 A H ) = 2 π λ d h R 0 ( y 1 cos ϕ y 3 sin ϕ )
In other words, the coordinates of each scatterer can be expressed as a function of the phase differences. Particularly, the component y 3 denotes the height of the scatterer with respect to the image plane. From this, we can proceed to:
y 1 = λ R 0 2 π ( Δ ϑ A V d v sin ϕ + Δ ϑ A H d h cos ϕ ) y 3 = λ R 0 2 π ( Δ ϑ A V d v cos ϕ Δ ϑ A H d h sin ϕ )
Furthermore, with respect to T x , the term c can be represented as:
c = Ω T × y [ c 1 , c 2 , c 3 ] = | 0 Ω T 2 Ω e f f y 1 y 2 y 3 | c 2 = y 1 Ω e f f
where Ω e f f is the modulus of Ω e f f . It should be noted that the term c 2 can be estimated by considering the Doppler component measured from the AC channel:
f m A C = K 1 A C λ = 2 c 2 λ
Substituting Equations (20) and (21) into Equation (22) yields:
f m A C = R 0 π ( Δ ϑ A V d v Ω e f f sin ϕ + Δ ϑ A H d h Ω e f f cos ϕ )
It is worth noting that the unknown parameters Ω e f f and ϕ can be jointly estimated using Equation (23). For simplicity, Equation (23) can be rewritten by defining the following quantities:
Z = f m A C , Y = R 0 π Δ ϑ A V d v , X = R 0 π Δ ϑ A H d h , α = Ω e f f sin ϕ , β = Ω e f f cos ϕ .
If there are a total of P scatterers extracted from the ISAR images, Equation (24) can be rewritten as:
Z = α Y + β X
where Z = [ Z 1 , Z 2 , , Z P ] T represents the Doppler information of the extracted scatterers for the central channel, Y = [ Y 1 , Y 2 , , Y P ] T and X = [ X 1 , X 2 , , X P ] T correspond to the vectors related to the interferometric phase matrices Δ ϑ A V and Δ ϑ A H only in the effective range and Doppler cells where the scatterers are extracted. In this respect, the elements in terms Z , Y and X are real values and Equation (25) denotes the equation of a plane, which can be done by evaluating the regression plane.
Consequently, the two parameters Ω e f f and ϕ can be estimated by first estimating α and β , and this problem can be solved by minimizing the function:
( α , β ) = p = 1 P [ Z p ( α Y p + β X p ) ]
Finally, the estimation of Ω e f f and ϕ can be calculated from the estimation of the estimated α ^ and β ^ as described by:
Ω ^ e f f = α ^ 2 + β ^ 2 ,   ϕ ^ = a tan ( α ^ β ^ )

4. Performance Analysis of the Proposed Algorithm

In Section 3, a 3D InISAR reconstruction algorithm with Dechirping processing is proposed. There are three crucial points to guarantee good reconstruction performance, namely image registration, computational cost, and soft assignment. In order to accurately define and evaluate the estimation and reconstruction errors, a simple soft assignment method is performed to assign each scatterer of the reconstructed target to the corresponding scatterer of the model.

4.1. Mismatch Problem of Image Registration

In an InISAR imaging system, the interferometric processing among three ISAR images is based on accurate image registration. Due to the symmetry of the AV and AH channels, this paper takes the AV and AC receiver pair as an example to research the mismatching of two ISAR images.
Compared to Equation (14), the pixel positions of the p t h scatterer in the 2D ISAR images from the receivers AV and AC are:
( γ c K 0 A V , 1 λ K 1 A V ) = { γ c ( 2 y 2 d v R 0 ( y 1 sin ϕ + y 3 cos ϕ ) ) , 1 λ ( 2 c 2 d v R 0 ( c 1 sin ϕ + c 3 cos ϕ ) ) } ( γ c K 0 A C , 1 λ K 1 A C ) = { γ c ( 2 y 2 ) , 1 λ ( 2 c 2 ) }
From Equation (28), it can be found that the baseline and direction of the effective rotation vector directly result in mismatches in the range frequency–Doppler domain. The range frequency offset in the range direction is:
Δ f r = γ c ( K 0 A V K 0 A C ) = γ c d v R 0 ( y 1 sin ϕ + y 3 cos ϕ )
Thus, in a pulse repetition period, the number of the range resolution cell corresponding to Δ f r is:
Δ n = Δ f r T s = B c d v R 0 ( y 1 sin ϕ + y 3 cos ϕ )
where B stands for the bandwidth of the transmitted signal.
Equally, the Doppler offset in the cross-range direction is:
Δ f m = 1 λ ( K 1 A V K 1 A C ) = 1 λ d v R 0 ( c 1 sin ϕ + c 3 cos ϕ )
Thus, in ISAR imaging observation time, the number of the cross-range resolution cell corresponding to Δ f m is:
Δ m = Δ f m T a = T a λ d v R 0 ( c 1 sin ϕ + c 3 cos ϕ )
Based on the above analysis, the offset between two ISAR images from different interferometric receivers is correlated with the initial distance, baseline length, target’s rotation, and other parameters. Therefore, according to the geometric relationship between the radar and the target, a quantitative analysis of the offset is given below.
For convenience, the parameters in this paper are utilized to evaluate the image mismatch. Suppose that in a far-field condition, R 0 = 10   km , d v = d h = 1   m , B = 200   Hz , f c = 10   GHz , T a = 0.5   s ,   ϕ = 45 , and the target size does not exceed 60 m. In this condition:
Δ n = B c d v R 0 ( y 1 sin ϕ + y 3 cos ϕ ) < 1 Δ m = T a λ d v R 0 ( c 1 sin ϕ + c 3 cos ϕ ) < 1
where c = [ c 1 , c 2 , c 3 ] is related to the total target rotation vector Ω T , which will not be too large [43].
Obviously, in a far-field condition, image misregistration does not appear after Dechirp processing and Fourier transformation in the range frequency–Doppler dimension. However, the mismatch problem in InISAR imaging may occur when the baseline is too long or the target’s size is too large. In this case, some image registration methods are needed to solve or mitigate the misregistration, which is beyond the scope of this paper.

4.2. Analysis of the Computational Cost

Based on the performance analysis above, the computational burden for 3D target reconstruction mainly results from dominant scatterer extraction. Here, assume that the number of the resolution cell in the range frequency–Doppler dimension and the number of the scatterers on the target are N and P , respectively.
For a 3D InISAR imaging algorithm based on the joint cross modified Wigner–Ville distribution (jc-MWVD) in [30], its implementation procedure in a certain range cell included the defined symmetric cross-correlation function [ N 2 ] , scaled transform [ 3 N 2 l o g 2 N ] , and Fourier transform [ N 2 l o g 2 N ] . Obviously, the computational cost of the algorithm proposed in [30] is on the order of [ ( N 2 + 4 N 2 l o g 2 N ) P ] .
For the method in [29], due to the application of the searching procedure, the computational cost is on the order of [ ( N 2 + M N l o g 2 N ) P ] , where M is the number of searching points and always greater than N for high-resolution imaging [6]. In this method, the searching steps and initial parameters are difficult to control. Therefore, this method is less suitable for realistic application.
In [18], the MC-CLEAN technique is based on the optimization problem and iteration, so the computational cost will increase linearly with the total number of scatterers. As the number of the scatterers on the target (such as an airplane or ship) may be up to hundreds in practice, this may result in a high computational burden.
Although the method in [21] does not use a searching procedure with respect to scattering center extraction, this method still requires approximately [ ( 6 N 2 ( N p 3 + 1 ) + 5 N 2 ) ] multiplication operations due to a N p × N p moving window.
The proposed C3D only employs multiply–accumulate operations (MAs) to extract the dominant scatterers. The procedure for cross-channel coherence among three ISAR images includes [ 12 N 2 ] MAs in Equations (16) and (17). The mask in Equation (18) requires [ 5 N 2 ] MAs. The computational cost is on the order of [ 12 N 2 + 5 N 2 ] . More importantly, the performance of the C3D is not affected by the total number of scatterers.
Under the assumptions of M = 8 N , N p = 4 , Figure 3 shows the computational complexities of the proposed method and three other methods. From Figure 3a, it can be observed that the computational costs of the three other methods increases rapidly with increases in the number of the resolution cell in the range frequency–Doppler dimension. Meanwhile, compared with the method in [29] and the jc-MWVD in Figure 3b, the computational cost of the proposed method is independent of the total number of scatterers, which makes this method better than the other methods of multi-scatterer target imaging. Based on Figure 3, it can be concluded that the proposed method is more computationally efficient and has a wider applicability than the other three methods in ISAR imaging applications.

4.3. Soft Assignment Method

In order to assign each scatterer of the 3D reconstructed target to the corresponding scatterer of the model, a soft assignment method based on the probabilistic least squares (PLS) approach [44] is proposed. Accordingly, the soft assignment matrix of dimensions P × K can be defined by:
κ = [ κ 1 , 1 κ 1 , K κ p , k κ P , 1 κ P , K ]
where P and K are the number of the scatterers in the model and the number of the extracted scatterers, respectively. Each element of the matrix κ p , k denotes the probability that the k t h extracted scatterer belongs to the p t h scatterers of the model. The sum of all possible assignments for a given scatterer is equal to the unity. The solution for the soft assignment problem is given as follows:
κ p , k = ( ε p , k T ε p , k ) 1 l = 1 P ( ε l , k T ε l , k ) 1 , p = 1 P κ p , k = k = 1 , K
where ε p , k is the Euclidean distance between the p t h scatterer of the model and the k t h extracted scatterer of the reconstructed target.
Therefore, each extracted scatterer can be assigned to the scatterer of the reference model with highest probability. In order to guarantee the precision of soft assignment, we can easily distinguish the scatterer of the reference model and the matching one by setting the appropriate probability threshold Λ κ . In this way, the corresponding height and distance errors between the reference model and the reconstructed target, which are used to evaluate the performance of the proposed algorithm, can be calculated. Since the sum of all possible assignments for a given scatterer adds to the unity, the maximum value of each column is used here as the assignment indicator, namely Λ κ = max [ κ k ] , k = 1 K . In real world applications, when signal-to-noise ratio (SNR) is very low, each value of κ k tends to be decentralized, which is detrimental to soft assignment.

5. Simulation Results

In this section, simulated radar data will be adopted to verify the effectiveness of the proposed 3D InISAR imaging algorithm. To simulate a given SNR, Gaussian noise is added to the raw data (in the data domain). Several simulations were run to analyze the performance of our target reconstruction method.

5.1. Verification with Simple Airplane Model

The simulated 3D model of the airplane target, which is composed of P = 31 ideal scatterers, is shown in Figure 4a. The simulation parameters are listed in Table 1. Based on the traditional range frequency–Doppler (RD) algorithm in Equation (14), Figure 4b gives the 2D ISAR imaging result from central channel AC.
Figure 5 and Figure 6 provide a clearer and simpler visualization, in which the positions of the scatterers on the airplane are depicted in the planes y 1 y 2 , y 2 y 3 , and y 1 y 3 . The reconstructed target is superimposed onto the model so as to present a more direct viewing comparison. The regression planes of parameter estimation in Equation (25) are shown in Figure 5a and Figure 6a, from which the estimates of Ω e f f and ϕ can be indirectly obtained by estimating the coefficients of the multiple linear regression [45]. In Figure 5b and Figure 6b, it is evident by visual inspection that the 3D targets have been reconstructed, while, for baseline d v = d h = 0.5   m , the reconstruction result in Figure 5b is not as good as that in Figure 6b, where the baseline length is 4 m and the reconstructed target matches the reference model perfectly. Below, the performance of the proposed algorithm is analyzed with respect to different SNR and different baseline length.
In the given scenarios, Figure 7 shows the behaviors of the mean distance error and its standard deviation (SD) expressed in meters with respect to the SNR and the baseline length. The same performance analysis for height is depicted in Figure 8. In particular, the algorithm performance for five different baseline lengths is tested with increasing SNR, and the raw data is contaminated with zero-mean white Gaussian noise. The mean error χ ¯ and the SD δ for distance and height are defined by:
χ ¯ Θ = 1 S s = 1 S [ χ Θ ( s ) ] ,   δ Θ = 1 S s = 1 S [ χ Θ ( s ) χ ¯ Θ ] 2 Θ = D , H
where S is the number of Monte Carlo, and Θ denotes the distance or height error. After soft assignment, each scatterer of the reconstructed target is assigned to the corresponding scatterer of the model:
χ D ( s ) = k = 1 K ε D ( s , k ) ,   χ H ( s ) = k = 1 K | h m ( s , k ) h r ( s , k ) |
where ε D ( s , k ) is the Euclidean distance between the scatterer of the model and corresponding scatterer of the reconstructed target. h m ( s , k ) and h r ( s , k ) are the height with respect to the image plane referring to the model and to the reconstructed target, respectively. The effective rotation vector can be obtained via:
χ ¯ = 1 S s = 1 S | ^ ( s ) | ,   δ = 1 S s = 1 S [ ^ ( s ) E ( ^ ( s ) ) ] 2 , = Ω e f f , ϕ
where E (   ) indicates the expectation. Ω e f f and ϕ are a priori known.
It is known that a shorter baseline will produce larger phase errors in measurement and decrease the accuracy of the height estimation [19]. Therefore, larger distance errors and height estimation errors will occur with a decreasing baseline. In Figure 7 and Figure 8, the curves of the mean distance error, mean height error, and their standard deviations follow a decreasing trend when the SNR or the baseline length increases, which is in accordance with theoretical predictions.
Consistent with the previous conclusions, the mean errors and standard deviations of the estimates of Ω e f f and ϕ are given in Figure 9 and Figure 10, where the same decreasing pattern can be observed with increasing SNR and increasing baseline length. Compared to the method in [18], the same results can be obtained using the proposed method without any search processing or iteration procedures. Although the results above demonstrate that the longer baseline contributes to minimizing phase error measurement, there are conflicting requirements on the baseline length. According to Equation (19), the phase difference between the return echoes received by three antennas is a periodic function with period 2 π . As a result, the altitude measurement is unambiguous if the phase difference exceeds 2 π .

5.2. Verification with Complex Ship Model

In order to simulate a realistic echo return, the ship target model in Figure 11, whose dimensions are approximately 90 ( L ) × 27 ( W ) × 27 ( H ) m , is shaped by 1274 scattering points depicted with their reflection coefficients and positions in the target coordinate system. The radar parameters are shown in Table 2.
Similar to [14], this paper uses a standard tessellation language (STL) model shown in Figure 12 as an input to the simulation. Figure 13 shows the 2D ISAR images created via Equation (14) by processing the signal acquired from channels AV, AC, and AH, where the 2D ISAR images of the ship target are quite clearly visible. In Figure 13, the bounding box indicates a noise area of interest where the target is not present, which can be used to estimate the SNR of the bounding box to determine the power detection threshold. Based on the analysis and computation of the SNR of the bounding box in three ISAR images, values of P noise , 1 = 19.84   dB , P noise , 2 = 19.79   dB , and P noise , 3 = 19.82   dB are obtained.
The interferograms in Figure 14 are computed by processing the channel pairs AV–AC and AH-AC, where noise is present randomly, whereas the phase values of the target are scattered in an orderly manner and within the interval [ π , π ] , indicating that no phase wrapping occurs. The phenomena of the phase wrapping may occur when the phase differences in Equation (19) exceed 2 π . From Equation (19), it is known that the interferometric phases are mainly determined by the target size, wavelength, and baseline length. Thus, users of the algorithm can set appropriate parameters to avoid this problem and refer to measures to deal with phase wrapping [18].
By observing Figure 13, the SNRs estimated from the three ISAR images are P noise , 1 = 19.84   dB , P noise , 2 = 19.79   dB , and P noise , 3 = 19.82   dB , respectively. For three-channel coherence processing in this simulation, the threshold in image power depicted in Figure 15 can be obtained by calculating P t h = P noise , 1 + P noise , 2 + P noise , 3 + 45   dB . Thereby, the coherence threshold and power threshold can be set as C t h 15   dB , P t h 15   dB to remove the low SNR resolution cells.
The C3D detection mask from Equation (18) and 2D ISAR image received from the AC channel after detection mask processing are depicted in Figure 16a,b respectively. In Figure 16b, areas of no interest are removed and the effective information is preserved, after which the detected interferograms are passed to the 3D reconstruction process.
For comparison, the left column of Figure 17 shows the ideal model of the ship target in Figure 11, and the 3D InISAR reconstructed results are shown in the right column of Figure 17. The reconstructed ship target’s size almost matches the actual one depicted in Figure 17.
Figure 17 verifies the viability of the proposed InISAR imaging algorithm. In Figure 18, an InISAR image based on the methods in [18,21] is chosen to compare with the imaging result based on the proposed method under the condition of SNR = 0 dB. The simulation times of the proposed 3D InISAR imaging method and the methods in [18] and [21] are 0.2 s and 12.6 s, demonstrating that our proposed algorithm is more efficient than the other methods. The reason is that the proposed C3D-based InISAR imaging algorithm employs holistic multiply–accumulate operations to extract the dominant scatterers instead of iterating or searching for pixels.
In order to quantitatively evaluate the performance of InISAR imaging, the relative mean error (RME) of the estimator is defined by:
R M E = 1 H h = 1 H | ϒ ^ h ϒ ϒ | where ϒ = ϕ , y 1 o r y 3 ,
where ϒ ^ and ϒ represent the estimated parameters and given parameters, respectively. H denotes the total number of extracted effective scattering points, and h denotes the h t h effective scattering point. The tested input SNRs were S N R = [ 5 , 0 , 5 ]   dB , and 400 trials were performed for each SNR value.
After soft assignment, the RMEs of ϕ ^ , y ^ 1 -coordinate, and y ^ 3 -coordinate using the proposed method and the method in [21] are shown in Table 3. As depicted, the RMEs of the estimated parameters with the proposed method in this paper are relatively small and within the acceptable range in a real scenario [46], which further demonstrates the effectiveness of our InISAR imaging algorithm. Theoretically, test losses may arise from approximation operations during the modeling of the echo signal and the extraction of the scattering center. Compared with the method in [21], the proposed method can acquire higher precision in parameter estimation and coordinate reconstruction. Due to the influence of error propagation in Equation (20), the RMEs of the reconstructed coordinates are larger than those of the rotational parameter estimation.
There are several artificial points in Figure 17 and Figure 18, which may be induced by noise or sidelobes of some scatterers. Therefore, in the future, more attention should be paid to reducing the effects of noise and sidelobes.

6. Conclusions

In this paper, a novel 3D target reconstruction algorithm based on an L-shape InISAR imaging system is presented, which makes use of Dechirp processing and interferometry. This imaging system consists of three antennas lying on two orthogonal baselines. The analytical expressions for estimating a target’s effective rotation vector and extracting 3D coordinates of the target are derived in detail. Moreover, a nonsearching detection algorithm, referred to as C3D, is introduced to remove the effects of noise and sidelobes while retaining areas of interest where the total power and coherence derived from the three channels are above thresholds.
Given the significance of pixel registration in 3D InISAR imaging, the paper also discusses and quantitatively analyzes the mismatching of two ISAR images and their influence on target reconstruction. In order to verify the robustness of the proposed algorithm, several experiments are conducted using simulated data in different scenarios. The simulation results are consistent with the theory that performance improves with increasing SNR and baseline length. At the same time, the present algorithm is proved to reconstruct the 3D coordinates of multi-scattering point targets with high accuracy and efficiency in the case of low signal-to-noise ratio, which is very suitable for practical applications.
Finally, the application and effectiveness of our InISAR imaging algorithm is validated by means of a simulated ship, with satisfactory results.

Author Contributions

Conceptualization, Q.L. and S.Z.; methodology, S.Z.; software, Q.L.; validation, Q.L. and S.Z.; formal analysis, Q.L.; investigation, Q.L.; resources, S.Z.; data curation, Q.L.; writing—original draft preparation, Q.L.; writing—review and editing, S.Z.; visualization, Q.L.; supervision, Q.L.; project administration, Q.L.; funding acquisition, S.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Basic Research Program of Shaanxi, grant number 2019JQ-669, and the Fundamental Research Funds for the Central Universities, grant number GK201903018.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Berizzi, F.; Mese, E.D.; Diani, M.; Martorella, M. High-resolution ISAR imaging of maneuvering targets by means of the range instantaneous Doppler technique: Modeling and performance analysis. IEEE Trans. Image Process. 2011, 10, 1880–1890. [Google Scholar] [CrossRef]
  2. Lv, Q.; Su, T.; He, X. An ISAR Imaging Algorithm for Nonuniformly Rotating Targets With Low SNR Based on Modified Bilinear Parameter Estimation of Cubic Phase Signal. IEEE Trans. Aerosp. Electron. Syst. 2018, 54, 3108–3124. [Google Scholar] [CrossRef]
  3. Ausherman, D.A.; Kozma, A.; Walker, J.L.; Jones, H.M.; Poggio, E.C. Developments in radar imaging. IEEE Trans. Aerosp. Electron. Syst. 1984, 20, 363–400. [Google Scholar] [CrossRef]
  4. Zhang, S.; Liu, Y.; Li, X. Fast Sparse Aperture ISAR Autofocusing and imaging via ADMM based Sparse Bayesian Learning. IEEE Trans. Image Process. 2019, 29, 3213–3226. [Google Scholar] [CrossRef] [PubMed]
  5. Zheng, J.; Liu, H.; Liu, J.; Du, X.; Liu, Q.H. Radar high-speed maneuvering target detection based on three-dimensional scaled transform. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2018, 11, 2821–2833. [Google Scholar] [CrossRef]
  6. Zuo, L.; Wang, B. ISAR Imaging of Non-Uniform Rotating Targets Based on Optimized Matching Fourier Transform. IEEE Access 2020, 8, 64324–64330. [Google Scholar] [CrossRef]
  7. Li, X.; Cui, G.; Yi, W.; Kong, L. Radar maneuvering target detection and motion parameter estimation based TRT-SGRFT. Signal Process. 2017, 133, 107–116. [Google Scholar] [CrossRef]
  8. Ma, C.; Yeo, T.S.; Tan, C.S.; Liu, Z. Three-dimensional imaging of targets using colocated MIMO radar. IEEE Trans. Geosci. Remote Sens. 2011, 49, 3009–3021. [Google Scholar] [CrossRef]
  9. Liu, Y.; Song, M.; Wu, K.; Wang, R.; Deng, Y. High-quality 3-D InISAR imaging of maneuvering target based on a combined processing approach. IEEE Geosci. Remote Sens. Lett. 2013, 10, 1036–1040. [Google Scholar] [CrossRef]
  10. Kang, B.S.; Lee, K.; Kim, K.T. Image Registration for 3-D Interferometric-ISAR Imaging Through Joint-Channel Phase Difference Functions. IEEE Trans. Aerosp. Electron. Syst. 2021, 57, 22–38. [Google Scholar] [CrossRef]
  11. Pastina, D.; Spina, C. Slope-based frame selection and scaling technique for ship ISAR imaging. IET Signal Process 2008, 2, 265–276. [Google Scholar] [CrossRef]
  12. Xu, J.; Teng, L.; Liu, F.; Xia, X.G.; Wang, Y.P.; Peng, S.B.; Peng, Y.N. Radar target imaging using three-dimensional space Radon-Fourier transform. In Proceedings of the 2014 International Radar Conference, Lille, France, 13–17 October 2014; pp. 1–6. [Google Scholar] [CrossRef]
  13. Martorella, M.; Salvetti, F.; Stagliano, D. 3D target reconstruction by means of 2D-ISAR imaging and interferometry. In Proceedings of the 2013 IEEE Radar Conference, Ottawa, ON, Canada, 29 April–3 May 2013; pp. 1–6. [Google Scholar] [CrossRef]
  14. Felguera-Martin, D.; Gonzalez-Partida, J.T.; Burgos-Garcia, M. Interferometric ISAR imaging on maritime target application: Simulation of realistic targets and dynamics. Prog. Electromagn. Res. 2012, 132, 571–586. [Google Scholar] [CrossRef] [Green Version]
  15. Battisti, N.; Martorella, M. Interferometric phase and target motion estimation for accurate 3D reflectivity reconstruction in ISAR systems. In Proceedings of the 2010 IEEE Radar Conference, Arlington, VA, USA, 10–14 May 2010; pp. 108–112. [Google Scholar] [CrossRef]
  16. Zhao, L.; Gao, M.; Martorella, M.; Stagliano, D. Bistatic three-dimensional interferometric ISAR image reconstruction. IEEE Trans. Aerosp. Electron. Syst. 2015, 51, 951–961. [Google Scholar] [CrossRef]
  17. Zhang, Q.; Yeo, T.S.; Du, G.; Zhang, S. Estimation of three-dimensional motion parameters in interferometric ISAR imaging. IEEE Trans. Geosci. Remote Sens. 2004, 42, 292–300. [Google Scholar] [CrossRef]
  18. Martorella, M.; Stagliano, D.; Salvetti, F.; Battisti, N. 3D interferometric ISAR imaging of noncooperative targets. IEEE Trans. Aerosp. Electron. Syst. 2014, 50, 3102–3114. [Google Scholar] [CrossRef]
  19. Staglianò, D.; Giusti, E.; Lischi, S.; Martorella, M. Bistatic three-dimensional interferometric ISAR. IET Radar Sonar Navig. 2016, 10, 63–75. [Google Scholar] [CrossRef]
  20. Staglianò, D.; Giusti, E.; Lischi, S.; Martorella, M. 3D InISAR-based target reconstruction algorithm by using a multi-channel ground-based radar demonstrator. In Proceedings of the 2014 International Radar Conference, Lille, France, 13–17 October 2014; pp. 1–6. [Google Scholar] [CrossRef]
  21. Stagliano, D.; Lischi, S.; Massini, R.; Musetti, L.; Martorella, M.; Berizzi, F. Soft 3D-ISAR image reconstruction using a dual interferometric radar. In Proceedings of the 2013 IEEE Radar Conference, Arlington, VA, USA, 10–15 May 2015; pp. 0572–0576. [Google Scholar] [CrossRef]
  22. Lischi, S.; Massini, R.; Staglianó, D.; Musetti, L.; Berizzi, F.; Neri, B.; Martorella, M. X-band compact low cost multi-channel radar prototype for short range high resolution 3D-InISAR. In Proceedings of the 2014 Radar Conference (EuRAD), Rome, Italy, 8–10 October 2014; pp. 157–160. [Google Scholar] [CrossRef]
  23. Wang, Y.; Li, X. Three-Dimensional Interferometric ISAR Imaging for the Ship Target Under the Bi-Static Configuration. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2016, 9, 1505–1520. [Google Scholar] [CrossRef]
  24. Rong, J.; Wang, Y.; Han, T. Interferometric ISAR Imaging of Maneuvering Targets With Arbitrary Three-Antenna Configuration. IEEE Trans. Geosci. Remote Sens. 2020, 58, 2020. [Google Scholar] [CrossRef]
  25. Wang, Y.; Rong, J.; Han, T. Novel Approach for High Resolution ISAR/InISAR Sensors Imaging of Maneuvering Target Based on Peak Extraction Technique. IEEE Sens. J. 2019, 19, 5541–5558. [Google Scholar] [CrossRef]
  26. Wang, Y.; Chen, X. Three-Dimensional Interferometric ISAR Sensors Imaging for the Ship Target with Two-Dimensional Sparsity. J. Harbin Inst. Technol. 2019, 26, 23–35. [Google Scholar]
  27. Stuff, M.; Biancalana, M.; Arnold, G.; Garbarino, J. Imaging moving objects in 3D from single aperture synthetic aperture radar. In Proceedings of the 2004 IEEE Radar Conference, Philadelphia, PA, USA, 29 April 2004; pp. 94–98. [Google Scholar] [CrossRef]
  28. Liu, Y.; Li, N.; Wang, R.; Deng, Y. Achieving high-quality three-dimensional InISAR imageries of maneuvering target via super-resolution ISAR imaging by exploiting sparseness. IEEE Geosci. Remote Sens. Lett. 2014, 11, 828–832. [Google Scholar] [CrossRef]
  29. Wang, G.; Xia, X.G.; Chen, V.C. Three-dimensional ISAR imaging of maneuvering targets using three receivers. IEEE Trans. Image Process. 2001, 10, 436–447. [Google Scholar] [CrossRef] [PubMed]
  30. Lv, Q.; Su, T.; Zheng, J.; Zhang, J. Three-dimensional interferometric inverse synthetic aperture radar imaging of maneuvering target based on the joint cross modified Wigner-Ville distribution. J. Appl. Remote Sens. 2016, 10, 015007. [Google Scholar] [CrossRef] [Green Version]
  31. Ma, C.; Yeo, T.S.; Zhang, Q.; Tan, H.S.; Wang, J. Three-dimensional ISAR imaging based on antenna array. IEEE Trans. Geosci. Remote Sens. 2008, 46, 504–515. [Google Scholar] [CrossRef]
  32. Ma, C.; Yeo, T.S.; Tan, H.S.; Wang, J.; Chen, B. Three-dimensional ISAR imaging using a two-dimensional sparse antenna array. IEEE Trans. Geosci. Remote Sens. Lett. 2008, 5, 378–382. [Google Scholar] [CrossRef]
  33. Ma, C.; Yeo, T.S.; Tan, C.S.; Tan, H.S. Sparse array 3-D ISAR imaging based on maximum likelihood estimation and CLEAN technique. IEEE Trans. Image Process. 2010, 19, 2127–2142. [Google Scholar] [CrossRef]
  34. Felguera-Martín, D.; González-Partida, J.T.; Almorox-González, P.; Burgos-García, M.; Dorta-Naranjo, B.P. Interferometric inverse synthetic aperture radar experiment using an interferometric linear frequency modulated continuous wave millimeter-wave radar. IET Radar Sonar Navig. 2011, 5, 39–47. [Google Scholar] [CrossRef] [Green Version]
  35. Xu, G.; Xing, M.; Xia, X.; Zhang, L.; Chen, Q.; Bao, Z. 3D Geometry and Motion Estimations of Maneuvering Targets for Interferometric ISAR With Sparse Aperture. IEEE Trans. Image Process. 2016, 25, 2005–2020. [Google Scholar] [CrossRef] [PubMed]
  36. Wang, Y.; Kang, J.; Jiang, Y. ISAR imaging of maneuvering target based on the local polynomial Wigner distribution and integrated high-order ambiguity function for cubic phase signal model. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2014, 7, 2971–2991. [Google Scholar] [CrossRef]
  37. Wang, Y.; Zhao, B.; Kang, J. Asymptotic statistical performance of local polynomial Wigner distribution for the parameters estimation of cubic-phase signal with application in ISAR imaging of ship target. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2015, 8, 1087–1098. [Google Scholar] [CrossRef]
  38. Rong, J.; Wang, Y.; Han, T. Iterative Optimization-Based ISAR Imaging With Sparse Aperture and Its Application in Interferometric ISAR Imaging. IEEE Sens. J. 2019, 19, 8681–8693. [Google Scholar] [CrossRef]
  39. Wang, J.; Cai, D.; Wen, Y. Comparison of matched filter and Dechirp processing used in Linear Frequency Modulation. In Proceedings of the 2011 IEEE 2nd International Conference on Computing, Control and Industrial Engineering, Wuhan, China, 20–21 August 2011; pp. 70–73. [Google Scholar] [CrossRef]
  40. Zhang, S.; Zhang, W.; Zong, Z.; Tian, Z.; Yeo, T.S. High-resolution bistatic ISAR imaging based on two-dimensional compressed sensing. IEEE Trans. Image Process. 2015, 63, 2098–2111. [Google Scholar] [CrossRef]
  41. Pu, W. Shuffle GAN With Autoencoder: A Deep Learning Approach to Separate Moving and Stationary Targets in SAR Imagery. IEEE Trans. Neural Netw. Learn. Syst. 2021. [Google Scholar] [CrossRef] [PubMed]
  42. Pu, W. Deep SAR Imaging and Motion Compensation. IEEE Trans. Image Process. 2021, 30, 2232–2247. [Google Scholar] [CrossRef]
  43. Wang, Y.; Jiang, Y. A novel algorithm for estimating the rotation angle in ISAR imaging. IEEE Geosci. Remote Sens. Lett. 2008, 5, 608–609. [Google Scholar] [CrossRef]
  44. Giusti, E.; Martorella, M.; Capria, A. Polarimetrically-persistent-scatterer-based automatic target recognition. IEEE Trans. Geosci. Remote Sens. 2011, 49, 4588–4599. [Google Scholar] [CrossRef]
  45. Chatterjee, S.; Hadi, A.S. Influential observations, high leverage points, and outliers in linear regression. Stat. Sci. 1986, 1, 379–416. [Google Scholar]
  46. Wu, Y.; Zhang, S.; Kang, H.; Yeo, T.S. Fast marginalized sparse bayesian learning for 3-D interferometric ISAR image formation via super-resolution ISAR imaging. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2015, 8, 4942–4951. [Google Scholar] [CrossRef]
Figure 1. Overall flowchart of 3D target reconstruction with the proposed algorithm.
Figure 1. Overall flowchart of 3D target reconstruction with the proposed algorithm.
Sensors 21 05073 g001
Figure 2. 3D InISAR system. (a) General scene for InISAR imaging. (b) Geometry of radar and target of interest.
Figure 2. 3D InISAR system. (a) General scene for InISAR imaging. (b) Geometry of radar and target of interest.
Sensors 21 05073 g002
Figure 3. Computational cost comparison. (a) Computational cost versus resolution cell ( P = 31 ). (b) Computational cost versus total scatterers ( N = 512 ).
Figure 3. Computational cost comparison. (a) Computational cost versus resolution cell ( P = 31 ). (b) Computational cost versus total scatterers ( N = 512 ).
Sensors 21 05073 g003
Figure 4. 3D airplane model and ISAR image. (a) 3D airplane model. (b) Traditional 2D ISAR imaging result.
Figure 4. 3D airplane model and ISAR image. (a) 3D airplane model. (b) Traditional 2D ISAR imaging result.
Sensors 21 05073 g004
Figure 5. 3D reconstruction results d v , d h = 0.5   m . (a) Regression plane of parameter estimation in Equation (25). (b) 3D reconstruction result of the airplane. (c) Range–horizontal direction view ( y 1 y 2 ). (d) Range–vertical direction view ( y 2 y 3 ). (e) Vertical–horizontal direction view ( y 1 y 3 ).
Figure 5. 3D reconstruction results d v , d h = 0.5   m . (a) Regression plane of parameter estimation in Equation (25). (b) 3D reconstruction result of the airplane. (c) Range–horizontal direction view ( y 1 y 2 ). (d) Range–vertical direction view ( y 2 y 3 ). (e) Vertical–horizontal direction view ( y 1 y 3 ).
Sensors 21 05073 g005
Figure 6. 3D reconstruction results d v , d h = 4   m . (a) Regression plane of parameter estimation in Equation (25). (b) 3D reconstruction result of the airplane. (c) Range–horizontal direction view ( y 1 y 2 ). (d) Range–vertical direction view ( y 2 y 3 ). (e) Vertical–horizontal direction view ( y 1 y 3 ).
Figure 6. 3D reconstruction results d v , d h = 4   m . (a) Regression plane of parameter estimation in Equation (25). (b) 3D reconstruction result of the airplane. (c) Range–horizontal direction view ( y 1 y 2 ). (d) Range–vertical direction view ( y 2 y 3 ). (e) Vertical–horizontal direction view ( y 1 y 3 ).
Sensors 21 05073 g006
Figure 7. Simulation result. (a) Mean distance error. (b) Standard deviation of distance error.
Figure 7. Simulation result. (a) Mean distance error. (b) Standard deviation of distance error.
Sensors 21 05073 g007
Figure 8. Simulation result. (a) Mean distance error. (b) Standard deviation of distance error.
Figure 8. Simulation result. (a) Mean distance error. (b) Standard deviation of distance error.
Sensors 21 05073 g008
Figure 9. Simulation result. (a) Mean error of estimate of Ω e f f . (b) Standard deviation of estimate of Ω e f f .
Figure 9. Simulation result. (a) Mean error of estimate of Ω e f f . (b) Standard deviation of estimate of Ω e f f .
Sensors 21 05073 g009
Figure 10. Simulation result. (a) Mean error of estimate of ϕ . (b) Standard deviation of estimate of ϕ .
Figure 10. Simulation result. (a) Mean error of estimate of ϕ . (b) Standard deviation of estimate of ϕ .
Sensors 21 05073 g010
Figure 11. Ship target of interest.
Figure 11. Ship target of interest.
Sensors 21 05073 g011
Figure 12. Three-dimensional tessellation ship target model. (a) Top view. (b) Front view. (c) Left view.
Figure 12. Three-dimensional tessellation ship target model. (a) Top view. (b) Front view. (c) Left view.
Sensors 21 05073 g012
Figure 13. ISAR imaging results. (a) AV-Channel. (b) AC-Channel. (c) AH-Channel.
Figure 13. ISAR imaging results. (a) AV-Channel. (b) AC-Channel. (c) AH-Channel.
Sensors 21 05073 g013
Figure 14. Interferograms. (a) AC–AV. (b) AC–AH.
Figure 14. Interferograms. (a) AC–AV. (b) AC–AH.
Sensors 21 05073 g014
Figure 15. Coherence and power maps. (a) Power map. (b) Coherence map.
Figure 15. Coherence and power maps. (a) Power map. (b) Coherence map.
Sensors 21 05073 g015
Figure 16. Detected mask and results after detection mask processing. (a) Detected mask. (b) 2D ISAR image received from the AC channel after detection processing.
Figure 16. Detected mask and results after detection mask processing. (a) Detected mask. (b) 2D ISAR image received from the AC channel after detection processing.
Sensors 21 05073 g016
Figure 17. Model and reconstructed results using the proposed method. (a) Range–horizontal direction view of ship model. (b) Range–horizontal direction view of reconstructed result. (c) Range–vertical direction view of ship model. (d) Range–vertical direction view of reconstructed result. (e) Vertical–horizontal direction view of ship model. (f) Vertical–horizontal direction view of reconstructed result.
Figure 17. Model and reconstructed results using the proposed method. (a) Range–horizontal direction view of ship model. (b) Range–horizontal direction view of reconstructed result. (c) Range–vertical direction view of ship model. (d) Range–vertical direction view of reconstructed result. (e) Vertical–horizontal direction view of ship model. (f) Vertical–horizontal direction view of reconstructed result.
Sensors 21 05073 g017
Figure 18. The reconstructed results using the method in [21] and the proposed method. (a) Range–horizontal direction view of result using the method in [21]. (b) Range–horizontal direction view of proposed method results. (c) Range–vertical direction view of result using the method in [21]. (d) Range–vertical direction view of proposed method results. (e) Vertical–horizontal direction view of result using the method in [21]. (f) Vertical–horizontal direction view of proposed method results.
Figure 18. The reconstructed results using the method in [21] and the proposed method. (a) Range–horizontal direction view of result using the method in [21]. (b) Range–horizontal direction view of proposed method results. (c) Range–vertical direction view of result using the method in [21]. (d) Range–vertical direction view of proposed method results. (e) Vertical–horizontal direction view of result using the method in [21]. (f) Vertical–horizontal direction view of proposed method results.
Sensors 21 05073 g018
Table 1. Simulation Parameters—Airplane.
Table 1. Simulation Parameters—Airplane.
Carrier Frequency10 GHzBaseline Length d v   =   0.5   m d h   =   0.5   m } in   Figure   5 d v   =   4.0   m d h   =   4.0   m } in   Figure   6
F s 200 MHzRadar-target distance10 km
Bandwidth200 MHzTarget velocity60 m/s
Pulsewidth2.56 us SNR0 dB
T o b s 0.5 sPitch/Roll/Yaw amplitude ( )[1,2,10]
PRF512 HzPitch/Roll/Yaw period ( s )[5,8,9]
Table 2. Simulation Parameters—Ship.
Table 2. Simulation Parameters—Ship.
Carrier Frequency10 GHzBaseline Length d v   =   2 m d h   =   2 m
F s 200 MHzRadar-target distance16 km
Bandwidth200 MHzTarget velocity30 m/s
Pulsewidth2.56 us SNR10 dB
T o b s 1 sPitch/Roll/Yaw amplitude ( )[1,1,5]
PRF512 HzPitch/Roll/Yaw period ( s )[6,8,10]
Table 3. Reconstruction performance comparison.
Table 3. Reconstruction performance comparison.
RME (%) 5   d B 0   d B 5   d B
ProposedMethod in [21]ProposedMethod in [21]ProposedMethod in [21]
ϕ ^ 0.420.541.081.612.894.63
y ^ 1 12.2112.9015.0918.0921.2828.61
y ^ 3 4.626.836.3810.6010.2016.37
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lv, Q.; Zhang, S. Three-Dimensional Interferometric ISAR Imaging Algorithm Based on Cross Coherence Processing. Sensors 2021, 21, 5073. https://0-doi-org.brum.beds.ac.uk/10.3390/s21155073

AMA Style

Lv Q, Zhang S. Three-Dimensional Interferometric ISAR Imaging Algorithm Based on Cross Coherence Processing. Sensors. 2021; 21(15):5073. https://0-doi-org.brum.beds.ac.uk/10.3390/s21155073

Chicago/Turabian Style

Lv, Qian, and Shaozhe Zhang. 2021. "Three-Dimensional Interferometric ISAR Imaging Algorithm Based on Cross Coherence Processing" Sensors 21, no. 15: 5073. https://0-doi-org.brum.beds.ac.uk/10.3390/s21155073

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop