Next Article in Journal
Two-Age-Structured COVID-19 Epidemic Model: Estimation of Virulence Parameters to Interpret Effects of National and Regional Feedback Interventions and Vaccination
Next Article in Special Issue
Proposal of the Dichotomous STATIS DUAL Method: Software and Application for the Analysis of Dichotomous Data, Applied to the Test of Learning Styles in University Students
Previous Article in Journal
A New Generalized t Distribution Based on a Distribution Construction Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Attention-TCN-BiGRU: An Air Target Combat Intention Recognition Model

Air and Missile Defense College, Air Force Engineering University, Xi’an 710051, China
*
Author to whom correspondence should be addressed.
Submission received: 6 September 2021 / Revised: 22 September 2021 / Accepted: 24 September 2021 / Published: 28 September 2021
(This article belongs to the Special Issue Advances of Machine Learning and Their Applications)

Abstract

:
The prerequisite for victory in war is the rapid and accurate identification of the tactical intention of the target on the battlefield. The efficiency of manual recognition of the combat intention of air targets is becoming less and less effective with the advent of information warfare. Moreover, if the traditional method of combat intention of air targets is based only on data from a single moment in time, the characteristic information on the time-series data is difficult to capture effectively. In this context, we design a new deep learning method attention mechanism with temporal convolutional network and bidirectional gated recurrent unit (Attention-TCN-BiGRU) to improve the recognition of the combat intent of air targets. Specifically, suitable characteristics are selected based on the combat mission and air posture to construct a characteristic set of air target intentions and encode them into temporal characteristics. Each characteristic in the characteristic set is given an appropriate weight through the attention mechanism. In addition, temporal convolutional network (TCN) is used to mine the data for latent characteristics and bidirectional gated recurrent unit (BiGRU) is used to capture long-term dependencies in the data. Experiments comparing with other methods and ablation demonstrate that Attention-TCN-BiGRU outperforms state-of-the-art methods in terms of accuracy in recognizing target intent in the air.

1. Introduction

With the development of military technology and aviation technology, informationization has gradually become the core of the modern battlefield, and future wars will also be informationized. At the same time, because the continuous development of modern technology has led to a dramatic increase in the dynamics and complexity of the environment of the modern battlefield, it has become increasingly difficult for military experts to identify the enemy’s combat intentions in a timely and effective manner from the vast amount of information, especially from the air targets of large-scale electronic countermeasures [1]. Therefore, it is urgent to construct a decision-making aided combat system to improve the accuracy and speed of intention recognition of air target combat, so as to help decision-makers seize air superiority and even win the war.
In order to meet the needs of decision support system, many researchers have conducted research on the problem of intention identification, and air target tactical intention recognition has gradually become one of the research hotspots in modern air combat. The existing studies on the identification of tactical intentions against enemy targets mainly include evidence theory [2,3,4], template matching [5], expert systems [6], Bayesian networks [7,8,9], and neural networks [10,11,12,13,14,15,16,17]. The establishment of a standard template base of template matching technology, the collection of information and the construction of probability distribution function in evidence theory, the determination of the structure of Bayesian network and probability distribution parameters, and the construction of knowledge base and inference engine of the expert system all need to organize, abstract, and explicitly describe the empirical knowledge and knowledge representation of experts in related fields and the great difficulty of engineering implementation. References [10,11,12,13,14] use different deep learning methods, build and simulate the deep neural network of the human brain, and use data-driven methods to extract the semantic characteristics from the original data from low-level to high-level, from concrete to abstract, from general to specific, simulating the memory mechanism and reasoning mode of commander when judging the battlefield situation, overcoming the difficulties of traditional models in knowledge expression [15]. However, the target’s tactical intention is realized through a series of tactical actions. Therefore, the dynamic attributes and battlefield environment of the target will show the characteristics of changing with time, and the enemy target has certain concealment and deception when performing combat operations. As a result, it is not sound enough for the above-mentioned deep learning method to judge the combat intention of the enemy target by using the characteristic information at a single time. In response to the drawbacks of the above methods, Liu et al. [16] established an air combat target intention prediction model based on long short-term memory (LSTM) network with incomplete information, introduced the cubic spline interpolation function fitting and the mean padding to repair the incomplete data, and used the adaptive moment estimation (Adam) optimization algorithm to accelerate the training speed of the target intention prediction model, so as to effectively prevent the local optimum problem. Xue et al. [17] designed panoramic convolutional long short-term memory neural network (PCLSTM), a new deep learning method, to improve the ability of intention recognition, and designed a time series pooling layer to reduce the parameters of the neural network. The above two methods based on LSTM network are verified by experiments to have a certain effect on air target tactical intention recognition. However, because LSTM is a one-way transmission network, it can only use past information to make judgments and cannot effectively use future information. At the same time, because the network shares the weight, the weight of different characteristics in identifying the combat intention of air targets is the same, which will have certain negative effects. Especially, when the intentions presented by different characteristics are contradictory, the weight proportion of key characteristics should be strengthened.
In view of the above problems, we propose an air target tactical intention recognition model based on Attention-TCN-BiGRU. Attention mechanism can learn the weight proportion of different characteristics to further highlight the key information affecting the intention, which has been successfully used in image caption generation, document classification, etc. In 2018, Bai et al. [18] proposed the model Temporal convolutional network (TCN), which has been shown to achieve state-of-the-art standards in many timing problems such as natural speech processing and audio synthesis. Compared with standard recurrent networks such as LSTM network and Gated Recurrent Unit (GRU), TCN has more accurate output results and a simpler structure. However, LSTM and GRU also have their advantages. TCN can effectively extract high-frequency and low-frequency information from the sequence, while LSTM and GRU are good at capturing long-term dependence in a series [19]. Therefore, our strategy is to explore the advantages of both networks rather than choose which one of them to use. In addition to this, we replace the GRU network with the Bidirectional Gated Recurrent Unit (BiGRU). BiGRU compares with GRU and TCN in its ability to learn not only from historical moment information but also from future moment information together for the current moment data [20]. The experimental results show that the Attention-TCN-BiGRU model is superior to other methods in the accuracy of intention recognition and has theoretical significance and reference value for auxiliary combat systems.
The rest of this paper is organized as follows: Section 2 describes in detail the air target intention recognition and how to select the characteristics and types of the tactical intention of air targets; Section 3 describes in detail the model proposed in this paper; Section 4 gives the experimental results and analysis; Section 5 concludes the paper.

2. Description of Aerial Target Combat Intention Recognition Problem

Air target combat intention recognition is a process of judging the enemy’s combat intentions. The state of enemy air targets will vary with different combat intentions. We use radar and other sensors to obtain the static attributes and real-time dynamic information of these targets from the complex battlefield environment. Combining the prior knowledge and the experience of experts, we analyze the data and propose predictions of combat intentions [11]. The target representation and reasoning process are shown in Figure 1.
Air target tactical intention recognition is a typical pattern recognition problem that can be described as a mapping of characteristics of air target tactical intention to types of air target tactical intention. Because the environment of the actual air battlefield is complex and changeable, and the air target will induce the decision-maker to wrong reasoning, it is easy to be deceived in identifying the enemy’s tactical intention by relying on the characteristic information of the air target intention at a single moment. Therefore, it is more accurate and sounder to infer its tactical intention from the characteristic information of the enemy air target at multiple continuous moments [15]. Define I T as the air combat target time series feature set composed of T consecutive air combat real-time characteristic information from t 1 to t T , that is, I T = ( I ( t 1 ) , I ( t 2 ) , , I ( t T ) ) . Vector Q is defined as the space set of air target combat intention type, thereby determining the mapping function from the spatial set to the times series characteristic set of enemy air target tactical intention types as:
Q = f ( I T ) = f ( I ( t 1 ) , I ( t 2 ) , , I ( t T ) )
To accurately identify the tactical intention of air targets, it is necessary to combine the tactical experience and military knowledge of experts in related fields and realize it through complex thinking activities such as key characteristic extraction, comparative analysis, association, and reasoning of air battlefield environment, so it is difficult to summarize it with simple mathematical formulas [13]. We extract the key characteristics of air targets to form the tactical intention characteristic set and use experts in related fields to label each sample to train the Attention-TCN-BiGRU network, thereby implicitly establishing the mapping relationship between the combat intention type space set and the time series characteristic data set. The whole process of air target combat intention recognition is shown in Figure 2.
As shown in Figure 2, the reasoning process [21,22] of air target tactical intention is as follows: Step 1—select and extract data information related to target tactical intention from real-time battlefield information, including battlefield environment information, target attribute information, and real-time state information of the target in the corresponding local time and space domain, and form an initial characteristic set; Step 2—Collect the state characteristic data of the target at N consecutive moments ( I ( t n ) I ( t n + N ) ) and store them in a stack mode to form a characteristic vector set; Step 3—Integrate the characteristic data of ( I ( t n ) I ( t n + N ) ) moments, encode and normalize them to form a standardized temporal characteristic set; Step 4—Divide the times series characteristic set into the training set and testing set; Step 5—Input the training set to the Attention-TCN-BiGRU network training model; Step 6—Input the testing set to the trained air target intention recognition model output to get the intention recognition results.

2.1. Selection of Tactical Intention Type

For different combat forms, different enemy entities, and different setting contexts, there are differences in the target combat intention space set corresponding to the target combat intention set. Therefore, we need to define the combat intention space set of the enemy target according to the corresponding combat background, the attributes of the enemy target, and the possible combat missions. For example, in the target intention space set by Lu et al. [23], the potential threat of underwater target is {evasion, patrol, and attack}; Chen et al. [24] established the combat intent space set for the enemy’s single-group maritime fleet as {retreat, cover, attack, reconnaissance}. In this paper, taking the enemy air target near-shore air attack on military buildings as the research background, we establish the tactical intention space set of enemy targets, including seven intention types {attack, penetration, surveillance, reconnaissance, feint, retreat, and electronic interference}. Please see Appendix A Table A1 for a detailed description of the seven intention types.
After determining the enemy’s combat intention space set, how to transform the human cognitive pattern into labels that intelligent models can train and that correspond to the intention types in the combat intention space set is the key to applying the intelligent recognition model proposed in this paper to combat intention recognition [16]. Therefore, we can encapsulate the cognitive experience of experts in air combat into labels to train the intelligent recognition model of this paper. For the seven intention types in the enemy target tactical intention space established in this paper, {0,1,2,3,4,5,6}, a total of seven label values is set, respectively. The corresponding tactical intention type coding and model analysis mechanism are shown in Figure 3. For example, if the prediction result of intention output by the proposed model is 0, it can be considered that the enemy target’s combat intention against intended target is attacked. Therefore, using the above knowledge encapsulation and model analysis can clearly and easily describe human experience and knowledge, which is convenient for model training.

2.2. Character Selection and Pre-Processing Process

The air target’s tactical intention is related to its combat mission and air situation. In order to identify different combat missions and air situations and to collect the actual characteristics of the sensors according to air targets such as radar stations, it is necessary to select appropriate air target intention characteristics.
From the perspective of a combat mission, when an air target performs a mission, some characteristic information needs to meet certain requirements. For example, when performing the task of penetration, it is divided into high-altitude penetration and low-altitude penetration, with a corresponding height of 10   km 11   km and above 50   m 200   m . Reconnaissance aircraft generally use low-altitude and high-altitude reconnaissance, with a corresponding height of 100   m 1000   m and above 15 km . There is also a certain relationship between radar signal status of air targets and combat missions. For example, air-to-air radars are usually kept on during air combat, and air-to-air radars and marine radars are kept on during reconnaissance missions, and the radar status is coded so that 1 means the radar is on and 0 means the radar is off [11]. Different types of air targets have different application values and strategic significance. For example, fighter aircraft have strong offensiveness, and reconnaissance aircraft have strong mobile reconnaissance capabilities. Therefore, the type of enemy aircraft can also be used as the characteristic of combat intention. However, because it is difficult to directly identify the type of enemy aircraft, the radar cross-section (RCS) is used as an alternative to the type of enemy aircraft. RCS can represent the type of enemy aircraft to a certain extent. Under normal circumstances, the RCS of stealth aircraft and small aircraft is less than 1   m 2 , the RCS between 1   m 2 and 10   m 2 is a medium-sized aircraft, and the RCS greater than 10   m 2 is a large-scale aircraft [17].
From the analysis of the air situation, many factors affect the superiority or inferiority of the air target. Considering the actual characteristics that can be collected, we mainly consider the enemy air target flight speed, altitude, acceleration, heading angle, azimuth angle, and the distance between the enemy, as shown in Figure 4. The air combat capability factor is also an important factor affecting the air situation. For the air combat capability of a fighter, a single machine air combat capability threat function is constructed according to the Reference [25]:
C = [ l n ε 1 + l n ( ε 2 + 1 ) + l n ( ε 3 + 1 ) ] ε 4 ε 5 ε 6 ε 7
In the formula, ε 1 ε 7 respectively represent fighter maneuverability, airborne weapon performance, airborne detection capability, fighter operation performance, fighter survivability, fighter combat range, and electronic information countermeasure capabilities. Air combat capability is an inherent attribute of aircraft. The air combat capability factor of enemy aircraft in a certain period can also be calculated by the above formula and then stored in the database and updated regularly according to the mastered information [26].
In addition, public opinion analysis is a worthy external characteristic to be selected [27]. When analyzing the target tactical intention, we often only focus on the attribute factors of the battlefield itself, while ignoring the influence of public opinion tendency in the war, which leads to inaccurate results of tactical intention analysis. There are many sources of public opinion, such as the government, the military, the Ministry of National Defense, the Ministry of Foreign Affairs, embassies, media, and civil organizations. Combining public opinion from different sources and considering the public opinion of combat environment of both sides, according to Miller’s 9-level quantification theory [28], public opinion information can be quantified into 5 levels, including 1, 3, 5, 7, and 9. To sum up, we have selected 12 kinds of characteristic information related to combating intention, including 8 kinds of numerical characteristics and 4 kinds of non-numerical characteristics. Table 1 and Table 2 describe each characteristic in detail, showing 8 numerical data forms and 4 non-numerical data encoding forms of a sample at a certain time.
Because of the different units of various characteristics, their distribution ranges are quite different, and characteristics with a large variance will have a greater impact, which leads to the inability to accurately obtain core characteristics from the original data. In order to eliminate the influence of data dimension and improve the convergence efficiency and accuracy of the network, we now perform Min–Max Normalization on the above 12 air target combat intention characteristics. For the i-th type of numerical data X i = ( x i 1 , x i 2 , x i r , , x i k ) ( i = 1 ,   2 , , 12 ) ; k is the number of data of the i-th type. The result of mapping the r-th original data value x i r in the i-th characteristic to the interval [0, 1] is x i r , and its formula is
x i r = x i r m i n X i m a x X i m i n X i
where m i n X i is the minimum value of the i-th characteristic X i ; m a x X i is the maximum value of the i-th characteristic X i .

3. Model Framework

In this section, we introduce our Attention-TCN-BiGRU model in detail. Attention mechanism assigns greater weight to key characteristics in the process of intention recognition. TCN extracts short-term local characteristics, whereas LSTM can capture the long-term dependence in a series. In addition, BiGRU can effectively use future information to make up for the shortcomings of the causal sequence in the TCN structure. To take full advantage of their merits, we combine them into a new hybrid model to further improve the accuracy of intention recognition. We introduce Attention mechanism, TCN, and BiGRU in Section 3.1, Section 3.2 and Section 3.3, respectively, and give the details of the overall framework of the proposed model in Section 3.4.

3.1. Attention Mechanism

Attention mechanism is a very hot topic at present, and there are various Attention mechanisms, such as Bottom-up Attention [29], Global attention [30], Self Attention [31], and so on. The essence of the attention mechanism comes from the human visual attention mechanism. When people perceive something visually, they generally do not see all of a scene from beginning to end but tend to observe and pay attention to a specific part of it according to their needs. Moreover, when people find that something they want to observe often appears in a certain part of a scene, they will learn and pay attention to that part when similar scenes appear in the future [32]. The distributed attention mechanism we use is a hierarchical attention network (HAN) proposed by Yang et al. [33] for text classification tasks. Not every word in a sentence is useful for classification. For example, when we do the emotion classification of text, we will focus on the words “happy” and “sad”. In order to enable the recurrent neural network to automatically focus “attention” on these words, the author designs a word-based attention model. Firstly, the sentence composed of a sequence of words is transformed into a word vector through a word embedding matrix, and then the hidden layer output hi is obtained through BiGRU network. Then a linear layer is transformed to obtain ui, using the softmax function to get the importance weight δ i of each word, and finally, the information representation of the sentence vector s is obtained by a weighted average of the output of BiGRU [34]. The formulas are as follows:
u i = Tanh ( W w h i + b i )
δ i = e x p ( u i T u w ) i e x p ( u i T u w )
s = i δ i h i
Formula (4) is a linear change process, Ww and bi are the parameter vectors of the hidden layer; formula (5) is defined for softmax function, and uw is a random initialization vector, which is continuously learned during training. formula (6) is the process of calculating sentence vector s.
In our proposed network, Attention mechanism is used to weight the contribution of different air target combat intention characteristics and extend the influence of the key characteristics, as shown in Figure 5. The implementation process is as follows. First, define the eigenvector X t = ( x 1 , t , x 2 , t , , x m , t ) at time step t, and calculate the proportion of characteristic weight according to the following softmax function formula.
α t = e x p ( X t u w 1 ) e x p ( X t u w 1 )
where α t = ( α 1 , t , α 2 , t , , α m , t ) , and u w 1 is the hidden layer vector to be learned during the training process. Next, the weight ratio of the i-th characteristics is α i = ( α i , t 1 , α i , t 2 , , α i , t n ) , and the average weight ratio is
α ¯ i = 1 t n t = t n t = t 1 α i , t
Finally, the weighted value of the i-th characteristics of the output of Attention mechanism is
R i = α ¯ i X i = ( α ¯ i x i , t 1 , α ¯ i x i , t 2 , , α ¯ i x i , t n )

3.2. TCN

Temporal convolutional network (TCN) is based on the convolutional neural network (CNN) model, and its structure combines causal convolution and dilated convolution. It can be used in a time series model, which can reduce the amount of calculation and keep a large receptive field for data. The proposed TCN effectively avoids the gradient disappearance or gradient explosion of RNN and has the advantages of parallel computing and controlling the sequence memory length by changing the size of the receptive field [35]. The following describes its structural components.

3.2.1. Causal Convolutions

Causal convolution can be visually represented in Figure 6. The output at time t is only related to the elements at time t and before in the previous layer [36]. Different from the traditional convolutional neural network, causal convolution cannot see the future data, and it is a one-way structure, not a two-way one. That is to say, only with the previous cause can there be the subsequent result, which is a strict time constraint model, so it is called causal convolution.

3.2.2. Dilated Convolutions

To address the problem of linear stacking of multiple convolutions in traditional neural networks to dilate the modeling of time series, TCN reduces the number of convolutional layers by using dilated convolutions to increase the range of the receptive field of each layer [37]. The difference between dilated convolutions and ordinary convolutions is that dilated convolution allows interval sampling of input during convolution, and the sampling rate depends on dilation factors. The formula of the receptive field is
R F = ( k 1 ) d + 1
where k is the kernel size, and d is the dilation factor. There are two ways to increase the TCN receptive field: increasing the dilation factor and choosing a larger kernel size, as shown in Figure 7. In the operation of dilation convolution, the dilation factor will increase exponentially with the increase of network depth, so the convolution network can still obtain a large receptive field with fewer layers [38].

3.2.3. Residual Connections

Another important network structure in the TCN network is the residual module. The residual module, shown in Figure 8, contains two layers of dilation causal convolution and nonlinear mapping. Moreover, the residual network has added the identity mapping of cross-layer connections, so that the network can transmit information in a cross-layer manner, which can solve the problem of gradient disappearance and make the shallow network easy to dilate into the deep network [39]. At the same time, Dropout and batch normalization are added to prevent the model from overfitting and speeding up training.

3.3. BiGRU

As a variant of Recurrent Neural Network (RNN), Gated Recurrent Unit (GRU) has a recursive structure similar to RNN and has the “memory” function of processing time-series data. Meanwhile, GRU can effectively alleviate the gradient disappearance and gradient explosion that may occur during RNN training, thus effectively solving the problem of long-term memory. LSTM network is also a variant of RNN [40]. Its performance is almost the same as that of GRU, but GRU is simpler in structure, which can reduce the amount of calculation and improve the training efficiency [41]. The internal structure of GRU is shown in Figure 9. GRU has two inputs: the output state h t 1 of the previous time and the input sequence value x t of the current time, and the output is the state h t of the current time. It mainly updates the model state through resetting gate r t and updating gate z t , and the gate r t reset controls the degree of forgetting of historical state information, so that the network can lose unimportant information. The gate z t update controls the proportion of the previous state information brought into the current state, helping the network to remember long-term information [42]. Its internal calculation formulas are as follows:
r t = σ ( W r x t + U r h t 1 )
z t = σ ( W z x t + U z h t 1 )
h ˜ t = Tanh ( W h ˜ x t + U h ˜ ( r t h t 1 ) )
h t = ( 1 z t ) h t 1 + z t h ˜ t
In the formulas and in Figure 9, σ is the sigmoid activation function, which changes the intermediate state to the range of [0, 1]; h t 1 and h t are the output states at moments t − 1 and t, respectively; x t is the input sequence value at moment t; h ˜ t is the candidate output state. W r , W z , W h ˜ , U r , U z and U h ˜ are the weight coefficient matrix corresponding to each part; Tanh is the hyperbolic tangent function; is the Hadamard product of the matrix.
The traditional GRU structure usually spreads in one direction along with the sequence transmission, and the information it obtains is the historical information before the current time, which leads to ignoring the future information. However, BiGRU structure is composed of forward GRU and backward GRU, which can capture the characteristics of front and back information [43]. Its model structure is shown in Figure 10.
As can be seen from Figure 10, the BiGRU hidden layer state h t at the moment t can be found in two parts: the forward hidden layer state h t and the backward hidden layer state h t . The forward hidden layer state h t is determined by the current input x t and the forward hidden layer state h t 1 at moment t 1 . The backward hidden layer state h t is determined by the current input x t and the backward hidden layer state h t 1 at moment t + 1 . The calculation formulas are as follows, where w i ( i = 1 ,   2 , 6 ) represents the weight from one unit layer to another.
h t = f ( w 1 x t + w 2 h t 1 )
h t = f ( w 3 x t + w 5 h t + 1 )
h t = g ( w 4 h t + w 6 h t )

3.4. Attention-TCN-BiGRU

Our strategy is to make full use of these three algorithms to solve our problems: the ability of Attention mechanism to select key characteristics, the ability of TCN to extract characteristics, the ability of BiGRU to capture long-term dependence, and the ability to obtain future information. Therefore, the model is named Attention-TCN-BiGRU. Figure 11 illustrates the general framework of our proposed model.
The input of the model is the data X after data preprocessing in Section 2.2, and its shape is (sample size, time step, characteristic size). In the first stage, the characteristic vector X i of each category in X are given different weights α ¯ i ( i = 1 , 2 , , m ) through the attention mechanism in Section 3.1. In the second stage, we perform characteristic extraction over T j = ( α ¯ 1 x 1 , t j , α ¯ 2 x 2 , t j , , α ¯ m x m , t j ) ( j = 1 , 2 , , n ) through TCN, which consists of two residual blocks. The first residual module consists of two dilation causal convolution layers, with kernel size of 4, dilation factor of 1, and filter number of 7. The dilation factor of the second residual module is 2, and the remaining parameters are the same as those of the first residual module. In the third stage, we use the output vector T j of the TCN as the input of the BiGRU network to extract the long-term correlation in the time series and the correlation between future information and current information. We take the output h t 64 at the last moment of the second layer of BiGRU network as the input of the full connection layer and output the intention type label with the highest probability by the following formula.
y = softmax ( w h t 64 + b )
where y is the intention type label, w is the weight matrix to be learned, and b is the bias vector.

4. Experimental Analysis

4.1. Experimental Data and Environment

The experimental data were obtained from the Air Combat Maneuvering Instrumentation (ACMI), whose labels were revised by experts in the field of air warfare. There were 10,500 samples in total, including 8400 training samples and 2100 test samples, with 14.3% of intentions per category. Time step was 12, and characteristic size was 12.
The experiments were conducted using the Python 3.8 language, in a Quadro RTX 5000/PCle/SSE2 GPU and CUDA11.0 accelerated environment, using the Keras 2.4.3 deep learning framework, with a computer configuration of an x86-64 CentOS7 PC system, Intel® Xeon(R) Sliver 4110 CPU @ 2.10GHz, and 64GB of RAM. The number of epochs set for the experiment was 100, and the batch size was 256.

4.2. Evaluation Metric

We propose the performance of the Attention-TCN-BiGRU air target recognition model of combat intention for validation. Five metrics are used to evaluate the classification of the network, namely, Loss, Accuracy, Precision, Recall, and F1-score. They are calculated as follows.
L o s s = i = 1 i = q y i 1 log y ^ i 1 + y i 2 log y ^ i 2 + + y i d log y ^ i d
A c c u r a c y = T P + T N T P + F N + F P + T N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1   s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
where q is the number of samples, d is the number of categories, y i d is the one-hot coded value (0 or 1), y ^ i d is the output value of the Softmax function ( d = 1 d = 7 y ^ i d = 1 ), TP is the True Positive, FN is the False Negative, FP is the False Positive, and TN is the True Negative.

4.3. Parameter Tuning

There are also many hyperparameters in Attention-TCN-BiGRU that have a significant impact on intention recognition performance. These include Filter numbers and Kernel size in the TCN network, Neurons size in the BiGRU network, Optimizer, and Learning rate of the whole network.
Appropriate size and number of convolutional kernels help to learn the rich characteristics in the data. As can be seen in Table 3 and Table 4, the highest accuracy and lowest loss values are achieved when Filter numbers = 7 and Kernel size = 4, so they are set to 7 and 4, respectively.
The number of neuron nodes is not as good as possible. Too many neuron nodes are prone to “overfitting”. The appropriate number of BiGRU neurons has a significant impact on the intention recognition performance of Attention-TCN-BiGRU. We considered both single-layer and two-layer BiGRU networks and tried different numbers of neurons. As can be seen from Table 5, the highest accuracy and lowest loss values were achieved when the BiGRU was set to two layers with 128 and 64 neurons, respectively.
Regarding optimizers, we compared five candidate algorithms; Stochastic Gradient Descent (SGD) [44], Root Mean Square prop (RMSprop) [45], Adaptive Moment Estimation (Adam) [46], Adamax [46], and Nesterov-accelerated Adaptive Moment Estimation (Nadam) [47] were used. As can be seen from Table 6, the Nadam optimizer has the highest accuracy and lowest loss values compared to the other optimizers. Therefore, Nadam was chosen for the optimization of our model. The default learning rate of Nadam in the Keras 2.4.3 deep learning framework is 0.002, which we have adjusted. As can be seen from Table 7, the accuracy and loss value of intention recognition is optimal when the learning rate is 0.004, so the learning rate we choose is 0.004.
The parameter settings for each network layer of our Attention-TCN-BiGRU model are shown in Table 8.

4.4. Results and Analysis

4.4.1. Attention-TCN-BiGRU Intention Recognition Result Analysis

The experimental results of the Attention-TCN-BiGRU model are shown in Figure 12. The convergence reached around 40 epochs, and the accuracy of the test set fluctuated around 94% and could reach up to 94.6%, with a loss value around 0.15 which could reach as low as 0.138.
The confusion matrix for the test set was produced to allow further observation of the relationship between the recognition intentions. The diagonal line indicates the number of correctly identified samples, as shown in Figure 13. As can be seen from Figure 13, the accuracy of the model is high for all seven intentions, with the highest accuracy of 98.7% for retreat and the lowest accuracy of 88.7% for the feint. In addition, it can be found that there are some mutual recognition errors for attack and feint intentions and a few mutual recognition errors for reconnaissance and surveillance. The network was unable to learn the obvious difference between the two intentions, which led to the misrecognition of the two intentions, in line with the actual situation.

4.4.2. Analysis of Characteristic Weight Proportion

The weights assigned to the aerial target characteristic set after the Attention layer are now analyzed in order to verify the effectiveness of the Attention mechanism. Its results are shown in Figure 14, with 12 characteristics on the horizontal axis and 7 intentions on the vertical axis.
From Figure 14, it is derived that the four characteristics of distance, altitude, speed and heading angle received the most attention among the characteristics. By reviewing the relevant reference [48,49] and consulting with relevant experts, it is clear that angle and altitude play the most critical role in the execution of the mission by warplanes, which is consistent with the experimental results. Combined with the analysis of the confusion matrix in Figure 14, there exist three types of intentions with more mutual recognition errors, attack, feint, and penetration, and many degrees of similarity in the attention given to various types of characteristics in their Attention layer, and the same situation exists for surveillance and reconnaissance. The two corroborate each other, further demonstrating that the Attention mechanism is indeed effective in focusing on key characteristics.

4.4.3. GRU, LSTM, BiLSTM, and BiGRU Selection Experiments

The Attention-TCN network was combined with BiGRU, BiLSTM, GRU, and LSTM, respectively, in order to verify that using the BiGRU network was superior, and the experimental results are shown in Table 9. The intention recognition accuracy of Attention-TCN-BiGRU and Attention-TCN-BiLSTM were similar, and both were higher than that of Attention-TCN-GRU and Attention-TCN-LSTM networks, thus verifying that the introduction of bidirectional propagation mechanism can effectively improve the intention recognition accuracy. In terms of intention recognition time for a single sample, the lowest time was 43.2 ms for Attention-TCN-GRU, and the longest time was 54.2 ms for Attention-TCN-BiLSTM, because the LSTM structure consists of four parts and the GRU structure consists of three parts, and the number of GRU parameters is about one quarter less than that of LSTM. Therefore, we chose the Attention-TCN-BiGRU network because it has lower time complexity when the performance is similar to Attention-TCN-BiLSTM.

4.4.4. Comparative Analysis of Intention Recognition Methods

In order to verify the superiority of our proposed method, we now compare it with Zhou et al. [11] proposed the use of ReLU function and Adam algorithm to optimize the deep back propagation (DBP) neural network air target combat intention recognition model; Zhai et al. [12] proposed a standardized fully connected residual (BN-FC-RES) network model for the analysis of air target combat intentions; Liu et al. [13] proposed a target intention recognition model based on radial basis function (RBF) neural network; Ou et al. [14] proposed a stack autoencoder (SAE) tactical intent intelligent recognition model; Liu et al. [16] proposed a long-short-term memory (LSTM) network-based air combat target intention prediction model under incomplete information, and Xue et al. [17] proposed a panoramic convolutional long short-term memory network (PCLSTM) for air target combat intention recognition. The experimental results are shown in Table 10.

4.4.5. Ablation Experiment

A comparative experimental analysis was conducted through ablation experiments and three assessment metrics; Precision, Recall, and F1 Score, were introduced to assess each type of intention in order to further demonstrate the superiority of our proposed method. The experimental results are shown in Table 11 and Table 12 and Figure 15.
From Table 11 and Table 12 and Figure 15, we can see that Attention-TCN-BiGRU is the most superior, with high accuracy and low loss value low. The accuracy, recall and F1 Score for each type of intention recognition are also the highest, which indicates that combining the Attention, TCN and BiGRU networks can indeed improve the network’s ability to recognize intention. TCN is also significantly better than BiGRU, indicating that the recently developed TCN networks are more capable of learning data characteristics. The lowest accuracy rate is for attack intention, which is due to the fact that some of the feint intention samples and surprise intention samples have similar characteristics to attack intention samples, resulting in too many feint and surprise samples being identified as false positives. The lowest recall rate for feint intention is due to the fact that feint intention is too deceptive, and the data characteristics are less easy to learn compared to feint intention. The highest accuracy and recall rate was for retreat intentions, which were relatively easy to learn due to the distinctive characteristics of retreat intention.

5. Conclusions

A temporal network for improving intention recognition is designed, which addresses the important role of air target combat intention recognition in modern air warfare. Firstly, by analyzing the combat characteristics of air targets, 12 easily accessible characteristics focusing on both the combat mission and the air posture are extracted. Moreover, the non-numerical characteristics of the 12 characteristics are numerically processed. Secondly, the Attention mechanism is used to attach attention weights to the various characteristics; the TCN network is used to extract potential characteristics from the data, and the long-term dependencies present in the sequence are captured by the BiGRU network. Finally, generic simulation experiments are conducted to verify the intention recognition capability of our proposed model.
The higher intention recognition accuracy of our proposed Attention-TCN-BiGRU generic is validated in comparison with DBP [11], BN-FC-RES [12], RBF [13], SAE [14], LSTM [16], and PCLSTM [17] methods as well as its ablation experiments. It compares well with other methods in its ability to perform the task of intent recognition of air targets, and it has theoretical implications and reference value for combat systems that aid decision-making.
We will focus on two issues in our future work. The first is the problem of recognizing intentions with high characteristic similarity and high deceptiveness. The second is how to effectively recognize intentions when they change in the detected time series.

Author Contributions

Conceptualization, F.T. and Y.S.; methodology, F.T.; software, F.T.; validation, X.G.; formal analysis, F.T.; investigation, Y.S.; resources, Y.S.; data curation, X.G.; writing—original draft preparation, F.T.; writing—review and editing, F.T.; visualization, F.T.; supervision, Y.S.; funding acquisition, Y.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (under Grants No. 61703426); Young Talent fund of University Association for Science and Technology in Shaanxi, China, under Grant No. 2019038, and the Innovation Capability Support Plan of Shaanxi, China, under Grant No. 2020KJXX-065.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Detailed description of seven intention types.
Table A1. Detailed description of seven intention types.
Intention TypeDescription
AttackAttack ground targets by launching bullets, missiles, and bombs from aerial targets.
PenetrationUse earth curvature and terrain as cover to effectively avoid various threats and achieve sudden attacks on targets.
SurveillancePassively monitor the situation in a certain space through aerial targets, usually continuously.
ReconnaissanceActively detect a certain space through air targets, which can be continuous or intermittent.
FeintA military means to deceive the enemy. Only make a limited attack, making the enemy’s judgment wrong.
RetreatWithdraw from the battlefield or conflict.
Electronic
interference
Air targets interfere with radar and communication systems by carrying electronic jamming equipment.
Table A2. Symbol description.
Table A2. Symbol description.
SymbolMeaning
QAir target combat intent space set
I ( t n ) Real-time feature information of air targets at time t n
I T Real-time characteristic information of air targets at T consecutive times from t1 to tT
CAir combat capability factor
x i r The r-th data in class i air target characteristics
X i Type i air target characteristics
uiLinear layer output vector in HAN
δ i Importance weight of the i-th word
sSentence vector obtained after weighted average of hierarchical attention network
α i , t The weight of the i-th type feature at time t
α t Importance weight at time t
α ¯ i The average value of the weight of the i-th type feature
RiThe weighted value of the i-th type feature output by the Attention mechanism
RFReceptive field
kKernel size
dDilation factor
xtInput sequence value at time t
rtReset gate at time t
σ Sigmoid activation function
Hadamard product of the matrix
ztUpdate gate at time t
h ˜ t Candidate output state at time t
htHidden state at time t
h t BiGRU forward hidden layer state at time t
h t BiGRU backward hidden layer state at time t
y Air target intent type label

References

  1. Yao, Q.K.; Liu, S.J.; He, X.Y.; Ou, W. Research and Prospect of Battlefield Target Operational Intention Recognition. J. Command. Control 2017, 3, 127–131. [Google Scholar] [CrossRef]
  2. Xia, P.; Chen, M.; Zou, J.; Feng, X. Prediction of Air Target Intention Utilizing Incomplete Information; Springer: Singapore, 2016. [Google Scholar] [CrossRef]
  3. Sun, Y.L.; Bao, L. Study on Recognition Technique of Targets’ Tactical Intentions in Sea Battle Field based on D-S Evidence Theory. Ship Electron. Eng. 2012, 32, 48–51. [Google Scholar] [CrossRef]
  4. Zhao, F.J.; Zhou, Z.J.; Hu, C.; Wang, L.; Liu, T. Aerial Target Intention Recognition Approach based on Belief-Rule-Base and Evidential Reasoning. Electron. Opt. Control 2017, 24, 15–19, 50. [Google Scholar] [CrossRef]
  5. Chen, Y.M.; Li, Y.C. Simulation of Target Tactical Intention Recognition Based on Knowledge Map. Comput. Simul. 2019, 36, 1–4, 19. [Google Scholar] [CrossRef]
  6. Yin, X.; Zhang, M.; Chen, M.Q. Combat Intention Recognition of the Target in the Air Based on Discriminant Analysis. J. Proj. Rocket. Missiles Guid. 2018, 38, 46–50. [Google Scholar] [CrossRef]
  7. Xu, Y.; Cheng, S.; Zhang, H.; Chen, Z. Air Target Combat Intention Identification Based on IE-DSBN. In Proceedings of the 2020 International Workshop on Electronic Communication and Artificial Intelligence (IWECAI), Shanghai, China, 12–14 June 2020. [Google Scholar] [CrossRef]
  8. Xu, X.M.; Yang, R.; Fu, Y. Situation assessment for air combat based on novel semi-supervised naive Bayes. J. Syst. Eng. Electron. 2018, 29, 768–779. [Google Scholar] [CrossRef]
  9. Qing, J.; Xiantai, G.; Weidong, J.; Nanfang, W. Intention recognition of aerial targets based on Bayesian optimization algorithm. In Proceedings of the 2017 2nd IEEE International Conference on Intelligent Transportation Engineering (ICITE), Singapore, 1–3 September 2017. [Google Scholar] [CrossRef]
  10. Zhou, T.G.; Chen, M.; Wang, Y.H.; He, J.L.; Yang, C.G. Information Entropy Based Intention Prediction of Aerial Targets under Uncertain and Incomplete Information. Entropy 2020, 22, 279. [Google Scholar] [CrossRef] [Green Version]
  11. Zhou, W.W.; Yao, P.Y.; Zhang, J.Y.; Wang, X.; Wei, S. Combat Intention Recognition Foraerial Targets based on Deep Neural Network. Acta Aeronaut. Astronaut. Sinca 2018, 39, 322468. [Google Scholar] [CrossRef]
  12. Zhai, X.Y.; Yang, F.B.; Ji, L.N.; Li, S.Q.; Lv, H.L. Air Combat Targets Threat Assessment based on Standardized Fully Connected Network and Residual Network. Fire Control Command Control 2020, 45, 39–44. [Google Scholar] [CrossRef]
  13. Liu, Y.J.; Kou, G.H.; Song, J.H. Target Recognition based on RBF Neural Network. Fire Control Command Control 2015, 40, 9–13. [Google Scholar] [CrossRef]
  14. Ou, W.; Liu, S.J.; He, X.Y.; Guo, S.M. Tactical Intention Recognition Algorithm based on Encoded Temporal Features. Command Control Simul. 2016, 38, 36–41. [Google Scholar] [CrossRef]
  15. Ou, W.; Liu, S.J.; He, X.Y.; Cao, Z.G. Study on Intelligent Recognition Model of Enemy Target’s Tactical Intention on Battlefield. Comput. Simul. 2017, 34, 10–14, 19. [Google Scholar] [CrossRef]
  16. Liu, Z.D.; Chen, M.; Wu, Q.X.; Chen, S.D. Prediction of Unmanned Aerial Vehicle Target Intention Under Incomplete Information. Sci. Sin. Inf. 2020, 50, 704–717. [Google Scholar] [CrossRef]
  17. Xue, J.; Zhu, J.; Xiao, J.; Tong, S.; Huang, L. Panoramic Convolutional Long Short-Term Memory Networks for Combat Intention Recognition of Aerial Targets. IEEE Access 2020, 8, 183312–183323. [Google Scholar] [CrossRef]
  18. Bai, S.; Kolter, J.Z.; Koltun, V. An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling. arXiv 2018, arXiv:1803.01271. [Google Scholar]
  19. Bi, J.; Zhang, X.; Yuan, H.; Zhang, J.; Zhou, M. A Hybrid Prediction Method for Realistic Network Traffic With Temporal Convolutional Network and LSTM. IEEE Trans. Autom. Sci. Eng. 2021. [Google Scholar] [CrossRef]
  20. Yan, J.; Liu, J.; Yu, Y.; Xu, H. Water Quality Prediction in the Luan River Based on 1-DRCNN and BiGRU Hybrid Neural Network Model. Water 2021, 13, 1273. [Google Scholar] [CrossRef]
  21. Lorenc, A.; Burinskiene, A. Improve the orders picking in e-commerce by using WMS data and BigData analysis. FME Trans. 2021, 49, 233–243. [Google Scholar] [CrossRef]
  22. Lorenc, A.; Kunar, M. The most common type of disruption in the supply chain—Evaluation based on the method using artificial neural networks. Int. J. Shipp. Transp. Logist. 2021, 13, 1–24. [Google Scholar] [CrossRef]
  23. Lu, G.Y.; Ding, Y.Y. Study on Intention Recognition to Foe of Underwater Platform. Command Control Simul. 2012, 34, 100–102. [Google Scholar]
  24. Chen, H.; Ren, Q.L.; Hua, Y.; Qiu, Y.N. Fuzzy Neural Network based Tactical Intention Recognition for Sea Targets. Syst. Eng. Electron. 2016; 38, 1847–1853. [Google Scholar] [CrossRef]
  25. Kojadinovic, I.; Marichal, J.L. Entropy of bi-capacities. Eur. J. Oper. Res. 2007, 178, 168–184. [Google Scholar] [CrossRef]
  26. Xi, Z.F.; Xu, A.; Kou, Y.X.; Li, Z.W.; Yang, A.W. Target threat assessment in air combat based on PCA-MPSO-ELM algorithm. Acta Aeronaut. AstroNautica Sin. 2020, 41, 323895. [Google Scholar] [CrossRef]
  27. Zhai, X.Y.; Yang, F.B.; Ji, L.N.; Lv, H.L.; Bai, Y.Q. Standardized Fully Connected Network and Residual Network Model for Combat Intention Analysis of Air Targets. Foreign Electron. Meas. Technol. 2019, 38, 1–6. [Google Scholar] [CrossRef]
  28. Kulak, O. A decision support system for fuzzy multi-attribute selection of material handling equipments—ScienceDirect. Expert Syst. Appl. 2005, 29, 310–319. [Google Scholar] [CrossRef]
  29. Anderson, P.; He, X.; Buehler, C.; Teney, D.; Johnson, M.; Gould, S.; Zhang, L. Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering. arXiv 2017, arXiv:1707.07998. [Google Scholar]
  30. Luong, M.T.; Pham, H.; Manning, C.D. Effective Approaches to Attention-based Neural Machine Translation. Comput. Sci. 2015. [Google Scholar] [CrossRef] [Green Version]
  31. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is All You Need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  32. Wang, W.; Sun, Y.X.; Qi, Q.J.; Meng, X.F. Text Sentiment Classification Model based on BiGRU-attention Neural Network. Appl. Res. Comput. 2019, 36, 3558–3564. [Google Scholar] [CrossRef]
  33. Yang, Z.; Yang, D.; Dyer, C.; He, X.; Smola, A.; Hovy, E. Hierarchical Attention Networks for Document Classification. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, CA, USA, 12–17 June 2016. [Google Scholar] [CrossRef] [Green Version]
  34. Song, Y.; Gao, S.; Li, Y.; Jia, L.; Li, Q.; Pang, F. Distributed Attention-Based Temporal Convolutional Network for Remaining Useful Life Prediction. IEEE Internet Things J. 2020, 8, 9594–9602. [Google Scholar] [CrossRef]
  35. Fan, Y.Y.; Li, C.J.; Yi, Q.; Li, B.Q. Classification of Field Moving Targets Based on Improved TCN Network. Comput. Eng. 2021, 47, 106–112. [Google Scholar] [CrossRef]
  36. Gao, Y.; Wang, W.; Lin, Q.; Cai, F.; Chai, Q. Fault Diagnosis for Power Converters Based on Optimized Temporal Convolutional Network. IEEE Trans. Instrum. Meas. 2020, 70. [Google Scholar] [CrossRef]
  37. Huang, Q.; Hain, T. Improving Audio Anomalies Recognition Using Temporal Convolutional Attention Network. In Proceedings of the ICASSP 2021–2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021. [Google Scholar] [CrossRef]
  38. Shi, H.; Wang, L.; Scherer, R.; Woźniak, M.; Zhang, P.; Wei, W. Short-term Load Forecasting Based on AdaBelief Optimized Temporal Convolutional Network and Gated Recurrent Unit Hybrid Neural Network. IEEE Access 2021, 9, 66965–66981. [Google Scholar] [CrossRef]
  39. Zhu, R.; Liao, W.; Wang, Y. Short-term prediction for wind power based on temporal convolutional network. Energy Rep. 2020, 6, 424–429. [Google Scholar] [CrossRef]
  40. Xu, Z.; Zeng, W.; Chu, X.; Cao, P. Multi-Aircraft Trajectory Collaborative Prediction Based on Social Long Short-Term Memory Network. Aerospace 2021, 8, 115. [Google Scholar] [CrossRef]
  41. Liu, Y.; Ma, J.; Tao, Y.; Shi, L.; Wei, L.; Li, L. Hybrid Neural Network Text Classification Combining TCN and GRU. In Proceedings of the 2020 IEEE 23rd International Conference on Computational Science and Engineering (CSE), Guangzhou, China, 29 December–1 January 2020. [Google Scholar] [CrossRef]
  42. Sun, Y.C.; Tian, R.L.; Wang, X.F. Emitter signal recognition based on improved CLDNN. Syst. Eng. Electron. 2021, 43, 42–47. [Google Scholar] [CrossRef]
  43. Chen, J.X.; Jiang, D.M.; Zhang, Y.N. A Hierarchical Bidirectional GRU Model With Attention for EEG-Based Emotion Classification. IEEE Access 2019, 7, 118530–118540. [Google Scholar] [CrossRef]
  44. Kiefer, J.; Wolfowitz, J. Stochastic Estimation of the Maximum of a Regression Function. Ann. Math. Stat. 1952, 23, 462–466. [Google Scholar] [CrossRef]
  45. Tieleman, T.; Hinton, G. Lecture 6.5-rmsprop: Divide the Gradient by a Running Average of its Recent Magnitude. COURSERA Neural Netw. Mach. Learn. 2012, 4, 26–30. [Google Scholar]
  46. Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  47. Dozat, T. Incorporating Nesterov Momentum into Adam. 2016. Available online: https://openreview.net/forum?id=OM0jvwB8jIp57ZJjtNEZ (accessed on 28 September 2021).
  48. Xi, Z.F.; Xu, A.; Kou, Y.X.; Li, Z.W.; Yang, A.W. Air Combat Threat Assessment Based on Improved GRA-TOPSIS. J. Beijing Univ. Aeronaut. Astronaut. 2020, 46, 388–397. [Google Scholar] [CrossRef]
  49. Yang, A.W.; Li, Z.W.; Xu, A.; Xi, Z.F.; Chang, Y.Z. Threat Level Assessment of the air Combat target Based on Weighted Cloudy Dynamic Bayesian Networks. Flight Dyn. 2020, 38, 87–94. [Google Scholar] [CrossRef]
Figure 1. Hierarchical representation and reasoning process of intention.
Figure 1. Hierarchical representation and reasoning process of intention.
Mathematics 09 02412 g001
Figure 2. Process of air target tactical intention recognition.
Figure 2. Process of air target tactical intention recognition.
Mathematics 09 02412 g002
Figure 3. Schematic diagram of combat intention coding and analysis. The same color is a set of correspondences.
Figure 3. Schematic diagram of combat intention coding and analysis. The same color is a set of correspondences.
Mathematics 09 02412 g003
Figure 4. Relative geometric position of air combat. H is the flight height, D is the distance, V is the flight speed, θ is the heading angle, and β is the azimuth angle.
Figure 4. Relative geometric position of air combat. H is the flight height, D is the distance, V is the flight speed, θ is the heading angle, and β is the azimuth angle.
Mathematics 09 02412 g004
Figure 5. Attention mechanism structure used in this article.
Figure 5. Attention mechanism structure used in this article.
Mathematics 09 02412 g005
Figure 6. Causal convolution. Kernel size = 3.
Figure 6. Causal convolution. Kernel size = 3.
Mathematics 09 02412 g006
Figure 7. Dilated convolution. Dilation factors = 2, kernel size = 3.
Figure 7. Dilated convolution. Dilation factors = 2, kernel size = 3.
Mathematics 09 02412 g007
Figure 8. TCN residual block. k is the kernel size, and d is the dilation factor.
Figure 8. TCN residual block. k is the kernel size, and d is the dilation factor.
Mathematics 09 02412 g008
Figure 9. Gated recurrent unit structure.
Figure 9. Gated recurrent unit structure.
Mathematics 09 02412 g009
Figure 10. Bidirectional gated recurrent unit structure.
Figure 10. Bidirectional gated recurrent unit structure.
Mathematics 09 02412 g010
Figure 11. Proposed Attention-TCN-BiGRU framework.
Figure 11. Proposed Attention-TCN-BiGRU framework.
Mathematics 09 02412 g011
Figure 12. Accuracy and loss value of the Attention-TCN-BiGRU model test set and training set.
Figure 12. Accuracy and loss value of the Attention-TCN-BiGRU model test set and training set.
Mathematics 09 02412 g012
Figure 13. Confusion matrix of Attention-TCN-BiGRU intention recognition results.
Figure 13. Confusion matrix of Attention-TCN-BiGRU intention recognition results.
Mathematics 09 02412 g013
Figure 14. Attention mechanism assigns weight to characteristics.
Figure 14. Attention mechanism assigns weight to characteristics.
Mathematics 09 02412 g014
Figure 15. Changes in accuracy and loss in ablation experiments: (a) Accuracy rate of ablation experiment; (b) loss value of ablation experiment.
Figure 15. Changes in accuracy and loss in ablation experiments: (a) Accuracy rate of ablation experiment; (b) loss value of ablation experiment.
Mathematics 09 02412 g015
Table 1. Numerical air target intelligence data.
Table 1. Numerical air target intelligence data.
CharacteristicDescriptionUnitNumerical Value
Distancethe distance between enemy plane and targeted buildingkm176.0
Velocitythe velocity of enemy planem/s260.0
RCSthe size of the radar reflection cross-sectional aream25.6
Heightthe height of the enemy planekm21.2
Azimuth anglethe angle from targeted building to the direction of the enemy aircraft (the true north is 0 mil, and the clockwise direction is divided into 6400 mil)mil1456.0
Accelerationthe acceleration of the enemy planem/s25.6
Heading angleThe flight direction of the enemy plane (the true north is 0 degrees, and the clockwise direction is 360 degrees)°45.3
Capability factorthe aerial combat capability of enemy aircraft/9.4
Table 2. Non-numerical air target intelligence data.
Table 2. Non-numerical air target intelligence data.
CharacteristicDescriptionCategorical Value
Electronic interference statuswhether the electronic jamming device is turned on0
Air-to-air radar statewhether the air-to-air radar is switched on1
Air-to-surface radar statewhether the air-to-surface radar is switched on1
Public opinion analysisCurrent international public opinion trends5
Table 3. Comparison of different filter numbers in the Attention-TCN-BiGRU.
Table 3. Comparison of different filter numbers in the Attention-TCN-BiGRU.
Filter NumbersAccuracyLoss
593.20.186
794.60.138
1093.60.171
1592.50.207
2092.30.238
Table 4. Comparison of different kernel sizes in the Attention-TCN-BiGRU.
Table 4. Comparison of different kernel sizes in the Attention-TCN-BiGRU.
Kernel SizeAccuracyLoss
292.50.209
393.70.172
494.60.138
592.60.210
692.30.227
Table 5. Comparison of different neurons size in the Attention-TCN-BiGRU.
Table 5. Comparison of different neurons size in the Attention-TCN-BiGRU.
Number of NeuronsAccuracyLoss
12894.00.162
25694.00.170
51292.30.201
64 6493.70.185
128 6494.60.138
128 12893.30.181
Table 6. Comparison of the different optimizers in the Attention-TCN-BiGRU.
Table 6. Comparison of the different optimizers in the Attention-TCN-BiGRU.
OptimizerAccuracyLoss
Nadam94.60.138
Adam94.30.151
Adamax92.00.203
RMSprop94.10.159
SGD90.10.276
Table 7. Comparison of different learning rates in the Attention-TCN-BiGRU.
Table 7. Comparison of different learning rates in the Attention-TCN-BiGRU.
Learning RateAccuracyLoss
0.00293.90.153
0.00494.60.138
0.00693.20.172
0.00892.80.197
0.01091.00.231
Table 8. Parameters of each layer of the proposed method.
Table 8. Parameters of each layer of the proposed method.
LayersParameters
Residual
block 1
Conv1D layer
Activation function
Normalization method
Conv1D layer
Activation function
Normalization method
Conv1D layer
Filter numbers = 7, Kernel size = 4, Dilation_rate = 1, Padding = ‘causal’
Rectified Linear Unit
Batch Normalization
Filter numbers = 7, Kernel size = 4, Dilation_rate = 1, Padding = ‘causal’
Rectified Linear Unit
Batch Normalization
Filter numbers = 7, Kernel size = 1, Dilation_rate = 1, Padding = ‘same’
Residual
block 2
Conv1D layer
Activation function
Normalization method
Conv1D layer
Activation function
Normalization method
Conv1D layer
Filter numbers = 7, Kernel size = 4, Dilation_rate = 2, Padding = ‘causal’
Rectified Linear Unit
Batch Normalization
Filter numbers = 7, Kernel size = 4, Dilation_rate = 2, Padding = ‘causal’
Rectified Linear Unit
Batch Normalization
Filter numbers = 7, Kernel size = 1, Dilation_rate = 2, Padding = ‘same’
Bidirectional GRU 1Hidden size = 128
Bidirectional GRU 2Hidden size = 64
DenseHidden size = 7, activation = ‘softmax’
Table 9. Comparison of four models.
Table 9. Comparison of four models.
ModelAccuracy (%)LossTime (ms)
Attention-TCN-BiGRU94.60.13850.4
Attention-TCN-BiLSTM94.70.12954.2
Attention-TCN-GRU93.90.14843.2
Attention-TCN-LSTM94.10.15146.9
Table 10. Comparison of different intention recognition models.
Table 10. Comparison of different intention recognition models.
ModelAccuracy (%)Loss
Attention-TCN-BiGRU94.60.138
PCLSTM88.70.338
LSTM84.50.445
BN-FC-RES81.60.498
SAE76.80.553
DBP75.90.548
RBF71.30.632
Table 11. Results of ablation experiment.
Table 11. Results of ablation experiment.
Model Composition StructureAccuracy (%)Loss
AttentionTCNBiGRU
94.60.138
93.30.185
93.50.173
90.80.247
90.20.298
85.10.435
Table 12. Results of evaluation indexes of ablation experiment. ①, ②, ③, ④, ⑤, and ⑥ respectively represent the Attention-TCN-BiGRU, Attention-TCN, Attention-BiGRU, TCN-BiGRU, TCN, and BiGRU air target intention recognition models.
Table 12. Results of evaluation indexes of ablation experiment. ①, ②, ③, ④, ⑤, and ⑥ respectively represent the Attention-TCN-BiGRU, Attention-TCN, Attention-BiGRU, TCN-BiGRU, TCN, and BiGRU air target intention recognition models.
Evaluation IndexPrecision (%)Recall (%)F1 Score (%)
Intention Style
Attack89.687.988.284.883.878.491.790.089.787.386.380.090.689.089.086.085.179.2
Penetration97.095.393.892.792.084.595.794.795.392.791.787.396.395.094.592.791.985.9
Surveillance96.395.395.391.390.982.795.394.795.090.789.786.395.895.095.291.090.384.5
Reconnaissance93.091.793.888.588.183.897.095.396.092.391.084.795.093.594.990.489.584.2
Feint90.888.788.687.486.480.488.786.085.783.384.776.389.787.387.185.385.578.3
Retreat98.398.098.096.796.396.098.798.798.797.096.795.098.598.398.396.896.595.5
Electronic interference97.396.296.694.594.290.295.093.794.092.391.386.096.195.095.393.492.788.1
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Teng, F.; Song, Y.; Guo, X. Attention-TCN-BiGRU: An Air Target Combat Intention Recognition Model. Mathematics 2021, 9, 2412. https://0-doi-org.brum.beds.ac.uk/10.3390/math9192412

AMA Style

Teng F, Song Y, Guo X. Attention-TCN-BiGRU: An Air Target Combat Intention Recognition Model. Mathematics. 2021; 9(19):2412. https://0-doi-org.brum.beds.ac.uk/10.3390/math9192412

Chicago/Turabian Style

Teng, Fei, Yafei Song, and Xinpeng Guo. 2021. "Attention-TCN-BiGRU: An Air Target Combat Intention Recognition Model" Mathematics 9, no. 19: 2412. https://0-doi-org.brum.beds.ac.uk/10.3390/math9192412

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop