Next Article in Journal
AI-Based Prediction of Myocardial Infarction Risk as an Element of Preventive Medicine
Previous Article in Journal
Behavior of Sedimentary Rock Tunnel against Rigid Projectile Impact
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimal Defense Strategy Selection Algorithm Based on Reinforcement Learning and Opposition-Based Learning

Shandong Provincial Key Laboratory of Computer Networks, Shandong Computer Science Center (National Supercomputer Center in Jinan), Qilu University of Technology (Shandong Academy of Sciences), Jinan 250014, China
*
Authors to whom correspondence should be addressed.
Submission received: 26 August 2022 / Revised: 19 September 2022 / Accepted: 19 September 2022 / Published: 24 September 2022

Abstract

:
Industrial control systems (ICS) are facing increasing cybersecurity issues, leading to enormous threats and risks to numerous industrial infrastructures. In order to resist such threats and risks, it is particularly important to scientifically construct security strategies before an attack occurs. The characteristics of evolutionary algorithms are very suitable for finding optimal strategies. However, the more common evolutionary algorithms currently used have relatively large limitations in convergence accuracy and convergence speed, such as PSO, DE, GA, etc. Therefore, this paper proposes a hybrid strategy differential evolution algorithm based on reinforcement learning and opposition-based learning to construct the optimal security strategy. It greatly improved the common problems of evolutionary algorithms. This paper first scans the vulnerabilities of the water distribution system and generates an attack graph. Then, in order to solve the balance problem of cost and benefit, a cost–benefit-based objective function is constructed. Finally, the optimal security strategy set is constructed using the algorithm proposed in this paper. Through experiments, it is found that in the problem of security strategy construction, the algorithm in this paper has obvious advantages in convergence speed and convergence accuracy compared with some other intelligent strategy selection algorithms.

1. Introduction

Now with the development of the Industrial Internet, the network attacks on industrial control systems are more widespread, and the network attacks faced by industrial control systems are very different from those faced by traditional IT systems. The biggest possible loss after an attack is information leakage, data tampering, etc., and major dangerous accidents often do not occur directly, because the most important thing in traditional IT information systems is the confidentiality of information. Unlike IT systems, cyber attacks against ICS often lead to serious consequences, such as environmental pollution, equipment damage, and even casualties. Among the well-known cyber attacks on ICS include the Iranian “Stuxnet” [1] incident in 2010 and the Ukraine power grid cyber attack [2] in 2015. Therefore, the network security problem of ICS needs to be urgently solved. Due to the dynamic nature of ICS systems, the size of the attack surface and the diversity of attack vectors, addressing the cybersecurity of ICS is a particularly difficult task. New system vulnerabilities are reported daily, while cybersecurity problems are often not bounded by physical boundaries, as the source of malicious action can be anywhere in the world. Given these threats, it is extremely important to have appropriate security strategies in place in advance [3,4,5]. Therefore, how to find the best strategy set among many security strategies has a wide range of application scenarios in industrial control prevention and response.
Evolutionary algorithms have strong scalability and are not limited by the nature of the problem, and they have great advantages in the field of finding the best security strategy. In the aspect of evolutionary algorithm, to find the best security policy set, predecessors have conducted a lot of related research. Dewri proposed Genetic Algorithm (GA) [6] and Poolsappasit et al. proposed an optimal security protection strategy selection model based on a Bayesian attack graph and used a particle swarm algorithm to find the optimal strategy set [7]. The evolutionary algorithms used above are relatively easy to fall into the local optimal solution, they are greatly affected by the initial number of sample populations, and there is a problem in that the convergence speed is not fast enough.
In addition, inspired by the predation behavior of bats, Yang proposed a novel bionic algorithm—the Bat Algorithm (BA) in 2010 [8]. Due to its simple structure and strong applicability, BA has been widely used. However, similar to many intelligent algorithms, BA itself also has problems such as easily falling into local optimum, slow convergence in later stages, having a single coding scheme, lack of a theoretical basis for parameter determination. Liu et al. proposed a differential evolution algorithm with a two-stage optimization mechanism [9]. In the first stage, the search range of the algorithm can be increased to avoid premature convergence, and the convergence speed can be accelerated in the second stage. Good results are obtained, but the effect of the size of the initial population on the algorithm is not fully accounted for.
Dixit et al. proposed an algorithm to improve DE by using PSO [10] and achieved good results. Among them, a sigmoid function is used to control the mutation strategy, which is actually very similar to our improvement on CR in this paper, which ensures population diversity to a certain extent. However, they do not have guidelines for setting control parameters, and they use the trial and error method to randomly select values. This can indeed reduce the ability to fall into local optimization, but it will reduce the convergence speed to a certain extent. Moreover, the NP in the experiment is a fixed value of 100, which can not guarantee that the performance of the algorithm can still remain stable in the case of very small samples.
Therefore, in terms of using evolutionary algorithms to select the best strategy, there is not yet a good and general evolutionary algorithm. With the deepening of the research on reinforcement learning [11,12,13] and opposition-based learning [14,15,16] in recent years, the method of using opposition-based learning combined with evolutionary algorithm has appeared [17,18,19], but it has not been widely applied in practice. There are also examples of combining reinforcement learning and evolutionary algorithms to solve problems [20]. Huzhenzhen et al. used the differential evolution algorithm improved by reinforcement learning to extract the parameters of the photovoltaic model [21]. Huynh used reinforcement learning to improve the differential evolution algorithm [22]. Five benchmark examples for minimizing the weight of truss structures are verified, and good results have been achieved. We draw on the ideas in [21,22] to apply reinforcement learning to the differential evolution algorithm, and we find that the methods proposed in [21,22] can complement each other’s shortcomings. So, we make improvements on this basis and form the prototype of the algorithm in this paper. However, the main function of reinforcement learning is to improve its convergence speed, and it is not a good solution to the problem of premature convergence, so we need an algorithm that can improve both the convergence speed and the convergence accuracy.
Therefore, this paper proposes a differential evolution algorithm (RODE) based on reinforcement learning and opposition-based learning. The modified algorithm can accurately and quickly select the optimal security strategy set. Before using it, we must first evaluate the security status of ICS. We constructed the Bayesian attack graph [7] of ICS, calculated the probability of each attribute node being attacked, and obtained the attack benefit. That is the value at risk. In order to weigh the benefits and costs, we quantify the attack benefits and protection costs [23,24], and we establish an objective function: minimize the attack benefits and protection costs, and the protection cost should maximize the allowable protection cost. Finally, security strategy selection is performed by using a hybrid differential evolution algorithm based on reinforcement learning and opposition-based learning. We compared the effects of similar algorithms and finally verified the feasibility and effectiveness of the method in the real industrial control environment. In summary, the main contributions of this paper are:
  • A differential evolution algorithm based on reinforcement learning and opposition-based learning is proposed and successfully applied to industrial control strategy selection.
  • Improve the existing enhanced differential evolution algorithm, shorten the convergence speed of the algorithm, and better avoid it falling into the local optimal solution, which is more suitable for strategy selection.
  • Compared with other similar algorithms, it is proved that the differential evolution algorithm based on reinforcement learning and opposition-based learning proposed in this paper has more advantages in the field of industrial control strategy selection.
The rest of the paper is organized as follows. Section 2 describes the method architecture for implementing optimal strategy selection. Section 3 describes how to assess the risk of industrial control systems. Section 4 describes the construction of security measures and objective functions, and in Section 5, the differential evolution algorithm based on mixed strategy proposed in this paper is introduced in detail. Section 6 analyzes the experimental results in detail, and Section 7 is the final summary.

2. Method Architectures

We build a simple framework for the realization of optimal security strategy selection, which is mainly composed of three steps: namely, risk assessment, constructing an objective function of attack benefit–protection cost, and optimal security strategy selection. The risk assessment phase begins by modeling the cyber attack using an attack graph (AG). Then, the time probability of successful exploitation of each vulnerability in the network is calculated. After that, the attack graph is converted into a Bayesian attack graph (BAG) to make it a quantitative model. The attack benefit is calculated according to the information provided by the BAG, and then, in order to solve the balance problem between the attack benefit and the protection cost, an objective function based on attack benefit–protection cost is constructed. Finally, the objective function is minimized by the differential evolution algorithm based on mixed strategy so as to find the optimal protection strategy set. Figure 1 depicts the main stages of the optimal decision selection framework, and in the following sections, each step shown in Figure 1 will be explained in detail.

3. ICS Security Assessment

3.1. Attack Graph

An attack graph ( A G ) is a tuple A G = ( S , S s t a r , S e n d , τ ) , where:
  • S is the set of attribute nodes in the attack graph. Its state is 1 or 0. When S = 1 , it means that the attribute S is broken by the attacker. When S = 0 , it means that the attribute S has not been broken by the attacker.
  • S s t a r S is the starting node of A G . It represents the node occupied by the attacker when the attack started.
  • S e n d is a set of root nodes in an A G . It represents the final target that the attacker wants to break, and it can have multiple existences.
  • τ S × S is a collection of state transitions. Each transition between the two states represents an exploit that represents a network state change from S p r e to S p o s .

3.2. The Probability of Attacking Behavior

AGs are effective modeling tools for multi-step attacks and are widely used in network security risk assessment, but the attack graph itself cannot quantitatively represent the security and risk level of the network [25]. Therefore, we need to obtain the success rate of the corresponding vulnerability utilization. The vulnerability availability refers to the probability that the attacker will succeed in the atomic attack by exploiting the vulnerability. The vulnerability availability is related to the difficulty of the vulnerability.
This paper uses the combination of CVSS [26] and expert experience to evaluate the vulnerability availability. CVSS provides three vulnerability assessment indicators, namely base, temporal, and environment. Each indicator is quantified with a numerical score from 0 to 10. The sub-items of the base attribute include Access Vector (AV), Access Complexity (AC), and Authentication (AU), and the sub-items of the temporal attribute include the Exploit Code Maturity (E), the Remediation Level (RL), and the Report Confidence (RC). The base score describes the intrinsic characteristics of the vulnerability, and the temporal score describes the characteristics of the vulnerability as a function of temporal stages. This paper uses base and temporal in CVSS and combines relevant expert experience to comprehensively evaluate the exploitability of vulnerabilities. Therefore, the probability obtained in this way is more in line with the actual situation at the time of evaluation.
The exploitability of a vulnerability can be expressed as Equation (1).
E x p l o i t a b i l i t y = 2 × A V × A C × A U
where A V is the access vector, A C is the access complexity, and A U is the authentication, as shown in Table 1.
Since vulnerabilities in different time stages have different properties, we use the temporal metric of CVSS. These metrics adjust the value of exploitability (Equation (1)). The final exploitability can be expressed as Equation (2).
P E = ( E × R L × R C ) × E x p l o i t a b i l i t y
where P E represents the probability of exploiting a vulnerability during risk assessment. The specific content is shown in Table 2.

3.3. Bayesian Attack Graph

The Bayesian attack graph adds the conditional probability table (CPT) of the node to the attack graph, which can not only describe the causal relationship of the multi-step atomic attack of the attack behavior but also calculate the potential various attack paths, as well as the node and overall risk values, transforming the attack graph from a qualitative structure to a quantitative structure. The added conditional probability distribution table is in the form of a table representing the conditional probability distribution of the value of Pr ( S i | P a [ S i ] ) . It reflects the probability of a state node being attacked given its set of parent nodes, where S i represents the state in the BAG and P a [ S i ] represents its parent node. Each node in the BAG has a corresponding CPT that specifies the probability of compromising the state of the network given different combinations of states of its parent node.
A Bayesian attack graph is a directed acyclic graph that can be formally defined as a tuple B A G = ( S , E , ε , P ) [25], where s represents the set of nodes. The edges of the Bayesian attack graph are represented by a set of ordered pairs E. A conjunction or disjunction between edges pointing to a node is represented by ε with possible values { AND , O R } . P is the conditional probability table of BAG nodes. Define e i as the exploitability that can transform S i P a [ S j ] from state S i to S j , whose value can be given by Equation (2). The conditional probability distribution function of S j is Pr ( S j | P a [ S j ] ) , which is defined as follows.

3.4. Probability Calculation

If S j has multiple parents, and the relationship between the incoming edges of node S j is AND ( ε = AND), it means that all parents of S j are to be used, and its value can be expressed by Equation (3).
Pr ( s j | P a [ s j ] ) = 0 , s i P a [ s j ] , s i = 0 S i = 1 P E ( e i ) , e l s e
If the relationship between the input edges to node S j is OR ( ε = OR), it means that at least one parent node of S j is used, and its value can be expressed by Equation (4).
Pr ( s j | P a [ s j ] ) = 0 , s i P a [ s j ] | s i = 0 1 S i = 1 [ 1 P E ( e i ) ] , e l s e
where P E ( e i ) is the exploit success probability of the current attribute node.
The prior probability of attribute node S i refers to the probability that attribute node S i can be reached under static conditions. It is expressed as attribute node S i , and the joint probability of all attribute nodes in the path to attribute node S i . The prior probability of attribute node S i can be expressed by Equation (5).
P ( S i ) = i = 1 n P ( S i | P a r ( S i ) )

4. Constructing the Attack Benefit–Defense Cost Objective Function

4.1. Security Strategy Analysis

A safety strategy (ST) is a set of measures to reduce risk, and the state of S T i is defined as a Boolean variable with possible states of { 1 , 0 } . The True state means that the S T i has been implemented, and the False state means that the S T i has not been implemented. The safety strategy ST is expressed as S T = { S T i | i = 1 , 2 , , n } , where S T i = { 1 , 0 } , and the implementation of each S T i also has to pay the corresponding implementation cost C O S T ( S T i ) . The cost of each S T i is a factor that depends on the entire organization [19], including direct and indirect costs, and it should be determined by an expert on a case-by-case basis.
Implementing S T i against a vulnerability would increase the effort required by an attacker to exploit the vulnerability, resulting in a decrease in the exploitability of the vulnerability. The ability of ST to reduce the probability of vulnerability being exploited is measured by P ( S T i ) , and its value can be obtained from relevant security documents or expert experience. When protective countermeasures are enabled, the exploitability of the vulnerability is bound to be reduced, and the degree of reduction is denoted by P ( S T i ) . The prior probability of its attribute node will also be reduced; that is, the probability of the node being attacked by the attacker will be reduced, and the corresponding attack benefit will also be reduced. There are the following inequalities (6) for this.
P ( S i | S T i = 0 ) > P ( S i | S T i = 1 )
Among them, P ( S i | S T i = 0 ) represents the probability that the node S i will be successfully used when the protection countermeasure is not activated, and P ( S i | S T i = 1 ) represents the probability that the node S i will be successfully used when the protection countermeasure is activated.

4.2. Cost–Benefit Analysis

4.2.1. Cost Analysis

Implementing a security strategy inevitably incurs protection costs. The cost of protection is expressed as COST, C O S T = { C O S T 1 , C O S T 2 , , C O S T i } , where C O S T i represents the cost of implementing the security Strategy S T i . Therefore, the total protection cost of implementing the security strategy S T n = ( S T 1 , S T 2 , . , S T n ) can be expressed by Equation (7).
C ( S T n ) = i = 1 n S T i × C O S T i

4.2.2. Benefit Analysis

The attacker’s attack benefit A G ( S j ) , which indicates the profit obtained by the attacker after successfully attacking the attribute node, can be expressed by Equation (8).
A G ( S j ) = v a l u e × P ( S j )
Among them, v a l u e refers to the value of the attacked node S j , which can be determined by experts according to its importance and its own asset value, and P ( S j ) represents the prior probability of the attribute node S j . The attack gains of all attribute nodes can be expressed by Equation (9).
A G s ( S T n ) = S j S n P ( S j | S T j ) × A G ( S j )
Among them, S T n is the security strategy set, and P ( S j ) is the prior probability of the node after taking the security strategy.

4.3. Building a Cost–Benefit Function

Our goal is to select the most cost-effective strategy set to reduce the risk of attack, so the main problem is how to balance the benefits and costs of constructing strategies. Since adopting a security strategy will reduce the attacker’s attack benefit, the risk of the system being attacked will decrease, and at the same time, the corresponding cost will inevitably increase. For this reason, the following fitness function is used:
min ( α A G s ( S T n ) + β ( S T n ) )
C ( S T n ) < B
Among all strategies, a sub-strategy set is selected to minimize the sum of the attack benefit and the strategy cost without exceeding the cost budget.
Among them, α and β are the preference weights of the total attack benefit and the total cost of the security strategy, respectively, 0 α ,   β 1 ,   α + β = 1 .

5. Build an Optimal Security Strategy

5.1. Differential Evolution Algorithm

The differential evolution algorithm, as an efficient global optimization algorithm based on the natural evolution of species, is simple and efficient in practical applications [21]. It evolves a population of Np D-dimensional parameter vectors, mainly through mutation, crossover, and environmental selection operations. That is, first randomly generate initial test solutions; then, generate new test solutions through mutation and crossover operations. Next, determine which test solutions enter the next generation population to become candidate solutions through selection operations, and then repeat the iterative operations of mutation, crossover, and selection until all the experimental solutions are selected. The solution stops when the accuracy of the current solution meets the requirements or when a certain number of iterations is reached.
  • initialization
    Randomly select Np D-dimensional population individuals in the search space as the initial solution. The random initialization of the i-th individual X i t = 0 is a D-dimensional vector, which is denoted as X i t = 0 = { X i , 1 0 , X i , 2 0 , , X i , D 0 } , and it can be expressed by Equation (12) [22]
    X i , j t = 0 = X min , j + r a n d i , j [ 0 , 1 ] ( X max , j X min , j ) ,
    where X min , j and X max , j are the predetermined lower and upper bounds of the search space for the j-th design variable, respectively, and r a n d i , j [ 0 , 1 ] is a random number generated in [0, 1].
  • Mutations
    In the individual population X i t = { X 1 t , X 2 t , , X N p t } , three individuals are selected according to the mutation strategy to create a mutation vector V i t , and the three individuals are different from each other. There are several basic evolution strategies [27,28] in DE. The mutation strategy selected in this paper is “DE/best/1/bin”, and it can be expressed as Equation (13):
    V i t = X b e s t t + F ( X r 2 t X r 3 t ) ,
    where i = 1 , 2 , N p , r 2 , r 3 are arbitrary and unequal random integers in the set { 1 , 2 , N p } , and X b e s t t is the best individual. F is the scaling factor, which is a positive control parameter for scaling the difference vector, and it is often selected as 0.8. Then, we check whether the mutation vector violates the boundary constraint, and the violated individual returns the corresponding value according to the condition. The formula can be expressed by Equation (14).
    V 2 x min , j v i , j t , i f v i , j t < x min , j 2 x max , j v i , j t , i f v i , j t > x max , j v i , j t , o t h e r w i s e
    where v i , j t is the j-th component of v i t ; x min , j and x max , j are the maximum upper limit and the minimum lower limit of the search space of the j-dimensional variable, respectively.
  • Crossover
    The crossover operation is very similar to the crossover operation of the classical genetic algorithm, and its purpose is to diversify the individuals of the current population. However, the crossover here is different from the crossover operation of the genetic algorithm. The crossover operation here is for a certain dimension of the entire population, while the crossover in the genetic algorithm is for each individual in the population, and the experimental solution generated after the crossover operation is:
    u i , j t = v i , j t , i f ran d i , j [ 0 , 1 ] C r r a n d i ( 1 , D ) = j x i , j t , otherwise
  • Selection
    The experimental solution produced after cross mutation may not be better than the initial solution. Therefore, it is necessary to compare the fitness values of the two objective functions and select individuals with better fitness values as offspring, which can be expressed as:
    x t t + 1 = u i t , if f ( u i t ) f ( x i t ) x i t , otherwise

5.2. Reinforcement Learning and Q-Learning

Reinforcement learning is that the agent learns in the way of “trial and error” [29], and it interacts with the environment by sending out actions so as to obtain corresponding rewards or punishments. If you receive rewards, it will encourage you to continue this action, and if you receive punishments, you will avoid continuing this action. The goal of reinforcement learning is to let the agent continue to try and learn and finally obtain the maximum reward. Figure 2 shows the main components of reinforcement learning, including learning subject, action, environment, state and reward.
A very common method in reinforcement learning, Q-learning [30], is a model-free reinforcement learning algorithm that achieves the maximum cumulative expected value by telling the agent what action to take under what circumstances. Q-learning is a value-based learning algorithm that usually uses the Bellman equation [31] for updating. The update Equation (17) can be expressed as follows:
Q t + 1 ( s t , a t ) = Q ( s t , a t ) + α [ r t + γ max a Q ( s t + 1 , a ) Q ( s t , a t ) ]
Among them, r t is the reward or penalty value returned by the environment after the agent performs the action, α is the learning rate, γ is the discount factor, and the value range of the two is [0, 1], and Q t + 1 ( s t , a t ) is the total cumulative reward obtained by the agent at time t. The learning rate α determines how aggressively the agent reacts after receiving the reward; the higher the value, the more the Q value in the Q table fluctuates at each learning stage. So, the value of α should not be fixed; we dynamically adjust the value of α according to the literature [22] so that the value of α will be set to a higher value in the early stage of the evolution process, with the iteration—the number of times it increases and decreases—is linear, because the optimal action strategy will gradually converge as the iteration progresses. Its adjustment formula is:
α ( t ) = 1 0.9 g e n G
where gen and G are the number of function evaluations used and the maximum number of function evaluations, respectively.

5.3. Opposition-Based Learning

Since opposition-based learning was proposed, it has been widely used in various evolutionary algorithms [32] and has been proved to be an effective search method.
At present, there are two main purposes of opposition-based learning. One is to speed up the convergence speed [33]. The reason is that when solving the problem, the initial value is often set by random setting or accumulated experience. Therefore, we can try to obtain a better solution x * by introducing the reverse solution of x. If the reverse solution x * of x has a better effect, let x * replace x and enter the next generation. This will have a nearly 50 % higher probability of being closer to the global optimal solution, thus speeding up the convergence of the optimization algorithm. The second purpose is to jump out of the local optimal solution. The reason is that when the population evolves to a certain extent, it may have fallen into the local optimal solution. At this time, taking the reverse solution of the population has a certain probability of jumping out of the local optimal solution. At present, most intelligent optimization algorithms combined with opposition-based learning ignore its second function. This paper integrates the two functions into the proposed algorithm, and it has achieved very obvious results. The inverse solution can be calculated as follows:
x * = a + b x
where a is the minimum value and b is the maximum value. Again, this definition can be extended to higher dimensions with the formula:
x j * = a j + b j x j
Suppose f ( x ) is the fitness function of a given problem, P i ( x i 1 , x i 2 , x i n ) is the i-th individual of the current population, O P i ( x i 1 * , x i 2 * , , x i n * ) is the reverse individual corresponding to P i ; if f ( O P i ) is better than f ( P i ) , then O P i is used to replace P i , and we enter the next generation of the population.

5.4. Mixed Strategy Differential Evolution Algorithm (RODE)

Differential evolution algorithms have shown excellent performance in terms of final accuracy, computational speed and robustness when optimizing various objective functions. Canonical DE requires very few control parameters (three to be precise: scale factor, crossover rate, and population size). According to paper [34], the performance of the differential evolution algorithm largely depends on the choice of control parameters, such as scale factor (F) and crossover probability (CR). For this reason, there have been a lot of discussions and suggestions on the parameter control problem, but the relationship between parameter control and performance optimization is very complicated, and it has not been completely solved so far. The main reason is that there is no fixed parameter that can be applied to various problems or even different evolution stages of a single problem [21]. Reinforcement learning happens to be model-free; it can continuously adjust parameter values to adapt to different stages by allowing the agent to keep trial and error, keep learning and through feedback. Therefore, we will introduce how to combine reinforcement learning to choose the scale factor F and the crossover probability CR without the user setting fixed parameters.
It should be noted that this paper makes some improvements to the problem of industrial control strategy selection through the enhanced differential evolution algorithm in Reference [21] and Reference [22]. In Reference [21], the enhanced differential evolution learning is used to extract the parameters of the photovoltaic model, but only the scale factor (F) is learned, the cross probability (CR) is still a fixed value, and the reward values are only 1 and 0, which cannot reflect the advantages of actions with greater contribution. Reference [22] uses the enhanced differential evolution algorithm to verify the weight minimization of five benchmark examples of truss structures, thereby proving the effectiveness and robustness of the differential evolution algorithm based on reinforcement learning. In the article, F and CR are listed in the action set at the same time, but the selection action given is that seven groups contain fixed combinations of F and CR, and the reasons for the combination are not explained. At the same time, the reward and punishment strategy is also fixed at 1 and −1. In addition, even if reinforcement learning is introduced, the differential evolution algorithm has a faster convergence speed and better ability to jump out of the local optimal solution, but in the face of a small initial population and insufficient initial sample diversity, its convergence speed and the ability to jump out of the local optimal solution cannot achieve a very ideal effect, so we introduce opposition-based learning to solve this problem. The opposition-based learning is first applied to the initial population, so that the optimization algorithm can approach the optimal solution faster. This has been proved in the literature [35]. We added opposition-based learning after the successful crossover of the population individuals, and we set a very low probability to trigger the effect of oppositional learning. After many experiments, we found that when the initial population is large and the diversity is rich, setting the probability to 0.1 has the most obvious effect; when the initial population is small and the population lacks diversity, setting the probability to 0.2–0.4 has the most obvious effect. Figure 3, Figure 4 and Figure 5 below show the hybrid differential evolution algorithm (RODE) and the flow of comparing DE and RLDE.
The method proposed in this paper improves some of the deficiencies in the above literature methods so that it can better select industrial control strategies. The specific details of the algorithm framework are given below.

Parameter Setting in Q-Learning

In Q-learning, actions will affect the income, the state describes whether the individual mutates successfully, and the reward value reflects the impact of the action. F can be obtained [21] by the following Equation (21):
F = F + f
Among them, f is the action, and the possible values are −0.1, 0, 0.1. The action selection strategy adopts the greedy strategy. After choosing action F, mutate to generate new offspring, and then evaluate the fitness value of the objective function. We use R to represent reward or punishment. If the fitness value of the offspring is better than that of its parents, then R > 0, it is a reward; otherwise, R < 0, and it is a punishment. The value of R is obtained by the following formula (22).
R t = f ( x i t ) f ( u i t )
Among them, f ( u i t ) is the fitness value of the mutant individual i of the t generation; f ( x i t ) is the fitness value of the parents of the mutant i of the t generation.
In the algorithm of this paper, the state setting refers to reference [21], where state = { 1 , 2 } . In order to increase the individual diversity of the population, we make the initial state and F of each individual randomly assigned. If the solution of the offspring is worse than that of the parent during the evolution, it indicates that the selected action plays a positive role, and the state changes to 2; if the solution of the child is not as good as that of the parent, it means that the selected action plays a negative role, and the state changes to 1. Since the state and F of each individual are different, each individual has a Q table to update independently, and updating the Q table independently for each individual does not increase the time complexity of the algorithm. The Q table created in this paper is shown in Table 3 below.

5.5. Setting of Parameter CR

For CR, considering that if we continue to use Q l e a r n i n g to select it, the overall algorithm complexity will be too high, and the Q table will be too complicated. Therefore, we propose a formula to adjust it dynamically. In the initial stage, the CR should be at a high probability, and crossover operations are encouraged to better increase the diversity of the population. With the increase of the number of iterations, the CR should gradually decrease, because the strategy at this time has gradually converged to the optimal; our proposed formula is as follows.
C R = 0.9 0.5 g e n G

5.6. Framework of RODE

Algorithm 1 gives the framework of the mixed strategy-based differential evolution algorithm (RODE).
Algorithm 1 The pseudo of RODE
  • Input: population N, maximum number of iterations G
  • Output: optimal protection strategy
  1:
Initialize the population randomly, actions, states, Q table and other parameters;
  2:
Opposition-based learning: Find the reverse solution for the initial population
  3:
Calculate the initial population objective function
  4:
for g e n < G 1 do
  5:
    for e = 1:Np do
  6:
        choose action, calculate F;
  7:
        Mutation: DE/best/1
  8:
         V i t = X b e s t t + F ( X r 2 t X r 3 t )
  9:
        Crossover:
10:
         r = randi ( [ 1 , D ] , 1 , 1 )
11:
        for  n = 1 D  do
12:
            c r = r a n d
13:
            CR ( e ) = 0.9 0.5 ( gen / G ) ;
14:
           if  ( cr < = CR ( e ) | | ( n = = r ) )  then
15:
                u ( e , n ) = v ( e , n )
16:
           else
17:
                u ( e , n ) = x ( e , n ) ;
18:
           end if
19:
        end for
20:
        Selection:
21:
        Compute objective function of NP new candidates u e
22:
        Calculate Reward
23:
        if  f ( u e ) f ( x e )  then State = 2
24:
           Opposition-based learning:
25:
           if  c r < { 0.1 , 0.2 , 0.3 , 0.4 } then u ( e ) = a + b u ( e )
26:
                x ( e ) = u ( e )
27:
           else
28:
                x ( e ) = u ( e )
29:
           end if
30:
        else
31:
           State = 1
32:
        end if
33:
        update Q-table
34:
    end for
35:
end for

6. Experiment and Discussion

6.1. Experimental Scene

The experimental scene in this paper is a water distribution system, in which there are three PLCs, which correspond to the water pump, pipeline and water tanker in the virtual system, which simulates the real industrial control system, as shown in Figure 6.

6.2. Generating a Bayesian Attack Graph

We use Nessuss [36] to scan the vulnerabilities of the water distribution system, and according to the scores in Table 1, Formula (2) is used to calculate the successful probability of exploiting each vulnerability. Then, the network configuration and network connectivity information of the vulnerability are input into the input.P file of the MulVAL tool, and finally, the attack graph is obtained. The visualization of the attack graph is as shown in Figure 7.

6.3. Attack Benefits and Protection Costs

The prior probability of attacking all nodes in the graph and the cost of the security strategy can be shown in Table 4 and Appendix A Table A1.

6.4. Result Analysis

In this paper, we set the fitness function α to 0.5, which means that the attack benefit and protection cost are equally weighted. Cost B is set to have no upper limit. The strategy set selected by the evolutionary algorithm is used as the input of the fitness function, and the fitness value is obtained.
In order to prove the effectiveness of the proposed algorithm, we compare four other evolutionary algorithms of the same type. They are differential evolution algorithm (DE), particle swarm algorithm (PSO) [37], particle swarm algorithm based on reinforcement learning (RLPSO) [38], differential evolution algorithm based on reinforcement learning (RLDE) improved in this paper, and differential evolution algorithm based on mixed strategy proposed (RODE) in this paper. The relevant parameters of its setting are shown in Table 5. We uniformly set the initial population number of these five algorithms to 200 or 20, and we set the number of iterations to 20 uniformly.
Figure 8 shows the experimental results when the initial population is 200. It can be seen that when the initial population is relatively sufficient, most of the algorithms have achieved good results, among which DE, RLDE and RODE converge to the optimal solution of 459.4463. RLPSO and PSO converge to 532.6584 and 478.0463, respectively. When the initial population is 200, although DE, RODE, and RLDE converge to the optimal solution, the convergence speed of RODE and RLDE is obviously better than that of DE.
Figure 9 shows the experimental results when the initial population number is 20. It can be seen that when the initial population size is particularly small, that is, when the population diversity is not rich, the performance of the algorithm is inhibited to a certain extent. Among them, the effects of PSO, DE, RLPSO and RLDE are not as good as those of the first experiment, and relative to the first experiment, RLDE and DE failed to jump out of the local optimal solution even in the case of 100 iterations in this experiment. In the thirtieth generation, RODE jumped out of the local optimal solution and still converged to the optimal solution 459.4463, which is the same as the experimental result of the initial population number of 200.
In order to further prove the good robustness of the RODE algorithm, we conducted the second type of experiments. This experiment is based on the initial population size of 200 and 10, and we let each algorithm run 100 times. Our main purpose is to observe the convergence speed and convergence accuracy of various algorithms under different initial population sizes.
The experimental results when the initial population size is 200 are shown in Figure 10. Among them, PSO falls into the local optimal solution 91 times, DE falls into the local optimal solution zero times, RLPSO falls into the local optimal solution 76 times, RLDE falls into the local optimal solution four times, and RODE falls into the local optimal solution zero times. Although neither DE nor the RODE algorithm has fallen into local optimization, it is obvious that RODE has converged near the optimal solution in the second generation, and its convergence speed is significantly faster than that of DE. At the same time, it can be seen that the convergence speed of the algorithm that introduced reinforcement learning has been significantly improved.
The experimental results when the initial population size is 10 are shown in Figure 11.
When the initial population size is 10, PSO falls into the local optimal solution 100 times, DE falls into the local optimal solution 75 times, RLPSO falls into the local optimal solution 99 times, and RLDE falls into the local optimal solution 99 times. There are 16 times the RODE becomes stuck in a local optimal solution. For this reason, we can see that when the initial population size is only 10, except for the RODE algorithm that introduces opposition-based learning, the convergence accuracy of the remaining algorithms is extremely poor. The comparison results of RODE and other algorithms are shown in the following Table 6.
Through the comparison of two experiments, we can find that when the initial population size is large, the convergence accuracy of the algorithm is high, and the convergence speed of the algorithm incorporating reinforcement learning is significantly improved. When the population size is very small, the convergence performance of DE, PSO, RLPSO and RLDE algorithms becomes particularly poor, and it is difficult for them to jump out of the local optimal solution effectively under the condition of 100 iterations. However, RODE can still maintain a high convergence accuracy. This is because after the introduction of opposition-based learning, there is a high probability of jumping out of the local optimal solution in subsequent iterations, which is not available in other algorithms.
Therefore, the experimental results can prove that the RODE algorithm in this paper can largely resist the influence of the initial population, has strong robustness, and can quickly and accurately find the optimal strategy set. The disadvantage of this method is that it can only passively prevent attacks [39,40] in advance. If faced with frequent and changeable attack methods, the effect will be greatly discounted. For this reason, the main direction in the future is to study how to apply RODE to dynamic protection [41].

6.5. Discussion

As we all know, in view of the characteristics of industrial control security defense, the accuracy and speed of the selection strategy are particularly critical, which requires that the selected evolutionary algorithm not only solve the problem of convergence accuracy but also solve the problem of convergence speed, and consider the impact of population size. However, the previous evolutionary algorithms seem to focus on solving one of the problems, and they do not consider the effect of population size, because their tests are all carried out under the premise that the population size remains unchanged [7], while RODE considers the effect of population size. At the same time, reinforcement learning and opposition-based learning are also used to solve the problems of convergence speed and convergence accuracy, respectively. Furthermore, we verify the effectiveness of RODE at different population sizes through the above two sets of experiments.
Last but not least, in the second experiment, comparing RLDE and RODE in Figure 10, that is, when the population is 200, the performance of RODE and RLDE is almost the same, indicating that RODE has played a role in reinforcement learning. Comparing RLDE and RODE in Figure 11, that is, when the population is 10, the effect of RODE is much higher than that of RODE, because RODE has the effect of opposition-based learning at this time. So, we have reason to believe that RODE greatly improves the performance of the DE algorithm, and our method also provides a good idea for improving other evolutionary algorithms.

7. Conclusions

The scientific deployment of security strategies is the key to reducing the risk of ICS being attacked. In this paper, a differential evolution algorithm based on reinforcement learning and opposition-based learning is proposed to obtain the optimal security strategy by combining fitness functions. First, a Bayesian attack graph is used to model the potential attack paths of the industrial control system. Then, we calculate the attack benefit of each node and the cost of implementing the security strategy, and we build a fitness function based on the attack benefit and security cost. The fitness function can effectively evaluate the value of the selected solution. Finally, the RODE algorithm is used to select the security strategy scheme. The RODE algorithm is suitable for populations of various sizes. When the population size is large enough, the reinforcement learning in RODE can speed up the convergence rate; when the population size is relatively small, the opposition-based learning in RODE can increase the diversity of the population, thereby improving the ability to jump out of the local optimal solution and avoid the situation of premature convergence. A simplified water distribution system is used to validate the RODE algorithm. The experimental results show that the algorithm has better robustness than other algorithms of the same type, and it can find the optimal security strategy accurately and quickly. At present, our research is suitable for policy construction before the attack to minimize the loss of the attack, but it still has great shortcomings in dealing with the dynamic attack problem. To this end, in the future, we work on combining the RODE algorithm with policy selection in the face of dynamic attacks.

Author Contributions

Conceptualization, D.Z.; methodology, Y.Y.; software, Y.Y.; validation, L.X. and Y.Z.; formal analysis, L.X. and Y.Z.; writing—original draft preparation, Y.Y.; writing—review and editing, L.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key Research and Development Project of China (2020AAA0107700), the National Natural Science Foundation of China (62172244), the Shandong Provincial Natural Science Foundation (ZR2020YQ06, ZR2021MF132), and the Young innovation team of colleges and universities in Shandong province (2021KJ001).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Detailed parameters of protective strategies.
Table A1. Detailed parameters of protective strategies.
Protective StrategiesProtective ActionCost P ( S T i )
S T 1 Disconnect 192.168.0.1-192.168.0.10200.25
S T 2 Disconnect Disconnect 192.168.0.2-192.168.0.10200.25
S T 3 Disconnect Internet 192.168.0.10300.30
S T 4 Disconnect Internet 192.168.0.2300.30
S T 5 Disconnect Internet700.40
S T 6 Disable Internet600.35
S T 7 Disable Internet direct network access550.32
S T 8 Disconnect Internet 192.168.0.1300.30
S T 9 Disable multi-hop access 192.168.0.10250.20
S T 10 Disable udp900.60
S T 11 Disable direct network access 192.168.0.10350.33
S T 12 Disable direct network access 192.168.0.2300.30
S T 13 Disable direct network access 192.168.0.1300.30
S T 14 Disable netAccess 192.168.0.10310.40
S T 15 Disable networkService 192.168.0.10310.40
S T 16 Patch CVE-1999-0517 192.168.0.10800.75
S T 17 Disable service programs180.20
S T 18 Disconnect 192.168.0.10-192.168.0.1200.25
S T 19 Disable execCode 192.168.0.10180.20
S T 20 Disconnect 192.168.0.10200.30
S T 21 Disconnect 192.168.0.10-192.168.0.2380.40
S T 22 Disconnect 192.168.0.1-192.168.0.2380.40
S T 23 Disable multi-hop access 192.168.0.1260.35
S T 24 Disable multi-hop access 192.168.0.2260.35
S T 25 Disable netAccess 192.168.0.2200.30
S T 26 Disable netAccess350.45
S T 27 Disable networkService 192.168.0.2380.40
S T 28 Patch CVE-1999-0517 192.168.0.21100.75
S T 29 Disable execCode 192.168.0.2600.55
S T 30 Disable multi-hop access100.20
S T 31 Disconnect 192.168.0.2-192.168.0.1900.38
S T 32 Disable netAccess 192.168.0.1800.36
S T 33 Disable service program 192.168.0.1800.36
S T 34 Disable networkService 192.168.0.11900.48
S T 35 Patch CVE-1999-0517 192.168.0.12000.70
S T 36 Install ids3000.60

References

  1. Chen, T.M.; Abu-Nimeh, S. Lessons from Stuxnet. Computer 2011, 44, 91–93. Available online: https://0-ieeexplore-ieee-org.brum.beds.ac.uk/abstract/document/5742014 (accessed on 17 September 2022). [CrossRef]
  2. Case, D.U. Analysis of the cyber attack on the Ukrainian power grid. Electr. Inf. Shar. Anal. Cent. (E-ISAC) 2016, 388, 1–29. Available online: https://africautc.org/wp-content/uploads/2018/05/E-ISAC_SANS_Ukraine_DUC_5.pdf (accessed on 17 September 2022).
  3. Nespoli, P.; Papamartzivanos, D.; Mármol, F.G.; Kambourakis, G. Optimal countermeasures selection against cyber attacks: A comprehensive survey on reaction frameworks. IEEE Commun. Surv. Tutorials 2017, 20, 1361–1396. Available online: https://0-ieeexplore-ieee-org.brum.beds.ac.uk/abstract/document/8169023 (accessed on 17 September 2022). [CrossRef]
  4. Zhao, D.; Wang, L.; Wang, Z.; Xiao, G. Virus propagation and patch distribution in multiplex networks: Modeling, analysis, and optimal allocation. IEEE Trans. Inf. Forensics Secur. 2018, 14, 1755–1767. Available online: https://0-ieeexplore-ieee-org.brum.beds.ac.uk/abstract/document/8565928 (accessed on 17 September 2022). [CrossRef]
  5. Lee, C.; Han, S.M.; Chae, Y.H.; Seong, P.H. Development of a cyberattack response planning method for nuclear power plants by using the Markov decision process model. Ann. Nucl. Energy 2022, 166, 108725. Available online: https://0-www-sciencedirect-com.brum.beds.ac.uk/science/article/pii/S0306454921006010 (accessed on 17 September 2022). [CrossRef]
  6. Dewri, R.; Ray, I.; Poolsappasit, N.; Whitley, D. Optimal security hardening on attack tree models of networks: A cost-benefit analysis. Int. J. Inf. Secur. 2012, 11, 167–188. Available online: https://linkspringer.53yu.com/article/10.1007/s10207-012-0160-y (accessed on 17 September 2022). [CrossRef]
  7. Poolsappasit, N.; Dewri, R.; Ray, I. Dynamic security risk management using bayesian attack graphs. IEEE Trans. Dependable Secur. Comput. 2011, 9, 61–74. Available online: https://0-ieeexplore-ieee-org.brum.beds.ac.uk/abstract/document/5936075 (accessed on 17 September 2022). [CrossRef]
  8. Yang, X.S. A new metaheuristic bat-inspired algorithm. In Nature Inspired Cooperative Strategies for Optimization (NICSO 2010); Springer: Berlin/Heidelberg, Germany, 2010; pp. 65–74. Available online: https://linkspringer.53yu.com/chapter/10.1007/978-3-642-12538-6_6 (accessed on 17 September 2022).
  9. Meng, Z.; Yang, C. Two-stage differential evolution with novel parameter control. Inf. Sci. 2022, 596, 321–342. Available online: https://0-www-sciencedirect-com.brum.beds.ac.uk/science/article/pii/S0020025522002547 (accessed on 17 September 2022). [CrossRef]
  10. Dixit, A.; Mani, A.; Bansal, R. An adaptive mutation strategy for differential evolution algorithm based on particle swarm optimization. Evol. Intell. 2022, 15, 1571–1585. Available online: https://linkspringer.53yu.com/article/10.1007/s12065-021-00568-z (accessed on 17 September 2022). [CrossRef]
  11. Kaelbling, L.P.; Littman, M.L.; Moore, A.W. Reinforcement learning: A survey. J. Artif. Intell. Res. 1996, 4, 237–285. Available online: https://www.jair.org/index.php/jair/article/view/10166 (accessed on 17 September 2022). [CrossRef]
  12. Xu, Z.; Han, G.; Liu, L.; Martínez-García, M.; Wang, Z. Multi-energy scheduling of an industrial integrated energy system by reinforcement learning-based differential evolution. IEEE Trans. Green Commun. Netw. 2021, 5, 1077–1090. Available online: https://0-ieeexplore-ieee-org.brum.beds.ac.uk/abstract/document/9361643 (accessed on 17 September 2022). [CrossRef]
  13. Liao, Z.; Li, S. Solving Nonlinear Equations Systems with an Enhanced Reinforcement Learning Based Differential Evolution. Complex Syst. Model. Simul. 2022, 2, 78–95. Available online: https://0-ieeexplore-ieee-org.brum.beds.ac.uk/abstract/document/9770100 (accessed on 17 September 2022). [CrossRef]
  14. Tizhoosh, H.R. Opposition-based learning: A new scheme for machine intelligence. In Proceedings of the International Conference on Computational Intelligence for Modelling, Control and Automation and International Conference on Intelligent Agents, Web Technologies and Internet Commerce (CIMCA-IAWTIC’06, Vienna, Austria, 28–30 November 2005; Volume 1, pp. 695–701. Available online: https://0-ieeexplore-ieee-org.brum.beds.ac.uk/abstract/document/1631345 (accessed on 17 September 2022).
  15. Deng, W.; Ni, H.; Liu, Y.; Chen, H.; Zhao, H. An adaptive differential evolution algorithm based on belief space and generalized opposition-based learning for resource allocation. Appl. Soft Comput. 2022, 127, 109419. Available online: https://0-www-sciencedirect-com.brum.beds.ac.uk/science/article/pii/S1568494622005518 (accessed on 17 September 2022). [CrossRef]
  16. Abed-alguni, B.H.; Paul, D. Island-based Cuckoo Search with elite opposition-based learning and multiple mutation methods for solving optimization problems. Soft Comput. 2022, 26, 3293–3312. Available online: https://linkspringer.53yu.com/article/10.1007/s00500-021-06665-6 (accessed on 17 September 2022). [CrossRef]
  17. Tubishat, M.; Idris, N.; Shuib, L.; Abushariah, M.A.; Mirjalili, S. Improved Salp Swarm Algorithm based on opposition based learning and novel local search algorithm for feature selection. Expert Syst. Appl. 2020, 145, 113122. Available online: https://0-www-sciencedirect-com.brum.beds.ac.uk/science/article/abs/pii/S0957417419308395 (accessed on 17 September 2022). [CrossRef]
  18. Hussien, A.G.; Amin, M. A self-adaptive Harris Hawks optimization algorithm with opposition-based learning and chaotic local search strategy for global optimization and feature selection. Int. J. Mach. Learn. Cybern. 2022, 13, 309–336. Available online: https://linkspringer.53yu.com/article/10.1007/s13042-021-01326-4 (accessed on 17 September 2022). [CrossRef]
  19. Rahnamayan, S.; Tizhoosh, H.R.; Salama, M.M. Opposition-based differential evolution algorithms. In Proceedings of the 2006 IEEE International Conference on Evolutionary Computation, Vancouver, BC, Canada, 16–21 July 2006; pp. 2010–2017. Available online: https://0-ieeexplore-ieee-org.brum.beds.ac.uk/abstract/document/1688554 (accessed on 17 September 2022).
  20. Fister, I.; Fister, D. Reinforcement Learning-Based Differential Evolution for Global Optimization. In Differential Evolution: From Theory to Practice; Springer: Berlin/Heidelberg, Germany, 2022; pp. 43–75. Available online: https://linkspringer.53yu.com/chapter/10.1007/978-981-16-8082-3_3 (accessed on 17 September 2022).
  21. Hu, Z.; Gong, W.; Li, S. Reinforcement learning-based differential evolution for parameters extraction of photovoltaic models. Energy Rep. 2021, 7, 916–928. Available online: https://0-www-sciencedirect-com.brum.beds.ac.uk/science/article/pii/S2352484721000974 (accessed on 17 September 2022). [CrossRef]
  22. Huynh, T.N.; Do, D.T.; Lee, J. Q-Learning-based parameter control in differential evolution for structural optimization. Appl. Soft Comput. 2021, 107, 107464. Available online: https://0-www-sciencedirect-com.brum.beds.ac.uk/science/article/abs/pii/S1568494621003872 (accessed on 17 September 2022). [CrossRef]
  23. Roy, A.; Kim, D.S.; Trivedi, K.S. Scalable optimal countermeasure selection using implicit enumeration on attack countermeasure trees. In Proceedings of the IEEE/IFIP International Conference on Dependable Systems and Networks (DSN 2012), Boston, MA, USA, 25–28 June 2012; Available online: https://0-ieeexplore-ieee-org.brum.beds.ac.uk/abstract/document/6263940 (accessed on 17 September 2022).
  24. Wang, S.; Zhang, Z.; Kadobayashi, Y. Exploring attack graph for cost-benefit security hardening: A probabilistic approach. Comput. Secur. 2013, 32, 158–169. Available online: https://0-www-sciencedirect-com.brum.beds.ac.uk/science/article/pii/S0167404812001496 (accessed on 17 September 2022). [CrossRef]
  25. Khosravi-Farmad, M.; Ghaemi-Bafghi, A. Bayesian decision network-based security risk management framework. J. Netw. Syst. Manag. 2020, 28, 1794–1819. Available online: https://linkspringer.53yu.com/article/10.1007/s10922-020-09558-5 (accessed on 17 September 2022). [CrossRef]
  26. Gallon, L.; Bascou, J.J. Using CVSS in attack graphs. In Proceedings of the 2011 Sixth International Conference on Availability, Reliability and Security, Vienna, Austria, 22–26 August 2011; pp. 59–66. Available online: https://0-ieeexplore-ieee-org.brum.beds.ac.uk/abstract/document/6045939 (accessed on 17 September 2022).
  27. Qin, A.K.; Huang, V.L.; Suganthan, P.N. Differential evolution algorithm with strategy adaptation for global numerical optimization. IEEE Trans. Evol. Comput. 2008, 13, 398–417. Available online: https://0-ieeexplore-ieee-org.brum.beds.ac.uk/abstract/document/4632146 (accessed on 17 September 2022). [CrossRef]
  28. Hansen, N.; Ostermeier, A. Completely derandomized self-adaptation in evolution strategies. Evol. Comput. 2001, 9, 159–195. Available online: https://0-ieeexplore-ieee-org.brum.beds.ac.uk/abstract/document/6790628 (accessed on 17 September 2022). [CrossRef]
  29. Samma, H.; Lim, C.P.; Saleh, J.M. A new reinforcement learning-based memetic particle swarm optimizer. Appl. Soft Comput. 2016, 43, 276–297. Available online: https://0-www-sciencedirect-com.brum.beds.ac.uk/science/article/abs/pii/S1568494616000132 (accessed on 17 September 2022). [CrossRef]
  30. Watkins, C.J.; Dayan, P. Q-learning. Mach. Learn. 1992, 8, 279–292. Available online: https://linkspringer.53yu.com/article/10.1007/BF00992698 (accessed on 17 September 2022). [CrossRef]
  31. O’Donoghue, B.; Osband, I.; Munos, R.; Mnih, V. The uncertainty bellman equation and exploration. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 3836–3845. Available online: http://proceedings.mlr.press/v80/o-donoghue18a/o-donoghue18a.pdf (accessed on 17 September 2022).
  32. Ming, H.; Wang, M.; Liang, X. An improved genetic algorithm using opposition-based learning for flexible job-shop scheduling problem. In Proceedings of the 2016 2nd International Conference on Cloud Computing and Internet of Things (CCIOT), Dalian, China, 22–23 October 2016; pp. 8–15. Available online: https://0-ieeexplore-ieee-org.brum.beds.ac.uk/abstract/document/7868294 (accessed on 17 September 2022).
  33. Agarwal, M.; Srivastava, G.M.S. Opposition-based learning inspired particle swarm optimization (OPSO) scheme for task scheduling problem in cloud computing. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 9855–9875. Available online: https://linkspringer.53yu.com/article/10.1007/s12652-020-02730-4 (accessed on 17 September 2022). [CrossRef]
  34. Gämperle, R.; Müller, S.D.; Koumoutsakos, P. A parameter study for differential evolution. Adv. Intell. Syst. Fuzzy Syst. Evol. Comput. 2002, 10, 293–298. Available online: https://www.cse-lab.ethz.ch/wp-content/uploads/2008/04/gmk_wseas_2002.pdf (accessed on 17 September 2022).
  35. Si, T.; Miranda, P.B.; Bhattacharya, D. Novel enhanced Salp Swarm Algorithms using opposition-based learning schemes for global optimization problems. Expert Syst. Appl. 2022, 207, 117961. Available online: https://0-www-sciencedirect-com.brum.beds.ac.uk/science/article/abs/pii/S0957417422011940 (accessed on 17 September 2022). [CrossRef]
  36. Anderson, H. Introduction to Nessus. SecurityFocus Printable INFOCUS 2003. Available online: http://cryptomex.org/SlidesSeguRedes/TutNessus.pdf (accessed on 17 September 2022).
  37. Marini, F.; Walczak, B. Particle swarm optimization (PSO). A tutorial. Chemom. Intell. Lab. Syst. 2015, 149, 153–165. Available online: https://0-www-sciencedirect-com.brum.beds.ac.uk/science/article/abs/pii/S0169743915002117 (accessed on 17 September 2022). [CrossRef]
  38. Liu, Y.; Lu, H.; Cheng, S.; Shi, Y. An adaptive online parameter control algorithm for particle swarm optimization based on reinforcement learning. In Proceedings of the 2019 IEEE Congress on Evolutionary Computation (CEC), Wellington, New Zealand, 10–13 June 2019; pp. 815–822. Available online: https://0-ieeexplore-ieee-org.brum.beds.ac.uk/abstract/document/8790035 (accessed on 17 September 2022).
  39. Ades, S.; Herrera, D.A.; Lahey, T.; Thomas, A.A.; Jasra, S.; Barry, M.; Sprague, J.; Dittus, K.; Plante, T.B.; Kelly, J.; et al. Cancer care in the wake of a cyberattack: How to prepare and what to expect. JCO Oncol. Pract. 2022, 18, 23–34. Available online: https://0-www-sciencedirect-com.brum.beds.ac.uk/science/article/abs/pii/S1879850021002757 (accessed on 17 September 2022). [CrossRef]
  40. Teoh, A.A.; Ghani, N.B.A.; Ahmad, M.; Jhanjhi, N.; Alzain, M.A.; Masud, M. Organizational data breach: Building conscious care behavior in incident response. Comput. Syst. Sci. Eng. 2022, 40, 505–515. Available online: https://eprints.um.edu.my/33609/ (accessed on 17 September 2022). [CrossRef]
  41. Li, X.; Zhou, C.; Tian, Y.C.; Qin, Y. A dynamic decision-making approach for intrusion response in industrial control systems. IEEE Trans. Ind. Inform. 2018, 15, 2544–2554. Available online: https://0-ieeexplore-ieee-org.brum.beds.ac.uk/abstract/document/8443136 (accessed on 17 September 2022). [CrossRef] [Green Version]
Figure 1. Optimal strategy selection framework for industrial control.
Figure 1. Optimal strategy selection framework for industrial control.
Applsci 12 09594 g001
Figure 2. Reinforcement learning description graph.
Figure 2. Reinforcement learning description graph.
Applsci 12 09594 g002
Figure 3. DE.
Figure 3. DE.
Applsci 12 09594 g003
Figure 4. RLDE.
Figure 4. RLDE.
Applsci 12 09594 g004
Figure 5. RODE.
Figure 5. RODE.
Applsci 12 09594 g005
Figure 6. Experimental scene.
Figure 6. Experimental scene.
Applsci 12 09594 g006
Figure 7. Attack Graph.
Figure 7. Attack Graph.
Applsci 12 09594 g007
Figure 8. Experimental comparison with an initial population of 200.
Figure 8. Experimental comparison with an initial population of 200.
Applsci 12 09594 g008
Figure 9. Experimental comparison with an initial population of 20.
Figure 9. Experimental comparison with an initial population of 20.
Applsci 12 09594 g009
Figure 10. Algorithm convergence test when the initial population size is 200. (a) DE algorithm runs 100 times. (b) PSO algorithm runs 100 times. (c) RLDE algorithm runs 100 times. (d) RLPSO algorithm runs 100 times. (e) RODE algorithm runs 100 times.
Figure 10. Algorithm convergence test when the initial population size is 200. (a) DE algorithm runs 100 times. (b) PSO algorithm runs 100 times. (c) RLDE algorithm runs 100 times. (d) RLPSO algorithm runs 100 times. (e) RODE algorithm runs 100 times.
Applsci 12 09594 g010
Figure 11. Algorithm convergence test when the initial population size is 10. (a) DE algorithm runs 100 times. (b) PSO algorithm runs 100 times. (c) RLDE algorithm runs 100 times. (d) RLPSO algorithm runs 100 times. (e) RODE algorithm runs 100 times.
Figure 11. Algorithm convergence test when the initial population size is 10. (a) DE algorithm runs 100 times. (b) PSO algorithm runs 100 times. (c) RLDE algorithm runs 100 times. (d) RLPSO algorithm runs 100 times. (e) RODE algorithm runs 100 times.
Applsci 12 09594 g011
Table 1. CVSS Score (base metric).
Table 1. CVSS Score (base metric).
IndexRankScore
Access Vector (AV)local access0.55
adjacent network access0.62
network accessible0.85
Access Complexity (AC)high0.44
low0.77
Authentication (AU)multiple instances of authentication0.50
single instance of authentication0.62
no authentication0.85
Table 2. CVSS Score (temporal metric).
Table 2. CVSS Score (temporal metric).
IndexRankScore
Exploit Code Maturity (E)Unproven (U)0.91
Proof-of-Conc (POC)0.94
Functional (F)0.97
High (H)1
Remediation Level (RL)Official (O)0.95
Temporary Fix (T)0.96
Workaround (W)0.97
Unavailable (U)1
Report Confidence (RC)Unknown (U)0.92
Reasonable (R)0.96
Confirmed (C)1
Table 3. Q-table for a single individual.
Table 3. Q-table for a single individual.
StateA1A2A3
1 Q ( s 1 , a 1 ) Q ( s 1 , a 2 ) Q ( s 1 , a 3 )
2 Q ( s 2 , a 1 ) Q ( s 2 , a 2 ) Q ( s 2 , a 3 )
Table 4. Prior Probability.
Table 4. Prior Probability.
Attribute NotesPriori ProbabilityAttribute NotesPriori ProbabilityAttribute NotesPriori Probability
S10.5000S20.1811S30.5000
S41.0000S51.0000S61.0000
S70.3967S80.5100S91.0000
S101.0000S111.0000S121.0000
S131.0000S140.3890S150.5297
S161.0000S171.0000S181.0000
S191.0000S201.0000S211.0000
S221.0000S231.0000S240.0801
S251.0000S261.0000S271.0000
S280.0800S291.0000S301.0000
S311.0000S321.0000S331.0000
S340.0336
Table 5. Parameter settings of investigated algorithms.
Table 5. Parameter settings of investigated algorithms.
AlgorithmParameter SettingsValue
PSOw0.9
c10.5
c22.5
RLPSO α 0.9
γ 0.1
DEF0.8
CR0.6
RLDE γ 0.8
RODE γ 0.8
Table 6. The number of times each algorithm falls into the local optimal solution.
Table 6. The number of times each algorithm falls into the local optimal solution.
Algorithm20010
PSO91100
RLPSO7699
DE075
RLDE499
RODE016
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yue, Y.; Zhou, Y.; Xu, L.; Zhao, D. Optimal Defense Strategy Selection Algorithm Based on Reinforcement Learning and Opposition-Based Learning. Appl. Sci. 2022, 12, 9594. https://0-doi-org.brum.beds.ac.uk/10.3390/app12199594

AMA Style

Yue Y, Zhou Y, Xu L, Zhao D. Optimal Defense Strategy Selection Algorithm Based on Reinforcement Learning and Opposition-Based Learning. Applied Sciences. 2022; 12(19):9594. https://0-doi-org.brum.beds.ac.uk/10.3390/app12199594

Chicago/Turabian Style

Yue, Yiqun, Yang Zhou, Lijuan Xu, and Dawei Zhao. 2022. "Optimal Defense Strategy Selection Algorithm Based on Reinforcement Learning and Opposition-Based Learning" Applied Sciences 12, no. 19: 9594. https://0-doi-org.brum.beds.ac.uk/10.3390/app12199594

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop