Next Article in Journal
A Practical Method for the Automatic Recognition of Rock Structures in Panoramic Borehole Image during Deep-Hole Drilling Engineering
Previous Article in Journal
Systematic Experimental Evaluation of Function Based Cellular Lattice Structure Manufactured by 3D Printing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Distributed Biased Boundary Attack Method in Black-Box Attack

1
College of Intelligence Science and Technology, National University of Defense and Technology, Changsha 410000, China
2
School of Civil Engineering, Central South University, Changsha 410000, China
*
Authors to whom correspondence should be addressed.
Submission received: 29 September 2021 / Revised: 2 November 2021 / Accepted: 4 November 2021 / Published: 8 November 2021

Abstract

:

Featured Application

The paper is about black-box attack methods and the generation of adversarial samples.

Abstract

The adversarial samples threaten the effectiveness of machine learning (ML) models and algorithms in many applications. In particular, black-box attack methods are quite close to actual scenarios. Research on black-box attack methods and the generation of adversarial samples is helpful to discover the defects of machine learning models. It can strengthen the robustness of machine learning algorithms models. Such methods require queries frequently, which are less efficient. This paper has made improvements in the initial generation and the search for the most effective adversarial examples. Besides, it is found that some indicators can be used to detect attacks, which is a new foundation compared with our previous studies. Firstly, the paper proposed an algorithm to generate initial adversarial samples with a smaller L2 norm; secondly, a combination between particle swarm optimization (PSO) and biased boundary adversarial attack (BBA) is proposed. It is the PSO-BBA. Experiments are conducted on the ImageNet. The PSO-BBA is compared with the baseline method. Experimental comparison results certificate that: (1) A distributed framework for adversarial attack methods is proposed; (2) The proposed initial point selection method can reduces query numbers effectively; (3) Compared to the original BBA, the proposed PSO-BBA algorithm accelerates the convergence speed and improves the accuracy of attack accuracy; (4) The improved PSO-BBA algorithm has preferable performance on targeted and non-targeted attacks; (5) The mean structural similarity (MSSIM) can be used as the indicators of adversarial attack.

1. Introduction

Deep learning is occupying the core of the rapidly developing field of machine learning and artificial intelligence research. It also demonstrates good performance on many tasks, especially in the field of computer vision. However, modern deep networks are very vulnerable to adversarial samples, which presents a great threat to the effectiveness and stabilization of the community [1,2,3,4,5]. It is shown that image classification methods based on deep neural networks (DNN) are always fragile. They will make mistakes when only small disturbances occur, which are invisible to the human eye. This indicates that many machine learning models trained with a large number of samples are still vulnerable to small adversarial disturbances. The algorithm to find this kind of anti-interference to the input image is usually called an anti-attack. The mathematical expression is as Equation (1). The C(x) is the classifier. In machine learning, the function of the classifier is to judge the category of a new input sample based on the labeled training data. Value (x, y) is an input sample and its label. The attack is to produce an adversarial sample xadv of x. When xadv is misclassified by ML model. The Lp norm of the difference between it and the original one is below a set threshold ε. It means that the attack succeeded.
C ( x a d v ) y , s . t . x a d v x p < ε
where ||•||p is Lp norm.
The generation of adversarial samples of the image classification model is the process of adding subtle interference, which is difficult for humans to recognize in the original image. It is inputted into the deep neural network. The deep neural network methods make wrong classification results. In Figure 1, slight interferences are added to the original image. The deep neural network methods give wrong judgment. The monkey is identified as a hand sanitizer.
Adversarial attack methods generally fall into two categories on the basis of attack effect and method. In the light of attack effects, these come in two classes: non-targeted and targeted attack methods. Non-targeted attack refers to the generation of adversarial samples by adding interference. It causes the classification algorithm to mistakenly classify the original image into images of different categories. Contrarily, targeted attack causes the classification algorithm to misclassify the original image as an image of a specified category.
On the basis of methods, these come in two classes: white and black-box methods. The white-box attack methods know about attacking deep neural network methods and model parameters in advance, such as the construction of networks, the situation of layers, neurons, etc. Additionally, they know the kinds of hyperparameters about the neural network, and the weights of connected neurons, for example. The attacker can design a targeted adversarial sample on the basis of the known information with regard to the neural network. This can make the neural network method lose efficacy. Such methods comprise the fast gradient attack proposed by Ian Goodfellow in 2014 [6] (Fast Gradient Sign Method, FGSM), Basic Iterative Method [7] (BIM), and DeepFool [8], Jacobian-based Saliency Map [9] (JSMA), Houdini [10] and Carlini & Wagner [11], and so on. Such attack algorithms mainly rely on gradient-based attack strategies.
Conversely, black-box attack methods do not know any information about the network, such as architecture or various hyperparameters. These only have the output results of the network and corresponding data labels. These ones are closer to the real application scenario than white ones. Such methods are harder to implement than white-box methods. Meanwhile, the attack effect is relatively poor. In terms of safety, the robustness of the neural network can be improved by the research of black-box methods. The network is trained with the corresponding adversarial samples. It is very helpful in improving the security of the network application.
The black-box attack methods can be generally divided into three categories: score-based attack, transfer-based attack and decision-based attack methods. Score-based attacks rely on predicted scores, such as the class probability or distribution of the model. These attacks use predicted values to estimate the gradient, mainly including the black-box variants of JSMA [12], zero-order Optimization attacks [13], predictive adversarial generative networks [14], and simple black-box attacks [15,16,17]. The disadvantage of this type of method is that it is difficult to obtain the score and probability distribution of the output sample in real applications, so it cannot be practically used.
A transfer-based attack does not rely on model information. It requires information about training data, which are used to train a fully observable agent model. The adversarial disturbances can be synthesized [18]. It relies on empirical observations. The adversarial examples are often transferred between models. If adversarial examples are created on a set of proxy models, the success rate of attacking models can reach 100% in some cases [19]. Dong proposed a prior stochastic gradient free method [20] to improve the performance of black-box adversarial attacks. The algorithm also uses transfer based on the priori and score-based information. The attack success rate is higher than only one method. If the data sets are inconsistent, the attack effect may decrease a lot, and the stability is not robust [21].
Decision-based attacks are direct attacks that only rely on the final decision of the model, such as Top-1 classification labels. These attack methods are aimed at the situation where the black-box output classification results can be obtained. In an actual situation, the confidence scores or model distribution are difficult to obtain. Compared with score-based attacks, decision-based attacks are more consistent with real applications. Decision-based attacks are more robust to standard defenses than other types of attacks, such as gradient masking, inherent randomness or robustness training. Compared with transfer-based attacks, decision-based attacks require much less information about the model: neither architecture nor training data. They are simpler to apply. The disadvantage of such methods is that the queries are much higher than in other scenarios. At present, decision-based attacks mainly include the boundary attack proposed by Brendal et al. in 2018 [12] (Boundary Attack, BA) and its variants [22]. Thomas Brunner is equivalent to the biased boundary attack proposed in 2019 [13] (Biased Boundary Attack, BBA). On the basis of the attack algorithm, low-frequency disturbance and surrogate model disturbance are added. The cutting plane attack proposed by Ren [23] (Cutting plane attack) improves the convergence method based on the boundary attack algorithm. Comparative analysis of the above three types of black-box attack methods is shown in Table 1.
Comprehensive comparison of various black-box attack methods shows that black-box adversarial methods can accomplish targeted and non-targeted scenarios. However, in score and transfer-based methods, information on training data and models are required. Strictly speaking, they are not a complete black-box method. By contrast, decision-based boundary methods are simple and flexible concepts. They simply need to know about the labels of results. These are a type of “complete black-box attack method”, and keep pace with real scenarios. They have good performance in target and non-target attacks with great application potential.
Currently, boundary attack methods and improved ones are best performance decision-based methods. A boundary attack algorithm is a black-box attack algorithm that meets the requirements of the real environment. It has superior effects in scenarios where only the input image label is obtained, and it is simple, easy to implement and robust. The improved one won the runner-up in the 2018 NeurIPS competition, named Biased Boundary Attack (BBA). It showed the BBA’s superiority in actual adversarial scenarios. However, BBA still has the following two challenges: (1) query numbers are high on the premise of guaranteeing the success rate; (2) The algorithm converges slowly.
This paper studies and analyzes the advantages and disadvantages of various current black-box attack methods. The decision-based methods are the research focus. The paper combines this with the swarm intelligence algorithm based on the biased boundary attack black-box adversarial method. A biased boundary attack adversarial algorithm based on particle swarm optimization is proposed. The methods mainly have the following innovations:
Firstly, the method of initial point generation is proposed for improvement. Initial point selection has great influence on the properties of BA and BBA. The original method for selection of the initial point cannot achieve the optimal effect. This paper proposes a new algorithm for the generation of the initial attack point. The experiment shows that it can reduce the number of attack queries and the convergence speed of the algorithm.
Secondly, the particle swarm optimization algorithm is used to search the optimal adversarial samples from multiple initial points. In the framework of the BA algorithm, the generation of optimal adversarial samples can be modeled as a regional search optimization problem. The swarm intelligence algorithm has advantages of fast convergence and a good optimization effect on this kind of problem. It is feasible to introduce swarm intelligence algorithms as the direction of improvement. We try to apply the particle swarm optimization algorithm to the algorithm-updating process. Multiple initial points are adopted to find the best adversarial sample points. It can not only expand the search space of the algorithm, but also improve the convergence rate. It avoids falling into a local optimal solution effectively. It also promotes the dependability and robustness of attack applications.
With the above corrective measure, on the premise of complete black-box with only input sample query labels, the proposed algorithm achieves target/non-target attack. Compared with BA, BBA and other baseline algorithms, query numbers are reduced. The success rate of adversarial attacks is improved. The convergence speed of the algorithm is better.
The rest of the paper is organized as follows. In Section 2, the related works on Boundary Attack (BA) and Biased boundary attack algorithm (BBA) are introduced. Section 3 proposes a distributed framework for adversarial attack methods. The experimental analysis is also given in Section 4. Finally, the disadvantages and future development directions are discussed concerning future work.

2. Related Works

2.1. Boundary Attack (BA)

Boundary attack algorithm is a kind of typical decision-based method. Starting from an initial adversarial image, a binary search is used to find a sample point, which is near the classification frontier. Random walk is performed along the frontier between two opposite areas. It reduces the distance from the target image. That is to say, the classification result obtained by input classifier query is always the category we want to misclassify. According to this step, we continue to iterate and gradually reduce the distance from the original image. The reason why this kind of algorithm is named “boundary attack” is that it generates adversarial samples by searching along the boundary until it converges to obtain the optimal or satisfactory solution.
The results obtained by this kind of method can meet the requirements for misclassification of the black-box model. The overall disturbance that it increases relative to the original image varies with the performance of the algorithm. Therefore, the criterion for measuring a successful adversarial sample is that the difference ||xadvx||p between it and the original image should be less than the given threshold ε. Take the ImageNet data set as an example: the image size is 299 × 299. The pixel value interval is (0, 255). When p = 2 (that is, the L2 norm) ε takes 25.89. This measure is used in our research and experimental results. As shown in Figure 2a, the generation of optimal adversarial samples can be considered as a search and optimization problem along the classification boundary between adversarial space and original space. Those boundary points that are both close to the original image and on the adversarial sample space are the feasible solutions that we want to find. However, in the decision-based black-box adversarial scene, the position and gradient direction of the real classification boundary are unknown. Therefore, it can be regarded as a high-dimensional space search problem in an uncertain environment. In the context of this complex problem, the BA algorithm adopts the method of distributing P from an appropriate pixel space. An adversarial disturbance is found closer to the original image according to a given adversarial criterion. At the same time, in order to overcome the uncertainty of the classification boundary, it needs to consume several queries to obtain a successful adversarial disturbance. The basic flow of the algorithm is as follows.
The first step is initialization. In the non-target scene, each pixel in the initial image is sampled from the uniform distribution U (0, 255), while the non-adversarial samples are rejected. In the target scene, we start with any sample that the black-box model can recognize as the target class.
The second step is to determine the disturbance. In order to ensure that the disturbance of each sampling is improved along the direction of the classification boundary as much as possible, as shown in Figure 2b, firstly the disturbance ηk with the sampling step size of δ is sampled on the hypersphere with the original image as the ball center, which is called the orthogonal disturbance. The disturbance of one step length of ε is sampled towards the direction of the original input image. Then the next adversarial sample point is obtained.
Boundary attack is easy to use. It just needs to adjust the two parameters of step size δ and ε. It has a good effect on target and non-target attack. This method can find adversarial samples that are similar to the effect of gradient-based white box attacks. It does not rely on the proxy model trained on the data, which is similar to the attack model. Besides, it has stronger robustness to the common defense methods, such as gradient masking, inherent randomness and robust training. The disadvantage is high query times and the efficiency is low. The black box model needs to be queried many times to generate successful adversarial samples. A distributed particle swarm optimization (PSO) algorithm is proposed to improve the initialization and optimization process. The improved details are introduced below.

2.2. Biased Boundary Attack Algorithm (BBA)

However, the BA samples the perturbation direction randomly (unbiased). The number of queries is too high and the convergence speed is slow in BA. Three biased perturbation directions are used in BBA algorithm: low-frequency perturbation, regional mask and gradient of surrogate model. The purpose is to reduce the size of the disturbed sampling space by using the three kinds of prior knowledge. It is proved feasible in practice by reducing the number of queries.

2.2.1. Low-Frequency Disturbance

The disturbance of the original attack is sampled from the Gaussian distribution. Guo proposed a method, which has better attack properties by adding low-frequency disturbance [24]. Then Brunner improved the BA algorithm by sampling disturbance from Perlin noise distribution.

2.2.2. Regional Mask

The boundary attack is the interpolation from the target image to the attacked image. In some areas, these images may already be similar, while in others they are very different. By creating the region mask of the image, we can reduce the search space by taking larger disturbances in the different regions. The specific measures are below.
Firstly, an image mask M is created according to the pixel difference between the adversary image and the original image.
M = |OadvOorigin|
where Oadv is the pixel value of the adversary image. Oorigin is the pixel value of the adversary image.
Then, the mask is recalculated according to the current position at each step. It is applied to the quadrature perturbation η k i of the previous sampling according to the element-wise.
η b i a s e d k = M η k ;   η b i a s e d k = η b i a s e d k η b i a s e d k
The distortion with large difference pixels is enlarged. A similar one is suppressed. Meanwhile, the disturbance vector is still unchanged.

2.2.3. Gradient of Surrogate Model

The transfer-based attack algorithm has achieved good performance in black-box attack applications. However, it will fail when the agent model is not consistent with the decision boundary of the defender closely. The transfer gradient information still works even in this case. The attack can be implemented by combining the original disturbance with limited gradient.
Firstly, the gradient of the surrogate model is calculated. Since the current position is already adversarial, a short distance is moved to the original image to ensure that the gradient is calculated from the non-adversarial region. Then the gradient is orthogonally projected to the original image direction. This projection is on the same hyperplane as the candidate points of the orthogonal step. Then, the candidate perturbations are biased to the projection gradient. After normalizing all vectors, the results are obtained below.
η b i a s e d k = ( 1 w ) η k + w η P G k
where w regulates the bias strength. It is tuned-up on the basis of the property of the agent. The alienability is large, the w value is increased. With above biased disturbances, sampling space has effective reduction. The query numbers are also reduced.

3. The Proposed Distributed Framework

3.1. Improvement of Initial Point Generation Algorithm

In BA methods, the generation of adversarial samples is equivalent to a search problem. One of the directions that can be improved in this problem is to define a heuristic information with a guiding function. The BBA is to introduce heuristic knowledge for the BA algorithm to speed up the search speed and efficiency. In addition, another important improvement direction is reducing the size of search space. In both BA and BBA algorithms, the approach taken begins with a single initial point. They search in a huge boundary space in a random or heuristic manner. This leads to the difficulty of too large search space, and the search gets stuck at local optimum easily. Therefore, this paper proposes the idea of combining particle swarm optimization (PSO) in the swarm intelligence algorithm to decompose the single-point search optimization into multiple points to search in their respective subspaces. Particle swarm optimization (PSO) algorithm is a random search algorithm based on group cooperation, which is developed by simulating the foraging behavior of birds. The purpose is to improve the possibility of searching for a practicable solution here. It can improve the success rate of adversarial attacks while speeding up the convergence rate. It can overcome shortcomings caused by the local optimal solution.
In this part, the initial point selection is discussed regarding the influence on the BA algorithm. Experiments have proved that the performance of the BBA algorithm has different sensitivity to the initial point selection [23,25]. Based on this, we further proposed a swarm intelligence search algorithm combined with multi-point starting.
In Figure 3, the selection of the initial point for the BBA algorithm is to find the adversarial image corresponding to the minimum L2 norm. It is calculated between the adversarial image 1 and the red star (original image). It is used to perform a binary search for seeking out the point X. It is the start of the BA algorithm. However, X is not the best start. In Figure 3, it can be seen that the boundary Y of the image 3 has a larger norm. It is nearer to the red star. That the Y converges to the best solution faster than X is very likely. It can be believed that the initial point calculation method of BBA may not find the best starting point.
The improvement of the initial point selection algorithm is proposed. The steps are as below: At first, the boundary points are found with the origin of multiple samples, respectively. One or several (when the PSO is used) points are chosen as the start point. Theoretically, it is difficult to find the optimal method. One method is to seek out the L2 norm among boundary points with the original. The closest distance is selected. The pseudo code is shown as Algorithm 1.
Algorithm 1. Improved algorithm of initial point selection
Input: Black-box classifier f, Adversarial image list X _ a d v , Step size d, Direction μ , Number of queries t, Threshold ε
Output: Initial adversarial image list X _ s t a r t
1:
d l = d , d r = d ;
2:
for i = 1 to len( X _ a d v ):
3:
while f ( X _ a d v [ i ] + d r μ ) = f ( X _ a d v [ i ] ) do
4:
d r = ( 1 + t ) d r ;
5:
while f ( X _ a d v [ i ] + d l μ ) f ( X _ a d v [ i ] ) do
6:
d l = ( 1 t ) d l ;
7:
while d r d l > ε do
8:
d m = ( d r + d l ) / 2
9:
if f ( X _ a d v [ i ] + d m μ ) = f ( X _ a d v [ i ] ) then
10:
   d l = d m
11:
else
12:
   d r = d m
13:
X _ s t a r t [ i ] = X _ a d v [ i ] + d r m
14:
end
15:
Output initial adversarial image list.
16:
return X _ s t a r t

3.2. BBA Algorithm with PSO Improvement

In this part, the particle swarm optimization algorithm is combined with the BBA algorithm. The advantages of multi-point starting are used to seek out the best solution. The efficiency is improved for searching the best adversarial samples. Query times are reduced with ensuring the success rate. The BBA-PSO is as below.
(1)
The X _ s t a r t is the start particle. The c a n d i d a t e _ l i s t is the initial velocity;
(2)
The particle i fitness function is below.
f i t n e s s i k = l 2 X i k X o r i g i n
where k is iteration numbers, Xorigin is the original image.
The update formula of velocity is:
V i k + 1 = w V i k + c 1 r 1 ( P i k X i k ) + c 2 r 2 ( P g k X i k )
where Vik+1 is velocity of next iteration and ω is a parameter of inertial velocity. Pik is the current optimal position. Xik is a position. Pgk is the current best position. c1 and c2 are acceleration coefficients, which are used to control the influence on the flight direction of particles. r1 and r2 are independent random numbers between [0, 1]. The characteristics of PSO are reflected which can keep the previous direction of velocity. Simultaneously, it also moves towards the best orientation. For the BBA problem, PSO can be used to update the disturbance, which retains the current optimal disturbance direction and moves towards the global optimal direction at the same time. The updated formula of position is as below:
X i k + 1 = X i k + V i k + 1
In adversarial environments, the perturbation space is searched and optimized by the continuous update of the particle swarm. When the updating position is outside the scope of sample space, the perturbation created by Equation (3) is the velocity updating to guarantee that the particle position stays in it. The procedure is shown in Figure 4 of PSO to seek out the best sample. The blue star is searched as the original particle. The green arrowhead is the disturbance of PSO updating by the iteration in Equation (6). The process does not waste any number of queries. Instead, it updates autonomously on the basis of the current orientation of each particle, and the best one. The query is needed for each update to judge only once that a particle has exceeded the boundary or not. If it is judged to exceed the space, a generated velocity is conducted as the correction by Equation (3). As the red arrowhead shows, the best sample is finally found through search, that is, the position of the yellow star. The adversarial sample search problem essentially has multiple solutions. Sometimes, it only needs to seek out one solution quickly. It can then be considered successful. The PSO algorithm based on multiple initial points can seek out the optimization among subspaces parallelly, which accelerates the speed of solution. One of the cost tradeoffs involved is that, per particle, it has to use up certain query numbers. The determination of particle numbers and its priority requires lots of experimental research. In the experiment of this article, the initial particle numbers are taken as five. Firstly, the priority is determined synchronously. Then it is updated asynchronously. The PSO_BBA is shown as Algorithm 2.

4. Experiments and Analysis

This paper evaluates the effectiveness of the PSO_BBA on dataset ImageNet comparatively [26]. It consists of 1000 classes with 299 × 299 size. Compared with the baseline algorithm, this paper adopts the same experiments and hyperparameter settings as those published in previous studies [20]. The selected attack object model is a pre-trained InceptionV3 network. In order to ensure the effectiveness of the attack, the algorithms automatically skipped over the images that could not be correctly recognized by the model itself.
In the experiments, 100 images are selected randomly from the verification dataset to attack. The average L2 distance is calculated for evaluating the effectiveness. The comparative methods attack on the same images to guarantee the effectiveness at the same time because the non targeted attack is relatively easy. This paper mainly focuses on the targeted attack task. The experimental results are shown in Table 2. It can be seen that our method needs less queries than BBA and BA for successful attacks.
Algorithm 2. BBA algorithm based on PSO (PSO_BBA)
Input: Initial adversarial image list X _ s t a r t , List of sampling disturbances c a n d i d a t e _ l i s t , The original image X o r i g i n , Number of particles n u m _ p a r t i c l e s
Output: Optimal adversarial image X _ a d v _ b e s t
1:
for i = 1 to len( X _ s t a r t ):
2:
Initial position Xi, velocity Vi:
3:
X i = X _ s t a r t [ i ] ; V i = c a n d i d a t e _ l i s t [ i ] ;
4:
Calculate the fitness function value and set the current local optimal position:
5:
f i t n e s s i k = l 2 X i k X o r i g i n ;
6:
p B e s t i = X i ;
7:
Calculate the global optimal location:
8:
g B e s t = m i n { p B e s t i } ;
9:
while query<=max_ queries = 15,000:
10:
for i = 1 to n:
11:
V i k + 1 = w V i k + c 1 r 1 ( P i k X i k ) + c 2 r 2 ( P g k X i k )
12:
X i k + 1 = X i k + V i k + 1
13:
if X i k + 1 == is_ adversarial
14:
   f i t n e s s i k = l 2 x i k x o r i g i n ;
15:
if f i t n e s s i k p B e s t i
16:
   p B e s t i = f i t n e s s i k ;
17:
else X i i + 1 = X i + h b i a s e d k
18:
   V i i + 1 = h b i a s e d k ;
19:
k = k + 1;
20:
g B e s t = m i n { p B e s t i } ;
21:
end
22:
Output optimal adversarial image:
23:
return X _ a d v _ b e s t = g B e s t
For details, the proposed method with BBA in the iteration step is compared. Figure 5 gives the variation of the average distortion relative to the original image with the number of queries. The ordinate decreases faster, the algorithm converges quicker. As shown in the figure, the improved algorithm needs to spend more queries to select the initial point for each particle in the early stage. As long as the initialization process is completed, the algorithm convergence is faster. The result is pretty well comparative to benchmarks. A comparative case of the specific display of the adversarial attack process is shown in Figure 6. The proposed PSO-BBA and BBA algorithms are, respectively, applied on the same original image to better compare and show the improved effect of the algorithm in the paper. The experimental results show that the proposed PSO-BBA algorithm can generate adversarial samples more effectively.
In addition to the L2 distance, this paper also uses the mean structural similarity (MSSIM) [27] to measure the similarity between the attacked image and the original one. MSSIM is a function of evaluating image fidelity based on human visual characteristics.
M S S I M = 1 L i = 1 L [ l ( f , f ^ ) ] a g [ c ( f , f ^ ) ] b g [ s ( f , f ^ ) ] g
where L is the number of image blocks, l(•), c(•) and s(•) represent the luminance, contrast and structural contrast functions, respectively. α , β and γ are the weight parameters to regulate the relative importance of the three components. MSSIM mainly considers the structure information of the image itself adequately, with a value range of [0, 1]. The closer the value of MSSIM is to 1, the higher the fidelity between images used to compare. In the experiment, the values of the relevant parameters of the evaluation index are recommended by the original paper [24].
A specific black-box attack example comparison experiment is shown in Figure 7. It can be seen that the improved PSO-BBA algorithm can effectively reduce the number of queries while ensuring the success rate of the attack. Although the similarity between the initial interference image generated by each particle and the original is poor. By expanding search space, the improved PSO-BBA algorithm generates a higher degree of similarity to the original image based on all particles. It also improves the convergence speed of the algorithm in comparison to the original method. The local optimal situation is avoided effectively. The performance of the attack algorithm is also improved. Figure 8shows an example of a failed BBA attack. The improved PSO-BBA algorithm succeeds and the number of attack queries is significantly reduced. However, the comparison results of MSSIM show that the two methods have changed the structural information of the original image. So, some evaluation indexes can be used as the indicators of detecting attacks.

5. Conclusions and Future Work

Aiming at the black-box adversarial attack application, hackers can obtain given input labels only. The paper researched existing problems in the black-box attack method. The PSO-BBA is put forward, which avoids the current BBA method falling into the situation of a local optimal solution. The proposed PSO-BBA method improves the convergence speed of the BBA and ensures the performance of the black-box attack. Through the experimental comparison with the baseline method, it is proved that the proposed PSO-BBA effectively reduces query numbers. It produced adversarial samples at a lower self-cost and faster efficiency successfully. Our next research will focus on conducting detailed theoretical and experimental analyses on the selection of initial particles in the improved algorithm [28].

Author Contributions

Conceptualization, W.Z.; Methodology, F.X.; Project administration, W.W.; Software, J.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. U1734208, 61806212) and Natural Science Foundation of Hunan Province (No. 2021JJ40693, 2019JJ50724).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

The authors are very grateful to the reviewers for their valuable and constructive feedbacks on this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Singh, A.K.; Liu, X.; Wang, H.; Ko, H. Recent advances in multimedia security and information hiding. Trans. Emerg. Telecommun. Technol. 2020, 32, e4193. [Google Scholar]
  2. Xu, G.; Li, H.; Liu, S.; Yang, K.; Lin, X. VerifyNet: Secure and Verifiable Federated Learning. IEEE Trans. Inf. Forensics Secur. 2020, 15, 911–926. [Google Scholar] [CrossRef]
  3. Phong, L.T.; Aono, Y.; Hayashi, T.; Wang, L.; Moriai, S. Privacy-Preserving Deep Learning via Additively Homomorphic Encryption. IEEE Trans. Inf. Forensics Secur. 2018, 13, 1333–1345. [Google Scholar] [CrossRef]
  4. Lv, Z.; Qiao, L.; Song, H. Analysis of the Security of Internet of Multimedia Things. ACM Trans. Multimed. Comput. Commun. Appl. 2021, 16, 1–16. [Google Scholar] [CrossRef]
  5. Xu, J.; Wang, C.; Li, T.; Xiang, F. Improved Adversarial Attack against Black-box Machine Learning Models. In Proceedings of the 2020 Chinese Automation Congress (CAC 2020), Shanghai, China, 6–8 November 2020; pp. 5907–5912. [Google Scholar]
  6. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and Harnessing Adversarial Examples. arXiv 2014, arXiv:1412.6572, 1–11. [Google Scholar]
  7. Kurakin, A.; Goodfellow, I.J.; Bengio, S. Adversarial Examples in the Physical World. In Artificial Intelligence Safety and Security; CRC Press: Boca Raton, FL, USA, 2018; pp. 99–112. [Google Scholar]
  8. Moosavi-Dezfooli, S.M.; Fawzi, A.; Frossard, P. DeepFool: A simple and accurate method to fool deep neural networks. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 27–30 June 2016; pp. 2574–2582. [Google Scholar]
  9. Papernot, N.; McDaniel, P.; Jha, S.; Fredrikson, M.; Celik, Z.B.; Swami, A. The Limitations of Deep Learning in Adversarial Settings. In Proceedings of the 2016 IEEE European Symposium on Security and Privacy (EuroS&P), Saarbruecken, Germany, 21–24 March 2016; pp. 372–387. [Google Scholar]
  10. Cisse, M.; Adi, Y.; Neverova, N.; Keshet, J. Houdini: Fooling deep structured prediction models. arXiv 2017, arXiv:1707.05373. [Google Scholar]
  11. Carlini, N.; Wagner, D. Towards Evaluating the Robustness of Neural Networks. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–26 May 2017; pp. 39–57. [Google Scholar]
  12. Brendel, W.; Rauber, J.; Bethge, M. Decision-Based Adversarial Attacks: Reliable Attacks Against Black-Box Machine Learning Models. In Proceedings of the 6th International Conference on Learning Representations (ICLR), Vancouver, BC, Canada, 30 April–3 May 2018; pp. 1–29. [Google Scholar]
  13. Brunner, T.; Diehl, F.; Le, M.T.; Knoll, A. Guessing Smart: Biased Sampling for Efficient Black-Box Adversarial Attacks. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 4957–4965. [Google Scholar]
  14. Hayes, J.; Danezis, G. Learning Universal Adversarial Perturbations with Generative Models. In Proceedings of the 2018 IEEE Symposium on Security and Privacy Workshops, SPW 2018, San Francisco, CA, USA, 24–24 May 2018; pp. 43–49. [Google Scholar]
  15. Guo, C.; Gardner, J.; You, Y.; Wilson, A.G.; Weinberger, K. Simple Black-box Adversarial Attacks. In Proceedings of the 36th International Conference on Machine Learning (ICML), Long Beach, CA, USA, 9–15 June 2019; pp. 4410–4423. [Google Scholar]
  16. Cheng, M.; Le, T.; Chen, P.Y.; Yi, J.; Zhang, H.; Hsieh, C.J. Query-Efficient Hard-label Black-box Attack: An Optimization-based Approach, ICLR (Poster). arXiv 2019, arXiv:1807.04457. [Google Scholar]
  17. Ilyas, A.; Engstrom, L.; Athalye, A.; Lin, J. Black-box Adversarial Attacks with Limited Queries and Information. In Proceedings of the International Conference on Machine Learning (ICML), Stockholm, Sweden, 10–15 July 2018. [Google Scholar]
  18. Papernot, N.; McDaniel, P.; Wu, X.; Jha, S.; Swami, A. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks. In Proceedings of the 2016 IEEE Symposium on Security and Privacy, San Jose, CA, USA, 22–26 May 2016; pp. 582–597. [Google Scholar]
  19. Liu, Y.; Chen, X.; Liu, C.; Song, D. Delving into transferable adversarial examples and black-box attacks. arXiv 2016, arXiv:1611.02770. [Google Scholar]
  20. Cheng, S.; Dong, Y.; Pang, T.; Su, H.; Zhu, J. Improving black-box adversarial attacks with a transfer-based prior. arXiv 2019, arXiv:1906.06919. [Google Scholar]
  21. Narodytska, N.; Kasiviswanathan, S.P. Simple black-box adversarial perturbations for deep networks. arXiv 2016, arXiv:1612.06299. [Google Scholar]
  22. Li, H.; Xu, X.; Zhang, X.; Yang, S.; Li, B. QEBA: Query-Efficient Boundary-Based Blackbox Attack. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 1218–1227. [Google Scholar]
  23. Ren, Y.; Zhou, Q.; Wang, Z.; Wu, T.; Wu, G.; Choo, K.-K.R. Query-efficient label-only attacks against black-box machine learning models. Comput. Secur. 2019, 90, 101698. [Google Scholar] [CrossRef]
  24. Guo, C.; Frank, J.S.; Weinberger, K.Q. Low frequency adversarial perturbation. arXiv 2018, arXiv:1809.08758. [Google Scholar]
  25. Chen, P.Y.; Zhang, H.; Sharma, Y.; Yi, J.; Hsieh, C.J. Zoo: Zeroth order optimization based black-box attacks to deep neural networks without training substitute models. In Proceedings of the 10th ACM workshop on artificial intelligence and security, Dallas, TX, USA, 3 November 2017; pp. 15–26. [Google Scholar]
  26. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Li, F.F. Imagenet: A Large-Scale Hierarchical Image Database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  27. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Kwon, H.; Kim, Y.; Yoon, H.; Choi, D. Classification score approach for detecting adversarial example in deep neural network. Multimed. Tools Appl. 2021, 80, 10339–10360. [Google Scholar] [CrossRef]
Figure 1. An example of an image adversarial examples attack.
Figure 1. An example of an image adversarial examples attack.
Applsci 11 10479 g001
Figure 2. Schematic diagram of boundary attack algorithm. (a) Initial adversarial image optimization along boundary (b) Move along the orthogonal direction and close to the original image.
Figure 2. Schematic diagram of boundary attack algorithm. (a) Initial adversarial image optimization along boundary (b) Move along the orthogonal direction and close to the original image.
Applsci 11 10479 g002
Figure 3. General view about the initial point of selection.
Figure 3. General view about the initial point of selection.
Applsci 11 10479 g003
Figure 4. The process of particle swarm searching for optimal adversarial examples.
Figure 4. The process of particle swarm searching for optimal adversarial examples.
Applsci 11 10479 g004
Figure 5. Comparison of the average distortion (L2 distance) in each iteration between the proposed PSO-BBA and the original BBA algorithm.
Figure 5. Comparison of the average distortion (L2 distance) in each iteration between the proposed PSO-BBA and the original BBA algorithm.
Applsci 11 10479 g005
Figure 6. Comparison results of PSO-BBA (top) and BBA (bottom) generated adversarial examples (Query times and the distortion (L2 distance)).
Figure 6. Comparison results of PSO-BBA (top) and BBA (bottom) generated adversarial examples (Query times and the distortion (L2 distance)).
Applsci 11 10479 g006aApplsci 11 10479 g006b
Figure 7. Comparison Experiment 1 of specific adversarial process and results between PSO-BBA and BBA.
Figure 7. Comparison Experiment 1 of specific adversarial process and results between PSO-BBA and BBA.
Applsci 11 10479 g007
Figure 8. Comparison Experiment 2 of specific adversarial process and results between PSO-BBA and BBA.
Figure 8. Comparison Experiment 2 of specific adversarial process and results between PSO-BBA and BBA.
Applsci 11 10479 g008
Table 1. Comparative analysis of three types black-box attack methods.
Table 1. Comparative analysis of three types black-box attack methods.
Black-Box Attack MethodsAttack TargetAttack ConditionsAlgorithm Characteristics
Score-based attackWith target/no targetThe class probability or distribution of the model is required.Use predicted values to estimate the gradient.
Transfer-based attackWith target/no targetDoes not rely on model information, but requires information on model training data.Train a fully observable agent model from which adversarial disturbances can be synthesized.
Decision-based attacksWith target/no targetOnly need the output labelsThe number of queries required for the attack may be higher.
Table 2. Comparison of the average query numbers.
Table 2. Comparison of the average query numbers.
Attack MethodsAverage Number of Queries on Successful Attack (Per Image)
BA>10,000
BBA5432
Proposed PSO_BBA4596
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xiang, F.; Xu, J.; Zhang, W.; Wang, W. A Distributed Biased Boundary Attack Method in Black-Box Attack. Appl. Sci. 2021, 11, 10479. https://0-doi-org.brum.beds.ac.uk/10.3390/app112110479

AMA Style

Xiang F, Xu J, Zhang W, Wang W. A Distributed Biased Boundary Attack Method in Black-Box Attack. Applied Sciences. 2021; 11(21):10479. https://0-doi-org.brum.beds.ac.uk/10.3390/app112110479

Chicago/Turabian Style

Xiang, Fengtao, Jiahui Xu, Wanpeng Zhang, and Weidong Wang. 2021. "A Distributed Biased Boundary Attack Method in Black-Box Attack" Applied Sciences 11, no. 21: 10479. https://0-doi-org.brum.beds.ac.uk/10.3390/app112110479

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop