Next Article in Journal
Geochemical Evolution of Mg-Bentonite Affected by the Contact of Carbon Steel and a Hydrothermal Gradient
Previous Article in Journal
Fast State-of-Charge-Balancing Strategy for Distributed Energy Storage Units Interfacing with DC–DC Boost Converters
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Attacking Robot Vision Models Efficiently Based on Improved Fast Gradient Sign Method

1
Hubei Province Key Laboratory of Intelligent Robot, Wuhan Institute of Technology, Wuhan 430079, China
2
School of Computer, Hubei University of Arts and Science, Xiangyang 441054, China
3
School of Science, Wuhan University of Technology, Wuhan 430062, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(3), 1257; https://doi.org/10.3390/app14031257
Submission received: 17 December 2023 / Revised: 31 January 2024 / Accepted: 1 February 2024 / Published: 2 February 2024

Abstract

:
The robot vision model is the basis for the robot to perceive and understand the environment and make correct decisions. However, the security and stability of robot vision models are seriously threatened by adversarial examples. In this study, we propose an adversarial attack algorithm, RMS-FGSM, for robot vision models based on root-mean-square propagation (RMSProp). RMS-FGSM uses an exponentially weighted moving average (EWMA) to reduce the weight of the historical cumulative squared gradient. Additionally, it can suppress the gradient growth based on an adaptive learning rate. By integrating with the RMSProp, RMS-FGSM is more likely to generate optimal adversarial examples, and a high attack success rate can be achieved. Experiments on two datasets (MNIST and CIFAR-100) and several models (LeNet, Alexnet, and Resnet-101) show that the attack success rate of RMS-FGSM is higher than the state-of-the-art methods. Above all, our generated adversarial examples have a smaller perturbation than those generated by existing methods under the same attack success rate.

1. Introduction

Nowadays, the robot vision model is widely used in the field of robot vision, such as image classification [1], target detection [2], speech recognition [3], autonomous vehicles [4], etc. Although the robot vision model has achieved success in the field of robot vision, it may suffer from security and robustness problems when feeding with adversarial examples. Adversarial examples are inputs to robot vision models that have been intentionally designed by adding small, imperceptible perturbations [5]. They will cause robot vision models to achieve erroneous results. Influenced by adversarial examples, autonomous vehicles may experience disruptions in their driving state, which can lead to traffic safety issues. Militarily, when using UAVs to detect unknown areas, the adversarial examples will lead to UAV target-identification errors. Apart from these, adversarial examples have a serious impact on the security of all kinds of intelligent system software based on deep neural networks. Memristive chaotic circuits based on CNNs may also be affected [6]. Therefore, the generation method of adversarial examples (adversarial attack method) has gained much attention from researchers. Based on adversarial examples, we can design adversarial defense methods [7,8,9,10] and develop secure and robust intelligent systems.
Currently, adversarial attack methods can be divided into three categories: (1) the gradient-based methods, (2) the optimization-based methods, and (3) the generative adversarial network (GAN)-based methods. The most typical gradient-based method is called the fast gradient sign method (FGSM), which was proposed by Goodfellow et al. [11] in 2015. It computes adversarial examples by maximizing the loss function of the attacked neural network model with the gradient ascent optimizer. Since it can generate adversarial examples efficiently, much research effort has been paid, and many variants have been proposed, including the basic iterative method (BIM, also known as I-FGSM) [12], momentum iterative fast gradient sign method (MI-FGSM) [13], etc. The optimization-based methods [14] directly optimize the distance between the real samples and adversarial examples subject to the misclassification of adversarial examples. The problem with this method is that the added perturbation cannot be constrained properly and may result in invalid adversarial examples [13]. The GAN-based methods [15,16] generate adversarial examples based on generative adversarial networks. These methods train (i) a feed-forward generator network that generates perturbations to create diverse adversarial examples and (ii) a discriminator network to ensure that the generated examples are realistic; once the generator network is trained, it can generate perturbations efficiently for any instance, so as to potentially accelerate adversarial training as defenses. Experiments show that the adversarial examples constructed via the generative adversarial networks have a high attack success rate in the case of defense. However, these kinds of methods are resource-intensive and time-consuming.
In this study, we focus on the gradient-based methods. Compared with other methods, the gradient-based methods are resource-saving and time-efficient. However, since the principle behind the gradient-based methods is the gradient ascent algorithm, just like the gradient descent algorithm, a large learning rate will cause the gradient of the objective function to grow too fast and miss the global optimal solution. A low value will lead to extra iterations and low efficiency. The improper learning rate will cause the problems of large perturbation and low attack success rate. To address this problem, Geoffrey Hinton [17] introduced an adaptive learning rate calculation method, root-mean-square propagation (RMSProp). This method performs a learning rate correction strategy and the exponentially weighted moving average (EWMA) during the gradient update, which can mitigate the problem of rapid gradient growth and keep it efficient. Inspired by the RMSProp algorithm, we efficiently propose an improved RMSProp-based fast gradient sign method, RMS-FGSM, to attack robot vision models. RMS-FGSM generates adversarial examples in an iterative manner. At each iteration, an adaptive learning rate is used to update adversarial examples. Our learning rate is computed based on the EWMA algorithm, which smoothes the change of gradients by considering historical data. Since our method can mitigate the problem of the gradient’s steep growth, adversarial examples can be updated along a stable optimization direction, and the optimal value is highly likely to be achieved; that is, adversarial examples with imperceptible perturbations and a high attack success rate.
The experiments in this study are carried out with several common classification models in robot vision and the environment of Python 3.9, torch 1.13.1, and TorchVision 0.14.1. Two commonly used datasets, MNIST [18] and CIFAR-100 [19], were used. To validate the method, we attacked LeNet [20], Alexnet [1], and Resnet-101 [21]. Experimental results persuasively verify that RMS-FGSM has better performance than other advanced attack methods. Above all, the perturbation of the adversarial examples generated using RMS-FGSM is smaller than other methods under the same attack success rate.
The rest of this study is arranged as follows. Section 2 introduces related work, including the FGSM, BIM, MI-FGSM, and NI-FGSM algorithms. The attack algorithm RMS-FGSM is described in Section 3. In Section 4, we present and analyze the experimental results. And the last section concludes the study.

2. Related Work

In recent years, many gradient-based attack algorithms have been proposed. The process of adversarial attacks can generally be defined as follows: let M be a pre-trained neural network, which can output predicted label y for an input sample x, that is, M(x) = y. J(x, W, y) denotes the loss function of M, where W is the set of weights of model M. Since model M is pre-trained (W is available), we can rewrite J(x, W, y) as J(x, y). The goal of adversarial attacks is to seek an example x * = x + η such that M x * M x . The noise η added to x* should be imperceptible. To this end, the Lp norm of η is required to be less than an allowed value ϵ as x * x p ϵ , where p could be 0, 1, 2, and .

2.1. Fast Gradient Sign Method

The fast gradient sign method (FGSM) is a one-step gradient-based approach that finds an adversarial example x* by maximizing the loss function J(x, y), which can be formalized as:
arg max x ( J ( x , y ) )
In order to solve the above problem, the FGSM uses the gradient ascent algorithm, which computes the adversarial noise η based on Equation (2):
η = ε sign ( x J ( x , y ) )
where ∇x denotes the gradient of loss function J w.r.t. x, s i g n · is the sign function, and ε is the perturbation factor. The FGSM employs the L norm to validate generated adversarial examples; that is, η ϵ should be satisfied.
The FGSM algorithm modifies the image only once based on the gradient calculation, which is time-efficient. However, due to the manually set improper learning rate, in case of the same attack success rate, the perturbation added to the adversarial examples generated using this method will be larger than those of other methods, which will violate the imperceptible requirement.

2.2. Basic Iterative Method

In 2016, Kurakin proposed a FGSM-based basic iterative method (BIM, also known as I-FGSM) to rapidly generate adversarial examples. Different from the one-step attack of the FGSM, the BIM iteratively applies the fast gradient multiple times with a small step size α as shown in Equation (3):
x t + 1 * = C l i p { x t * + α s i g n ( x J ( θ , x t * , y ) ) }
where x t * denotes the adversarial example obtained at the t-th iteration. In order to make the generated adversarial examples satisfy the L bound, the BIM uses the C l i p · function to clip x t * into the ϵ vicinity of x.
Intuitively, the BIM is as easy to understand as the FGSM, concise, and efficient, and the attack effect is obviously better than the FGSM. The author asserted that this approach ensures an optimal pace for gradient ascent, and it is, at worst, equivalent to the FGSM. However, further research has indicated that the samples generated using the BIM lack transferability, resulting in a weak black box attack [22,23,24] effect.

2.3. Momentum Iterative Fast Gradient Sign Method

Dong et al. [25] incorporated the momentum term into the BIM and proposed the MI-FGSM algorithm. As shown in Equation (4), it computes the cumulative gradient gt+1 by a weighted sum of gradients obtained in the past t-th iterations, where µ is a decay factor. Based on the cumulative gradient, it computes the adversarial example x t + 1 * by adding a small perturbation to x t * , as shown in Equation (5), where α is a step size. At each iteration, the current gradient x J ( x t * , y ) is normalized by the L1 distance of itself.
g t + 1 = μ g t + x J ( x t * , y ) | | x J ( x t * , y ) | | 1
x t + 1 * = x t * + α s i g n ( g t + 1 )
According to the principle of momentum in physics, the iterative calculation process of adversarial examples can be accelerated. In addition, with the help of momentum, the optimization process can go beyond local maxima and is more likely to reach a global maximum. Since the update direction becomes more stable, the MI-FGSM is more efficient and has better transferability [26,27] across different neural networks. However, this method may increase the gradient too rapidly and introduce a large perturbation to adversarial examples.

2.4. Nesterov Iterative Fast Gradient Sign Method

Lin et al. [28] leveraged the Nesterov accelerated gradient (NAG) [29] and presented the NI-FGSM algorithm so that it can improve transferability. The NI-FGSM makes a jump in the direction of previously accumulated gradients before computing the gradients in each iteration. It substitutes x t * in Equation (4) with x t a d v + α · μ · g t to leverage the looking-ahead property of the NAG and build a robust adversarial attack. Such looking ahead property of the NAG can help to escape from poor local maxima easier and faster, resulting in the improvement in transferability.

3. Our Technique

3.1. Method Framework

The process of the RMS-FGSM adversarial attack is shown in Figure 1; once the image dataset is input, preprocess the data and calculate the current gradient gt of the loss function. Next, compute the exponential average rt of the squared gradients with EWMA to prepare for the next step, and the parameter ρ is used as a weight parameter to control the acquisition of a historical cumulative squared gradient and a new squared gradient. Then, the learning rate τ is calculated, where the exponential average rt from the previous step controls the size of the learning rate. When gt increases, the learning rate decreases; conversely, when gt decreases, the learning rate increases to achieve the adaptive learning rate. Then, calculate the current perturbation value using the sign function and add it to the original example to generate a new adversarial example and proceed to the next iteration. After the number of iterations reaches a fixed value T, we can obtain the optimal adversarial example, and the perturbation of the adversarial example is imperceptible for humans. Finally, the adversarial example is input into the attacked neural network model, which will lead to the wrong classification results.

3.2. Fast Gradient Sign Method Based on Adaptive Learning Rate

The fast gradient sign method (FGSM) is able to generate adversarial examples efficiently based on the gradient ascend approach. However, as shown in Equation (2), a fixed learning rate ε is used to compute the perturbation, which may lead to slow convergence or even failure to converge. Additionally, a fixed learning rate may fail to adapt to different samples, which may result in insufficient updates for some samples to reach optimal solutions. Above all, the missing of the optimal solution may introduce a large perturbation to adversarial examples and decrease the attack success rate. Although some improved variant methods such as I_FGSM and MI_FGSM have been proposed, they also employ a fixed step size α as shown in Equations (3) and (5). In order to further improve the performance of the gradient-based adversarial attack methods, we propose the RMS-FGSM, which utilizes the adaptive learning rate calculation method RMSProp.
Let f x ; W , b = y be a neural network model, such as the Alexnet for image classification tasks. W and b are the weight matrix and offset vector, respectively. When feeding with an image sample x, the model will output prediction result y. In the training process of neural networks, we always construct a loss function of the network L ( f ( x ; W , b ) , y ) and reduce the training of neural networks to an optimization problem as follows:
arg min W , b ( L ( f ( x ; W , b ) , y ) )
In order to resolve the optimization problem, a frequently used approach is the gradient descent, which calculates the optimal W and b along the direction of the gradients W L ( f ( x ; W , b ) , y ) and b L ( f ( x ; W , b ) , y ) , respectively. During this process, a labeled training dataset D = ( x , y )   |   x   i s   a n   i n p u t   i m a g e   s a m p l e ,   y   i s   t h e   c o r r e s p o n d i n g   c l a s s i f i c a t i o n   l a b e l will be used.
Different from the above training process, the white-box adversarial attack methods assume that the neural network model f x ; W , b is pre-trained. Therefore, the weight matrix W and offset vector b are both available. The purpose of the adversarial attack is to generate difference-inducing input samples. Formally, given an image sample x with a label y, an adversarial attack aims to generate an adversarial example x* by adding an imperceptible perturbation to x, such that f ( x ; W , b ) y . This problem can be addressed by maximizing the loss between label y and the prediction f ( x ; W , b ) . Formally, we have the following optimization problem (we neglect W and b because they are available):
arg max x ( L ( f ( x ) , y ) )
In this study, we address the problem with the gradient ascent algorithm. Different from the FGSM and its variants, the RMS-FGSM calculates the adversarial examples based on an adaptive learning rate given by RMSProp. RMSProp is a variant of the gradient descent algorithm used to adjust model parameters. The most distinctive feature of RMSProp is that it can adaptively adjust the learning rate of different parameters by combining the first-order and second-order moments. By utilizing the smooth nature of the square root function, it can avoid the problems of drastic changes and unstable behavior in the parameter space. The RMS-FGSM computes adversarial examples in an iterative manner as follows:
g t = x L ( f ( x t * ) , y )
r t = ρ r t 1 + ( 1 ρ ) g t 2
τ = ϕ r t + σ
x t + 1 * = c l i p { x t * + τ s i g n ( g t ) }
where gt is the gradient of loss function L ( f ( x t * ) , y ) w.r.t x at the t-th iteration. x t * denotes the adversarial example achieved at the t-th iteration. Specifically, x 0 * = x , f(x) = y. In order to adjust the learning rate of adversarial examples, the RMS-FGSM computes an exponential average rt of the squared gradients based on the EWMA at each iteration according to Equation (9). This is essentially an exponentially decreasing weighted moving average, where the weight of each squared gradient decreases exponentially over time. The square of gt aims to eliminate the effect of the gradient sign. ρ is a constant that is typically set to 0.9. It determines the speed of weight decay of historical gradients, with smaller values resulting in faster decay. The aim of the decay factor ρ is to decrease the influence of historical gradients on the computation of the exponential average. Based on the exponential average, the RMS-FGSM calculates the learning rate τ according to Equation (10), where φ represents the initial learning rate, and σ is usually set to 10−6 to prevent the denominator from being zero. As we can see, if the gradient is large, the learning rate will be suppressed; if the gradient is small, the learning rate will be increased. Different from the usual use case of RMSProp, the essence of our adversarial attack method is the gradient ascent rather than the gradient descent algorithm. Consequently, the RMS-FGSM updates the adversarial examples by adding (rather than subtracting) the computed perturbation τ s i g n ( g t ) according to Equation (11). Additionally, in order to achieve imperceptible adversarial examples, we use the c l i p · function to impose the bounding constraint x * x 2 ϵ . The algorithm stops when the specified number of iterations is reached; then, we obtain the adversarial examples.
In conclusion, our approach has the following advantages: (1) Different adversarial examples have different learning rates, and the learning rate varies according to gradients during the iterative calculation process. This is completely different from the previous FGSM and variant algorithms with a fixed learning rate. (2) The problems of drastic changes and unstable behavior incurred by the fixed learning rate can be mitigated. (3) Based on the adaptive learning rate, it is more likely to achieve optimal adversarial examples while keeping a high convergence speed. (4) Our adversarial examples may have a small perturbation. (5) Since optimal adversarial examples with a small perturbation can be obtained, our approach may have a high attack success rate. (6) The high convergence speed makes our method time-efficient.

3.3. RMS-FGSM Algorithm

The process of generating adversarial samples using the RMS-FGSM algorithm is as follows: 1. Given an original image x; 2. Input x to classifier f(x) and calculate its gradient gt of loss function w.r.t x.; 3. According to gt, compute an exponential average rt of the squared gradients based on the EWMA; 4. Calculate the adaptive learning rate τ according to rt; 5. Calculate the perturbation pixel by pixel, add it to x, and use the c l i p · function to prevent pixels from going out of range; 6. Repeat steps 2 to 5 until the specified number of iterations is reached.
The outline of our the RMS-FGSM algorithm framework is shown in Algorithm 1.
Algorithm 1: RMS-FGSM.
Input: A classifier f with loss function L; original image x; ground-truth label y; decay rate ρ; number of iterations T; initial learning rate φ ;
Output: Adversarial example x*;
1: Initialize the variable r0 = 0; x 0 * = x ;
2: For t = 0 to T-1 do:
3: Input x t * to classifier f;
4: Calculate the current gradient: g t = x L ( f ( x t * ) , y ) ;
5: Cumulative squared gradient: r t = ρ r t 1 + ( 1 ρ ) g t 2 ;
6: Compute the adaptive learning rate: τ = ϕ r t + σ ;
7: Add perturbation to the sample: x t + 1 * = c l i p { x t * + τ s i g n ( g t ) } ;
8: End for
9: return x * = x T * .

4. Experiments

4.1. Experimental Setup

Datasets. We evaluate our method based on two commonly used datasets, MNIST and CIFAR-100. The MNIST dataset is a subset of the NIST (National Institute of Standards and Technology) dataset. It has a total of 70,000 images, including 60,000 in the training set and 10,000 in the test set. Each image is a 28 × 28 pixel image of 0–9 gray scale handwritten digits with white text on a black background. The CIFAR-100 dataset is a small dataset for recognizing pervasive objects. Consisting of 60,000 32 × 32 pixel images, it has 100 classes, and each class contains 600 images with 500 training images and 100 test images. The 100 classes in CIFAR-100 are divided into 20 superclasses. Each image has a “fine” label (the class it belongs to) and a “coarse” label (the superclass it belongs to).
Implementation details. A set of different epsilon values (EPS) represents the initial learning rate in the BIM, MI-FGSM, NI-FGSM, and RMS-FGSM, and it is actually the perturbation factor in the FGSM method. Our experimental results are based on the L2 norm and non-target attack. The decay factor ρ in our RMS-FGSM is set to 0.9, and the number of iterations T in the BIM, MI-FGSM, NI-FGSM, and RMS-FGSM is set to 10. Our experiments run on a machine with an Intel Core i7-8700K CPU, 32 GB memory, and an Nvidia GeForce GTX 1080Ti graphics card (11 GB memory). We use Python 3.9, Torch 1.13.1, TorchVision 0.14.1, and related libraries.
Evaluation metrics. For the performance evaluation and comparison, we use three different metrics, including the perturbation value (l2) calculated for different initial learning rates, average attack success rate (Avg asr), and average perturbation value (Avg per).

4.2. Experimental Method

Several common classification models in robot vision are used for our experiments. In order to verify the advantages of the proposed method, we designed two experiments to verify that the proposed method is superior to other methods in terms of effectiveness and imperceptibility. (1) In experiment 1, we set MNIST as the dataset and the LeNet model as the experimental model. Firstly, a set of different epsilon values (0, 0.05, 0.1, 0.15, 0.2, 0.25, 0.3, 0.35, and 0.4) were used as the initial learning rate to generate adversarial examples on the MNIST with the FGSM, BIM, MI-FGSM, NI-FGSM, and RMS-FGSM algorithm, respectively. Then, we analyzed the experimental results to compare the effectiveness and imperceptibility of the adversarial examples generated by these different methods with different initial learning rates. (2) In experiment 2, we set CIFAR-100 as the dataset and Alexnet and Resnet-101 as the experimental models. Generate adversarial examples on both models using the same methods as in experiment 1. And analyze the experimental results to compare the effectiveness and imperceptibility of the adversarial examples generated using these different methods on different models.

4.3. Adversarial Attack Based on MNIST

We first perform a white-box attack on the LeNet network using different methods, including the FGSM, BIM, MI-FGSM, NI-FGSM, and RMS-FGSM algorithms, to demonstrate the effectiveness of our method. Table 1 presents the results of experiment 1 on the MNIST dataset. In addition, we also calculate and plot the prediction accuracy of LeNet with the FGSM and RMS-FGSM methods under different epsilon values in Figure 2, which represents the proportion of test samples that predicted correctly over all test samples.
As we can see from Table 1, as the EPS increases, the l2 of the adversarial examples with the FGSM also increases. The EPS is actually the perturbation factor of Equation (2) in the FGSM. At the same time, the l2 values in the BIM, MI-FGSM, and NI-FGSM methods are slowly increasing, while the l2 in the RMS-FGSM has been maintained at around 0.8. From the images, the perturbation of the adversarial examples in the FGSM can be easily recognized by humans, but the perturbations in the BIM, MI-FGSM, NI-FGSM, and RMS-FGSM methods are small. Furthermore, we also calculate the average attack success rate (Avg asr) and average perturbation value (Avg per) in these methods. The results show that the FGSM, BIM, MI-FGSM, and NI-FGSM methods obtain average attack success rates of 72.6%, 84.5%, 86.4%, and 89.6%, respectively, and the average attack success rate of the RMS-FGSM reaches 91.6%, which is the highest of these methods. The average perturbation value of the RMS-FGSM is 0.705, which is also the lowest among the methods.
In Figure 2, we can confirm that the accuracy of the model gradually decreases as the EPS increases in the FGSM method. However, the accuracy of the model in the RMS-FGSM is not affected by the initial learning rate, and finally maintains at about 0.01, which is much lower than the accuracy in the FGSM. This means that the RMS-FGSM is able to find the global optimal solution using the exponentially weighted moving average and adaptive learning rate, and it is not affected by the initial learning rate.

4.4. Adversarial Attack Based on CIFAR-100

In experiment 2, we also perform a white-box attack with the methods in experiment 1, but we use the CIFAR-100 dataset on Alexnet and Resnet-101, respectively. Table 2 and Table 3 show the results of experiment 2 on the CIFAR-100 dataset and two different models.
From the experimental results in Table 2, when the EPS is 0.05, the l2 of the FGSM is 1319.89, which is smaller than those of other methods. However, the l2 of the FGSM far exceeds those of other methods as the EPS changes, resulting in a large perturbation on the adversarial example images. The l2 in the BIM, MI-FGSM, and NI-FGSM methods slightly changes as EPS increases but stably remains around 2400 in the RMS-FGSM. In Table 2, we also observe that the RMS-FGSM has the highest average attack success rate of 94.6% on Alexnet, while the FGSM, BIM, MI-FGSM, and NI-FGSM methods obtain an average attack success rate of 72.4%, 83.5%, 90.2%, and 92.8%, respectively. The average perturbation value of the RMS-FGSM is 2140.7 on Alexnet, which has a smaller perturbation on adversarial images.
Then, Table 3 shows the results of experiment 2 on Resnet-101. It can be seen that the l2 of the FGSM increases rapidly as the EPS changes; there is also a slight increase for the BIM, MI-FGSM, and NI-FGSM methods, while the l2 of the RMS-FGSM is always maintained at a small value. In addition, the RMS-FGSM reaches a slightly higher average attack success rate of 99.9% on Resnet-101 compared with 99.1% and 99.6% in the MI-FGSM and NI-FGSM. However, our attack method obtains a significantly lower average perturbation value of 2148.39 compared with other methods.
Such experimental results suggest that the RMS-FGSM finds the global optimum point in the process of generating adversarial examples and produces an imperceptible perturbation so that there is only an extremely small difference between the adversarial example and the original example. In addition, with the change in the initial learning rate, the perturbation value generated using the previous methods also fluctuates, while the perturbation value generated using our method remains at a small value. The experimental results on two different models suggest that our RMS-FGSM method is more effective at misleading a white-box network.

5. Conclusions

In this study, we focus on the gradient-based method and improve the fast gradient sign method. The gradient-based method is a kind of efficient and typical attack approach. However, this kind of method is likely to suffer from the problems of large perturbation value and low attack success rate. In this study, we propose an adversarial attack algorithm based on root-mean-square propagation. By integrating with the RMSProp, RMS-FGSM is more likely to generate optimal adversarial examples, and a high attack success rate can be achieved. The experimental results convincingly verified that the performance of the RMS-FGSM is better than the other methods when generating adversarial examples on the MNIST and CIFAR-100 datasets. Furthermore, the adversarial examples generated using the RMS-FGSM have a smaller perturbation under the same attack success rate.

Author Contributions

Conceptualization, D.H. and D.C.; methodology, D.C.; software, D.H.; validation, D.H., Y.Z. and H.Z.; formal analysis, L.X.; investigation, L.X.; writing—original draft preparation, D.H.; writing—review and editing, D.C.; visualization, D.H.; supervision, H.Z.; project administration, Y.Z.; funding acquisition, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (No. 62171328, 62171327).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the 26th Annual Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012. [Google Scholar]
  2. Xie, C.; Wang, J.; Zhang, Z.; Zhou, Y.; Xie, L.; Yuille, A. Adversarial Examples for Semantic Segmentation and Object Detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  3. Cisse, M.; Adi, Y.; Neverova, N.; Keshet, J. Houdini: Fooling deep structured prediction models. arXiv 2017, arXiv:1707.05373. [Google Scholar]
  4. Bojarski, M.; Del Testa, D.; Dworakowski, D.; Firner, B.; Flepp, B.; Goyal, P.; Jackel, L.D.; Monfort, M.; Muller, U.; Zhang, J.; et al. End to end learning for self-driving cars. arXiv 2016, arXiv:1604.07316. [Google Scholar]
  5. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. In Proceedings of the 2nd International Conference on Learning Representations, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  6. Buscarino, A.; Fortuna, L.; Frasca, M.; Gambuzza, L.V.; Sciuto, G. Memristive chaotic circuits based on cellular nonlinear networks. Int. J. Bifurc. Chaos 2012, 22, 1250070. [Google Scholar] [CrossRef]
  7. Zhang, H.; Yu, Y.; Jiao, J.; Xing, E.; El Ghaoui, L.; Jordan, M. Theoretically principled trade-off between robustness and accuracy. In Proceedings of the International Conference on Machine Learning (ICML), Long Beach, CA, USA, 9–15 June 2019; pp. 7472–7482. [Google Scholar]
  8. Wong, E.; Kolter, Z. Provable defenses against adversarial examples via the convex outer adversarial polytope. In Proceedings of the International Conference on Machine Learning (ICML), San Diego, CA, USA, 3–5 May 2018; pp. 5286–5295. [Google Scholar]
  9. Tram’er, F.; Kurakin, A.; Papernot, N.; Goodfellow, I.J.; Boneh, D.; McDaniel, P.D. Ensemble adversarial training: Attacks and defenses. In Proceedings of the International Conference on Learning Representations (ICLR), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  10. Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards deep learning models resistant to adversarial attacks. In Proceedings of the International Conference on Learning Representations (ICLR), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  11. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  12. Kurakin, A.; Goodfellow, I.; Bengio, S. Adversarial examples in the physical world. In Proceedings of the International Conference on Learning Representations Workshop, Toulon, France, 24–26 April 2017. [Google Scholar]
  13. Dong, Y.; Liao, F.; Pang, T.; Su, H.; Hu, X.; Li, J.; Zhu, J. Boosting adversarial attacks with momentum. arXiv 2017, arXiv:1710.06081. [Google Scholar]
  14. Carlini, N.; Wagner, D. Towards Evaluating the Robustness of Neural Networks. In Proceedings of the IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–26 May 2017. [Google Scholar]
  15. Xie, P.; Shi, S.; Xie, W.; Qin, R.; Hai, J.; Wang, L.; Chen, J.; Hu, G.; Yan, B. Improving the Transferability of Adversarial Examples by Using Generative Adversarial Networks and Data Enhancement. J. Phys. Conf. Ser. 2022, 2203, 012026. [Google Scholar] [CrossRef]
  16. Hu, S.; Liu, X.; Zhang, Y.; Li, M.; Zhang, L.Y.; Jin, H.; Wu, L. Protecting Facial Privacy: Generating Adversarial Identity Masks via Style-robust Makeup Transfer. arXiv 2022, arXiv:2203.03121. [Google Scholar]
  17. Tieleman, T.; Hinton, G. Lecture 6.5-rmsprop: Divide the gradient by a running average of its recent magnitude. Coursera Neural Netw. Mach. Learn. 2012, 4, 26–31. [Google Scholar]
  18. The Mnist Database of Handwritten Digits. Available online: http://yann.lecun.com/exdb/mnist/ (accessed on 11 June 2023).
  19. The CIFAR-10 Dataset. Available online: http://www.cs.toronto.edu/~kriz/cifar.html (accessed on 2 July 2023).
  20. Lecun, Y.; Bottou, L. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  21. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  22. Akhtar, N.; Mian, A. Threat of adversarial attacks on deep learning in computer vision: A survey. IEEE Access 2018, 6, 14410–14430. [Google Scholar] [CrossRef]
  23. Wu, T.; Luo, T.; Ii, D.C.W. Black-Box Attack using Adversarial Examples: A New Method of Improving Transferability. World Sci. Annu. Rev. Artif. Intell. 2023, 1, 2250005. [Google Scholar] [CrossRef]
  24. Ji, Y.; Zhou, G. Improving Adversarial Attacks with Ensemble-Based Approaches. In Proceedings of the CAAI International Conference on Artificial Intelligence, Beijing, China, 27–28 August 2022; Springer: Cham, Switzerland, 2022. [Google Scholar]
  25. G’eron, A. Hands-on Machine Learning with Scikit-Learn and Tensorflow: Concepts, Tools, and Techniques to Build Intelligent Systems; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2017. [Google Scholar]
  26. Papernot, N.; McDaniel, P.; Goodfellow, I. Transferability in machine learning: From phenomena to black-box attacks using adversarial samples. arXiv 2016, arXiv:1605.07277. [Google Scholar]
  27. Naseer, M.; Khan, S.H.; Rahman, S.; Porikli, F. Distorting neural representations to generate highly transferable adversarial examples. arXiv 2018, arXiv:1811.09020. [Google Scholar]
  28. Lin, J.; Song, C.; He, K.; Wang, L.; Hopcroft, J.E. Nesterov accelerated gradient and scale invariance for adversarial attacks. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  29. Nesterov, Y. A method for unconstrained convex minimization problem with the rate of convergence o(1/k2). Dokl. Akad. Nauk. USSR 1983, 269, 543–547. [Google Scholar]
Figure 1. The overall framework of our method.
Figure 1. The overall framework of our method.
Applsci 14 01257 g001
Figure 2. The prediction accuracy in the FGSM (a) and RMS-FGSM (b) methods when attacking the LeNet model. Accuracy represents the proportion of test samples that predicted correctly over all test samples; the lower prediction accuracy of the model indicates the higher attack success rate of the adversarial method. As EPS increases, the accuracy in RMS-FGSM decreases faster than that in FGSM and tends to converge.
Figure 2. The prediction accuracy in the FGSM (a) and RMS-FGSM (b) methods when attacking the LeNet model. Accuracy represents the proportion of test samples that predicted correctly over all test samples; the lower prediction accuracy of the model indicates the higher attack success rate of the adversarial method. As EPS increases, the accuracy in RMS-FGSM decreases faster than that in FGSM and tends to converge.
Applsci 14 01257 g002
Table 1. The experimental results of experiment 1 based on MNIST and LeNet 1.
Table 1. The experimental results of experiment 1 based on MNIST and LeNet 1.
EPSInputLaFGSMBIMMI-FGSMNI-FGSMRMS-FGSM
AdvArl2AdvArl2AdvArl2AdvArl2AdvArl2
0Applsci 14 01257 i0011Applsci 14 01257 i00210Applsci 14 01257 i00310Applsci 14 01257 i00410Applsci 14 01257 i00510Applsci 14 01257 i00610
0.05Applsci 14 01257 i0072Applsci 14 01257 i00821.374Applsci 14 01257 i00950.876Applsci 14 01257 i01070.884Applsci 14 01257 i01140.816Applsci 14 01257 i01280.745
0.1Applsci 14 01257 i0133Applsci 14 01257 i01482.753Applsci 14 01257 i01580.982Applsci 14 01257 i01690.975Applsci 14 01257 i01770.828Applsci 14 01257 i01870.776
0.15Applsci 14 01257 i0194Applsci 14 01257 i02094.081Applsci 14 01257 i02191.078Applsci 14 01257 i02290.978Applsci 14 01257 i02360.854Applsci 14 01257 i02450.792
0.2Applsci 14 01257 i0255Applsci 14 01257 i02625.517Applsci 14 01257 i02781.146Applsci 14 01257 i02861.042Applsci 14 01257 i02990.873Applsci 14 01257 i03080.804
0.25Applsci 14 01257 i0316Applsci 14 01257 i03206.878Applsci 14 01257 i03301.177Applsci 14 01257 i03441.078Applsci 14 01257 i03580.883Applsci 14 01257 i03690.801
0.3Applsci 14 01257 i0377Applsci 14 01257 i03828.149Applsci 14 01257 i03951.216Applsci 14 01257 i04081.106Applsci 14 01257 i04130.907Applsci 14 01257 i04220.806
0.35Applsci 14 01257 i0438Applsci 14 01257 i04459.591Applsci 14 01257 i04561.223Applsci 14 01257 i04651.149Applsci 14 01257 i04790.915Applsci 14 01257 i04860.813
0.4Applsci 14 01257 i0499Applsci 14 01257 i050210.969Applsci 14 01257 i05111.275Applsci 14 01257 i05241.178Applsci 14 01257 i05320.902Applsci 14 01257 i05450.809
Avg asr72.6%84.5%86.4%89.6%91.6%
Avg per5.4790.9970.9320.7750.705
1 La: the true label of the original image; Adv: the generated adversarial example image; Ar: the prediction label given using the model.
Table 2. The experimental results of experiment 2 based on CIFAR-100 and Alexnet.
Table 2. The experimental results of experiment 2 based on CIFAR-100 and Alexnet.
EPSInputLaFGSMBIMMI-FGSMNI-FGSMRMS-FGSM
AdvArl2AdvArl2AdvArl2AdvArl2AdvArl2
0Applsci 14 01257 i05511Applsci 14 01257 i056110Applsci 14 01257 i057110Applsci 14 01257 i058110Applsci 14 01257 i059110Applsci 14 01257 i060110
0.05Applsci 14 01257 i06123Applsci 14 01257 i062831319.89Applsci 14 01257 i063522168.72Applsci 14 01257 i064432749.18Applsci 14 01257 i065712883.94Applsci 14 01257 i06672331.99
0.1Applsci 14 01257 i06738Applsci 14 01257 i068582631.59Applsci 14 01257 i069572373.98Applsci 14 01257 i070552950.84Applsci 14 01257 i071663004.96Applsci 14 01257 i072362422.59
0.15Applsci 14 01257 i07342Applsci 14 01257 i074283959.67Applsci 14 01257 i075942584.86Applsci 14 01257 i076672989.19Applsci 14 01257 i07743041.01Applsci 14 01257 i078462425.45
0.2Applsci 14 01257 i07951Applsci 14 01257 i080395278.10Applsci 14 01257 i081472832.27Applsci 14 01257 i082133055.10Applsci 14 01257 i083483076.21Applsci 14 01257 i084392422.92
0.25Applsci 14 01257 i08565Applsci 14 01257 i086486439.86Applsci 14 01257 i087553084.59Applsci 14 01257 i088733024.25Applsci 14 01257 i089523103.83Applsci 14 01257 i090602392.04
0.3Applsci 14 01257 i09174Applsci 14 01257 i092297917.18Applsci 14 01257 i093743113.83Applsci 14 01257 i094853078.15Applsci 14 01257 i095223064.81Applsci 14 01257 i096652413.65
0.35Applsci 14 01257 i09788Applsci 14 01257 i098419239.20Applsci 14 01257 i099543152.67Applsci 14 01257 i100933126.84Applsci 14 01257 i101323105.91Applsci 14 01257 i102832433.23
0.4Applsci 14 01257 i10396Applsci 14 01257 i104310,304.54Applsci 14 01257 i105793088.63Applsci 14 01257 i106663093.78Applsci 14 01257 i107523093.25Applsci 14 01257 i108452424.46
Avg asr72.4%83.5%90.2%92.8%94.6%
Avg per5232.222488.832674.142708.212140.7
Table 3. The experimental results of experiment 2 based on CIFAR-100 and Resnet-101.
Table 3. The experimental results of experiment 2 based on CIFAR-100 and Resnet-101.
EPSInputLaFGSMBIMMI-FGSMNI-FGSMRMS-FGSM
AdvArl2AdvArl2AdvArl2AdvArl2AdvArl2
0Applsci 14 01257 i10915Applsci 14 01257 i110150Applsci 14 01257 i111150Applsci 14 01257 i112150Applsci 14 01257 i113150Applsci 14 01257 i114150
0.05Applsci 14 01257 i11529Applsci 14 01257 i116861318.59Applsci 14 01257 i117962246.92Applsci 14 01257 i11872860.22Applsci 14 01257 i119962918.59Applsci 14 01257 i120782298.88
0.1Applsci 14 01257 i12136Applsci 14 01257 i122452629.47Applsci 14 01257 i123732414.55Applsci 14 01257 i124672967.01Applsci 14 01257 i125672967.01Applsci 14 01257 i126332507.00
0.15Applsci 14 01257 i12745Applsci 14 01257 i128133951.65Applsci 14 01257 i129762599.35Applsci 14 01257 i130373003.38Applsci 14 01257 i131183058.88Applsci 14 01257 i132612417.71
0.2Applsci 14 01257 i13357Applsci 14 01257 i134815279.56Applsci 14 01257 i135832828.45Applsci 14 01257 i136293052.83Applsci 14 01257 i137103072.11Applsci 14 01257 i138802425.84
0.25Applsci 14 01257 i13968Applsci 14 01257 i140486598.16Applsci 14 01257 i141113023.82Applsci 14 01257 i142623103.63Applsci 14 01257 i143593103.83Applsci 14 01257 i144432402.16
0.3Applsci 14 01257 i14575Applsci 14 01257 i146297883.54Applsci 14 01257 i14753089.71Applsci 14 01257 i148443099.17Applsci 14 01257 i149933100.97Applsci 14 01257 i150352422.22
0.35Applsci 14 01257 i15184Applsci 14 01257 i152239231.48Applsci 14 01257 i153783096.01Applsci 14 01257 i154803103.81Applsci 14 01257 i155943106.59Applsci 14 01257 i156982461.51
0.4Applsci 14 01257 i15792Applsci 14 01257 i1587610,459.62Applsci 14 01257 i159623112.46Applsci 14 01257 i160643109.20Applsci 14 01257 i16163101.57Applsci 14 01257 i162492400.20
Avg asr81.9%97.3%99.1%99.6%99.9%
Avg per5261.482499.032699.142714.392148.39
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hong, D.; Chen, D.; Zhang, Y.; Zhou, H.; Xie, L. Attacking Robot Vision Models Efficiently Based on Improved Fast Gradient Sign Method. Appl. Sci. 2024, 14, 1257. https://doi.org/10.3390/app14031257

AMA Style

Hong D, Chen D, Zhang Y, Zhou H, Xie L. Attacking Robot Vision Models Efficiently Based on Improved Fast Gradient Sign Method. Applied Sciences. 2024; 14(3):1257. https://doi.org/10.3390/app14031257

Chicago/Turabian Style

Hong, Dian, Deng Chen, Yanduo Zhang, Huabing Zhou, and Liang Xie. 2024. "Attacking Robot Vision Models Efficiently Based on Improved Fast Gradient Sign Method" Applied Sciences 14, no. 3: 1257. https://doi.org/10.3390/app14031257

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop