This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Looking From the Future: Multi-order Iterations Can Enhance Adversarial Attack Transferability

Zijian Ying, Qianmu Li, Tao Wang, Zhichao Lian, Shunmei Meng, Xuyun Zhang Zijian Ying, Zhichao Lian, Qianmu Li are with the School of Cyber Science and Technology, Nanjing University of Science and Technology, Nanjing 210094, China. E-mail: [email protected], [email protected], [email protected] Wang and Meng Shunmei are with the School of Computer Science and Technology, Nanjing University of Science and Technology, Nanjing 210094, China. E-mail: [email protected], [email protected] Xuyun is with the School of Computing, Faculty of Science and Engineering, Macquarie University, NSW 2109, Australia. E-mail: [email protected]
Abstract

Various methods try to enhance adversarial transferability by improving the generalization from different perspectives. In this paper, we rethink the optimization process and propose a novel sequence optimization concept, which is named Looking From the Future (LFF). LFF makes use of the original optimization process to refine the very first local optimization choice. Adapting the LFF concept to the adversarial attack task, we further propose an LFF attack as well as an MLFF attack with better generalization ability. Furthermore, guiding with the LFF concept, we propose an LLF𝒩LLF^{\mathcal{N}} attack which entends the LFF attack to a multi-order attack, further enhancing the transfer attack ability. All our proposed methods can be directly applied to the iteration-based attack methods. We evaluate our proposed method on ImageNet1k datasets by applying several SOTA adversarial attack methods under four kinds of tasks. Experimental results show that our proposed method can greatly enhance the attack transferability. Ablation experiments are also applied to verify the effectiveness of each component. The source code will be released after this paper is accepted.

Index Terms:
Adversarial attack, Optimization process, XAI

I Introduction

Deep Neural Networks(DNNs) have shown vulnerability to adversarial examples. By adding human-imperceptible perturbations to the clean input, DNNs will result in misclassification. At the same time, some adversarial examples generated on one network can also have an attack effect on another network. This phenomenon is called adversarial attack transferability. And, some methods are trying to attack a black-box network by generating adversarial examples on a white-box network. This kind of attack behavior is called the transfer attack. For the generated adversarial sample, the more black-box networks it can successfully attack, the stronger the transfer attack capability of the adversarial sample will be. This capability is also called adversarial transferability.

Nowadays, a lot of works try to enhance adversarial transferability with various kinds of approaches. One of the most classic methods is the MI-FGSM[1]. MI-FGSM, utilizing momentum to replace the gradient, inherits the iterative method in I-FGSM[5] to enhance the attack capability of the adversarial examples and enhance adversarial transferability by combining historical gradient information at the same time. This leads to a branch of adversarial attack methods with optimizing gradients themselves. These methods attempt to improve the adversarial transferability of adversarial examples by changing the gradient of the optimization process to avoid falling into local minima. Then another branch of adversarial methods is directly modifying the loss function to change the optimization terminal. Recently, input-transformation methods utilize data augmentation approaches, changing the input, to improve the diversity of the gradients. These methods indeed effectively enhance the adversarial transferability.

However the essential of these methods is solving the optimization problem, i.e. optimization methods. The optimization process of finding the optimal result through iteration can be regarded as a sequence. Then, each search point is a node in this sequence. Once the optimization method and the starting point of the search are determined, the sequence itself is relatively fixed. It can be found that most of the existing adversarial attack methods only generate the next node with information from the current node and previous nodes. This will prevent the optimization process itself from benefiting from information from subsequent nodes. Merely using the current node or previous nodes will make each node update only in the direction of local optimization with limited global information. Since the subsequent nodes belong to the future for the current node, we name the information from the subsequent nodes as future information. Similarly, the information from previous nodes is called historical information. Therefore, refining the optimization process with future information is necessary.

There are a few works that have made preliminary attempts to use this future information. [14] proposed the PI-FGSM method which looks ahead with one step to guide the momentum generation. Although this work somehow utilizes a little future information, i.e. the gradient from the next node, the terminal goal is to rich the local information by sampling around the current node. Another work that somehow utilizes future information is [9]. This work determines the starting state of momentum to optimize gradient consistency by pre-querying the perturbations after N rounds before starting the search. However, the future information is merely utilized once at the very beginning of the optimization. Also, with the search steps going greater, the future information will be diluted. Therefore, the above two works still only use local gradients to update nodes to a certain extent.

In this work, we start by drawing the concept of looking from the future (LFF). LFF rethinks the optimization process and regenerates the first node from the perspective of the optimization result. However, there are two problems for LFF directly applying to generate adversarial examples. The first one is the constraint of perturbation of each node makes the re-optimization result from LFF have worse attack capability. Another one is if the optimization process utilizes the historical information, LFF will repeatedly accumulate historical information, thus falling into local search again. To tackle the above two problems, we modify the LFF concept into the process of the original adversarial attack methods and formally propose the LFF attack. We reintegrate the weights of each piece of information in the optimization process so that no overfitting of local optimization will occur regardless of whether only the current node information or historical information is used. Therefore, the LFF attack can be easily applied to the existing adversarial attack methods, which are based on the iteration optimization process, to enhance the adversarial attack performance. At the same time, to reduce the complexity of the algorithm and further increase the adversarial transferability, we also combined the existing mechanisms to optimize the attack process. Furthermore, we also propose a multi-order LFF attack, under the guidance of the LFF concept, which further enhances the transfer attack performance.

To evaluate our method comprehensively, we evaluate the performance of our proposed method with several state-of-the-art adversarial attack methods on a wide range of high-performance deep neural networks, containing both CNN structure networks and Transformer structure networks. By covering the scenery of single model attacks, attack defensive networks, targetted attacks and ensemble network attacks, the comparison experiment can verify the effectiveness of our proposed method. The ablation experiment is also applied to check the effect of each component and each hyperparameter.

The remainder of this paper is structured as follows. Section II delves into related work on adversarial attack methods. Section III provides an overview of the notations used in adversarial attacks, the concept of looking from the future and the difficulty of applying LFF to the adversarial attack process. In Section IV, we present the LFF attack method and derivation process. The multi-order LFF attack method is also attached. Section V presents the results of our experiments and corresponding analysis. Lastly, in section VI, we offer some concluding remarks and outline potential avenues for future research.

II Related Work

Since [4] proposed the FGSM methods, the adversarial attack, especially the transfer attack, has been greatly improved. A lot of methods improve the adversarial attack based on the FGSM methods which are marked as FGSMs. One of the most classic methods is the I-FGSM[5]. I-FGSM enhances the adversarial attack by iteratively searching the perturbation, while FGSM merely generates the adversarial example with one step. Although I-FGSM can successfully attack the white-box networks, which means a high attack capability, it has a very poor transfer attack performance. Then various branches of methods try to enhance the adversarial transferability based on the I-FGSM.

II-A FGSMs based on gradients

One of the branches is modifying the gradients to obtain generalization adversarial examples. The most representative work is the MI-FGSM[1]. Via introducing momentum into to adversarial examples generation process, gradients from each previous node will be utilized to smooth the search direction for the current node. This will help the optimization process somehow avoid the local minima. NI-FGSM[7] method refined the momentum calculation process to further enhance the adversarial example generalization. VMI-FGSM and VNI-FGSM [12] utilize the variance tuning the momentum to further avoid the local minima. EMI-FGSM[14] sampling the local gradient information around the node from PI-FGSM and ensemble the information to enhance the adversarial transferability. PGN[3] tries to find the local maxima point to directly find the better transfer attack condition. MIG[8] utilizes the integrated gradients in XAI methods to enhance the gradient in MI-FGSM. Some other works inspired by the optimizers refine the gradients or the momentums with the latest optimizer [20, 18, 21].

The above methods mainly focus on refining the gradient itself, while some other methods try to enhance the gradient by modifying the loss function. [6] proposes the Po+Trip loss to enhance the targeted transfer attack inspired by the Poincare distance. [22] thinks that the simple logit loss also has good transfer attack performance. [19] considers the class interaction and proposes the relative Cross Entropy loss by raising the prediction probability of other classes to enhance the adversarial attack generalization.

II-B FGSMs based on data augmentation

Another one of the branches is utilizing data augmentation to enhance the diversity of input, which is also called input transformation. The most representative work is the DI2\text{DI}^{2}-FGSM [17], which uses resize and padding operations to enhance input diversity. TI-FGSM [2] utilizes shifting operation with a kernel matrix on the gradients. SI-FGSM [7] applies scale transformation to the input image to gain augmented data. SIA [16] divides the original image into blocks and applies a random image transformation onto each image block to craft a set of diverse images for gradient calculation. BSR [10] furthermore introduces the block shuffle and rotation operations into the input transformation.

The above methods mainly focus on the transformation of the input image itself, which can be called self-transformation. Some other methods try to enhance the input diversity by introducing information from other images. [13] extends the SI-FGSM with the mixup strategy, which is named Admix. This method chooses several other images and mixes them into the original images to enhance the input diversity. [11] improves the mixup strategy with a non-linear way to mask the chosen image into the original image. [15] rethinks the process of the admix and changes the mixing images process to the mixing gradients process.

III Preliminary

III-A Notations

Here we first give out some fundamental notations. xRNx\in R^{N} is the input data sample, where NN is the data dimension. H:RNRDH:R^{N}\to R^{D} is the feature extractor, where DD is the feature dimension. G:RDRKG:R^{D}\to R^{K} is the classifier, where KK is the number of classes. Then F=GH:RNRKF=G\circ H:R^{N}\to R^{K} is the entire classification network. The output of FF is zRKz\in R^{K} called logits. The output of softmax(z)\text{softmax}(z) is pp which is the predicted probability. pp is also marked as PF(x)RKP_{F}(x)\in R^{K} to emphsize the network and input data. The goal of the adversarial attack is to generate an adversarial example x^\hat{x} that can lead to ff fail, which is expressed as F(x)=ytF(x^)ytF(x)=y_{t}\wedge F(\hat{x})\neq y_{t}. Here yty_{t} indicates the truth label of xx as the specific predicted class index. To express without ambiguity, yy represents the index classification result for FF, zz represents the vector result for FF and hh represents the vector result for HH.

Then the adversarial attack transferability usually can be expressed as follows:

F1(x)=F2(x)=yF1(x^)yF2(x^)yF^{1}(x)=F^{2}(x)=y\wedge F^{1}(\hat{x})\neq y\wedge F^{2}(\hat{x})\neq y (1)

This expression can be understood that for an adversarial example x^\hat{x} which can successfully attack network F1F^{1}, x^\hat{x} can also successfully attack network F2F^{2}.

Then perturbation δ=x^x\delta=\hat{x}-x is generated by the adversarial example generation method which is marked as 𝒜()\mathcal{A}(\cdot). Under iteration scope, the perturbation δ\delta can be decomposed with the sum of the perturbation from each iteration which can be presented as δ=t=1Tδt\delta=\sum_{t=1}^{T}\delta_{t}. The constraint for each iteration perturbation δt\delta_{t} can be presented as δtpϵt{||\delta_{t}||}_{p}\leq\epsilon_{t}. In the I-FGSM-based method, ϵt\epsilon_{t} usually constantly equals a certain value, e.g. ϵt=ϵ/T\epsilon_{t}=\epsilon/T. To make a more clear statement, with no special instructions, δ\delta with subscript numbers represents the corresponding iteration perturbation, e.g. δt\delta_{t} is the tt-th round perturbation. And δ\delta with superscript numbers represents the sum of the corresponding iteration perturbation, e.g. δt=i=1tδi\delta^{t}=\sum_{i=1}^{t}\delta_{i}.

III-B Looking from the future (new)

The terminal goal of the adversarial attack is to find an optimal adversarial example that can successfully attack the target model with minor perturbation as much as it can. This task is usually transferred into an optimization task with setting a loss function as the optimization goal, e.g. the optimization goal for the untargeted attack task can be formulated as:

argmax𝛿(F(x+δ),yt),s.t.δpϵ,\underset{\delta}{\operatorname{argmax}}\quad\mathcal{L}(F(x+\delta),y_{t}),s.t.||\delta||_{p}\leq\epsilon, (2)

where ϵ\epsilon is the threshold for the perturbation which emphasizes the perturbation constraint. Previous methods try to enhance the adversarial transferability by optimizing the optimal process with various approaches. However, from the goal to look at the optimization process, if there is an optimal point δ𝒪\delta^{\mathcal{O}} for (2), then the shortest search routine can be determined, e.g. the straight line segment from 0 and δ𝒪\delta^{\mathcal{O}} in Euclidean space. Then the best first iteration result δ1𝔅\delta^{\mathfrak{B}}_{1} for the entire optimization task should satisfy:

argminδ1𝔅δ𝒪δ1𝔅p,s.t.δ1𝔅pϵ1.\underset{\delta^{\mathfrak{B}}_{1}}{\operatorname{argmin}}\quad||\delta^{\mathcal{O}}-\delta^{\mathfrak{B}}_{1}||_{p},s.t.||\delta^{\mathfrak{B}}_{1}||_{p}\leq\epsilon_{1}. (3)

The solution for this optimization is

δ1𝔅=αδ𝒪,\delta^{\mathfrak{B}}_{1}=\alpha\cdot\delta^{\mathcal{O}}, (4)

where α\alpha is a linear coefficient to limit δ1𝔅\delta^{\mathfrak{B}}_{1} satisfying the constraint. However, directly obtaining δ𝒪\delta^{\mathcal{O}} is a really tough task. Finding the optimization point through iteration is one of the most effective and commonly used methods. These methods continuously approach the optimization target in an iterative manner to obtain an approximation of the optimization result. After a certain round of search, the optimization result can be treated as an approximation point δ𝒪\delta^{\mathcal{O^{\prime}}} for the optimal point δ𝒪\delta^{\mathcal{O}}. Then, the shortest search routine from 0 to δ𝒪\delta^{\mathcal{O^{\prime}}} can be determined. When back to the very first iteration, the best first optimization search result should be

argminδ1𝔅δ𝒪δ1𝔅p,s.t.δ1𝔅p=ϵ1.\underset{\delta^{\mathfrak{B^{\prime}}}_{1}}{\operatorname{argmin}}\quad||\delta^{\mathcal{O}}-\delta^{\mathfrak{B^{\prime}}}_{1}||_{p},s.t.||\delta^{\mathfrak{B^{\prime}}}_{1}||_{p}=\epsilon_{1}. (5)

The corresponding solution for this optimization is

δ1𝔅=αδ𝒪=αTt=1δt𝒪.\delta^{\mathfrak{B^{\prime}}}_{1}=\alpha\cdot\delta^{\mathcal{O^{\prime}}}=\alpha\cdot{\sum_{T}^{t=1}\delta^{\mathcal{O^{\prime}}}_{t}}. (6)

From a general scope, for any optimization task, after obtaining the original search routine, the approximation for the best first iteration result can also be obtained with Eq. (6). Then with the same optimization process, the approximation for the best next iteration result (δ2𝔅\delta^{\mathfrak{B^{\prime}}}_{2}) as well as any tt-th iteration result (δt𝔅\delta^{\mathfrak{B^{\prime}}}_{t}) can be obtained. Considering that the δ𝒪\delta^{\mathcal{O}} usually cannot be directly obtained, δt𝔅\delta^{\mathfrak{B}}_{t} can also be represented with δt𝔅\delta^{\mathfrak{B^{\prime}}}_{t} in this manuscript. We name this process as Looking From the Future (LFF).

III-C LFF in transfer attack

I-FGSM-based attack methods obtain great success in transfer attacks. Intuitively, LFF can be directly used to enhance these kinds of attack methods. However, FGSM-based methods have the mechanism of symbolization of gradients which means perturbations of each iteration all satisfy δt=ϵ||\delta_{t}||_{\infty}=\epsilon_{\infty} and δt2=Nϵ2||\delta_{t}||_{2}=\sqrt{N\cdot\epsilon^{2}_{\infty}} at the same time. However, direcly applying Eq. (6) which refers to δ1𝔅=Tt=1δt/T\delta^{\mathfrak{B}}_{1}={\sum_{T}^{t=1}\delta^{t}}/T will cause δt𝔅ϵ||\delta^{\mathfrak{B}}_{t}||_{\infty}\leq\epsilon_{\infty} and δ1𝔅2δt2||\delta^{\mathfrak{B}}_{1}||_{2}\leq||\delta_{t}||_{2}. Only when i,jT,δi=δj\forall i,j\in T,\delta_{i}=\delta_{j}, δ1𝔅=δt||\delta^{\mathfrak{B}}_{1}||_{\infty}=||\delta_{t}||_{\infty} and δ1𝔅2=δt2||\delta^{\mathfrak{B}}_{1}||_{2}=||\delta_{t}||_{2}. This can only happen on a strictly linear function, while feature space in deep learning networks is usually nonlinear. The α\alpha in Eq. (6) can apply δ1𝔅\delta^{\mathfrak{B}}_{1} to satisfy δ1𝔅=δt||\delta^{\mathfrak{B}}_{1}||_{\infty}=||\delta_{t}||_{\infty} or δ1𝔅2=δt2||\delta^{\mathfrak{B}}_{1}||_{2}=||\delta_{t}||_{2}. However, simply applying LFF on LL_{\infty} constraint will not work. Keeping δ1𝔅=δt||\delta^{\mathfrak{B}}_{1}||_{\infty}=||\delta_{t}||_{\infty} will cause δ1𝔅2δt2||\delta^{\mathfrak{B}}_{1}||_{2}\leq||\delta_{t}||_{2} which means δt𝔅\delta^{\mathfrak{B}}_{t} will have a smaller Euclidean distance compared with δt\delta_{t}. A smaller Euclidean distance will lead to a slower convergence speed, which means it may even need more iterations to reach the optimal points.

Meanwhile, applying LFF on L2L_{2} will cause the δt𝔅ϵ||\delta^{\mathfrak{B}}_{t}||_{\infty}\geq\epsilon_{\infty}. When the search routine comes near TϵT\cdot\epsilon_{\infty} (the LL_{\infty} constraint to the final perturbation), a large number of effective updates for those dimensions that already reach TϵT\cdot\epsilon_{\infty} will be discarded. This will also make the optimization routine less effective.

Also, considering that greater adversarial transferability related to more generalized examples for each model, when 𝒜()\mathcal{A}(\cdot) itself gets overfitting to a certain model or repeatedly using history information, LFF will further exacerbate the degree of the overfitting.

IV Methodology

IV-A One Order LFF Attack

Giving an I-FGSM based attacking method 𝒜()\mathcal{A}(\cdot) and the clean data xx, the original tt-th iteration from the attack process is δt=𝒜(F,x+i=0t1δi)\delta_{t}=\mathcal{A}(F,x+\sum_{i=0}^{t-1}\delta_{i}), where tTt\in T and x+δ0=xx+\delta_{0}=x. The gradient corresponding to the δt\delta_{t} is gt=𝒜g(F,x+i=0t1δi)g_{t}=\mathcal{A}_{g}(F,x+\sum_{i=0}^{t-1}\delta_{i}), which means that δt=sign(gt)\delta_{t}=\textit{sign}(g_{t}), where sign()\textit{sign}(\cdot) is the symbolization function. 𝒬\mathcal{Q} is the quantity of the steps looking from the future, i.e. 𝒜()\mathcal{A}(\cdot) iters 𝒬\mathcal{Q} steps. Then the straightforward description for the tt-th perturbation from one order LFF attack is :

δt𝔅=αsign(q=1𝒬βq𝒜g(F,x+j=0t1δj𝔅+i=0q1δi)𝒜g(F,x+j=0t1δj𝔅+i=0q1δi)p),\delta^{\mathfrak{B}}_{t}=\alpha\cdot\textit{sign}(\sum_{q=1}^{\mathcal{Q}}\beta^{q}\cdot\frac{\mathcal{A}_{g}(F,x+\sum_{j=0}^{t-1}\delta^{\mathfrak{B}}_{j}+\sum_{i=0}^{q-1}\delta_{i})}{{||\mathcal{A}_{g}(F,x+\sum_{j=0}^{t-1}\delta^{\mathfrak{B}}_{j}+\sum_{i=0}^{q-1}\delta_{i})||_{p}}}), (7)

where α\alpha is the updating rate, δ0𝔅=δ0\delta^{\mathfrak{B}}_{0}=\delta_{0}, and β\beta is the future penalty coefficient.

When Eq. (7) is directly applied to the pure I-FGSM method, gtg_{t} equals the gradients of the corresponding input. The description is:

𝒜g(F,x+i=0t1δi)=x+i=0t1δi,\mathcal{A}_{g}(F,x+\sum_{i=0}^{t-1}\delta_{i})=\nabla_{x+\sum_{i=0}^{t-1}\delta_{i}}\mathcal{L}, (8)

where x\nabla_{x}\mathcal{L} is the gradient of xx for the loss function \mathcal{L}. Then Eq. (7) can be rewritten with the following description:

δt𝔅=αsign(q=1𝒬βqx+j=0t1δj𝔅+i=0q1δix+j=0t1δj𝔅+i=0q1δip),\delta^{\mathfrak{B}}_{t}=\alpha\cdot\textit{sign}(\sum_{q=1}^{\mathcal{Q}}\beta^{q}\cdot\frac{\nabla_{x+\sum_{j=0}^{t-1}\delta^{\mathfrak{B}}_{j}+\sum_{i=0}^{q-1}\delta_{i}}\mathcal{L}}{{||\nabla_{x+\sum_{j=0}^{t-1}\delta^{\mathfrak{B}}_{j}+\sum_{i=0}^{q-1}\delta_{i}}\mathcal{L}||_{p}}}), (9)

When Eq. (7) is applied to the MI-FGSM method, which is the most widely used method, gtg_{t} equals the momentums of the corresponding input. The description is

𝒜g(F,x+i=0t1δi)\displaystyle\mathcal{A}_{g}(F,x+\sum_{i=0}^{t-1}\delta_{i}) =M(x+i=0t1δi),\displaystyle=M({x+\sum_{i=0}^{t-1}\delta_{i}}), (10)
M(x+i=0t1δi)\displaystyle M({x+\sum_{i=0}^{t-1}\delta_{i}}) =μM(x+i=0t2δi)+x+i=0t1δix+i=0t1δi1,\displaystyle=\mu M({x+\sum_{i=0}^{t-2}\delta_{i}})+\frac{\nabla_{x+\sum_{i=0}^{t-1}\delta_{i}}\mathcal{L}}{||\nabla_{x+\sum_{i=0}^{t-1}\delta_{i}}\mathcal{L}||_{1}},

where M(x+δ0)=xx1M(x+\delta_{0})=\frac{\nabla_{x}\mathcal{L}}{||\nabla_{x}\mathcal{L}||_{1}} and μ\mu is the momentum decay factor. When Eq.(10) is introduced into Eq.(7), the formulation can be obtained as

δt𝔅\displaystyle\delta^{\mathfrak{B}}_{t} =αsign(q=1𝒬βqM(x+j=0t1δj𝔅+i=0q1δi)M(x+j=0t1δj𝔅+i=0q1δi)p)\displaystyle=\alpha\cdot\textit{sign}(\sum_{q=1}^{\mathcal{Q}}\beta^{q}\cdot\frac{M(x+\sum_{j=0}^{t-1}\delta^{\mathfrak{B}}_{j}+\sum_{i=0}^{q-1}\delta_{i})}{{||M(x+\sum_{j=0}^{t-1}\delta^{\mathfrak{B}}_{j}+\sum_{i=0}^{q-1}\delta_{i})||_{p}}}) (11)

The expansion of Eq.(11) (shorten x+j=0t1δj𝔅+i=0q1δix+\sum_{j=0}^{t-1}\delta^{\mathfrak{B}}_{j}+\sum_{i=0}^{q-1}\delta_{i} with x𝔅t1q1x_{\mathfrak{B}_{t-1}}^{q-1}) is

δt𝔅\displaystyle\delta^{\mathfrak{B}}_{t} =αsign(q=1𝒬βqM(x𝔅t1q1)M(x𝔅t1q1)p)\displaystyle=\alpha\cdot\textit{sign}(\sum_{q=1}^{\mathcal{Q}}\beta^{q}\cdot\frac{M(x_{\mathfrak{B}_{t-1}}^{q-1})}{{||M(x_{\mathfrak{B}_{t-1}}^{q-1})||_{p}}}) (12)
=αsign(q=1𝒬βqM(x𝔅t1q1)pi=0q1μq1ix𝔅t1ix𝔅t1i1)\displaystyle=\alpha\cdot\textit{sign}(\sum_{q=1}^{\mathcal{Q}}\frac{\beta^{q}}{{||M(x_{\mathfrak{B}_{t-1}}^{q-1})||_{p}}}\cdot\sum_{i=0}^{q-1}\mu^{q-1-i}\frac{\nabla_{x_{\mathfrak{B}_{t-1}}^{i}}\mathcal{L}}{||\nabla_{x_{\mathfrak{B}_{t-1}}^{i}}\mathcal{L}||_{1}})
=αsign(i=0𝒬1(l=i𝒬1βl+1μliM(x𝔅t1l)p)x𝔅t1ix𝔅t1i1).\displaystyle=\alpha\cdot\textit{sign}(\sum_{i=0}^{\mathcal{Q}-1}(\sum_{l=i}^{\mathcal{Q}-1}\frac{\beta^{l+1}\cdot\mu^{l-i}}{||M(x_{\mathfrak{B}_{t-1}}^{l})||_{p}})\cdot\frac{\nabla_{x_{\mathfrak{B}_{t-1}}^{i}}\mathcal{L}}{||\nabla_{x_{\mathfrak{B}_{t-1}}^{i}}\mathcal{L}||_{1}}).

Eq.(12) indicates that for a given optimization process, the δt𝔅\delta^{\mathfrak{B}}_{t} can be represented as a polynomial refers to the gradient of each optimization point using L1L_{1}-norm regularization. The coefficient refers to the ii-th optimization point can be marked as 𝒞i=l=i𝒬1βl+1μliM(x𝔅t1l)p\mathcal{C}_{i}=\sum_{l=i}^{\mathcal{Q}-1}\frac{\beta^{l+1}\cdot\mu^{l-i}}{||M(x_{\mathfrak{B}_{t-1}}^{l})||_{p}} and the gradient of ii-th optimization point with L1L_{1}-norm regularization can be marked as 𝒢i=x𝔅t1ix𝔅t1i1\mathcal{G}_{i}=\frac{\nabla_{x_{\mathfrak{B}_{t-1}}^{i}}\mathcal{L}}{||\nabla_{x_{\mathfrak{B}_{t-1}}^{i}}\mathcal{L}||_{1}}. Then Eq.(12) can be refined as

δt𝔅=αsign(𝒞,𝒢),\displaystyle\delta^{\mathfrak{B}}_{t}=\alpha\cdot\textit{sign}(\left\langle\vec{\mathcal{C}},\vec{\mathcal{G}}\right\rangle), (13)

where ,\left\langle\cdot,\cdot\right\rangle is the inner product, 𝒞=(𝒞0,𝒞1,,𝒞𝒬1)\vec{\mathcal{C}}=(\mathcal{C}_{0},\mathcal{C}_{1},\cdots,\mathcal{C}_{\mathcal{Q}-1}) and 𝒢=(𝒢0,𝒢1,,𝒢𝒬1)\vec{\mathcal{G}}=(\mathcal{G}_{0},\mathcal{G}_{1},\cdots,\mathcal{G}_{\mathcal{Q}-1}). Then Eq.(13) can be understood as a superposition of 𝒢i\mathcal{G}_{i}. 𝒢i\mathcal{G}_{i} with a smaller value of ii is closer to the starting point xx. The closer the gradient of a point is to xx, the more local optimization information it contains; the farther the gradient of a point is from xx, the more generalization information it contains. When coefficients of those 𝒢i\mathcal{G}_{i} with a smaller value of ii are too great, δt𝔅\delta^{\mathfrak{B}}_{t} will tend to x𝔅t11x_{\mathfrak{B}_{t-1}}^{1} which means overfitting. When coefficients of those 𝒢i\mathcal{G}_{i} with a greater value of ii are too great, δt𝔅\delta^{\mathfrak{B}}_{t} will reduce attack capability due to too much generalization. Looking back to 𝒞\vec{\mathcal{C}}, the result of dividing 𝒞0\mathcal{C}_{0} by 𝒞𝒬1\mathcal{C}_{\mathcal{Q}-1} is

𝒞0𝒞𝒬1\displaystyle\frac{\mathcal{C}_{0}}{\mathcal{C}_{\mathcal{Q}-1}} =l=0𝒬1βl+1μl0M(x𝔅t1l)pl=𝒬1𝒬1βl+1μl(𝒬1)M(x𝔅t1l)p\displaystyle=\frac{\sum_{l=0}^{\mathcal{Q}-1}\frac{\beta^{l+1}\cdot\mu^{l-0}}{||M(x_{\mathfrak{B}_{t-1}}^{l})||_{p}}}{\sum_{l=\mathcal{Q}-1}^{\mathcal{Q}-1}\frac{\beta^{l+1}\cdot\mu^{l-(\mathcal{Q}-1)}}{||M(x_{\mathfrak{B}_{t-1}}^{l})||_{p}}} (14)
=M(x𝔅t1𝒬1)pβ𝒬l=0𝒬1βl+1μlM(x𝔅t1l)p\displaystyle=\frac{||M(x_{\mathfrak{B}_{t-1}}^{\mathcal{Q}-1})||_{p}}{\beta^{\mathcal{Q}}}\sum_{l=0}^{\mathcal{Q}-1}\frac{\beta^{l+1}\cdot\mu^{l}}{||M(x_{\mathfrak{B}_{t-1}}^{l})||_{p}}
=l=0𝒬1μlM(x𝔅t1𝒬1)pM(x𝔅t1l)pβ𝒬l1\displaystyle=\sum_{l=0}^{\mathcal{Q}-1}\frac{\mu^{l}\cdot\frac{||M(x_{\mathfrak{B}_{t-1}}^{\mathcal{Q}-1})||_{p}}{||M(x_{\mathfrak{B}_{t-1}}^{l})||_{p}}}{\beta^{\mathcal{Q}-l-1}}

For an attack method without overfitting, it can be assumed that i,j𝒬1,i<j,M(x𝔅t1i)p<M(x𝔅t1j)p\forall i,j\in\mathcal{Q}-1,i<j,||M(x_{\mathfrak{B}_{t-1}}^{i})||_{p}<||M(x_{\mathfrak{B}_{t-1}}^{j})||_{p}. Then for given μ>0\mu>0 and β>0\beta>0, 𝒢0𝒢𝒬1\frac{\mathcal{G}_{0}}{\mathcal{G}_{\mathcal{Q}-1}} will be greater when 𝒬\mathcal{Q} becomes greater. For example, when μ=1\mu=1 and β=1\beta=1, the upper bound of Eq.(14) is

𝒞0𝒞𝒬1=l=0𝒬1M(x𝔅t1𝒬1)pM(x𝔅t1l)pl=0𝒬1𝒬l+1.\frac{\mathcal{C}_{0}}{\mathcal{C}_{\mathcal{Q}-1}}=\sum_{l=0}^{\mathcal{Q}-1}\frac{||M(x_{\mathfrak{B}_{t-1}}^{\mathcal{Q}-1})||_{p}}{||M(x_{\mathfrak{B}_{t-1}}^{l})||_{p}}\leq\sum_{l=0}^{\mathcal{Q}-1}\frac{\mathcal{Q}}{l+1}. (15)

Even if 𝒬\mathcal{Q} is not great, e.g. 𝒬=5\mathcal{Q}=5, 𝒢0𝒢𝒬1\frac{\mathcal{G}_{0}}{\mathcal{G}_{\mathcal{Q}-1}} will be 11.41711.417 which is much greater than 11. And, this will cause the LFF result to tend to the 𝒢i\mathcal{G}_{i} with a smaller index value. In extreme cases, this situation can be described as:

𝒬+,δt𝔅αsign(𝒢0).\mathcal{Q}\to+\infty,\delta^{\mathfrak{B}}_{t}\to\alpha\cdot\textit{sign}(\mathcal{G}_{0}). (16)

Then, LLF will degenerate to the original attack method. To avoid this situation, a very simple but effective method can be applied, which is directly applying future penalty coefficient to 𝒢\vec{\mathcal{G}}. Then the ii-th coefficient equals βi\beta^{i}. To avoid ambiguity in the expression, this new coefficient is marked as 𝒞i=βi{\mathcal{C}^{\prime}}_{i}=\beta^{i}. Then the coefficients vector is 𝒞=(β1,β2,,β𝒬)\vec{\mathcal{C}^{\prime}}=(\beta^{1},\beta^{2},\dots,\beta^{\mathcal{Q}}). The perturbation update process can be described as

δt𝔅=αsign(𝒞,𝒢).\displaystyle\delta^{\mathfrak{B}}_{t}=\alpha\cdot\textit{sign}(\left\langle\vec{\mathcal{C}^{\prime}},\vec{\mathcal{G}}\right\rangle). (17)

At the same time, δ=i=0𝒬δi\delta=\sum_{i=0}^{\mathcal{Q}}\delta_{i} is just the approximation to the optimal point δ𝒪\delta^{\mathcal{O}} with 𝒬\mathcal{Q} steps. Greater 𝒬\mathcal{Q} refers to greater complexity. Therefore, LFF can set 𝒬\mathcal{Q} with a certain value that is not too great, which can combine the advantages of LFF and the low calculation complexity with the original iteration process. Especially, when 𝒬=1\mathcal{Q}=1, the LFF attack will be degenerated to the corresponding IFGSM-based method. Naturally, the iteration between each δt𝔅\delta^{\mathfrak{B}}_{t} can also be optimized with a similar gradient-based method, e.g. the momentum method. Then a momentum version of the LFF attack, which is named MLFF, can be described as:

gt𝔅=𝒞,𝒢,\displaystyle g^{\mathfrak{B}}_{t}=\left\langle\vec{\mathcal{C}^{\prime}},\vec{\mathcal{G}}\right\rangle, (18)
Mt𝔅=ηMt1𝔅+gt𝔅gt𝔅1,\displaystyle M^{\mathfrak{B}}_{t}=\eta\cdot M^{\mathfrak{B}}_{t-1}+\frac{g^{\mathfrak{B}}_{t}}{{||g^{\mathfrak{B}}_{t}||}_{1}},
δt𝔅=αsign(Mt𝔅),\displaystyle\delta^{\mathfrak{B}}_{t}=\alpha\cdot\textit{sign}(M^{\mathfrak{B}}_{t}),

where η\eta is the momentum decay factor.

IV-B Multi-order LFF (LFF𝒩\textit{LFF}^{\mathcal{N}}) Attack

The sequence of the δt𝔅\delta^{\mathfrak{B}}_{t} from the one-order LFF attack can still be enhanced with the LFF mechanism. With the same guidelines, the enhancing process can be iterated to 𝒩\mathcal{N}-order LFF, which is named LFF𝒩\textit{LFF}^{\mathcal{N}} attack. The description for the LFF𝒩\textit{LFF}^{\mathcal{N}} attack is:

δt𝔅ν=αsign(𝒞ν,𝒢ν),\delta^{\mathfrak{B}^{\mathcal{\nu}}}_{t}=\alpha\cdot\textit{sign}(\left\langle\vec{{\mathcal{C}^{\mathcal{\nu}}}^{\prime}},\vec{\mathcal{G}^{\mathcal{\nu}}}\right\rangle), (19)

where ν𝒩\nu\in\mathcal{N} is the indicator for the order,

𝒞νi\displaystyle{\mathcal{C}^{\mathcal{\nu}}}^{\prime}_{i} =βνi,\displaystyle={\beta^{\mathcal{\nu}}}^{i}, (20)
𝒢tν\displaystyle\mathcal{G}^{\mathcal{\nu}}_{t} =x+j=0t1δj𝔅ν+i=0q1δi𝔅ν1x+j=0t1δj𝔅ν+i=0q1δi𝔅ν11,\displaystyle=\frac{\nabla_{x+\sum_{j=0}^{t-1}\delta^{\mathfrak{B}^{\nu}}_{j}+\sum_{i=0}^{q-1}\delta^{\mathfrak{B}^{\nu-1}}_{i}}\mathcal{L}}{||\nabla_{x+\sum_{j=0}^{t-1}\delta^{\mathfrak{B}^{\nu}}_{j}+\sum_{i=0}^{q-1}\delta^{\mathfrak{B}^{\nu-1}}_{i}}\mathcal{L}||_{1}},

and δi𝔅0=δi\delta^{\mathfrak{B}^{0}}_{i}=\delta_{i}. Especially, when 𝒩=1\mathcal{N}=1, Eq.(19) will be degenerated to Eq.(17). In the LFF𝒩\textit{LFF}^{\mathcal{N}} attack, there are multi-iteration processes. Therefore, there are many ways to combine operations that can further optimize the optimization process. Here, only a simple extension for MLFF with applying momentum mechanism for any ν\nu-th order, where ν𝒩\nu\in\mathcal{N}. This method is named MLFF𝒩\textit{MLFF}^{\mathcal{N}}. The corresponding formulation is shown as follows:

gt𝔅ν=𝒞ν,𝒢ν,\displaystyle g^{\mathfrak{B}^{\mathcal{\nu}}}_{t}=\left\langle\vec{{\mathcal{C}^{\mathcal{\nu}}}^{\prime}},\vec{\mathcal{G}^{\mathcal{\nu}}}\right\rangle, (21)
Mt𝔅ν=ηνMt1𝔅ν+gt𝔅νgt𝔅ν1,\displaystyle M^{\mathfrak{B}^{\mathcal{\nu}}}_{t}=\eta_{\mathcal{\nu}}\cdot M^{\mathfrak{B}^{\mathcal{\nu}}}_{t-1}+\frac{g^{\mathfrak{B}^{\mathcal{\nu}}}_{t}}{{||g^{\mathfrak{B}^{\mathcal{\nu}}}_{t}||}_{1}},
δt𝔅ν=αsign(Mt𝔅ν),\displaystyle\delta^{\mathfrak{B}^{\mathcal{\nu}}}_{t}=\alpha\cdot\textit{sign}(M^{\mathfrak{B}^{\mathcal{\nu}}}_{t}),

where ην\eta_{\mathcal{\nu}} is the momentum penalty factor for ν\nu-th order.

V Experiment

V-A Experiment Settings

Because most SOTA methods are realized on MI-FGSM, we conduct the comparison experiment using the MLLF method with those SOTA methods.

V-A1 Classifcation Networks

We conduct a wide range of classification networks within the Timm which is one of the greatest deep neural network libraries with various pre-trained models. We apply both CNN structure networks and Transformer structure networks. All classification networks include ResNet-50, BiT-50, Inception-v3, Inception-ResNet-v2, ConvNeXT-B, ViT-B, Swin-B, Deit-B. ResNet is one of the most classic deep learning networks. BiT-50 is one of the latest structures for ResNet models, which represents the best performance of the ResNet series. Inception-v3 is the pure inception structure CNNs, and Inception-ResNet is the combination of ResNet and inception structure. ConvNeXT is the most latest CNN structure network which represents the best performance of CNNs. Vision Transformer(ViT) is the typical Transformer structure network for image classification. Swin and Deit are two of the latest Transformer structure models.

V-A2 Dataset

We conduct the ImageNet1k dataset for all experiments. ImageNet1k dataset contains 1000 classes from the ImageNet dataset. All classification networks used in the experiments from Timm are pre-trained with the ImageNet1k dataset. Specifically, we conduct the ILSVRC2012 dataset as the testing dataset. The ILSVRC2012 dataset is one of the most classic datasets for adversarial attacks. The ILSVRC2012 dataset comes from the testing data of the ImageNet1k dataset. It still is widely used for evaluating adversarial attack performance, especially the transfer attack.

V-A3 Baselines

We conduct four attack methods as the baselines for the comparison experiment. They are MI-FGSM, EMI-FGSM, Admix and SIA. MI-FGSM is one of the most classic methods as well as one of the most widely used methods. The other baselines are all realized based on MI-FGSM. EMI-FGSM is one of the most recent gradient-based methods. Admix can typically represent the method with the mixup strategy. SIA is one of the latest self-transformation data augmentation methods. SIA is also one of the methods to achieve the best attack performance.

Model Attack Res50 BiT50 IncRes-v2 ConvneXT-B ViT-B Swin-B Deit-B Inc-v3
Inc-v3 MI 41.4 20.6 43.9 20.6 9.3 15.1 11.3 100.0
MLFF-MI 47.0 23.4 55.5 24.4 9.7 16.2 12.5 100.0
Inc. +5.6 +2.8 +11.6 +3.8 +0.4 +1.1 +1.2 +0.0
EMI 71.1 43.6 81.0 42.0 16.9 33.0 22.4 100.0
MLFF-EMI 76.2 49.4 85.1 48.0 21.9 37.8 26.6 100.0
Inc. +5.1 +5.8 +4.1 +6.0 +5.0 +4.8 +4.2 +0.0
Admix 80.0 52.6 86.8 45.9 23.3 39.3 30.5 100.0
MLFF-Admix 82.7 53.6 87.5 46.2 23.6 39.6 30.1 100.0
Inc. +2.7 +1.0 +0.7 +0.3 +0.3 +0.3 -0.4 +0.0
SIA 91.9 72.3 95.8 62.6 32.2 54.5 36.0 100.0
MLFF-SIA 95.9 81.6 99.0 74.0 40.0 63.9 43.2 100.0
Inc. +4.0 +9.3 +3.2 +11.4 +7.8 +9.4 +6.8 +0.0
ViT-B MI 51.3 37.7 32.6 41.2 100.0 54.4 66.9 39.8
MLFF-MI 53.1 37.4 32.9 38.3 100.0 52.0 68.4 42.4
Inc. +1.8 -0.3 +0.3 -2.9 +0.0 -2.2 +1.5 +2.6
EMI 71.8 62.2 55.8 66.7 100.0 77.8 89.9 62.5
MLFF-EMI 76.2 69.7 60.4 70.2 100.0 82.5 94.2 67.0
Inc. +4.4 +7.5 +4.6 +3.5 +0.0 +4.7 +4.3 +4.5
Admix 68.1 61.2 51.6 63.6 99.9 76.0 87.4 58.5
MLFF-Admix 72.9 64.2 53.7 64.7 100.0 79.4 91.2 61.8
Inc. +4.8 +3.0 +2.1 +1.1 +0.1 +3.4 +3.8 +3.3
SIA 82.6 80.7 71.2 82.5 99.6 87.5 89.0 75.6
MLFF-SIA 91.1 90.8 83.8 91.9 100.0 94.5 96.2 86.9
Inc. +8.5 +10.1 +12.6 +9.4 +0.4 +7.0 +7.2 +11.3

TABLE I: The attack success rates (%) against eight networks by baseline attacks and our method. The Increment (Inc. %) to the corresponding baselines that are greater than zero are marked in bold.
Model Attack Inc-v3adv IncRes-v2adv ConvNeXT-B+FD Swin-B+FD ConvNeXT-B+Bit-Red Swin-B+Bit-Red
Inc-v3 MI 19.7 7.8 16.5 15.2 20.8 14.7
MLFF-MI 23.4 8.8 16.7 16.1 23.7 16.4
Inc. +3.7 +1.0 +0.2 +0.9 +2.9 +1.7
EMI 38.5 15.6 27.8 25.4 42.2 33.0
MLFF-EMI 43.3 19.4 31.0 28.8 46.7 37.9
Inc. +4.8 +3.8 +3.2 +3.4 +4.5 +4.9
Admix 55.2 27.2 36.9 33.0 45.6 39.5
MLFF-Admix 55.8 26.9 37.2 34.0 46.6 40.9
Inc. +0.6 -0.3 +0.3 +1.0 +1.0 +1.4
SIA 54.0 28.2 39.3 38.6 61.8 53.2
MLFF-SIA 68.6 35.1 42.3 42.5 73.7 65.1
Inc. +14.6 +6.9 +3.0 +3.9 +11.9 +11.9

TABLE II: The attack success rates (%) against six defensive models by baseline attacks and our method. The Increment (Inc. %) to the corresponding baselines that are greater than zero are marked in bold.

V-A4 Hyperparameters

Following many previous works, the perturbation constraint is set to L=16L_{\infty}=16, the quantity of the iteration is 1616, the update rate is α=1255\alpha=\frac{1}{255}. For the very basic method MI-FGSM, the momentum decay factor is 11. The number of gradient collections in EMI-FGSM is 1111 with a linear sample and the radius value is 77. For Admix, the number of scale copies is set to 55 and the number of mix images is set to 33 while the mix ratio is 0.20.2. In SIA, we set the splitting number to 33 and the number of transformed images for gradient calculation is 1010.

In the comparison experiment, to fairly compare the performance of LFF and other methods, 𝒬\mathcal{Q} is set to 44, which means a low calculation complexity and only looking from a very near future. Because when 𝒬\mathcal{Q} is small LFF will tend to I-FGSM even if the attack method applied is the MI-FGSM-based method, the MLFF method is applied to the comparison experiment. The future penalty factor β\beta is set to 11 and the momentum decay for the MLFF is 11.

V-A5 Evaluation metrics

Attack Success Rate (ASR) is applied as the main evaluation metric for the performance of attack methods. ASR can present how many images have been successfully attacked, no matter for the white-box network (surrogate network) or the black-box network (victim networks). In addition, because the LFF attack is applied to the existing attack methods, the increment of the ASR, which is marked as Inc., is also considered to present the increment of the performance compared with the original attack method. The greater ASR and Inc. are, the better.

V-B Comparison Experiments

The comparison experiment is applied within four scopes. Firstly we evaluate baselines and our proposed method under the condition of a single surrogate network with the untargeted attack task. Secondly, we apply the comparison on the defensive networks as well as defense methods applied to pre-trained networks. Thirdly, the targeted attack task is applied. Fourthly, ensemble networks are applied as the surrogate network. The adversarial example from ensemble networks is tested by both original networks and defensive networks as well as defense methods.

Model Attack Res50 BiT50 IncRes-v2 ConvNeXT-B ViT-B Swin-B Deit-B Inc-v3
Inc-v3 MI 0.3 0.4 0.5 0.4 0.0 0.1 0.1 100.0
MLFF-MI 0.6 0.4 0.9 0.2 0.1 0.0 0.1 100.0
Inc. +0.3 +0.0 +0.4 -0.2 +0.1 -0.1 +0.0 +0.0
EMI 2.7 1.2 5.9 1.4 0.1 0.7 0.2 97.4
MLFF-EMI 5.0 2.0 12.1 2.0 0.4 1.1 1.0 100.0
Inc. +2.3 +0.8 +6.2 +0.6 +0.3 +0.4 +0.8 +2.6
Admix 2.9 0.7 5.1 1.2 0.4 0.5 0.3 98.7
MLFF-Admix 4.1 1.7 8.7 1.7 0.5 0.8 0.5 99.8
Inc. +1.2 +1.08 +3.6 +0.5 +0.1 +0.3 +0.2 +1.1
SIA 7.0 3.8 15.9 3.9 1.0 2.6 1.2 74.5
MLFF-SIA 26.7 13.7 49.3 13.2 2.3 6.8 3.5 97.0
Inc. +19.7 +9.9 +33.4 +8.3 +1.3 +4.2 +2.3 +22.5

TABLE III: The attack success rates (%) against eight models by baseline attacks and our method for the targeted attack task. The Increment (Inc. %) to the corresponding baselines that are greater than zero are marked in bold.
Model Attack Res50 BiT50 IncRes-v2 ConvNeXT-B ViT-B Swin-B Deit-B Inc-v3
Em MI 100.0 68.9 72.4 70.4 98.5 71.0 98.9 99.7
MLFF-MI 100.0 75.4 79.5 73.7 100.0 73.8 99.8 100.0
Inc. +0.0 +6.5 +7.1 +3.3 +1.5 +2.8 +0.9 +0.3
EMI 100.0 92.1 94.7 89.5 99.6 90.5 99.6 100.0
MLFF-EMI 100.0 95.2 96.5 92.8 100.0 93.3 100.0 100.0
Inc. +0.0 +3.1 +1.8 +3.3 +0.4 +2.8 +0.4 +0.0
Admix 100.0 93.8 96.8 88.7 97.6 90.2 97.7 100.0
MLFF-Admix 100.0 96.2 98.7 93.1 99.7 93.8 99.8 100.0
Inc. +0.0 +2.4 +1.9 +4.4 +2.1 +3.6 +2.1 +0.0
SIA 100.0 98.1 98.8 97.3 98.9 97.3 99.4 99.8
MLFF-SIA 100.0 99.8 99.9 99.6 99.8 99.4 99.8 100.0
Inc. +0.0 +1.7 +1.1 +2.3 +0.9 +2.1 +0.4 +0.2

TABLE IV: The attack success rates (%) against eight models by baseline attacks and our method with ensembling Res50, Inc-v3, ViT-B and Deit-B. The Increment (Inc. %) to the corresponding baselines that are greater than zero are marked in bold.
Model Attack Inc-v3adv IncRes-v2adv ConvNeXT-B+FD Swin-B+FD ConvNeXT-B+Bit-Red Swin-B+Bit-Red
Em MI 37.4 25.0 45.6 52.1 69.5 70.5
MLFF-MI 40.1 22.5 47.3 53.3 73.8 73.7
Inc. +2.7 +2.5 +1.7 +1.2 +4.3 +3.2
EMI 69.0 48.2 73.3 76.6 89.3 90.2
MLFF-EMI 76.0 55.3 78.7 80.1 92.5 93.2
Inc. +7.0 +7.1 +5.4 +3.5 +3.2 +3.0
Admix 80.6 64.0 77.6 79.9 89.1 89.8
MLFF-Admix 85.1 69.0 82.2 84.4 93.5 94.1
Inc. 4.5 +5.0 +4.6 +4.5 +4.4 +4.3
SIA 86.4 72.3 85.3 86.7 97.4 97.3
MLFF-SIA 94.0 84.0 91.5 93.4 99.6 99.4
Inc. +7.6 +11.7 +6.2 +6.7 +2.2 +2.1

TABLE V: The attack success rates (%) against six defensive models by baseline attacks and our method with ensembling Res50, Inc-v3, ViT-B and Deit-B. The Increment (Inc. %) to the corresponding baselines that are greater than zero are marked in bold.
Refer to caption
Figure 1: The experimental results of changing 𝒬\mathcal{Q} in LFF-I, LFF-MI, MLFF-I, and MLFF-MI respectively.
Refer to caption
Figure 2: The experimental results of changing β\beta in MLFF-MI with 𝒬=4\mathcal{Q}=4 and 𝒬=10\mathcal{Q}=10 respectively.

V-B1 Untargeted attack with single surrogate network

In this comparison experiment, Inception-v3 and Vit-B are chosen to be the surrogate networks individually. Inception-v3 represents the adversarial examples from the CNN structure, while ViT represents the adversarial examples from the Transformer structure. All other networks are applied as the victim networks. The experimental results are shown in Table I.

Overall, the MLFF method can greatly improve the performance of all baselines. No matter whether the surrogate network is the CNN structure or the Transformer structure, the ASRs for the victim networks almost have increased. Especially, some of Inc. are over 10%10\%. From the details, there are three conditions the Inc. is the negative value. One occurs by applying Admix as the baseline with generating adversarial examples from Inception-v3 and testing for the Deit-B networks. However, when observing the Inc. for the MLFF Admix, it can be noted that MLFF seems to have not very great performance improvement on Admix. This might refer to the randomness in Admix. For each iteration, Admix randomly chooses images to mix up. Therefore, the process of the optimization itself is not fixed. Especially when this randomness is great, the randomness of the original sequence obtained in the LFF mechanism will be very great, thereby reducing the optimization effect. The performance of the MLFF-Admix in ViT-B can also verify this phenomenon. Another two occur by applying MI-FGSM as the baseline with generating adversarial examples from ViT-B and testing for ConvNeXT-B and Swin-B. This may be due to the small value of 𝒬\mathcal{Q} we set. The small 𝒬\mathcal{Q} will cause the MLFF to tend to the original attack, i.e. the MI-FGSM itself. At the same time, when the attack method tramps into the overfitting, MLFF will accelerate the overfitting. This all might be the reason that there are conditions that MLFF has negative Inc. values. It should also be noted that MLFF has very great Inc. values for both EMI and SIA methods. Although SIA itself has performed relatively well on ASR, MLFF-SIA can further enhance the adversarial transferability with a larger degree of improvement.

V-B2 Untargeted attack with defensive models

In this comparison experiment, two adversarial training networks are applied, i.e. Adversarial Inception-v3 and Ensemble Adversarial Inception-ResNet-v2. Two defensive methods FD and Bit-Red are applied to the ConvNext and Swin individually. The surrogate network is the Inception-v3 network. The experimental results are shown in Table II.

The experimental results are still corresponding to the previous experimental results. The increment to MI-FGSM and Admix method are relatively small, while the increment to EMI and SIA are very great.

V-B3 Targeted attack

In this comparison experiment, the targeted label for the attack is randomly chosen without repeating. The loss function for the targeted attack is the Cross-Entropy loss. The surrogate model is the Inception-v3 network. The experimental results are shown in Table III.

V-B4 Untargeted attack with ensemble networks

In this comparison experiment, the ensembled network is treated as the surrogate network for adversarial example generation. The ensembled network contains four networks, i.e. ResNet50, Inception-v3, ViT-B and Deit-B. This ensemble network contains both CNN structure networks and Transformer structure networks. The attack success rates for the original networks are placed in Table IV. The attack success rates for the defensive networks as well as defense methods are placed in Table V.

V-C Ablation Experiment

The ablation experiment is applied within 2 scopes. Firstly, we verify the effectiveness of the LFF as well as the MLFF with different 𝒬\mathcal{Q}. Then, we verify the influence of the future penalty factor β\beta.

V-C1 Influence of the looking from the future steps

In this ablation experiment, the quantity of the looking from the future 𝒬\mathcal{Q} is changed to verify the mechanism of the LLF. With greater 𝒬\mathcal{Q}, LFF attack as well as MLFF attack will have better attack performance. With smaller 𝒬\mathcal{Q}, the LFF attack will have very poor attack performance but the MLFF attack will keep a relatively stable performance. We apply the LFF attack and the MLFF attack to MI-FGSM and I-FGSM respectively. The experimental results are shown in the Fig. 1.

Overall, the experimental results can sufficiently verify the effect of the 𝒬\mathcal{Q}. The greater value of 𝒬\mathcal{Q} is, LFF attack as well as MLFF attack performances better. This tendency is especially obvious when the baseline is the MI-FGSM. However, this tendency seems to be weak for the MLFF-I attack. We speculate that this phenomenon is caused by the I-FGSM. Because I-FGSM easily falls into the local minima, which means low adversarial transferability, MLFF can only help I-FGSM quickly approach this local optimal point, although MLFF-I has a relatively good performance compared with the MI-FGSM.

V-C2 Influence of the future penalty factor

In this ablation experiment, the future penalty factor β\beta is changed. With greater β\beta, the LFF attack will contain more future information to generalize the perturbation. When β\beta is too great, the attack performance will drop due to losing attack capability. With smaller β\beta, the LFF attack will contain more local attack information. When β\beta is too small, the attack performance will degenerate to the original attack. Hence we apply β\beta from 0.10.1 to 1.91.9 with 0.10.1 intervel under the condition of 𝒬=4\mathcal{Q}=4 and 𝒬=10\mathcal{Q}=10 respectively. Because 𝒬=4\mathcal{Q}=4 is relatively small, the MLFF method is applied. The baseline is the MI-FGSM method. The experimental results are shown in the Fig. 2.

Overall, it can be noted that the trend in both subfigures is that ASR gradually increases with the increment of β\beta value. After reaching the peak, ASR gradually decreases as the β\beta value decreases. This phenomenon verifies the guess about the effect of the β\beta, i.e. the effect of the gradients in Eq. (17). A closer look at the horizontal coordinates of the peaks of the two graphs reveals that the optimal value of β\beta is different when facing different 𝒬\mathcal{Q}. When 𝒬\mathcal{Q} becomes greater, the best β\beta seems to become smaller. When comparing the two subfigures horizontally, it can be found that the average performance is significantly better when the 𝒬\mathcal{Q} is greater than when the 𝒬\mathcal{Q} is smaller.

VI Conclusion and Discussion

In this paper, we propose a novel concept which is named Looking From the Future. We extend the LFF concept into the adversarial attack task and propose the LFF attack as well as the MLFF attack while eliminating the disadvantages of focusing the local information and overfitting. Furthermore, we extend the LFF attack to the multi-order LFF attack, which is named LFF𝒩\text{LFF}^{\mathcal{N}}. Comparison experiments on four tasks as well as ablation experiments have conducted the performance of our proposed method. Experimental results clearly show that the LFF attack can greatly increase existing adversarial attack methods.

References

  • [1] Y. Dong, F. Liao, T. Pang, H. Su, J. Zhu, X. Hu, and J. Li. Boosting adversarial attacks with momentum. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 9185–9193, 2018.
  • [2] Y. Dong, T. Pang, H. Su, and J. Zhu. Evading defenses to transferable adversarial examples by translation-invariant attacks. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 4312–4321, 2019.
  • [3] Z. Ge, H. Liu, W. Xiaosen, F. Shang, and Y. Liu. Boosting adversarial transferability by achieving flat local maxima. Advances in Neural Information Processing Systems, 36:70141–70161, 2023.
  • [4] I. J. Goodfellow, J. Shlens, and C. Szegedy. Explaining and harnessing adversarial examples. arXiv preprint arXiv:1412.6572, 2014.
  • [5] A. Kurakin, I. J. Goodfellow, and S. Bengio. Adversarial examples in the physical world. In Artificial intelligence safety and security, pages 99–112. Chapman and Hall/CRC, 2018.
  • [6] M. Li, C. Deng, T. Li, J. Yan, X. Gao, and H. Huang. Towards transferable targeted attack. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020.
  • [7] J. Lin, C. Song, K. He, L. Wang, and J. E. Hopcroft. Nesterov accelerated gradient and scale invariance for adversarial attacks. arXiv preprint arXiv:1908.06281, 2019.
  • [8] W. Ma, Y. Li, X. Jia, and W. Xu. Transferable adversarial attack for both vision transformers and convolutional networks via momentum integrated gradients. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 4630–4639, 2023.
  • [9] J. Wang, Z. Chen, K. Jiang, D. Yang, L. Hong, P. Guo, H. Guo, and W. Zhang. Boosting the transferability of adversarial attacks with global momentum initialization. arXiv preprint arXiv:2211.11236, 2022.
  • [10] K. Wang, X. He, W. Wang, and X. Wang. Boosting adversarial transferability by block shuffle and rotation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 24336–24346, 2024.
  • [11] T. Wang, Z. Ying, Q. Li, et al. Boost adversarial transferability by uniform scale and mix mask method. arXiv preprint arXiv:2311.12051, 2023.
  • [12] X. Wang and K. He. Enhancing the transferability of adversarial attacks through variance tuning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 1924–1933, 2021.
  • [13] X. Wang, X. He, J. Wang, and K. He. Admix: Enhancing the transferability of adversarial attacks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 16158–16167, 2021.
  • [14] X. Wang, J. Lin, H. Hu, J. Wang, and K. He. Boosting adversarial transferability through enhanced momentum. arXiv preprint arXiv:2103.10609, 2021.
  • [15] X. Wang and Z. Yin. Rethinking mixup for improving the adversarial transferability. arXiv preprint arXiv:2311.17087, 2023.
  • [16] X. Wang, Z. Zhang, and J. Zhang. Structure invariant transformation for better adversarial transferability. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 4607–4619, 2023.
  • [17] C. Xie, Z. Zhang, Y. Zhou, S. Bai, J. Wang, Z. Ren, and A. L. Yuille. Improving transferability of adversarial examples with input diversity. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 2730–2739, 2019.
  • [18] B. Yang, H. Zhang, Z. Li, Y. Zhang, K. Xu, and J. Wang. Adversarial example generation with adabelief optimizer and crop invariance. Applied Intelligence, 53(2):2332–2347, 2023.
  • [19] C. Zhang, P. Benz, A. Karjauv, J. W. Cho, K. Zhang, and I. S. Kweon. Investigating top-k white-box and transferable black-box attack. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 15085–15094, 2022.
  • [20] J. Zhang, W. Qian, R. Nie, J. Cao, and D. Xu. Generate adversarial examples by adaptive moment iterative fast gradient sign method. Applied Intelligence, 53(1):1101–1114, 2023.
  • [21] Q. Zhang, Y. Zhang, Y. Shao, M. Liu, J. Li, J. Yuan, and R. Wang. Boosting adversarial attacks with nadam optimizer. Electronics, 12(6):1464, 2023.
  • [22] Z. Zhao, Z. Liu, and M. Larson. On success and simplicity: A second look at transferable targeted attacks. Advances in Neural Information Processing Systems, 34:6115–6128, 2021.