Making Adversarial Examples More Transferable and Indistinguishable
Abstract
Fast gradient sign attack series are popular methods that are used to generate adversarial examples. However, most of the approaches based on fast gradient sign attack series cannot balance the indistinguishability and transferability due to the limitations of the basic sign structure. To address this problem, we propose a method, called Adam Iterative Fast Gradient Tanh Method (AI-FGTM), to generate indistinguishable adversarial examples with high transferability. Besides, smaller kernels and dynamic step size are also applied to generate adversarial examples for further increasing the attack success rates. Extensive experiments on an ImageNet-compatible dataset show that our method generates more indistinguishable adversarial examples and achieves higher attack success rates without extra running time and resource. Our best transfer-based attack NI-TI-DI-AITM can fool six classic defense models with an average success rate of 89.3% and three advanced defense models with an average success rate of 82.7%, which are higher than the state-of-the-art gradient-based attacks. Additionally, our method can also reduce nearly 20% mean perturbation. We expect that our method will serve as a new baseline for generating adversarial examples with better transferability and indistinguishability.
Introduction
Despite the great success on many tasks, deep neural networks (DNNs) have been shown that they are vulnerable to adversarial examples (Goodfellow, Shlens, and Szegedy 2015; Szegedy et al. 2014), i.e., the inputs with imperceptible perturbations can cause the incorrect results of DNNs. Moreover, a tougher problem terms transferability (Liu et al. 2017; Moosavi-Dezfooli et al. 2017) that the adversarial examples crafted by a known DNN can also fool other unknown DNNs. Consequently, adversarial examples present severe threats to the real-world applications (Athalye et al. 2018; Eykholt et al. 2018; Kurakin, Goodfellow, and Bengio 2017a) and have motivated extensive research on defense methods (Madry et al. 2018; Liao et al. 2018; Guo et al. 2018; Raghunathan, Steinhardt, and Liang 2018; Wong and Kolter 2018; Pang, Du, and Zhu 2018; Samangouei, Kabkab, and Chellappa 2018). Foolbox (Rauber et al. 2020) roughly categorizes attack methods into three types: the gradient-based methods (Dong et al. 2018; Goodfellow, Shlens, and Szegedy 2015; Kurakin, Goodfellow, and Bengio 2017b), the score-based methods (Narodytska and Kasiviswanathan 2017), and the decision-based methods (Brendel, Rauber, and Bethge 2018; Chen, Jordan, and Wainwright 2020). In this paper, we focus on the gradient-based methods. Although the adversarial examples crafted by using the gradient-based methods satisfy the bound and continually achieve higher black-box success rates, these examples can be identified easily. In addition, the adversarial examples generated by the approaches based on the basic sign structure are limited. Taking TI-MI-FGSM (the combination of translation-invariant method (Dong et al. 2019) and momentum iterative fast gradient sign method (Dong et al. 2018)) as an example, the gradient processing steps, such as Gaussian blur, the gradient normalization, and the sign function, severely damage the gradient information. Additionally, the sign function also increases the perturbation size.


In this paper, we propose a method, called Adam Iterative Fast Gradient Tanh Method (AI-FGTM), which improves the indistinguishability and transferability of adversarial examples. It is known that the fast gradient sign attack series iteratively process gradient information with transformation, normalization, and the sign function. To preserve the gradient information as much as possible, AI-FGTM modifies the major gradient processing steps. Still take TI-MI-FGSM as an example, to avoid the loss of gradient information and generate imperceptible perturbations, we replace the momentum algorithm and the sign function with Adam (Kingma and Ba 2015) and the tanh function, respectively. Then, we employ dynamic step size and smaller filters in Gaussian blur. The overview of AI-FGTM is shown in Fig. 1, and the detailed process will be given in Methodology. Furthermore, combining the existing attack methods with AI-FGTM can get much smaller perturbations and deliver state-of-the-art success rates. Fig. 2 shows the comparison of different examples, where the adversarial examples are crafted by three combined attacks, namely, DIM (Xie et al. 2019), TI-DIM (Dong et al. 2019) and TI-DI-AITM (the combination of TI-DIM and our method).
In summary, we make the following contributions:
-
1.
Inspired by the limitations of the fast gradient sign series, we propose AI-FGTM in which the major gradient processing steps are improved to boost the indistinguishability and transferability of adversarial examples.
-
2.
We show that AI-FGTM integrated with other transfer-based attacks can obtain much smaller perturbations and larger losses than the current sign attack series.
-
3.
The empirical experiments prove that without extra running time and resource, our best attack fools six classic defense models with an average success rate of 89.3% and three advanced defense models with an average success rate of 82.7%, which are higher than the state-of-the-art gradient-based attacks.
Review of Existing Attack Methods
Problem definition
Let be a set of pre-trained classifiers, where denotes the white-box classifiers and represents the unknown classifiers. Given a clean example , it can be correctly classified to the ground-truth label by all pre-trained classifiers. It is possible to craft an adversarial example that satisfies by using the white-box classifiers, where could be , , , , and is the perturbation size. In this paper, we focus on non-targeted attack with . Note that, the adversarial example can mislead the white-box classifiers and the unknown classifiers simultaneously.
The gradient-based methods
Here, we introduce the family of the gradient-based methods.
Fast Gradient Sign Method (FGSM) (Goodfellow, Shlens, and Szegedy 2015) establishes the basic framework of the gradient-based methods. It efficiently crafts an adversarial example by using one-step update while maximizing the loss function of a given classifier as
(1) |
where computes the gradient of the loss function w.r.t. , is the sign function, and is the given scalar value that restricts the norm of the perturbation.
Basic Iterative Method (BIM) (Kurakin, Goodfellow, and Bengio 2017a) is the iterative version of FGSM that performs better in white-box attack but less effective in transfer-based attack. It iteratively updates the adversarial example with a small step size as
(2) |
where with denoting the number of iterations. performs per-pixel clipping as
(3) |
Momentum Iterative Fast Gradient Sign Method (MI-FGSM) (Dong et al. 2018) enhances the transferability of adversarial examples by incorporating momentum term into gradient process, given as
(4) |
(5) |
where denotes the accumulated gradient at iteration , and is the decay factor of .
Nesterov Iterative Method (NIM) (Lin et al. 2020) integrates an anticipatory update into MI-FGSM and further increases the transferability of adversarial examples. The update procedures are expressed as
(6) |
(7) |
(8) |
Scale-Invariant Method (SIM) (Lin et al. 2020) applies the scale copies of the input image to further improve the transferability. However, SIM requires much more running time and resource.
Diverse Input Method (DIM) (Xie et al. 2019) applies random resizing and padding to the adversarial examples with the probability at each iteration. DIM can be easily integrated into other gradient-based methods to further boost the transferability of adversarial examples. The transformation function is
(9) |
Translation-Invariant Method (TIM) (Dong et al. 2019) optimizes an adversarial example by an ensemble of translated examples as
(10) |
where denotes the translation function that respectively shifts input by and pixels along the two-dimensions. And TIM calculates the gradient of the loss function at a point , convolves the gradient with a pre-defined Gaussian filter and updates as
(11) |
(12) |
Note that, with limited running time and computing resources, the combination of NIM, TIM and DIM (NI-TI-DIM) is strong transfer-based attack method so far.
Methodology
Motivations
Based on the contradiction that the adversarial examples achieve high success rates but can be identified easily, our observations are shown as follows:
-
1.
The sign function in gradient-based methods has two main disadvantages. On the one hand, the sign function normalizes all the gradient values to 1, -1 or 0, and thus leads to the loss of gradient information. On the other hand, the sign function normalizes some small gradient values to 1 or -1, which increases the perturbation size. While the tanh function not only normalizes the large gradient values as the sign function, but also maintains the small gradient values as function . Therefore, the tanh function can replace the sign function and reduce the perturbation size.
-
2.
With iterations , the applications of Nesterov’s accelerated gradient (NAG) (Lin et al. 2020) and the momentum algorithm (Dong et al. 2018) in adversarial attacks demonstrate that we can migrate other methods to generate adversarial examples. Moreover, the gradient is normalized by the distance of itself before the momentum algorithm. Intuitively, due to the performance of traditional convergence algorithms, Adam can achieve larger losses than the momentum algorithm in such small number of iterations. Additionally, Adam can normalize the gradient with , where denotes the first moment vector, is the second moment vector and .
-
3.
Traditional convergence algorithms employ learning rate decay to improve the model performance. Existing gradient-based methods set stable step size . In intuition, we can improve the transferability with the step size change. Different from the traditional convergence algorithms, the attack methods with the -ball restriction aim to maximize the loss function of the target models. Hence, we use the increasing step size with .
-
4.
Dong et al. (Dong et al. 2019) show that Gaussian blur with a large kernel improves the transferability of adversarial examples. However, Gaussian blur with larger kernels leads to the loss of the gradient information. Using the modifications mentioned above, the gradient information is preserved and plays a more important role in generating adversarial examples. Consequently, we apply smaller kernels in Gaussian blur to avoid the loss of the gradient information.
Based on the above four observations, we propose AI-FGTM to craft the adversarial examples which are expected to be more transferable and indistinguishable.
AI-FGTM
Adam (Kingma and Ba 2015) uses the exponential moving averages of squared past gradients to mitigate the rapid decay of learning rate. Essentially, this algorithm limits the reliance of update to only the past few gradients by the following simple recursion:
(13) |
(14) |
(15) |
where denotes the first moment vector, represents the second moment vector, and are the exponential decay rates.
Abbreviation | Definition |
TI-DIM | The combination of MI-FGSM, TIM and DIM |
NI-TI-DIM | The combination of MI-FGSM, NIM, TIM and DIM |
SI-NI-TI-DIM | The combination of MI-FGSM, NIM, SIM, TIM and DIM |
TI-DI-AITM | The combination of AI-FGTM, TIM and DIM |
NI-TI-DI-AITM | The combination of AI-FGTM, NIM, TIM and DIM |
SI-NI-TI-DI-AITM | The combination of AI-FGTM, NIM, SIM, TIM and DIM |
Attack | Inc-v3 | Inc-v4 | IncRes-v2 | Res-v2-101 | Model ensemble |
TI-DIM | 172.8 | 261.2 | 277.8 | 234.0 | 767.5 |
NI-TI-DIM | 174.5 | 238.9 | 291.8 | 243.0 | 830.2 |
SI-NI-TI-DIM | 608.2 | 1086.3 | 1156.2 | 1096.2 | 3490.2 |
TI-DI-AITM | 170.6 | 258.5 | 280.4 | 239.3 | 762.7 |
NI-TI-DI-AITM | 173.5 | 253.7 | 288.1 | 242.1 | 770.1 |
SI-NI-TI-DI-AITM | 603.6 | 1103.9 | 1119.4 | 1123.1 | 3341.6 |
Due to the opposite optimization objectives, we apply Adam into adversarial attack with some modifications. Starting with , and , the first moment estimate and the second moment estimate are presented as follows:
(16) |
(17) |
where and denote the first moment factor and second moment factor, respectively. We replace the sign function with the tanh function and update as
(18) |
(19) |
where and are exponential decay rates, and denotes the scale factor. Specifically, is the increasing step size with . Then the tanh function reduces the perturbations of adversarial examples without any success rate reduction. Furthermore, replaces the normalization and the first moment estimate of Eq. 4 due to the fact that Adam has faster divergence speed than momentum attack algorithm (as shown in Fig. 1(b)).
The combination of AI-FGTM and NIM
Inspired by NIM that integrates an anticipatory update into MI-FGSM. Similarly, we can also integrate an anticipatory update into AI-FGTM. We first calculate the step size in each iteration as Eq. 18, and the Nesterov term can be expressed as
(20) |
(21) |
(22) |
(23) |
We summarize NI-TI-DI-AITM as the combination of AI-FGTM, NIM, TIM and DIM, and the procedure is given in Algorithm 1.
Input:
A clean example and its ground-truth label ;
Parameters:
The perturbation size ; the iteration number ; the decay factors and ; the exponential decay rates and ; the scale factor ; the probability .
Output:
An adversarial example .
Experiments
In this section, we provide extensive experimental results on an ImageNet-compatible dataset to validate our method. First, we introduce the experimental setting. Then, we compare the running efficiency of different transfer-based attacks. Next, we present the ablation study of the effects of different parts of our method. Finally, we compare the results of the baseline attacks. Table 1 presents the definitions of the abbreviations used in this paper.
Experimental setting
Dataset. We utilize 1000 images 111https://github.com/tensorflow/cleverhans/tree/master/examples/nips17˙adversarial˙competition/dataset which are used in the NIPS 2017 adversarial competition to conduct the following experiments.
Models. In this paper, we employ thirteen models to perform the following experiments. Four non-defense models (Inception v3 (Inc-v3) (Szegedy et al. 2016), Inception v4 (Inc-v4), Inception ResNet v2 (IncRes-v2) (Szegedy et al. 2017), and ResNet v2-101 (Res-v2-101) (He et al. 2016)) are used as white-box models to craft adversarial examples. Six defense models (Inc-v3ens3, Inc-v3ens4, IncResv2ens (Tramèr et al. 2018), high-level representation guided denoiser (HGD) (Liao et al. 2018), input transformation through random resizing and padding (R&P) (Xie et al. 2018), and rank-3 submission 222https://github.com/anlthms/nips-2017/tree/master/mmd in the NIPS 2017 adversarial competition) are employed as classic models to evaluate the crafted adversarial examples. In addition, we also evaluate the attacks with three advanced defenses (Feature Distillation (Liu et al. 2019), Comdefend (Jia et al. 2019), and Randomized Smoothing (Cohen, Rosenfeld, and Kolter 2019)).
Baselines. We focus on the comparison of TI-DIM, NI-TI-DIM, TI-DI-AITM and NI-TI-DI-AITM, where TI-DIM and NI-TI-DIM are both the state-of-the-art methods.
Hyper-parameters. According to TI-DIM (Dong et al. 2019) and NI-FGSM (Lin et al. 2020), we set the maximum perturbation , and the number of iteration . Specifically, we set the kernel size to in normal TI-DIM and NI-TI-DIM while in TI-DI-AITM. The exploration of appropriate settings of our method is illustrated in Appendix.
The mean perturbation size. For an adversarial example with the size of , the mean perturbation size can be calculated as
(24) |
where denotes the value of channel of the image at coordinates .
Attack | tanh | Adam | smaller kernels | dynamic step size | mean success rate (%) | mean perturbation |
TI-DIM | 82.0 | 10.46 | ||||
✓ | 82.4 | 9.14 | ||||
✓ | 83.6 | 9.20 | ||||
✓ | ✓ | 83.1 | 7.86 | |||
✓ | ✓ | ✓ | 86.5 | 7.82 | ||
✓ | ✓ | ✓ | ✓ | 88.0 | 8.11 |
Attack | Inc-v3ens3 | Inc-v3ens4 | IncRes-v2ens | HGD | R&P | NIPS-r3 | |||||||||||||||||||||||||||||
Inc-v3 |
|
|
|
|
|
|
|
||||||||||||||||||||||||||||
Inc-v4 |
|
|
|
|
|
|
|
||||||||||||||||||||||||||||
IncRes-v2 |
|
|
|
|
|
|
|
||||||||||||||||||||||||||||
Res-v2-101 |
|
|
|
|
|
|
|

Attack | Inc-v3ens3 | Inc-v3ens4 | IncRes-v2ens | HGD | R&P | NIPS-r3 | Average |
TI-DIM | 83.9 | 83.2 | 78.4 | 81.9 | 81.2 | 83.6 | 82.0 |
TI-DI-AITM | 90.2 | 88.5 | 85.4 | 88.3 | 87.1 | 88.7 | 88.0 |
NI-TI-DIM | 85.5 | 85.9 | 80.1 | 83.6 | 82.9 | 84.3 | 83.7 |
NI-TI-DI-AITM | 91.8 | 90.3 | 85.8 | 89.4 | 88.6 | 90.1 | 89.3 |
Attack | Feature Distillation | Comdefend | Randomized Smoothing | Average |
TI-DIM | 83.1 | 78.2 | 49.9 | 70.4 |
TI-DI-AITM | 90.6 | 87.9 | 63.7 | 80.7 |
NI-TI-DIM | 82.1 | 84.7 | 58.6 | 75.1 |
NI-TI-DI-AITM | 91.4 | 90.3 | 66.4 | 82.7 |
The comparison of running efficiency
We compare the running time of each attack mentioned in Table 1 using a piece of Nvidia GPU GTX 1080 Ti. Table 2 shows the running time under single-model setting and model ensemble setting. It can be seen that attacks combined with our method AI-FGTM do not cost extra running. Additionally, SIM requires at least two pieces of GPUs under model ensemble setting and costs much more running time than other attacks under both single-model setting and model ensemble setting. Therefore, we exclude SIM in the following experiments.
Ablation study
Table 3 shows the ablation study of the effects of different parts of our method. We compare the mean perturbation and the mean success rates of the adversarial examples against six classic defense models. Our observations are shown as follow:
-
1.
Both of the tanh function and Adam can reduce the perturbation size. Additionally, Adam can also improve the transferability of adversarial examples.
-
2.
The combination of the tanh function and Adam can greatly reduce the perturbation size, but only slightly improve the transferability of adversarial examples.
-
3.
Using smaller kernels and dynamic step size can improve the transferability of adversarial examples even using dynamic step size slightly increase the perturbation size.
The validation results in the single-model attack scenario
In this section, we compare the success rates of AI-FGTM based attacks and the baseline attacks against six classic defenses. We generate adversarial attacks for Inc-v3, Incv4, IncRes-v2, and Res-v2-101 by severally using TI-DIM, TI-DI-AITM, NI-TI-DIM and NI-TI-DI-AITM.
The validation results in the model ensemble attack scenario
In this section, we present the success rates of adversarial examples generated for an ensemble of four non-defense models. Table 5 gives the results of transfer-based attacks against six classic defense models. It shows that our methods achieve higher success rates than baseline attacks. In particular, without extra running time and resource, TI-DI-AITM and NI-TI-DI-AITM fool six defense models with an average success rate of 88.0% and 89.3%, respectively, which are higher than the state-of-the-art gradient-based attacks.
We also validate our method by comparing the different results between DIM, TI-DIM and TI-DI-AI-FGTM in Fig. 5. Adversarial examples are generated for the ensemble of Inc-v3, Inc-v4, IncRes-v2 and Res-v2-101 using different attack methods. Fig. 5 (a) shows that the tanh function does not hurt the performance of adversarial examples and Adam can boost the attack success rates. Fig. 5 (b) shows that our method significantly reduce the mean perturbation size of adversarial examples. In particular, our method reduces 40% perturbation while delivering the stable performance. Fig. 5 (c) shows that our approach with obtains the largest loss of all the methods.
We evaluate the attacks with three more advanced defenses, namely Feature Distillation (Liu et al. 2019), Comdefend (Jia et al. 2019) and Randomized Smoothing (Cohen, Rosenfeld, and Kolter 2019). Table 6 shows the success rates of TI-DIM, TI-DI-AITM, NI-TI-DIM and NI-TI-DI-AITM against these defenses in the ensemble attack scenario.
In Table 6, we find that the attacks with AI-FGTM consistently outperform the attacks with MI-FGSM. In general, our methods can fool these defenses with high success rates.
Based on the above experimental results, it is reasonable to state that the proposed TI-DI-AITM and NI-TI-DI-AITM can generate adversarial examples with much better indistinguishability and transferability. Meanwhile, TI-DI-AITM and NI-TI-DI-AITM raise the security challenge for the development of more effective defense models.
Conclusion
In this paper, we propose AI-FGTM to craft adversarial examples that are much indistinguishable and transferable. AI-FGTM modifies the major gradient processing steps of the basic sign structure to address the limitations faced by the existing basic sign involved methods. Compared with the state-of-the-art attacks, extensive experiments on an ImageNet-compatible dataset show that our method generates more indistinguishable adversarial examples and achieves higher attack success rates without extra running time and resource. Our best attack NI-TI-DI-AITM can fool six classic defense models with an average success rate of 89.3% and fool three advanced defense models with an average success rate of 82.7%, which are higher than the state-of-the-art gradient-based attacks. Additionally, our method reduces nearly 20% mean perturbation. It is expected that our method serves as a new baseline for generating adversarial examples with higher transferability and indistinguishability.
Acknowledgements
This work was supported in part by National Natural Science Foundation of China under Grant (62106281). This paper was finished with the encouragement of Zou’s wife Maojuan Tian. Zou would like thank her and tell her:‘ the most romantic thing I can imagine is gradually getting old with you in scientific exploration.’
References
- Athalye et al. (2018) Athalye, A.; Engstrom, L.; Ilyas, A.; and Kwok, K. 2018. Synthesizing Robust Adversarial Examples. In Proceedings of the 35th International Conference on Machine Learning, 284–293.
- Brendel, Rauber, and Bethge (2018) Brendel, W.; Rauber, J.; and Bethge, M. 2018. Decision-Based Adversarial Attacks: Reliable Attacks Against Black-Box Machine Learning Models. In 6th International Conference on Learning Representations.
- Chen, Jordan, and Wainwright (2020) Chen, J.; Jordan, M. I.; and Wainwright, M. J. 2020. HopSkipJumpAttack: A Query-Efficient Decision-Based Attack. In 2020 IEEE Symposium on Security and Privacy, SP 2020, San Francisco, CA, USA, May 18-21, 2020, 1277–1294. IEEE.
- Cohen, Rosenfeld, and Kolter (2019) Cohen, J. M.; Rosenfeld, E.; and Kolter, J. Z. 2019. Certified Adversarial Robustness via Randomized Smoothing. In Proceedings of the 36th International Conference on Machine Learning, 1310–1320.
- Dong et al. (2018) Dong, Y.; Liao, F.; Pang, T.; Su, H.; Zhu, J.; Hu, X.; and Li, J. 2018. Boosting Adversarial Attacks With Momentum. In 2018 IEEE Conference on Computer Vision and Pattern Recognition, 9185–9193.
- Dong et al. (2019) Dong, Y.; Pang, T.; Su, H.; and Zhu, J. 2019. Evading Defenses to Transferable Adversarial Examples by Translation-Invariant Attacks. In IEEE Conference on Computer Vision and Pattern Recognition, 4312–4321.
- Eykholt et al. (2018) Eykholt, K.; Evtimov, I.; Fernandes, E.; Li, B.; Rahmati, A.; Xiao, C.; Prakash, A.; Kohno, T.; and Song, D. 2018. Robust Physical-World Attacks on Deep Learning Visual Classification. In 2018 IEEE Conference on Computer Vision and Pattern Recognition, 1625–1634.
- Goodfellow, Shlens, and Szegedy (2015) Goodfellow, I. J.; Shlens, J.; and Szegedy, C. 2015. Explaining and Harnessing Adversarial Examples. In 3rd International Conference on Learning Representations.
- Guo et al. (2018) Guo, C.; Rana, M.; Cissé, M.; and van der Maaten, L. 2018. Countering Adversarial Images using Input Transformations. In 6th International Conference on Learning Representations.
- He et al. (2016) He, K.; Zhang, X.; Ren, S.; and Sun, J. 2016. Identity Mappings in Deep Residual Networks. In Computer Vision - ECCV 2016 - 14th European Conference, 630–645.
- Jia et al. (2019) Jia, X.; Wei, X.; Cao, X.; and Foroosh, H. 2019. ComDefend: An Efficient Image Compression Model to Defend Adversarial Examples. In IEEE Conference on Computer Vision and Pattern Recognition, 6084–6092.
- Kingma and Ba (2015) Kingma, D. P.; and Ba, J. 2015. Adam: A Method for Stochastic Optimization. In 3rd International Conference on Learning Representations.
- Kurakin, Goodfellow, and Bengio (2017a) Kurakin, A.; Goodfellow, I. J.; and Bengio, S. 2017a. Adversarial examples in the physical world. In 5th International Conference on Learning Representations.
- Kurakin, Goodfellow, and Bengio (2017b) Kurakin, A.; Goodfellow, I. J.; and Bengio, S. 2017b. Adversarial Machine Learning at Scale. In 5th International Conference on Learning Representations.
- Liao et al. (2018) Liao, F.; Liang, M.; Dong, Y.; Pang, T.; Hu, X.; and Zhu, J. 2018. Defense Against Adversarial Attacks Using High-Level Representation Guided Denoiser. In 2018 IEEE Conference on Computer Vision and Pattern Recognition, 1778–1787.
- Lin et al. (2020) Lin, J.; Song, C.; He, K.; Wang, L.; and Hopcroft, J. E. 2020. Nesterov Accelerated Gradient and Scale Invariance for Adversarial Attacks. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. OpenReview.net.
- Liu et al. (2017) Liu, Y.; Chen, X.; Liu, C.; and Song, D. 2017. Delving into Transferable Adversarial Examples and Black-box Attacks. In 5th International Conference on Learning Representations.
- Liu et al. (2019) Liu, Z.; Liu, Q.; Liu, T.; Xu, N.; Lin, X.; Wang, Y.; and Wen, W. 2019. Feature Distillation: DNN-Oriented JPEG Compression Against Adversarial Examples. In IEEE Conference on Computer Vision and Pattern Recognition, 860–868.
- Madry et al. (2018) Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; and Vladu, A. 2018. Towards Deep Learning Models Resistant to Adversarial Attacks. In 6th International Conference on Learning Representations.
- Moosavi-Dezfooli et al. (2017) Moosavi-Dezfooli, S.; Fawzi, A.; Fawzi, O.; and Frossard, P. 2017. Universal Adversarial Perturbations. In 2017 IEEE Conference on Computer Vision and Pattern Recognition, 86–94.
- Narodytska and Kasiviswanathan (2017) Narodytska, N.; and Kasiviswanathan, S. P. 2017. Simple Black-Box Adversarial Attacks on Deep Neural Networks. In 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops, CVPR Workshops 2017, Honolulu, HI, USA, July 21-26, 2017, 1310–1318. IEEE Computer Society.
- Pang, Du, and Zhu (2018) Pang, T.; Du, C.; and Zhu, J. 2018. Max-Mahalanobis Linear Discriminant Analysis Networks. In Proceedings of the 35th International Conference on Machine Learning, 4013–4022.
- Raghunathan, Steinhardt, and Liang (2018) Raghunathan, A.; Steinhardt, J.; and Liang, P. 2018. Certified Defenses against Adversarial Examples. In 6th International Conference on Learning Representations.
- Rauber et al. (2020) Rauber, J.; Zimmermann, R. S.; Bethge, M.; and Brendel, W. 2020. Foolbox Native: Fast adversarial attacks to benchmark the robustness of machine learning models in PyTorch, TensorFlow, and JAX. J. Open Source Softw., 5(53): 2607.
- Samangouei, Kabkab, and Chellappa (2018) Samangouei, P.; Kabkab, M.; and Chellappa, R. 2018. Defense-GAN: Protecting Classifiers Against Adversarial Attacks Using Generative Models. In 6th International Conference on Learning Representations.
- Szegedy et al. (2017) Szegedy, C.; Ioffe, S.; Vanhoucke, V.; and Alemi, A. A. 2017. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, 4278–4284.
- Szegedy et al. (2016) Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; and Wojna, Z. 2016. Rethinking the Inception Architecture for Computer Vision. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, 2818–2826.
- Szegedy et al. (2014) Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I. J.; and Fergus, R. 2014. Intriguing properties of neural networks. In 2nd International Conference on Learning Representations.
- Tramèr et al. (2018) Tramèr, F.; Kurakin, A.; Papernot, N.; Goodfellow, I. J.; Boneh, D.; and McDaniel, P. D. 2018. Ensemble Adversarial Training: Attacks and Defenses. In 6th International Conference on Learning Representations.
- Wong and Kolter (2018) Wong, E.; and Kolter, J. Z. 2018. Provable Defenses against Adversarial Examples via the Convex Outer Adversarial Polytope. In Proceedings of the 35th International Conference on Machine Learning, 5283–5292.
- Xie et al. (2018) Xie, C.; Wang, J.; Zhang, Z.; Ren, Z.; and Yuille, A. L. 2018. Mitigating Adversarial Effects Through Randomization. In 6th International Conference on Learning Representations.
- Xie et al. (2019) Xie, C.; Zhang, Z.; Zhou, Y.; Bai, S.; Wang, J.; Ren, Z.; and Yuille, A. L. 2019. Improving Transferability of Adversarial Examples With Input Diversity. In IEEE Conference on Computer Vision and Pattern Recognition, 2730–2739.
Appendix
In this supplementary material, we provide more results in our experiments. In Sec. A, we investigate the effects of different hyper-parameters of AI-FGTM. In Sec. B, we report the success rates in white-box attack to show the effectiveness of AI-FGTM. In Sec. C, we present the visual examples generated by different attacks to show the better indistinguishability of the adversarial examples generated by our methods. In Sec. D, we finally present the value distributions of the accumulated gradients across iterations to demonstrate the limitation of the basic sign attack series. Codes are also available in the supplementary material. Codes are available at https://github.com/278287847/AI-FGTM.
A. The indistinguishability with PSNR and SSIM
We evaluate the indistinguishability with PSNR and SSIM (higher value indicates better indistinguishability). In detail, We generate adversarial examples for inc-v3 with TI-DIM, TI-DI-AITM, NI-TI-DIM and NI-TI-DI-AITM, and get the mean PSNR and SSIM between the adversarial examples and the clean examples. The results are shown in Table 7. We can find that the attacks with AI-FGTM can always perform better.
Attack | PSNR | SSIM | |||||||
|
|
|
|||||||
|
|
|
B. The effects of different hyper-parameters
We explore the effects of different hyper-parameters of AI-FGTM and aim to find the appropriate settings to balance the success rates of both white-box and black-box attacks. The adversarial examples are generated for the ensemble of Inc-v3, Inc-v4, IncRes-v2, and Res-v2-101 using TI-DI-AITM. We first show the results of white-box attacks against four known models, and then we present the performance of black-box attacks against three defense models in Fig. 4. It can be seen that the appropriate settings are , , , , , and the kernel length is 9.

C. The effectiveness in the white-box case
In order to demonstrate the effectiveness of our method, we report the success rates in white-box attack. We present the validation results in single-model attack scenario and model ensemble attack scenario. Table 8 presents the success rates of TI-DIM, TI-DI-AITM, NI-TI-DIM, and NI-TI-DI-AITM against white-box models. The details of the abbreviations are stated more precisely in Table 1 of our submitted paper.
As shown in Table 8, we can find that NI-TI-DI-AITM consistently outperform NI-TI-DIM in the single-model attack scenario. Compared with MI-FGSM, experimental results show that our method can improve the performance of NIM. Fig. 2 of our submitted paper also demonstrate that our method can generate adversarial examples with better indistinguishability.
Attack | Inc-v3ens3 | Inc-v3ens4 | IncRes-v2ens | HGD | R&P | NIPS-r3 | |||||||||||||||
Inc-v3 |
|
|
|
|
|
|
|
||||||||||||||
Inc-v4 |
|
|
|
|
|
|
|
||||||||||||||
IncRes-v2 |
|
|
|
|
|
|
|
||||||||||||||
Res-v2-101 |
|
|
|
|
|
|
|
D. Visualization of adversarial examples
We visualize six groups of adversarial examples generated by six different attacks in Fig 5. The adversarial examples are crafted on the ensemble models, including Inc-v3, Inc-v4, IncRes-v2 and Res-101, using TI-DIM, TI-DI-AITM (=1.3), TI-DI-AITM (=0.65), NI-TI-DIM, NI-TI-DI-AITM (=1.3) and NI-TI-DI-AITM (=0.65). We see that adversarial examples generates by our method have the better indistinguishability.

E. The value distributions of the accumulated gradients across iterations
We present the value distributions of the accumulated gradients across iterations to demonstrate the limitation of the basic sign attack series. With iteration number , we present the value distributions of the accumulated gradients in each iteration in Fig 6. As the number of iterations increases, values of the accumulated gradients tend to be greater than 1 or less than -1. However, a large number of values are in the range of in the first three iterations. With constant step size , perturbation size of adversarial examples is greatly enlarged. Hence, we replace the sign function with the tanh function and gradually increase the step size.
