WSAM: Visual Explanations from Style Augmentation as Adversarial Attacker and Their Influence in Image Classification
Abstract
Currently, style augmentation is capturing attention due to convolutional neural networks (CNN) being strongly biased toward recognizing textures rather than shapes. Most existing styling methods either perform a low-fidelity style transfer or a weak style representation in the embedding vector. This paper outlines a style augmentation algorithm using stochastic-based sampling with noise addition to improving randomization on a general linear transformation for style transfer. With our augmentation strategy, all models not only present incredible robustness against image stylizing but also outperform all previous methods and surpass the state-of-the-art performance for the STL-10 dataset. In addition, we present an analysis of the model interpretations under different style variations. At the same time, we compare comprehensive experiments demonstrating the performance when applied to deep neural architectures in training settings.
1 INTRODUCTION
Currently, deep learning neural nets require a large amount of data, usually annotated, to increase the generalization and obtain high performance. To deal with this problem, methods for artificial data generation are performed to increase the training samples; this common learning strategy is called data augmentation. In computer vision, data augmentation increases the number of images through pixel-level processing and transformations. For supervised tasks where labels are known, these operations perform label-preserving transformations controlled by the probability of applying the operation and usually a magnitude that intensifies the operation effects on the image [Szegedy et al., 2016, Tanaka and Aranha, 2019]. More recently, random erasing [DeVries and Taylor, 2017] and GAN-based augmentation [Tanaka and Aranha, 2019] improved the previous accuracy. In contrast, recent advances in style transfer [Ghiasi et al., 2017, Jackson et al., 2018] lead us to think about the influence of applying random styling and what deep networks learn from this.
Style augmentation is a technique that generates variations from an original set of images changing only the style information and keeping the main content. The style transformation applied to the image changes the image’s pixel information, generating a new diverse set of samples that follow the same original distribution. In contrast, content information remains equal [Ghiasi et al., 2017]. However, original style transfer techniques started with heavy computation to generate one stylized image. Experimentally, augmenting the training set randomly shows a new level of stochastic behavior, avoids overfitting in a small dataset, and stabilizes performance on large ones [Zheng et al., 2019]. Nowadays, some can work close to real-time performance while others can generate a batch of styles per image [Ghiasi et al., 2017, Jackson et al., 2018].
In Interpretable Machine Learning (IML), specifically in image-based models such as CNN, several methods exist to interpret and explain predictions. Usually, large and complex models like CNN are called “black-box” due to their vast number of parameters (hidden layers). So, to know the information shared through each layer, some methods were developed using information from layers and gradients such as Saliency Maps [Simonyan et al., 2013], and CAM-based methods [Zhou et al., 2016, Selvaraju et al., 2017]. These methods help explain complex “black box” image-based models and identify essential features in each sample prediction. In our approach, we will use these model explainers to highlight regions inside the input images to provide a visual interpretation of them.
In this work, we propose an augmentation strategy based on traditional augmentation plus style transformations. Besides, we implement new methods to visualize, explain, and interpret the behavior of our trained models. Also, we can understand which features are activating based on the style augmentation selected and study the influence of that style. Our main contributions in the present work are summarized as follows:
-
•
We give an explanation of the successful augmentation strategy based on interpretation methods.
-
•
We propose a Style Activation Map (SAM), Weighted Style Activation Map (WSAM), and WSAM Variance to visualize and understand the influence of style augmentation.
-
•
We outperform previous results on the STL-10 dataset using traditional and style augmentations.
2 RELATED WORKS
2.1 Style Transfer
In the first neural algorithm [Gatys et al., 2015], a content image and a style image are inputted to the neural network to obtain an output image with the original content but a new style. [Jing et al., 2017] employed the Gram matrices to model textures by encoding the correlations between convolutional features from different layers. Previous style transfer works [Ulyanov et al., 2017] improve the visual fidelity in which semantic structure was preserved for images with higher resolution. In [Geirhos et al., 2018] was concluded that neural networks have a strong bias with texture. Although the initial developments generated exciting results compared to the pioneer method, drawbacks such as weak texture synthesis and high computational cost were present [Ulyanov et al., 2017, Jing et al., 2017]. More recently, [Li et al., 2018, Ghiasi et al., 2017] solved the problem by relying on arbitrary styles without retraining the neural model. Also, other techniques adjusted a new parameter or inserted noise carefully to generate more style variations from one style input [Ghiasi et al., 2017, Kotovenko, Dmytro, adn Sanakoyeu, Artsiom, and Lang, Sabine, and Ommer, 2019]. Using these latter strategies, the first augmentation employing successfully style augmentation performing a cross-domain classification task [Jackson et al., 2018] follows the methodology adopted on [Ghiasi et al., 2017], which uses an Inception-v3 [Szegedy et al., 2015] architecture for the encoder and residual blocks for the decoder networks. However, the latent space is modified by a multivariate normal distribution which changes the style embedding. Other contemporary approaches [Zheng et al., 2019, Georgievski, 2019] used style augmentation and reported exciting results in classification tasks, specifically STL-10, CIFAR-100, and Tiny-ImageNet-200 datasets. Other interesting applications are extended to segmentation tasks [Hesse et al., 2019, Gkitsas et al., 2019].
Based on this literature review, we used a neural transfer model following a trade-off between edge preservation, flexibility to generate style variations, time processing, and best visual fidelity under different styles. We also compare our methodology to prior approaches used for style augmentation.
2.2 Deep Network Explanations
Explaining a CNN is focused on analyzing the information passed through each layer inside the network. Following this idea, several methods were proposed to visualize and obtain a notion about which features of a deep CNN were activated in one specific layer. In [Simonyan et al., 2013] (saliency maps) showed the convolutional activations, [Zeiler and Fergus, 2014] showed the impact of applying occlusion to the input image. In other methods, they use the gradients to visualize features and explain deep CNN networks such as DeepLIFT [Shrikumar et al., 2017], which computes scores for each feature; Integrated Gradients [Sundararajan et al., 2017], which computes features based on gradients; CAM [Zhou et al., 2016], and Grad-CAM [Selvaraju et al., 2017] which computes relevant regions using gradient and feature maps. Each method identifies features with high and strong activation representing the prediction for a specific predicted category.
Guided by this literature review, we propose a new method called Style Activation Maps (SAM) based on the Grad-CAM method applied to style augmentation. We choose this one due to better behavior and performance against adversarial attacks or noise-adding techniques [Adebayo et al., 2018, Gilpin et al., 2018]. Our main goal is to understand and interpret the impact of applying style augmentation in classification tasks and analyze their influence.
3 PROPOSED METHOD
In this section, we present theoretical formulation and some interpretation methods used.
3.1 Style Augmentation
For our experiments, we used the same methodology as [Jackson et al., 2018]; we nevertheless used a faster VGG-based network and added noise to diversify the style features. Specifically, we used an architecture composed of a generalized form of a linear transformation [Li et al., 2018]. Also, we compare with other related works [Jackson et al., 2018, Zheng et al., 2019] that use neural style augmentation.
Formally, let be the content image set and let be the precomputed style embedding set from , are used to feed the styling algorithm to generate the output set . Moreover, we denote zero-mean vectors and . Our style strategy transfers elements from the style set to a specific element from the content set .
The VGG (“r41”) architecture, denoted as , maps . and a non-linear function maps , where , and . Also, we denote , as the compress and uncompress CNN-based networks from the original paper [Li et al., 2018]. embeds the input image to an embedding vector that contains the semantic information of the image. More concisely, we use this non-linear function to map the original image to an embedding vector as shown in Eq. 1 for the content image and Eq. 2 for the style image. In our implementation, the function employs a CNN whose output is used to compute the covariance matrix and feed it to a fully-connected layer.
Since we use an architecture based on linear transformations, which is generalized from previous approaches [Ghiasi et al., 2017], the transformation matrix sets and preserves the feature affinity of the content image (determined by the covariance matrix of the content and the style). This is expressed in Eq. 3. In our implementation, we precomputed the style vectors and saved all textures in memory; thereby, our modifications are described in Eq. 4 and 5.
(1) | |||
(2) | |||
(3) |
In our implementation, we precompute the style vector and save all textures in memory; thereby, our modifications are expressed in Eq. 4 and 5.
(4) |
(5) |
Where is the interpolation hyper-parameter which controls the strength of the style transfer similarly to [Jackson et al., 2018], and , defined in Eq. 6, is the embedding vector of the style set with a noise addition for style randomization.
(6) |
As argued in prior methodologies, minor variations increase the randomization in the process; thereby, we apply noise instead of using a sampling strategy similar to applying a Gaussian noise in the latent space of generative networks during the training. In particular, we set this noise source as a multivariate normal distribution which means covariance scales and shifts into the embedding space. This is also useful for understanding the randomization process and the influence of the latent space.
3.2 Model Interpretation
In this work, we propose a new method Style Activation Map based on Grad-CAM to visualize the predictions and the highlighting regions with the most representative activated features from styled images. To do this, we extract from the penultimate layer the feature maps of width u and height v, with each element indexed by i,j. So refers to the activation at location of the feature map . We apply the GlobalAveragePooling (GAP) technique to the feature maps to get the neuron importance weights defined in Eq. 7.
(7) |
Where represents the neuron importance weights, c is the class, the size of the image, k is the k-th feature map, is the feature map, the score for class c, and is the gradient vector obtained via back-propagation. Next, we calculate the corresponding activation maps for each prediction using Eq. 7. From this point, we propose a new technique to visualize the highlighted regions in stylization and their variations. We present two methods: the Style Activation Map (SAM) defined as the relevant highlighted regions of the different styles in predictions and the Weighted Style Activation Map (WSAM) defined as the weighted sum of all styles applied in all samples per class.
We denote the SAM of a styled image with style , style intensity , and class c by . Where is the style intensity used, is the style used. Also, we have the k-th feature activation maps , and their class score for class c:
(8) |
We apply the ReLU function to the weighted linear combination of the feature maps because we are only interested in features with a positive influence. Then, we use this result to obtain the WSAM doing a weighted mean of and their predictions using all styles and all intensities. We define as the product of total styles and total intensities evaluated, so we have:
(9) |
Once we calculate in eq. 9 we will calculate the total variance region of samples to identify the most significant styles features for the classifier:
(10) |
Where is the i-th input sample stylized with (no style), the image size, and their class score for the class c. Our metric shows the highlighted region variance between an image and its styles with different s.


4 EXPERIMENTS AND RESULTS
We perform our experiments using the STL-10 () dataset, where samples are distributed in 5,000 and 8,000 labeled data for training and testing, respectively. We disregard the 100,000 unlabeled data for all our experiments. Besides, all experiments were performed using five different networks with high performance, such as Xception [Chollet, 2016], InceptionV3-299 [Szegedy et al., 2015], InceptionV4 [Szegedy et al., 2016], WideResNet-96 [Zagoruyko2016], and WideResNet-101 [Kabir et al., 2020]. We also compare our results with other state-of-the-art style augmentation like SWWAE [Zhao et al., 2015], Exemplar Convnet [Dosovitskiy et al., 2014], IIC [Ji et al., 2018], Ensemble [Thoma, 2017], WideResNet+cutout [DeVries and Taylor, 2017], InceptionV3 [Jackson et al., 2018], and STADA [Zheng et al., 2019].
4.1 Style Augmentation
First, we explore the effects of style augmentation through t-SNE visualization of images after applying the styler network to a subset of the test set Figure 1(a); we note some clusters of original images and their styles separate a bit of distance such as truck and horse. In Figure 1(b), we performed some styles using some values to find the best balance between style and content information as described above in Eq. 5. However, we emphasize the difference between traditional augmentation and the classical technique like rotation, mirroring, cutout [DeVries and Taylor, 2017], etc. With style augmentation, we increase the number of samples using about 80 000 styles and sampling for style intensity. Figure 1(b) shows different styles and different values (style intensity) from 0.0 to 1.0 by steps of 0.2.
Images with augmentation strategies for training deep models include traditional augmentations, cutouts, and our style augmentation method using a lower style effect (). At this point, we can consider style augmentation as a noise-adding technique or adversarial attacker due to the style distortion, which makes images more challenging to represent and be associated with the correct class.
4.2 Training Models
For experiments, we define four learning strategies, which are composed of no augmentation (None or N/A), traditional augmentation (Trad), style augmentation (SA), and both (Trad+SA) for each model. In Table 1, we present the quantitative comparisons between the state-of-the-art methods in style augmentation using styling and architectures for the STL-10 dataset; the Extra column means additional data is used to train that model, Trad column means traditional augmentation plus cutout, and Style column indicates our style augmentation.
Network | Extra | Trad | Style | Acc |
SWWAE | ✓ | ✓ | 74.33 | |
Exempla Conv | ✓ | ✓ | 75.40 | |
IIC | ✓ | ✓ | 88.80 | |
Baseline | ✓ | 75.67 | ||
Ensemble | ✓ | 77.62 | ||
STADA∗ | ✓ | ✓ | 75.31 | |
InceptionV3-299∗ | ✓ | ✓ | 80.80 | |
Xception-96∗ | ✓ | ✓ | 82.67 | |
Xception-128∗ | ✓ | ✓ | 85.11 | |
73.37 | ||||
✓ | 86.19 | |||
✓ | 74.89 | |||
Xception-256* | ✓ | ✓ | 86.85 | |
79.17 | ||||
✓ | 86.49 | |||
✓ | 80.52 | |||
InceptionV4-299* | ✓ | ✓ | 88.18 | |
77.28 | ||||
✓ | 87.26 | |||
✓ | 83.58 | |||
WideResNet-96* (WRN) | ✓ | ✓ | 88.83 | |
87.83 | ||||
✓ | 88.23 | |||
✓ | 92.23 | |||
WideResNet-101* (WRN) | ✓ | ✓ | 94.67 |


We note that in all cases, style augmentation helps to improve results. Besides, we found that models with higher input resolution reached higher accuracy after applying the styling method shown in Figure 2(a). Experiments on different input sizes support this affirmation [Chollet, 2016].
Furthermore, in Figure 2(b), we analyze the influence of style additions to a subset of the test set composed of 100 samples (10 samples per class), computing their average accuracy in each point on axis X consisting of an overall 20,000 random styles, these styles were sorted following from greater to lower accuracy. Note that the accuracy of the model trained without style augmentation decreased drastically for some styles. In contrast, the use of styles in training becomes the same architecture more robust to strong variations without losing accuracy.
5 STYLE ACTIVATION MAPS VISUALIZATION
Once our training step was finished, we evaluated and understood the stylization behavior in our models. First, in Figure 3(a), we show how our Style Activation Map works. Each row is the model, and each column is the learning strategy using no augmentation (N/A), using only style augmentation (SA), using traditional augmentation plus cutout (Trad), and using both (Trad+SA). We take a random sample with no style () to calculate their SAM (style activation map) for each model and each augmentation strategy. From this, we see how both Trad and Trad+SA help the models to focus on the plane instead other regions like no augmentation (N/A). Also, it is important to highlight that the better the prediction, the more accurate the region in the object (in this case, a plane).


On the other hand, using the best model WideResNet-101, we use the same random sample (a plane) to test the different learning strategies using the same style but varying the parameter. Let’s say, in this case, we will use as input the stylized sample. In Figure 3(b), we show the influence of image stylization. Each row means learning strategy N/A, SA, Trad, and Trad+SA. So, each column implies that the styled input sample by value varies from 1.0 (no intensity/style) to 0.0 (more intensity) before being evaluated by each network. We saw that a styled image tested in a model which does not use SA gets too bad results, but this did not happen in the model trained with SA. Also, the SAM-relevant regions in styled models tend to be constant along the variations.


In Figure 4(a), we show different samples, styles, and values. We show the influence of style in random samples with random styles; we note that some styles exist which don’t help to improve prediction. Otherwise, it gets worse. From this result, we say some styles can influence positively, negatively, or do not impact the input image. In addition, this result shows how the relevant regions for the network change depending on the style, and these two results are shown in Figure 4(b), also improving or not the confidence of the prediction.
6 DISCUSSIONS
We train, test, and visualize the impact of the style augmentation varying both values (from 0.0 to 1.0 in steps of 0.2) and learning strategies (N/A, SA, Trad, and Trad+SA) using the STL-10 dataset. We achieve high performance and the best result with WideResNet-101. We show the behavior of the style augmentation technique proposed (see Figure 1(a) and Figure 1(b)). We identify that some styles perturb the images more than others using the same sample, like adding noise. Also, we argue that by using larger input sizes and removing some complex styles, we probably remove the negative impacts on training (see Figure 2(a). Furthermore, our experiments showed interesting robustness to styles when styling is included in the training (see Figure 2(b)). Nonetheless, we also observed that the accuracy of models with Trad decreased drastically for some styles. Additionally, we found that some textures are more challenging to perform a style transfer using cutting-edge networks.
We explored more deeply the effects of particular styles and their influence on training and testing. In Figure 3(a), we present how style made a model more robust thanks to the different intensities of , which behaves as noise but does not apply to everyone. Specifically, we took the case of the plane evaluated in Figure 3(b). We got a low accuracy (0.341%) with higher style intensity (). Otherwise, we got the highest accuracy (0.988%) with . Furthermore, experiment results suggested that the best fit for could be between 0.3 and 0.8, similar results were found in [Jackson et al., 2018]. In Figure 4(a), we note that some styles have no effects, and for others, the network learns how to classify images correctly with higher intensities (noise). Also, style strengthens the correlation between predictions and styled features activation maps (see Figure 4(b)).

Category | Category | ||
airplane | 0.107 | horse | 0.269 |
truck | 0.129 | bird | 0.316 |
deer | 0.175 | dog | 0.338 |
cat | 0.193 | monkey | 0.380 |
car | 0.228 | ship | 0.456 |
We now calculate the WSAM variance and WSAM for each class sample, using all styles and s. In Table 2, we present the WSAM variance of all SAMs. Besides, in Figure 5, we show the result of WSAM for one sample per class. These results give us an idea about the impact of applying 79 424 styles with different intensities during the training phase and how the network learns to deal with those noisy samples (styled images), helping the robustness of the model. Finally, These results allow us to understand the influence of style augmentation in image classification. We can say that style augmentation can be used as a noise adder or adversarial attacker, making our model more robust against adversarial attacks.
7 CONCLUSIONS AND FUTURE WORK
In this work, we define a metric to explain by experimentation the behavior, the impact, and how the style augmentation may impact getting better results in the classification tasks. This metric is composed of three main outputs: Style Activation Map (SAM), Weighted Style Activation Map (WSAM), and ; this last one measures the variance of the regions of relevant features in styled samples. We outperform the state-of-the-art without extra data in style augmentation accuracy with WideResNet-101 trained on the STL-10 dataset; besides, our method gives robustness to input variations. From results and experiments, style augmentation has an impact on the model, and this impact can be visualized through SAM regions generated. We conclude that styles may modify and perturb different features from the input images (as an adversarial attacker), thus causing another set of images with slight variations in the distribution or becoming outliers making that prediction fail. In future directions, we will extend this study to more complex models with a higher number of parameters (like transformers) and higher images size like ImageNet and explain how style could influence their internal behavior. Also, we propose to understand more deeply which features are selected to be preserved in each style and which distortion they could generate through the network layers.
8 ACKNOWLEDGEMENTS
This work was supported by Carlos Chagas Filho Foundation for Research Support of Rio de Janeiro State (FAPERJ)-Brazil (grant #E-26/201.424/2021), São Paulo Research Foundation (FAPESP)-Brazil (grant #2021/07012-0), and the School of Applied Mathematics at Fundação Getulio Vargas (FGV/EMAp). Any opinions, findings, conclusions, or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the FAPESP, FAPERJ, or FGV.
REFERENCES
- Adebayo et al., 2018 Adebayo, J., Gilmer, J., Muelly, M., Goodfellow, I., Hardt, M., and Kim, B. (2018). Sanity checks for saliency maps.
- Chollet, 2016 Chollet, F. (2016). Xception: Deep Learning with Depthwise Separable Convolutions.
- DeVries and Taylor, 2017 DeVries, T. and Taylor, G. W. (2017). Improved Regularization of Convolutional Neural Networks with Cutout.
- Dosovitskiy et al., 2014 Dosovitskiy, A., Fischer, P., Springenberg, J. T., Riedmiller, M., and Brox, T. (2014). Discriminative Unsupervised Feature Learning with Exemplar Convolutional Neural Networks.
- Gatys et al., 2015 Gatys, L. A., Ecker, A. S., and Bethge, M. (2015). A Neural Algorithm of Artistic Style.
- Geirhos et al., 2018 Geirhos, R., Rubisch, P., Michaelis, C., Bethge, M., Wichmann, F. A., and Brendel, W. (2018). ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness.
- Georgievski, 2019 Georgievski, B. (2019). Image Augmentation with Neural Style Transfer. pages 212–224.
- Ghiasi et al., 2017 Ghiasi, G., Lee, H., Kudlur, M., Dumoulin, V., and Shlens, J. (2017). Exploring the structure of a real-time, arbitrary neural artistic stylization network.
- Gilpin et al., 2018 Gilpin, L. H., Bau, D., Yuan, B. Z., Bajwa, A., Specter, M., and Kagal, L. (2018). Explaining explanations: An overview of interpretability of machine learning.
- Gkitsas et al., 2019 Gkitsas, V., Karakottas, A., Zioulis, N., Zarpalas, D., and Daras, P. (2019). Restyling Data: Application to Unsupervised Domain Adaptation.
- Hesse et al., 2019 Hesse, L. S., Kuling, G., Veta, M., and Martel, A. L. (2019). Intensity augmentation for domain transfer of whole breast segmentation in MRI.
- Jackson et al., 2018 Jackson, P. T., Atapour-Abarghouei, A., Bonner, S., Breckon, T., and Obara, B. (2018). Style Augmentation: Data Augmentation via Style Randomization.
- Ji et al., 2018 Ji, X., Henriques, J. F., and Vedaldi, A. (2018). Invariant Information Clustering for Unsupervised Image Classification and Segmentation.
- Jing et al., 2017 Jing, Y., Yang, Y., Feng, Z., Ye, J., Yu, Y., and Song, M. (2017). Neural Style Transfer: A Review.
- Kabir et al., 2020 Kabir, H. M. D., Abdar, M., Jalali, S. M. J., Khosravi, A., Atiya, A. F., Nahavandi, S., and Srinivasan, D. (2020). Spinalnet: Deep neural network with gradual input.
- Kotovenko, Dmytro, adn Sanakoyeu, Artsiom, and Lang, Sabine, and Ommer, 2019 Kotovenko, Dmytro, adn Sanakoyeu, Artsiom, and Lang, Sabine, and Ommer, B. (2019). Content and Style Disentanglement for Artistic Style Transfer.
- Li et al., 2018 Li, X., Liu, S., Kautz, J., and Yang, M.-H. (2018). Learning Linear Transformations for Fast Arbitrary Style Transfer.
- Selvaraju et al., 2017 Selvaraju, R. R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., and Batra, D. (2017). Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, pages 618–626.
- Shrikumar et al., 2017 Shrikumar, A., Greenside, P., and Kundaje, A. (2017). Learning important features through propagating activation differences. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pages 3145–3153. JMLR. org.
- Simonyan et al., 2013 Simonyan, K., Vedaldi, A., and Zisserman, A. (2013). Deep inside convolutional networks: Visualising image classification models and saliency maps.
- Sundararajan et al., 2017 Sundararajan, M., Taly, A., and Yan, Q. (2017). Axiomatic attribution for deep networks. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pages 3319–3328. JMLR. org.
- Szegedy et al., 2016 Szegedy, C., Ioffe, S., Vanhoucke, V., and Alemi, A. (2016). Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning.
- Szegedy et al., 2015 Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., and Wojna, Z. (2015). Rethinking the Inception Architecture for Computer Vision.
- Tanaka and Aranha, 2019 Tanaka, F. H. K. d. S. and Aranha, C. (2019). Data Augmentation Using GANs.
- Thoma, 2017 Thoma, M. (2017). Analysis and Optimization of Convolutional Neural Network Architectures.
- Ulyanov et al., 2017 Ulyanov, D., Vedaldi, A., and Lempitsky, V. (2017). Improved Texture Networks: Maximizing Quality and Diversity in Feed-forward Stylization and Texture Synthesis.
- Zagoruyko and Komodakis, 2016 Zagoruyko, S. and Komodakis, N. (2016). Wide Residual Networks.
- Zeiler and Fergus, 2014 Zeiler, M. D. and Fergus, R. (2014). Visualizing and understanding convolutional networks. In European conference on computer vision, pages 818–833. Springer.
- Zhao et al., 2015 Zhao, J., Mathieu, M., Goroshin, R., and LeCun, Y. (2015). Stacked What-Where Auto-encoders.
- Zheng et al., 2019 Zheng, X., Chalasani, T., Ghosal, K., Lutz, S., and Smolic, A. (2019). STaDA: Style Transfer as Data Augmentation.
- Zhou et al., 2016 Zhou, B., Khosla, A., Lapedriza, A., Oliva, A., and Torralba, A. (2016). Learning deep features for discriminative localization. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2921–2929.