Date of publication xxxx 00, 0000, date of current version xxxx 00, 0000. 10.1109/ACCESS.2024.0429000
Corresponding author: Shohei Enomoto (e-mail: [email protected]).
Dynamic Test-Time Augmentation via Differentiable Functions
Abstract
Distribution shifts, which often occur in the real world, degrade the accuracy of deep learning systems, and thus improving robustness to distribution shifts is essential for practical applications. To improve robustness, we study an image enhancement method that generates recognition-friendly images without retraining the recognition model. We propose a novel image enhancement method, DynTTA, which is based on differentiable data augmentation techniques and generates a blended image from many augmented images to improve the recognition accuracy under distribution shifts. In addition to standard data augmentations, DynTTA also incorporates deep neural network-based image transformation, further improving the robustness. Because DynTTA is composed of differentiable functions, it can be directly trained with the classification loss of the recognition model. In experiments with widely used image recognition datasets using various classification models, DynTTA improves the robustness with almost no reduction in classification accuracy for clean images, thus outperforming the existing methods. Furthermore, the results show that robustness is significantly improved by estimating the training-time augmentations for distribution-shifted datasets using DynTTA and retraining the recognition model with the estimated augmentations. DynTTA is a promising approach for applications that require both clean accuracy and robustness. Our code is available at https://github.com/s-enmt/DynTTA.
Index Terms:
Distribution-shift, Image enhancement, Robustness, Test-time augmentation, Visual recognition.=-21pt
I Introduction
With the development of deep learning, the field of visual recognition has made great progress, and services using deep learning are becoming more practical. Many deep learning models are trained and validated on images from the same distribution. However, images in the real world are subject to distribution shifts due to various factors such as weather conditions, sensor noise, blurring, and compression artifacts. Deep learning models do not usually take these distribution shifts into account, causing the accuracy to reduce in the presence of such artifacts. This is fatal for safety-critical applications, such as automated driving, where the environment changes frequently.
There are two main approaches to solving this problem: training robust recognition models and image enhancement. The former uses data augmentation [13, 15, 19], or training algorithms such that the recognition model has high robustness. The latter uses test-time augmentation [20] or deep neural networks (DNNs) [47] to transform the distorted images into recognition-friendly images that are easily recognized by the recognition model. These approaches can be used in combination to achieve higher robustness. This paper focuses on the image enhancement approach, which is highly practical because it is used before inference by the pretrained recognition model without retraining the recognition model. We found that DNN-based image enhancement [47] overfits to particular transformation patterns, resulting in excessive image transformation even for clean images and reduced clean accuracy. Test-time augmentation-based image enhancement [20], which dynamically selects the best of several data augmentations for an input image, improves robustness while maintaining clean accuracy. However, its augmentation search space is limited, so the improvement in robustness is small.
To remove these limitations, we propose a novel image enhancement method, DynTTA. DynTTA uses differentiable data augmentation techniques [40, 9, 28] and image blending [13, 56, 15, 19] to dynamically generate the recognition-friendly image for an input image from a huge augmentation space. DynTTA incorporates a DNN-based image transformation by considering it as a data augmentation. Data augmentations work as a hint for learning diverse image transformations and avoid overfitting, resulting in significantly improving robustness without losing clean accuracy. An overview diagram of DynTTA is shown in Fig. 1. DynTTA takes an image as input and outputs the magnitude parameters and blend weights for predefined data augmentation. Each augmentation is performed on the basis of the magnitude parameters. These augmented images are linearly combined with blend weights to generate the output image. The recognition model takes DynTTA’s output images as input and performs inference. Because DynTTA is composed of differentiable functions, it can be directly trained with the loss of the recognition task. As a result, DynTTA transforms distorted images into recognizable ones to improve accuracy without needing model retraining. In addition, we hypothesize that training with the inverse operations of highly weighted augmentations by DynTTA would make the classification model robust against the given distribution. We thus propose a novel method using DynTTA: estimating effective training-time augmentations and retraining classification models with those augmentations.

DynTTA was evaluated on widely used image classification datasets and various classification models. We introduce a practical training and evaluation scenario, the blind setting, which does not assume the type of test-time distribution shifts and has not been used in existing literature. DynTTA improves the accuracy for distorted images while maintaining the accuracy for clean images, which is a better result than those of existing methods. Specifically, experimental results on the ImageNet dataset [43] show that DynTTA improves the robustness of pre-trained ResNet50 [10] by 5.30 percentage points while maintaining clean accuracy. Moreover, retraining with the estimated training-time augmentation on the PACS dataset [26] improves the accuracy of ResNet50 by 6.43 percentage points. DynTTA is a promising approach for applications that require both clean accuracy and robustness, such as automated driving.
The main contributions of this paper are as follows.
-
•
We propose DynTTA, a novel image enhancement method based on differentiable data augmentation techniques and image blending. DynTTA generates diverse images by transformations with various data augmentations, including a DNN, which significantly improves robustness while maintaining clean accuracy.
-
•
We propose a novel method to estimate effective training-time augmentations and show that retraining the classification model with the estimated augmentations significantly improves accuracy.
-
•
Our extensive experiments show that DynTTA improves robustness compared to existing methods while maintaining clean accuracy. In addition, experimental results in the blind setting show that DynTTA is practical under various distribution shifts.
II Related Work
II-A Robustness of DNNs
DNNs are known to be vulnerable to image distortions, which often occur in the real world. These image distortions have been studied in several works. Diamond et al. [3] showed that defects such as noise and blur in real-world sensors degrade performance in image recognition networks. Pei et al. [36] found that haze reduces the accuracy of image classification and experimented that the dehaze method for human visibility does not effectively improve image classification performance. Pei et al. [37] empirically studied real-world image degradation problems for nine kinds of degraded images:-— hazy, motion-blurred, fish-eye, underwater, low resolution, salt-and-peppered, white Gaussian noise, Gaussian-blurred, and out-of-focus. Various datasets have been proposed to evaluate the robustness of DNNs. Geirhos et al. [7] made a style-transformed dataset to demonstrate that DNNs recognize objects by their texture rather than their shape. Hendrycks and Dietterich [12] proposed 19 algorithmically generated corruption datasets in five levels, belonging to four categories: noise, blur, weather, and digital. They showed that DNNs trained on a clean dataset have lower accuracy on these corruption datasets. In addition, they introduced a dataset of naturally occurring adversarial examples in the real world [14] and a dataset containing distribution shifts that occur in the real world, such as image style, image blurriness, geographic location, camera operation, and more [11]. In this study, we evaluate our method on these datasets.
II-B Training Robust Classification Models
Many works have been studied to improve the robustness of DNNs to naturally occurring image distortions. One way to improve the robustness is the data augmentation. Mixing multiple data augmentations [13, 56, 5, 15, 19, 54] and adding noise during training [42, 53] have been proposed as effective data augmentations to improve robustness. Mintun et al. [33] found that the similarity between data augmentation and test-time image corruption is strongly correlated with performance. Training algorithms to improve robustness have also been studied. Disentangled learning via auxiliary batch normalization [58, 61, 32, 4], a method to improve adversarial training, improves not only robustness but also clean accuracy. Li et al. [27] proposed an algorithm to train the model paying complementary attention to the shape and texture of the objects. Since these methods are used when training recognition models, they can be used in combination with image enhancement methods, which assume that the recognition models are pretrained. Our experimental results show that combining the recognition models pretrained by these methods with image enhancement further improves the robustness.
II-C Image Enhancement
This section presents two image enhancement approaches, test-time augmentation and image transformation. Table I shows a summary of the pros and cons of these approaches and DynTTA.
Clean Accuracy | Robustness | |
---|---|---|
Test-time Augmentation | ||
Image Transformation | ||
DynTTA |
II-C1 Test-time Augmentation
Recognition accuracy can be improved by using data augmentation during not only the training time but also the test time. Various test-time augmentation methods have been proposed, using simple geometric transformations such as flip and crop [38, 59, 2, 23, 48, 10], mixup [24, 35], and augmentation in the embedding space [1]. Shanmugam et al. [44] experimentally analyzed why and when test-time augmentation works, and Kimura et al. [21] gave theoretical guarantees. These studies use only specific data augmentations for all test set images. Dynamically selecting the best data augmentation for each input image is expected to improve the accuracy. Lyzhov et al. [31] proposed a greedy search-based test-time augmentation method to find the augmentation policy on the test set, but it is not optimal for each input image. Kim et al. [20] proposed a module called Loss Predictor to predict the classification loss of augmented images, allowing dynamic selection of the best one for the input image from 12 data augmentations. However, despite the augmentation space actually being infinite, these methods have significant limitations on the augmentation space due to computational cost and the improvement in robustness is small. Our method eliminates this limitation and achieves better robustness than existing methods by generating the best augmented image from a large augmentation space. In addition, we propose the training-time augmentation estimation method for distribution-shifted datasets using test-time augmentation and show that retraining the classification model by estimated augmentations significantly improves accuracy.
II-C2 Image Transformation
Several methods have been proposed to improve the recognition accuracy by image transformation using DNNs. Sharma et al. [45] proposed convolutional neural network-based enhancement filters that enhance image-specific details to improve recognition accuracy. Subsequent work [29, 34, 47, 25, 6] used DNN to transform corrupted images into recognition-friendly ones. Milanfar et al. [49] showed that not only image enhancement but also resizing at the same time improves the recognition accuracy. These methods use DNNs to transform images, but tend to perform excessive transformations to remove distortions, which reduces the accuracy of clean images. By considering these image enhancement methods as a kind of data augmentation, we treat them as part of DynTTA. DNN-based transformations reduce clean accuracy because they overfit particular transformation patterns that removes distortions. We integrate DNN transformation and data augmentations to enable learning of diverse image transformations, which avoids overfitting and improves robustness while maintaining clean accuracy.
III DynTTA
III-A Key Ideas
Many data augmentations have continuous magnitude parameters (e.g., rotation has a continuous magnitude parameter from 0 to 360 degrees). In addition, there are many combinations of two or more data augmentations (e.g., augmentation is performed in order of rotation, contrast, and sharpness). These facts make the augmentation space infinite. To solve the test-time augmentation challenge of selecting the best combination of data augmentations and their optimal magnitude parameters for the test image from an infinite space, we introduce two key ideas. The first is to find the local optimal magnitude parameter by optimization. Since most data augmentations are differentiable with respect to the magnitude parameter [40, 9, 28, 41], the local optimal magnitude parameter is obtained by optimization algorithms, such as gradient descent. The second is to represent the combination of augmentation with image blending, which is the process of affine combination of multiple images. For example, if the weights are , only the 1st augmented image is selected, and if the weights are , the 1st and 3rd augmented images are combined. Note that image blending does not provide the performing order of the augmentation. We call these ideas “Optimization” and “Blending”, respectively. On the basis of these key ideas, we propose a novel image enhancement method, DynTTA.
III-B Overview of DynTTA
The overview of DynTTA is shown in Fig. 1. The differentiable data augmentations used in DynTTA and their range of magnitude parameters are defined. Most data augmentations, such as rotation, have a magnitude parameter. For data augmentations that do not have a magnitude parameter, such as auto-contrast, is not defined. Section III-C details the data augmentations of DynTTA.
Any DNN can be used as the backbone of DynTTA by modifying its output layer. DynTTA takes an image as input and outputs the magnitude parameters and blend weights . is the number of data augmentations. The magnitude parameter is mapped to a value in the range or using an activation function such as sigmoid function or hyperbolic tangent function. It is then multiplied by the predefined magnitude range . The blend weights are converted to weights that sum to 1 by the softmax function. and are shown in the following equation.
(3) | ||||
(4) |
By using the magnitude parameters and blend weights, DynTTA generates a recognition-friendly image. data augmentations are performed with magnitude parameters , then DynTTA generates a blended image by linearly combining augmented images with . DynTTA outputs the image using the following equation:
(5) |
where is the input image, is the -th augmentation operation, and is the output image.
III-C Augmentation Space
The data augmentations and their and used by DynTTA are shown in Table II. Because DynTTA obtains magnitude parameters through optimization, only differentiable data augmentations are used. This study uses the kornia library [40] and deals only with general data augmentation. We do not employ data augmentation such as shear because we believe it is ineffective when used during testing. Flip is a typical test-time augmentation, but it is not used because it is non-differentiable. Placing low-pass and high-pass filters within the DNN architecture [62, 39, 55, 16] or using them as training-time augmentation [17, 60, 46] is known to improve robustness. Inspired by these studies, we try to improve the robustness by using low-pass and high-pass filters as test-time augmentation. These filters have a filter size parameter that is not obtained via gradient descent. Therefore, this parameter is discretized, and multiple filters are prepared. 19 low-pass and 19 high-pass filters are made, each with a filter size from to in increments of . URIE uses a DNN to generate recognition-friendly images. Such an image transformation model is applied to DynTTA by considering it as a kind of augmentation. In this case, the image transformation model is trained simultaneously with DynTTA.
DynTTA handles many data augmentations but is also computationally expensive. By not executing data augmentation with small blend weights, the number of data augmentation executions is significantly reduced while maintaining accuracy. The details are described in the Appendix.
Augmentation | Inverse | Augmentation | Inverse | ||||
---|---|---|---|---|---|---|---|
Rotate | tanh | 30 | Rotate | Hue | tanh | 2.0 | Hue |
Scale | tanh | 0.3 | Scale | Equalize | - | - | Equalize |
Saturate | sigmoid | 5.0 | Saturate | Invert | - | - | Invert |
Contrast | sigmoid | 3.0 | Contrast | Gamma | sigmoid | 3.0 | Gamma |
Sharpness | sigmoid | 10 | Sharpness | LPFs () | - | - | HPFs |
Brightness | tanh | 0.6 | Brightness | HPFs () | - | - | LPFs |
A-Contrast | - | - | Contrast | URIE | - | - | - |
III-D Training and Testing
DynTTA is trained by applying it before inference by a pretrained classification model. When DynTTA is trained, the classification model is frozen. The output image of DynTTA is used as input to the classification model to calculate the cross entropy loss. Because DynTTA is differentiable, it minimizes this loss and outputs recognition-friendly images. During testing, the DynTTA is frozen and the DynTTA output image is input into the classification model to obtain the prediction results.
III-E Retraining with Estimated Training-time Augmentations
We propose a different use of DynTTA, improving accuracy by reproducing unknown distribution shifts with data augmentations and retraining the classification model with it. For a distribution-shifted dataset, DynTTA outputs the blend weights that weight effective test-time augmentations. We hypothesize that the distribution shift would consist of inverse operations of the highly weighted test-time augmentations. For example, consider an unknown distribution shift dataset for which DynTTA estimates that LPF is effective. Since LPF removes high-frequency components, this dataset is expected to contain high-frequency components. To make a classification model robust against high-frequency components, high-frequency components should be included in the training distribution. To achieve this, the inverse operation of LPF, i.e., HPF, is used to include the high-frequency components in the training dataset and retrain the model.
Algorithm 1 shows how to estimate effective augmentations at training-time. We rank the effectiveness of augmentation by the blend weights and obtain the top inverse operations of estimated augmentations. If the inverse operation does not exist, it is omitted. The inverse operation for each augmentation is defined in Table II. The estimated augmentations are included in the AugMix [13] augmentation space to make a classification model robust against a given unknown distribution.
IV Experiments
This section describes the experimental setup, experimental results comparing DynTTA to other image enhancement methods, and detailed experiments with DynTTA. It also shows that robustness is significantly improved by using DynTTA to estimate the effective training-time augmentations for distribution shift, which is then used to retrain the classification model. Some experiments are described in more detail in the Appendix.
IV-A Training and Evaluation Settings
In this study, two settings were used for training and evaluation: blind and non-blind. In the non-blind setting, which is used in the existing literature [20, 47], the corruption dataset [12] is used. This dataset consists of 19 types of corruption with five severity levels generated by the algorithm. 15 corruptions are used for training as Seen corruptions, and four corruptions are used for testing as Unseen corruptions. Seen and Unseen corruptions consist of the same four categories (noise, blur, weather, and digital). In this setting, the type of corruption (e.g. noise) in the test set is known in advance, so this knowledge is used to generate artificial corruption (e.g. Gaussian noise), which is then used as Seen for training. However, this setting is often impractical because test set corruption is often unknown. Therefore, we introduce a new training and evaluation setting, the blind setting, which has no assumption on test-time distribution shifts. In the blind setting, the distribution of the test set is unknown, so data augmentation is used such as AugMix, which improves robustness to unknown distributions. AugMix mixes nine augmentations by default, and these augmentations do not include the corruption datasets used for testing. In this setting, an image enhancement model trains to enhance the features needed to classify out-of-distribution data from diverse data generated by the data augmentation. This enables the image enhancement model to increase classification accuracy for an unknown distribution, even though it does not learn the distribution of test time.
IV-B Experimental Setup
Our method was evaluated on two image recognition datasets (CUB [57] and ImageNet [43]) and one domain generalization dataset (PACS [26]). All experiments except ImageNet were run three times, and the average results are reported. Loss Predictor (LP) () [20] and URIE [47] were used as comparison methods. Classification models were prepared in advance, and the backbone network of DynTTA and LP was prepared by fine-tuning the ResNet18 [10] pretrained on ImageNet. For training in the non-blind setting, corruption was randomly selected for each mini-batch to train the enhancement models. The training corruptions consist of 15 Seen corruptions in five levels and clean (no corruption). For testing, four Unseen corruptions were used. In the blind setting, the enhancement models were trained using AugMix+DeepAugment [11] for ImageNet and AugMix for the other datasets. For testing, all 19 corruptions were used. When evaluating on the corruptions dataset, the average accuracy was used as an evaluation metric. In addition to evaluating the robustness of these distribution shift datasets, the clean accuracy on the standard dataset was also evaluated. The baseline did not use the enhancement model, and the difference from the baseline is reported in the results. In the tables showing experimental results, the best result is shown in bold, and the second best result is shown in underline.
IV-C Performance Evaluation
IV-C1 Classification on the CUB
The results for classification accuracy on the CUB dataset are shown in Table III. In the non-blind setting, URIE improves robustness but reduces the clean accuracy by a maximum of about 3.3 percentage points. LP does not decrease the clean accuracy, but the improvement in robustness is small. DynTTA outperforms the comparison methods in terms of both clean accuracy and robustness and significantly improves the robustness with almost no decrease in clean accuracy. In the blind setting, similar to the non-blind setting, URIE tends to reduce the clean accuracy by a maximum of about 1.0 percentage points. LP improves both clean accuracy and robustness slightly. DynTTA has less degradation in clean accuracy than URIE, and it is also more robust. In particular, DynTTA significantly improves the robustness of Mixer-B16 [51] compared to the comparison methods. Our experimental results show that AugMix-trained enhancement models further improve the robustness of AugMix-trained classification models. The results indicate that our method can be used in conjunction with existing robustness improvement methods. Furthermore, experimental results on generalizability for different classification models at training and testing, the effects of the backbone network, augmentation space, and magnitude range, and comparison to simple baselines are provided in the Appendix.
Non-blind | |||
---|---|---|---|
Classifier | Enhancer | Clean | Unseen |
ResNet50 | URIE | 78.39 (-3.32) | 55.93 (7.53) |
LP | 81.70 (-0.01) | 50.88 (2.48) | |
DynTTA | 81.58 (-0.13) | 58.02 (9.62) | |
ResNet50† | URIE | 80.99 (-1.56) | 63.95 (3.95) |
LP | 82.59 (0.04) | 60.79 (0.79) | |
DynTTA | 82.64 (0.09) | 65.49 (5.49) | |
Mixer-B16 | URIE | 86.73 (-0.69) | 72.77 (3.93) |
LP | 87.39 (-0.03) | 70.06 (1.22) | |
DynTTA | 87.56 (0.14) | 73.89 (5.05) | |
Mixer-B16† | URIE | 86.61 (-0.47) | 75.88 (2.68) |
LP | 86.93 (0.05) | 74.20 (1.00) | |
DynTTA | 87.26 (0.38) | 76.58 (3.38) | |
DeiT-base | URIE | 83.91 (-1.25) | 68.61 (1.79) |
LP | 85.16 (0.00) | 67.47 (0.65) | |
DynTTA | 84.79 (-0.37) | 69.80 (2.98) | |
DeiT-base† | URIE | 83.88 (-0.29) | 70.56 (1.50) |
LP | 84.20 (0.03) | 69.64 (0.58) | |
DynTTA | 84.42 (0.25) | 71.76 (2.70) | |
Blind | |||
Classifier | Enhancer | Clean | Corruption |
ResNet50 | URIE | 80.68 (-1.03) | 51.47 (2.89) |
LP | 81.75 (0.04) | 48.61 (0.03) | |
DynTTA | 81.65 (-0.06) | 51.97 (3.39) | |
ResNet50† | URIE | 82.75 (0.20) | 62.05 (0.30) |
LP | 82.65 (0.10) | 61.76 (0.01) | |
DynTTA | 82.84 (0.29) | 62.39 (0.64) | |
Mixer-B16 | URIE | 87.04 (-0.38) | 71.27 (0.07) |
LP | 87.49 (0.07) | 71.78 (0.58) | |
DynTTA | 87.37 (-0.05) | 72.74 (1.54) | |
Mixer-B16† | URIE | 86.89 (0.01) | 74.41 (0.04) |
LP | 86.95 (0.07) | 74.44 (0.07) | |
DynTTA | 87.12 (0.24) | 75.46 (1.09) | |
DeiT-base | URIE | 84.71 (-0.45) | 67.44 (0.96) |
LP | 85.21 (0.05) | 66.49 (0.01) | |
DynTTA | 85.06 (-0.10) | 67.37 (0.89) | |
DeiT-base† | URIE | 84.21 (0.04) | 70.38 (0.10) |
LP | 84.21 (0.04) | 70.28 (0.00) | |
DynTTA | 84.31 (0.14) | 70.35 (0.07) |
IV-C2 Classification on the ImageNet
The results for classification accuracy on the ImageNet dataset are shown in Table IV. In the non-blind setting, URIE improves robustness but reduces the clean accuracy by a maximum of about 1.6 percentage points. LP does not decrease the ResNet50 clean accuracy, but the improvement in robustness is small. In DeiT [52], LP shows a decrease in both clean accuracy and robustness. DynTTA slightly reduces the clean accuracy, but it improves robustness more than the comparison methods. In the blind setting, URIE and LP experiments show almost the same trend in results as in the non-blind setting. DynTTA improves robustness over URIE without degrading the clean accuracy.
Non-blind | |||
---|---|---|---|
Classifier | Enhancer | Clean | Unseen |
ResNet50 | URIE | 74.56 (-1.57) | 49.05 (3.96) |
LP | 76.12 (-0.01) | 46.03 (0.94) | |
DynTTA | 75.89 (-0.24) | 50.55 (5.46) | |
DeiT-base | URIE | 81.01 (-0.73) | 67.50 (2.19) |
LP | 81.09 (-0.65) | 64.22 (-1.09) | |
DynTTA | 81.25 (-0.50) | 67.92 (2.61) | |
Blind | |||
Classifier | Enhancer | Clean | Corruption |
ResNet50 | URIE | 74.71 (-1.42) | 44.19 (4.62) |
LP | 76.13 (0.00) | 40.29 (0.72) | |
DynTTA | 76.04 (-0.09) | 44.87 (5.30) | |
DeiT-base | URIE | 80.69 (-1.06) | 62.69 (1.07) |
LP | 80.76 (-0.98) | 60.28 (-1.34) | |
DynTTA | 81.75 (0.01) | 62.71 (1.08) |
Furthermore, the DynTTA performance was evaluated on distribution shift datasets Stylized-ImageNet [7] (Stylized), ImageNet-A [14] (A), and ImageNet-R [11] (R) other than the corruption dataset in the blind setting. The results are shown in Table V. URIE significantly decreases accuracy when the dataset is A and the classification model is DeiT, indicating that URIE is overfitting to the Corruption and R distribution. LP barely affects the accuracy with ResNet50 and worsens the accuracy with DeiT. While URIE and LP sometimes show decreased accuracy, DynTTA consistently improves generalization performance on these complex distribution shift environments without overfitting a particular distribution.
ResNet50 | |||
---|---|---|---|
Enhancer | Stylized | A | R |
URIE | 9.38 (2.20) | 0.59 (0.59) | 38.17 (2.00) |
LP | 7.16 (-0.01) | 0.01 (0.01) | 36.18 (0.01) |
DynTTA | 9.80 (2.63) | 0.16 (0.16) | 37.48 (1.31) |
DeiT | |||
Enhancer | Stylized | A | R |
URIE | 19.93 (1.91) | 24.99 (-2.86) | 46.96 (1.61) |
LP | 16.23 (-1.80) | 25.73 (-2.12) | 44.17 (-1.18) |
DynTTA | 20.12 (2.09) | 28.61 (0.76) | 45.64 (0.29) |
IV-C3 Classification on the PACS
The effect of changing the training domain in the blind setting was evaluated, and the average accuracy is shown in Table VI. Other experimental settings followed DomainBed [8]. Image enhancement models trained in the blind setting (trained with AugMix) are also effective for the domain generalization dataset, and DynTTA outperforms URIE in terms of average accuracy.
Enhancer | Art | Cartoon | Photo | Sketch | Avg. |
---|---|---|---|---|---|
URIE | 77.81 (2.48) | 80.48 (3.30) | 51.06 (5.59) | 33.60 (8.90) | 60.74 (5.07) |
DynTTA | 78.65 (3.32) | 80.77 (3.59) | 50.33 (4.86) | 34.27 (9.57) | 61.01 (5.34) |
IV-C4 Retraining with Estimated Augmentations
The effect of retraining with estimated augmentations by DynTTA on the PACS dataset was evaluated. A small split of the test domain was treated as a given unknown dataset to estimate effective data augmentations. Note that a small split was not used in the final evaluation. For a fair comparison with an equal number of augmentations, a scenario involving original 13 augmentations (as outlined in the official AugMix code) was compared to a scenario comprising 9 default augmentations plus 4 estimated augmentations. The accuracy in each scenario is shown in Table VII. The estimated data augmentations consistently improve accuracy over the original.
Augmentation | Art | Cartoon | Photo | Sketch | Avg. |
---|---|---|---|---|---|
Original | 75.62 (0.29) | 79.01 (1.83) | 47.07 (1.60) | 32.80 (8.10) | 58.62 (2.95) |
Estimated | 78.03 (2.70) | 79.88 (2.70) | 52.84 (7.37) | 37.63 (12.93) | 62.10 (6.43) |
IV-D Detailed Experiments
This section describes experiments on the CUB dataset using ResNet50 as the classification model for an in-depth analysis of DynTTA
IV-D1 Effects of Key Ideas
We conducted an ablation study of our key ideas using ResNet50 on the CUB dataset and compared it to LP. LP selects the best augmentation for a test image by learning the classification losses of 12 augmented images111Identity; {-20, 20} Rotate; {0.8, 1.2} Zoom; {0.5, 2.0} Saturate; Auto-contrast; {0.2, 0.5, 2.0, 4.0} Sharpness as a label. One of our key ideas, “Blending”, improves upon LP by using not one image but a combination of many augmented images. “Blending” is lightweight because it does not require the computation of classification losses, which has the advantage of being easily extendable in augmentation space. Specifically, LP requires the same number of classification model inferences as the number of data augmentations used, while “Blending” requires only one. Here we define DynTTA (BL), which has the same augmentation and magnitude parameters as LP (see footnote) and outputs only blend weights. The effect of the “Blending” was measured by using DynTTA (BL). In addition, to measure the effect of “Optimization”, we define DynTTA (BLOPT) that extends the DynTTA (BL) and simultaneously outputs magnitude parameters. At this time, the magnitude parameter ranges were set the same as in Table II except for Zoom, and Identity was not used because it was included in some data augmentations (e.g., a rotation of 0 degrees). The results are shown in Table VIII. DynTTA (BL) is much more robust than LP, meaning that blending multiple images is better than choosing one best image. In the non-blind setting, DynTTA (BLOPT) is less robust than DynTTA (BL). This is because the magnitude parameters used by LP and “Blending” may have been tuned by the authors for their non-blind setting experiment. On the other hand, in the blind setting, DynTTA (BLOPT) is more robust than DynTTA (BL). This is because “Optimization” automatically finds the local optimal magnitude parameters for the blind setting using gradient descent (here we use Adam [22], see Section III and Appendix for details). Moreover, “Optimization” eliminates magnitude hyperparameters. For example, LP has 10-magnitude hyperparameters, but these are difficult to tune in the blind setting. “Optimization” eliminates these hyperparameters, making DynTTA more practical.
Non-blind | Blind | ||||
---|---|---|---|---|---|
Enhancer | Clean | Corruption | Clean | Corruption | |
DynTTA (BL) | 81.64 (-0.06) | 52.49 (1.61) | 81.71 (-0.04) | 51.12 (2.51) | |
DynTTA (BLOPT) | 81.68 (-0.02) | 52.15 (1.27) | 81.73 (-0.02) | 51.51 (2.90) |
IV-D2 Avoiding Overfitting of DNN-based Transformation
When a model overfits, it tend to use the same set of transformations more (less diverse), which happens in URIE. The mean squared error (MSE) between the raw and enhanced images was measured as the amount of image transformation to show that DynTTA reduces overfitting. The results are shown in Fig. 2. URIE almost always shows the same MSE, regardless of severity, and performs almost the same set of transformations on any image, resulting in useless transformations even on clean images. In contrast, DynTTA shows a large MSE at high severity, a small MSE at low severity, and a particularly small MSE for clean images. Furthermore, DynTTA has a large variance of MSEs, indicating that it learns to use different transformations depending on the severity and type of corruption. The data augmentations introduced by DynTTA work as a hint to learn diverse transformations, avoiding overfitting and enhancing images under various distributions.

IV-D3 Backbone Network of DynTTA
Because any neural network can be used as the backbone network of DynTTA, the effect of backbone network selection was examined. Four models were used as backbone network: ResNet18, ResNet50 [10], EfficientNet-b0 [50], and MobileNetV3 [18]. The results are shown in the Table IX. ResNet18 shows the highest robustness in the non-blind setting. EfficientNet-b0 and MobileNetV3 are more robust than ResNet in the blind setting, but their clean accuracy degrades more in the non-blind setting. ResNet50 has higher classification accuracy than ResNet18, but there is no correlation between the accuracy of the model as a classification model and the performance of the model as an enhancement model.
Non-blind | ||
---|---|---|
Backbone Network | Clean | Corruption |
ResNet18 | 81.58 (-0.13) | 58.02 (9.62) |
ResNet50 | 81.74 (0.03) | 57.56 (9.16) |
EfficientNet-b0 | 81.27 (-0.44) | 57.37 (8.97) |
MobileNetV3 | 81.08 (-0.63) | 57.65 (9.25) |
Blind | ||
Backbone Network | Clean | Corruption |
ResNet18 | 81.65 (-0.06) | 51.97 (3.39) |
ResNet50 | 81.73 (0.02) | 52.07 (3.49) |
EfficientNet-b0 | 81.73 (0.02) | 52.95 (4.37) |
MobileNetV3 | 81.73 (0.02) | 53.00 (4.42) |
IV-D4 Performance Evaluation Using a Classification Model Different from the One Used During Training
This section discusses the generalizability of image enhancement models to classification models other than what was used for its training. Usually, there is a one-to-one correspondence between image enhancement models and classification models. We investigated whether an image enhancement model trained together with one classification model is also beneficial for another classification model. The results of the image enhancement models trained with ResNet50, MLP-Mixer, and DeiT in the non-blind setting on the CUB dataset are shown in Table X. URIE trained on ResNet50 significantly reduces clean accuracy and does not improve robustness. URIE trained with MLP-Mixer or DeiT reduces clean accuracy by about 1-2 percentage points but improves robustness. LP improves robustness with almost no reduction in clean accuracy. DynTTA trained with ResNet50 reduces clean accuracy about 1 percentage point but improves robustness. DynTTA trained with MLP-Mixer or DeiT improves robustness over the comparison methods with almost no reduction in clean accuracy. Our experimental results show that when the coupled classification model is a high accuracy model such as MLP-Mixer or DeiT, the image enhancement model has high generalizability. In particular, DynTTA improves robustness over the comparison methods with almost no reduction in clean accuracy, which shows a high generalizability. DynTTA trained with a highly accurate classification model shows its potential to be applied to a variety of classification models.
Classifier at training | Classifier at testing | Enhancer | Clean | Corruption |
---|---|---|---|---|
ResNet50 | Mixer-B16 | URIE | 83.76 (-3.66) | 68.94 (0.10) |
LP | 87.38 (-0.04) | 69.91 (1.07) | ||
DynTTA | 86.55 (-0.87) | 71.86 (3.02) | ||
DeiT | URIE | 81.65 (-3.51) | 64.45 (-2.37) | |
LP | 85.13 (-0.03) | 67.25 (0.43) | ||
DynTTA | 84.25 (-1.01) | 67.81 (1.09) | ||
Mixer-B16 | ResNet50 | URIE | 80.80 (-1.01) | 52.99 (4.59) |
LP | 81.74 (0.03) | 50.67 (2.27) | ||
DynTTA | 82.03 (0.32) | 53.29 (4.89) | ||
DeiT | URIE | 84.35 (-0.81) | 67.61 (0.79) | |
LP | 85.13 (-0.03) | 67.25 (0.43) | ||
DynTTA | 85.02 (-0.14) | 69.29 (2.47) | ||
DeiT | ResNet50 | URIE | 79.85 (-1.86) | 51.51 (3.11) |
LP | 81.72 (0.01) | 49.95 (1.55) | ||
DynTTA | 81.60 (-0.11) | 53.23 (4.83) | ||
Mixer-B16 | URIE | 86.07 (-1.35) | 71.35 (2.51) | |
LP | 87.37 (-0.05) | 69.55 (0.71) | ||
DynTTA | 87.14 (-0.28) | 72.77 (3.93) |
IV-D5 Effects of Augmentation Space
In this study, the 14 data augmentations in Table II were used for DynTTA. We investigated the contribution of each data augmentation to classification accuracy in the non-blind setting using ResNet50 on the CUB dataset. Fig. 3 shows the accuracy when each augmentation was excluded from DynTTA before training. All data augmentations, except Equalize, contribute to improving robustness with almost no reduction in clean accuracy. In particular, URIE significantly contributed to improving robustness, but it tends to degrade clean accuracy. DynTTA blends URIE with other data augmentations to maintain clean accuracy. This result indicates that each data augmentation contributes differently to robustness, with some being effective and others being ineffective. Effective data augmentations can be predefined when the type of distribution shift is known in advance, as in the non-blind setting. For example, we estimate that Sharpness is effective in environments where blurring often occurs.

IV-D6 Effects of Magnitude Range
DynTTA requires predefinition of the magnitude range . We investigated the effect of different magnitude ranges on classification accuracy. In the experiments in the blind setting here, we also experimented with the stylized CUB dataset [7] in addition to the corruption dataset. The ranges in Table II multiplied by and are referred to as “Large” and “Small”. The results are shown in Table XI. In the blind setting, “Small” shows a better result for the corruption dataset and “Large” for the stylized dataset. The clean accuracy of “Large” is a little lower than the others. A large magnitude range is observed to be effective for large distribution shifts such as stylized datasets, but large image transformations degrade clean accuracy. In the non-blind setting, “Small” is more robust than the standard range. This may have been caused by “Small” having a smaller search space and the optimization being more stable.
Non-blind | ||
---|---|---|
Range | Clean | Unseen |
Small | 81.58 (0.00) | 58.41 (0.39) |
Large | 81.26 (-0.32) | 57.80 (-0.22) |
Blind | |||
---|---|---|---|
Range | Clean | Corruption | Stylized |
Small | 81.65 (0.00) | 52.62 (0.65) | 18.35 (0.19) |
Large | 81.61 (-0.04) | 52.32 (0.35) | 18.50 (0.34) |
IV-D7 Visualization of DynTTA Output
The output of DynTTA in the non-blind setting is visualized. Fig. 4 shows augmented images and output image by DynTTA for level-five Unseen corruptions. The bottom two rows of augmented images are almost black, except for the bottom right, so it is hard to see what they show, but this is because high-pass filters remove low-frequency domains that are visible to humans. Speckle Noise image is denoised and Gaussian Blur image is sharpened. The corruptions in the Spatter and Saturate images have not been removed, but the shape and texture of the object are enhanced.

V Limitations
Although DynTTA significantly improves robustness with almost no reduction in clean accuracy, it has computational overhead. First, to discuss network overhead, Table XII presents the number of parameters and multiply-accumulate operations (MACs). DynTTA exhibits nearly identical numbers of parameters and MACs as URIE, yet it significantly improves performance and achieves high robustness with fewer overheads than ResNet152. Moreover, URIE (Large), which increases the number of layers in URIE, shows no performance improvement. These results indicate that data augmentations integrated by DynTTA are much more effective with little overhead than increasing model size. Next, the computational cost of performing many data augmentation runs is discussed. The Appendix presents a technique designed to decrease the number of data augmentation runs required. While we acknowledge that we have not completely solved this issue, which remains a limitation of our work, we maintain that it does not affect the practical usability of DynTTA. As an example, utilizing the idle resources of edge devices for image enhancement effectively mitigates the concern over computational expenses [30].
Classifier | Enhancer | Clean | Corruption | Params (M) | MACs (G) |
---|---|---|---|---|---|
ResNet50 | – | 81.71 | 48.58 | 25.56 | 4.14 |
URIE | 80.68 | 51.47 | 26.23 | 7.16 | |
URIE (Large) | 80.61 | 51.47 | 37.43 | 9.96 | |
DynTTA | 81.73 | 53.00 | 28.77 | 7.22 | |
ResNet152 | – | 82.12 | 49.69 | 60.19 | 11.61 |
VI Conclusion
In this paper, we proposed DynTTA, a novel image enhancement method based on differentiable data augmentation techniques and image blending. DynTTA uses a gradient descent algorithm to find magnitude parameters and blend weights from a huge augmentation space. This augmentation space includes deep neural networks (DNNs) such as URIE as well as standard data augmentations. Image enhancement with DNN and data augmentation can learn diverse transformations and thus avoid overfitting. By applying DynTTA before inference by the pretrained classification model, DynTTA improves robustness while maintaining clean accuracy. In addition to the existing scenario evaluated under the assumption of knowing the type of test-time distribution shift, we introduced a practical training and evaluation scenario that does not assume the type of test-time distribution shifts. Our experimental results show that DynTTA outperforms existing methods and is effective in practical settings. Furthermore, DynTTA estimates effective training-time augmentations for the distribution-shifted datasets and shows that retraining with estimated augmentations significantly improves accuracy. However, the overhead during DynTTA inference has not been completely solved and is a limitation of this study. In this study, we experimented only with the image classification task. We will apply and evaluate DynTTA to other practical tasks such as object detection and segmentation. Future work includes exploring the optimal backbone network architecture for DynTTA and adding differentiable data augmentation to further improve robustness.
References
- [1] Arsenii Ashukha, Andrei Atanov, and Dmitry Vetrov. Mean embeddings with test-time data augmentation for ensembling of representations. arXiv preprint arXiv:2106.08038, 2021.
- [2] Christian Clausner, Apostolos Antonacopoulos, and Stefan Pletschacher. Icdar2019 competition on recognition of documents with complex layouts - rdcl2019. In International Conference on Document Analysis and Recognition (ICDAR), 2019.
- [3] Steven Diamond, Vincent Sitzmann, Frank Julca-Aguilar, Stephen Boyd, Gordon Wetzstein, and Felix Heide. Dirty pixels: Towards end-to-end image processing and perception. ACM Transactions on Graphics (TOG), 2021.
- [4] Shohei Enomoto. Entprop: High entropy propagation for improving accuracy and robustness. In Conference on Uncertainty in Artificial Intelligence (UAI), 2024.
- [5] Benjamin Erichson, Soon Hoe Lim, Winnie Xu, Francisco Utrera, Ziang Cao, and Michael Mahoney. NoisyMix: Boosting model robustness to common corruptions. In International Conference on Artificial Intelligence and Statistics (AISTATS), 2024.
- [6] Jin Gao, Jialing Zhang, Xihui Liu, Trevor Darrell, Evan Shelhamer, and Dequan Wang. Back to the source: Diffusion-driven test-time adaptation. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023.
- [7] Robert Geirhos, Patricia Rubisch, Claudio Michaelis, Matthias Bethge, Felix A Wichmann, and Wieland Brendel. Imagenet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. In International Conference on Learning Representations (ICLR), 2019.
- [8] Ishaan Gulrajani and David Lopez-Paz. In search of lost domain generalization. In International Conference on Learning Representations (ICLR), 2020.
- [9] Ryuichiro Hataya, Jan Zdenek, Kazuki Yoshizoe, and Hideki Nakayama. Faster AutoAugment: Learning Augmentation Strategies using Backpropagation. In European Conference on Computer Vision (ECCV), 2020.
- [10] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
- [11] Dan Hendrycks, Steven Basart, Norman Mu, Saurav Kadavath, Frank Wang, Evan Dorundo, Rahul Desai, Tyler Zhu, Samyak Parajuli, Mike Guo, et al. The many faces of robustness: A critical analysis of out-of-distribution generalization. In IEEE/CVF International Conference on Computer Vision (ICCV), 2021.
- [12] Dan Hendrycks and Thomas Dietterich. Benchmarking neural network robustness to common corruptions and perturbations. In International Conference on Learning Representations (ICLR), 2019.
- [13] Dan Hendrycks, Norman Mu, Ekin D. Cubuk, Barret Zoph, Justin Gilmer, and Balaji Lakshminarayanan. AugMix: A simple data processing method to improve robustness and uncertainty. In International Conference on Learning Representations (ICLR), 2020.
- [14] Dan Hendrycks, Kevin Zhao, Steven Basart, Jacob Steinhardt, and Dawn Song. Natural adversarial examples. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021.
- [15] Dan Hendrycks, Andy Zou, Mantas Mazeika, Leonard Tang, Bo Li, Dawn Song, and Jacob Steinhardt. Pixmix: Dreamlike pictures comprehensively improve safety measures. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022.
- [16] Md Tahmid Hossain, Shyh Wei Teng, Ferdous Sohel, and Guojun Lu. Robust image classification using a low-pass activation function and dct augmentation. IEEE Access, 2021.
- [17] Md Tahmid Hossain, Shyh Wei Teng, Dengsheng Zhang, Suryani Lim, and Guojun Lu. Distortion robust image classification using deep convolutional neural network with discrete cosine transform. In IEEE International Conference on Image Processing (ICIP), 2019.
- [18] Andrew Howard, Mark Sandler, Grace Chu, Liang-Chieh Chen, Bo Chen, Mingxing Tan, Weijun Wang, Yukun Zhu, Ruoming Pang, Vijay Vasudevan, et al. Searching for mobilenetv3. In IEEE/CVF International Conference on Computer Vision (ICCV), 2019.
- [19] Zhenglin Huang, Xiaoan Bao, Na Zhang, Qingqi Zhang, Xiao mei Tu, Biao Wu, and Xi Yang. Ipmix: Label-preserving data augmentation method for training robust classifiers. In Advances in Neural Information Processing Systems (NeurIPS), 2023.
- [20] Ildoo Kim, Younghoon Kim, and Sungwoong Kim. Learning loss for test-time augmentation. In Advances in Neural Information Processing Systems (NeurIPS), 2020.
- [21] Masanari Kimura. Understanding test-time augmentation. In International Conference on Neural Information Processing (ICONIP), 2021.
- [22] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. In International Conference on Learning Representations (ICLR), 2015.
- [23] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems (NeurIPS), 2012.
- [24] Hansang Lee, Haeil Lee, Helen Hong, and Junmo Kim. Test-time mixup augmentation for uncertainty estimation in skin lesion diagnosis. In Medical Imaging with Deep Learning (MIDL), 2021.
- [25] Younkwan Lee, Jihyo Jeon, Yeongmin Ko, Byunggwan Jeon, and Moongu Jeon. Task-driven deep image enhancement network for autonomous driving in bad weather. In IEEE International Conference on Robotics and Automation (ICRA), 2021.
- [26] Da Li, Yongxin Yang, Yi-Zhe Song, and Timothy M Hospedales. Deeper, broader and artier domain generalization. In IEEE/CVF International Conference on Computer Vision (ICCV), 2017.
- [27] Yingwei Li, Qihang Yu, Mingxing Tan, Jieru Mei, Peng Tang, Wei Shen, Alan Yuille, et al. Shape-texture debiased neural network training. In International Conference on Learning Representations (ICLR), 2020.
- [28] Yonggang Li, Guosheng Hu, Yongtao Wang, Timothy M. Hospedales, Neil Martin Robertson, and Yongxin Yang. DADA: differentiable automatic data augmentation. In European Conference on Computer Vision (ECCV), 2020.
- [29] Ding Liu, Bihan Wen, Xianming Liu, Zhangyang Wang, and Thomas S. Huang. When image denoising meets high-level vision tasks: A deep learning approach. In International Joint Conference on Artificial Intelligence (IJCAI), 2018.
- [30] Yan Lu, Shiqi Jiang, Ting Cao, and Yuanchao Shu. Turbo: Opportunistic enhancement for edge video analytics. In ACM Conference on Embedded Networked Sensor Systems, 2022.
- [31] Alexander Lyzhov, Yuliya Molchanova, Arsenii Ashukha, Dmitry Molchanov, and Dmitry Vetrov. Greedy policy search: A simple baseline for learnable test-time augmentation. In Conference on Uncertainty in Artificial Intelligence (UAI), 2020.
- [32] Jieru Mei, Yucheng Han, Yutong Bai, Yixiao Zhang, Yingwei Li, Xianhang Li, Alan Yuille, and Cihang Xie. Fast advprop. In International Conference on Learning Representations (ICLR), 2022.
- [33] Eric Mintun, Alexander Kirillov, and Saining Xie. On interaction between augmentations and corruptions in natural corruption robustness. In Advances in Neural Information Processing Systems (NeurIPS), 2021.
- [34] Takayuki Okatani, Xing Liu, and Masanori Suganuma. Improving generalization ability of deep neural networks for visual recognition tasks. In Computational Color Imaging Workshop (CCIW), 2019.
- [35] Tianyu Pang, Kun Xu, and Jun Zhu. Mixup inference: Better exploiting mixup to defend adversarial attacks. In International Conference on Learning Representations (ICLR), 2019.
- [36] Yanting Pei, Yaping Huang, Qi Zou, Yuhang Lu, and Song Wang. Does haze removal help cnn-based image classification? In European Conference on Computer Vision (ECCV), 2018.
- [37] Yanting Pei, Yaping Huang, Qi Zou, Xingyuan Zhang, and Song Wang. Effects of image degradation and degradation removal to cnn-based image classification. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2021.
- [38] Juan C Perez, Motasem Alfarra, Guillaume Jeanneret, Laura Rueda, Ali Thabet, Bernard Ghanem, and Pablo Arbelaez. Enhancing adversarial robustness via test-time transformation ensembling. In IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), 2021.
- [39] Yongming Rao, Wenliang Zhao, Zheng Zhu, Jiwen Lu, and Jie Zhou. Global filter networks for image classification. In Advances in Neural Information Processing Systems (NeurIPS), 2021.
- [40] Edgar Riba, Dmytro Mishkin, Daniel Ponsa, Ethan Rublee, and Gary Bradski. Kornia: an open source differentiable computer vision library for pytorch. In IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), 2020.
- [41] Cédric Rommel, Thomas Moreau, and Alexandre Gramfort. Deep invariant networks with differentiable augmentation layers. In Advances in Neural Information Processing Systems (NeurIPS), 2022.
- [42] Evgenia Rusak, Lukas Schott, Roland S Zimmermann, Julian Bitterwolf, Oliver Bringmann, Matthias Bethge, and Wieland Brendel. A simple way to make neural networks robust against diverse image corruptions. In European Conference on Computer Vision (ECCV), 2020.
- [43] Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, et al. Imagenet large scale visual recognition challenge. International journal of computer vision (IJCV), 2015.
- [44] Divya Shanmugam, Davis Blalock, Guha Balakrishnan, and John Guttag. Better aggregation in test-time augmentation. In IEEE/CVF International Conference on Computer Vision (ICCV), 2021.
- [45] Vivek Sharma, Ali Diba, Davy Neven, Michael S. Brown, Luc Van Gool, and Rainer Stiefelhagen. Classification-driven dynamic image enhancement. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2018.
- [46] Ryan Soklaski, Michael Yee, and Theodoros Tsiligkaridis. Fourier-based augmentations for improved robustness and uncertainty calibration. In Workshop on Distribution Shifts: Connecting Methods and Applications (DistShift), 2021.
- [47] Taeyoung Son, Juwon Kang, Namyup Kim, Sunghyun Cho, and Suha Kwak. Urie: Universal image enhancement for visual recognition in the wild. In European Conference on Computer Vision (ECCV), 2020.
- [48] Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, and Andrew Rabinovich. Going deeper with convolutions. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2015.
- [49] Hossein Talebi and Peyman Milanfar. Learning to resize images for computer vision tasks. In IEEE/CVF International Conference on Computer Vision (CVPR), 2021.
- [50] Mingxing Tan and Quoc Le. Efficientnet: Rethinking model scaling for convolutional neural networks. In International Conference on Machine Learning (ICML), 2019.
- [51] Ilya O Tolstikhin, Neil Houlsby, Alexander Kolesnikov, Lucas Beyer, Xiaohua Zhai, Thomas Unterthiner, Jessica Yung, Andreas Steiner, Daniel Keysers, Jakob Uszkoreit, et al. Mlp-mixer: An all-mlp architecture for vision. In Advances in Neural Information Processing Systems (NeurIPS), 2021.
- [52] Hugo Touvron, Matthieu Cord, Matthijs Douze, Francisco Massa, Alexandre Sablayrolles, and Herve Jegou. Training data-efficient image transformers & distillation through attention. In International Conference on Machine Learning (ICML), 2021.
- [53] Theodoros Tsiligkaridis and Athanasios Tsiligkaridis. Diverse gaussian noise consistency regularization for robustness and uncertainty calibration. In International Joint Conference on Neural Networks (IJCNN), 2023.
- [54] Puru Vaish, Shunxin Wang, and Nicola Strisciuglio. Fourier-basis functions to bridge augmentation gap: Rethinking frequency augmentation in image classification. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2024.
- [55] Cristina Vasconcelos, Hugo Larochelle, Vincent Dumoulin, Rob Romijnders, Nicolas Le Roux, and Ross Goroshin. Impact of aliasing on generalization in deep convolutional networks. In IEEE/CVF International Conference on Computer Vision (ICCV), 2021.
- [56] Haotao Wang, Chaowei Xiao, Jean Kossaifi, Zhiding Yu, Anima Anandkumar, and Zhangyang Wang. Augmax: Adversarial composition of random augmentations for robust training. In Advances in Neural Information Processing Systems (NeurIPS), 2021.
- [57] P. Welinder, S. Branson, T. Mita, C. Wah, F. Schroff, S. Belongie, and P. Perona. Caltech-UCSD Birds 200. Technical report, California Institute of Technology, 2010.
- [58] Cihang Xie, Mingxing Tan, Boqing Gong, Jiang Wang, Alan L Yuille, and Quoc V Le. Adversarial examples improve image recognition. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020.
- [59] Xulei Yang, Zeng Zeng, Sin G Teo, Li Wang, Vijay Chandrasekhar, and Steven Hoi. Deep learning for practical image recognition: Case study on kaggle competitions. In ACM SIGKDD international conference on knowledge discovery & data mining (KDD), 2018.
- [60] Dong Yin, Raphael Gontijo Lopes, Jon Shlens, Ekin Dogus Cubuk, and Justin Gilmer. A fourier perspective on model robustness in computer vision. In Advances in Neural Information Processing Systems (NeurIPS), 2019.
- [61] Jinghao Zhang, Zhenhua Feng, Guosheng Hu, Changbin Shao, and Yaochu Jin. Mixprop: Towards high-performance image recognition via dual batch normalisation. In British Machine Vision Conference (BMVC), 2022.
- [62] Xueyan Zou, Fanyi Xiao, Zhiding Yu, and Yong Jae Lee. Delving deeper into anti-aliasing in convnets. In British Machine Vision Conference (BMVC), 2020.
Shohei Enomoto received the B.S. degree from Tohoku University in 2014 and the M.S. degree from Tokyo Institute of Technology in 2016. Since 2016, he has been engaged in researching deep learning and computer vision at NTT. |
Monikka Roslianna Busto is a Researcher at NTT Software Innovation Center. She graduated from the Electrical and Electronics Engineering Institute, College of Engineering, the University of the Philippines in 2017, and received a master’s degree from the Department of Information and Communications Engineering, Tokyo Institute of Technology in 2021. She joined Nippon Telegraph and Telephone Corporation in the same year. Her research interests include computer vision, collaborative intelligence for edge computing, remote sensing image analysis and multi-modal AI. |
Takeharu Eda received his B.S. in Mathematics from Kyoto University in 2001 and his M.S. in Engineering from the Nara Institute of Science and Technology in 2003. He joined NTT Laboratories in 2003, focusing on research in various aspects of machine learning and systems. In 2011, he transitioned to NTT Communications, where he launched a web hosting service utilizing CloudStack/OpenStack-based infrastructure and migration tools. He managed international development teams with members from the US, Germany, and Japan. In 2015, he moved to the NTT Software Innovation Center, developing a scalable surveillance video system using deep learning-based computer vision techniques. Since 2022, he has been involved in the research and development of a space computing project with SpaceCompass. Eda is a member of the Information Processing Society of Japan (IPSJ), the Association for Computing Machinery (ACM), and IEEE. |