Test-time Fourier Style Calibration for Domain Generalization
Abstract
The topic of generalizing machine learning models learned on a collection of source domains to unknown target domains is challenging. While many domain generalization (DG) methods have achieved promising results, they primarily rely on the source domains at train-time without manipulating the target domains at test-time. Thus, it is still possible that those methods can overfit to source domains and perform poorly on target domains. Driven by the observation that domains are strongly related to styles, we argue that reducing the gap between source and target styles can boost models’ generalizability. To solve the dilemma of having no access to the target domain during training, we introduce Test-time Fourier Style Calibration (TF-Cal) for calibrating the target domain style on the fly during testing. To access styles, we utilize Fourier transformation to decompose features into amplitude (style) features and phase (semantic) features. Furthermore, we present an effective technique to Augment Amplitude Features (AAF) to complement TF-Cal. Extensive experiments on several popular DG benchmarks and a segmentation dataset for medical images demonstrate that our method outperforms state-of-the-art methods.
1 Introduction
Over the past few years, deep convolutional neural networks (CNNs) have been shown to perform excellently in a variety of computer vision applications. However, CNNs have demonstrated a drastic performance drop when both the training (source domain) and testing (target domain) data come from different distributions. This incapability causes numerous problems in real-world fields such as autonomous driving and medical imaging analysis. As a consequence, domain adaptation (DA) has been extensively researched as a mechanism of transferring knowledge from source domains to labeled/unlabeled target domains. Yet, DA approaches rely on known target domains at training time, which fails to generalize well to unknown testing target domains. Domain generalization (DG) has lately attracted much attention as a more challenging setting for enhancing CNNs’ generalizability to unknown testing data. Specifically, DG is to train models that are resistant to distribution shifts between domains (i.e., domain shifts) so that models can generalize well when evaluated on any previously unseen target domain, provided that several source domains are available for training. To address the problem, various methods were proposed and achieved promising results, including adversarial learning Zhou et al. (2020); Sicilia et al. (2021), meta learning Li et al. (2018a); Balaji et al. (2018), and data augmentation Xu et al. (2021); Zhou et al. (2021). Among these, approaches based on data augmentation have gained popularity due to their simplicity and effectiveness, which can expose models to a wide variety of source domains to learn generalizable representation. However, most existing data augmentation approaches cannot perform well on test data that are out of the augmented source data distribution. This begs the question of how we can reduce domain gaps when the augmented source domains are insufficient to cover the target domain.
In this paper, we propose a novel yet simple method called Test-time Fourier Style Calibration (TF-Cal) to calibrate the target style features at inference-time. Our method is driven by the assumption that the lack of generalizability is due to the difference in styles of images (e.g., sketches, cartoon, photo) between the source and target Zhou et al. (2021). Based on this idea, we employ Fourier transformation to decompose latent features from early CNN layers into amplitude features and phase features. The amplitude encodes the intensities (style information), while the phase encodes the spatial positions (semantic information) Piotrowski and Campbell (1982). To achieve test-time style calibration, we propose to convexly calibrate the target amplitude features using the source amplitude prototype estimated at train-time. Hence we narrow the style gap between the target and source domains on the fly when evaluating, while retaining the key semantic feature for classification. We additionally show that increasing the diversity of features can complement TF-Cal to further boost models’ generalizability based on the convex hull theory proposed by Albuquerque et al. (2019). Inspired by Xu et al. (2021) which augments amplitude in image space, we propose augmenting amplitude in feature space (AAF) that diversifies higher-level styles, while preserving semantics. Then we have the Test-time Augmentated Fourier Style Calibration (TAF-Cal) by combining TF-Cal and AAF. Moreover, our algorithm does not require domain labels, allowing for more data collection flexibility.
The proposed work is evaluated on multiple DG benchmarks with different backbones on image classification and medical image segmentation. Through extensive experiments, we demonstrate that our method significantly improves the generalizability of CNNs and outperforms multiple state-of-the-art methods. We also show that TF-Cal is useful in conjunction with augmentation techniques. Finally, we present qualitative analysis and quantitative ablation studies to verify the effectiveness of our method.
2 Related Work
2.1 Domain Generalization
DG works in recent years can be roughly categorized into: domain-invariant learning, meta-learning and data augmentation approaches.
Domain-invariant learning aims to map input data to features that are invariant to source domain shift in order to be resilient to unseen target domain shift. The domain adversarial neural network (DANN) Ganin et al. (2016), which was originally developed for DA, consists of a domain classifier trained to discriminate between the source and target distributions, and a feature extractor trained to learn features that are not only beneficial for the task (e.g., classifcation) but also capable of confusing the domain classifier. Albuquerque et al. (2019); Sicilia et al. (2021) leverage the DANN to align the representation of source domains to learn the domain invariance in DG. Li et al. (2018b) propose to use Maximum Mean Discrepancy with adversarial autoencoder to align domain distributions, mapping the aligned distribution to a prior distribution adversarially. Another line of works focus on minimizing contrastive loss to learn domain-invariant representation. MASF Dou et al. (2019) minimizes contrastive losses with positive pair of same class and negative pair of different class.
Meta-learning based approach MLDG Li et al. (2018a) simulates domain shift by dividing data from multiple sources into meta-train and meta-test sets. Balaji et al. (2018) propose a meta-regularizer learned to be robust to domain shift by meta-learning. Dou et al. (2019) regularize semantic structure of features while optimizing models by meta-learning. While meta-learning methods can simulate the domain shift during training by dividing source domains to meta-train and meta-test sets, it is still likely that the simulated domain shift cannot cover the target domain shift.
Data augmentation method is most related to our work, attempting to augment the training images or features to enhance the generalization performance on unseen data. Zhou et al. (2021) probabilistically mix the instance statistics of training features. Xu et al. (2021) propose to augment amplitude of images with a regularization to form consistency between original and augmented images. To generate diverse novel training images, Zhou et al. (2020) design an image generation network with domain-adversarial training plus optimal transport based metrics. A concurrent work Zhang et al. (2021) has a similar motivation as ours to augment/modify target data during evaluation, developing a multi-view regualizeried meta-learning that encourages test-time augmentation with multi-view. Although these methods achieve favorable results on unseen targets, they might fail when the augmented source cannot cover the unseen target. Our method mitigates this issue by calibrating the target with a source prototype, allowing the calibrated target to be covered by the source.

3 Methodology
3.1 Problem Definition
We define and as the input and label space. Given source domains that are available at train-time, we indicate that each contains input and label pairs where and is the number of inputs in the source domain . The goal of DG is to train a function learned from data of to perform well on an unseen target domain which share the same label space as . The function can be defined as , where is a task-specific network parameterized by and which is a feature encoder with layers parameterized by .
3.2 Background
To reduce the domain shift, we need to first understand the difference between two distributions which are measured by -divergence proposed by Ben-David et al. (2009):
(1) |
where classifier . Following that, Albuquerque et al. (2019) defines the convex hull of that is a set of mixture of source domain distributions:
(2) |
where the represents non-negative coefficient in the -dimensional simplex . Then, an ideal case is assumed that the ideal target lies in the source domain convex hull . Under this assumption, the risk on the target domain is upper-bounded Albuquerque et al. (2019) by:
(3) |
where, on the right hand side, the first term corresponds to the risks over all source domains, and the second term is the -divergence between the ideal target and the real target , and the third term which is the largest -divergence between any pair of source domains. corresponds to . The first term can be minimized by empirical risk minimization (ERM), and the second term is hard to minimized due to having no access to the target domain at training, and the third term can be minimized by removing the source domain-discrimintive information which is the style information in the context of this paper. According to Cha et al. (2021), ERM is a strong baseline to learn source domain invariance compared to other methods such as DANN Ganin et al. (2016).
3.3 Motivation
As the Eq.(3) states, a model can achieve acceptable performance on the target domain by using ERM to train on the source domains when the ideal target domain is assumed to be covered by source convex hull . However, in reality, this assumption typically does not hold. This is the reason why domain-adversarial based methods Zhou et al. (2020); Albuquerque et al. (2019) may fail in DG as they only reduce source divergence. Therefore, augmenting images or features of source domains can possibly extend the (i.e., ), which motivating a variety of works Zhou et al. (2021); Sicilia et al. (2021) to diversify training data. Still, as illustraed in Fig.1b, it is likely that the augmented source distributions (dashed circles) may not accurately reflect the previously unseen target domain (red rhombi). This difficulty arises mostly due to the absence of target data during training. Instead of extending arbitrarily, we aim to pull closer to with the help from source data.
Therefore, it motivates us to calibrate the target features at test-time based on what a model learned from source features. By doing so, we pull the target closer to the source convex hull , and reduce the difference between and . To further encourage the reduction in source-target divergence, we augment the source features to expand the convex hull.
3.4 Method Overview
Our method TF-Cal is to calibrate style features. To this end, we use Fourier transformation to decompose features into phase features that contain semantics and amplitude features that contain styles. Then we let a model to learn with a style calibrating function (TF-Cal) for reducing the gap between the target domain and source domains prototype. We further introduce an augmenting function (AAF) for expanding the source convex hull by diversifying style features. The model then is trained in a two-stage manner: in the early stage, we train the model with AAF to learn a great diversity of style features, and in the late stage, we train the model with both AAF and TF-Cal (i.e., TAF-Cal) that uses the estimated source prototype from diverse styles. During testing, we leverage the learned to calibrate target styles. Our method is easy to implement as a plug-and-play module without adding any additional network. The overall pipeline is illustrated in Fig.1a and TF-Cal is illustrated in Fig.1b, and the detail of the method will be explained as follows.
3.5 Fourier Transformation of Features
Given a latent feature tensor with computed from -th layer of the feature extractor, we want to extract style information from the feature. In our implementation, we choose the early layers since they can extract style information well Zhou et al. (2021). Fourier transformation , which can be implemented using Fast Fourier transformation (FFT) Nussbaumer (1982), is able to decompose the feature into the phase spectrum encoding the spatial information and amplitude spectrum encoding the intensity information. As illustrated in Fig.1, the amplitude component of the horse features depicts the intensity of texture, while phase component preserves the shape of object. Thus, it is natural to associate the amplitude with styles and the phase with semantics. Since our method only modifies the amplitude, the semantic information in the phase will not be affected.
The Fourier transformation is done on each channel of :
(4) |
where represents the value of the features in the frequency space corresponding to the coordinates and . With , we can retrieve the amplitude and phase features with real and imaginary parts:
(5) | ||||
To reconstruct phase and amplitude spectrum features back into the spatial feature space, we can use the inverse Fourier Transformation .
3.6 Test-time Fourier Style Calibration
Estimating Source Prototype
The first step in our algorithm TF-Cal is to estimate the source domain prototype that will be used for the calibrating function . The can be estimated by taking the average over the amplitude features of all data points from source domains along the batch dimension:
(6) |
Since the is estimated via all training samples, it will be a proper representation of the centroid of source styles. We leverage as the anchor to pull a style closer to the source convex hull . In practice, we store the from each mini-batch in a memory bank, and compute the mean of all amplitude features from the bank at the end of one epoch. We will verify calibrating style with source prototype instead of a random prototype in the ablation studies section.
Training with the Style Calibrating Function
The TF-Cal is proposed to reduce the divergence between source and target styles during testing. We need to first train a model with a calibrating function with the input of the source prototype and source amplitude features . Therefore, we let the model’s parameters be associated with the source prototype that is the common information shared by the target and source styles. The is used to train the model by linearly interpolating between and the :
(7) |
where is the hyperparameter to control the strength of calibration. The calibrating function is activated with a probability during training. Before the training is finished, we store the source prototype from the final epoch for the usage of calibrating target style at test-time.
Calibrating the Target Style on the Fly
During the evaluation, we leverage the function to calibrate the target style with the source prototype :
(8) |
where the is the hyperparamter to control the strength of calibration during testing. The reason for controlling is that, due to CNNs’ inductive biases toward styles Geirhos et al. (2019), target amplitude may contain useful style information for classification. We set the calibration probability to 1 during testing. Our method requires only one test sample for calibration which is realistic in practical applications. Through calibration, the target style is pulled closer to the source style, helping the model to achieve a similar performance on target data as it does on source data. The advantage of our method is that we let the target and source styles share the that is controllable and at hand. Thus, we can offset the source-target divergence with the guidance from the source prototype at test-time. Instead, other DG methods only rely on the source data, hoping to cover the target distribution or learn domain-invariance that overfits to source domains during training.
3.7 Augmenting Amplitude Features
As we can see from Eq.(7) and Eq.(8), the common part of the input is the and the different part is the amplitude features and . Therefore, to further boost the generalizability with TF-Cal, we want and to be as similar as possible. An effective way is to augment amplitude in feature space (AAF) hoping to expand the source convex hull to decrease the source-target divergence. Therefore, based on the strategy of mixing and swapping amplitude of images Xu et al. (2021), we propose to mix and swap amplitude feature pairs () randomly from any source domains:
(9) |
where . For swapping amplitude between two samples, we set to 0. Compared to Xu et al. (2021) that augment styles in the image space, AAF utilizes higher level style information from features with much more channels, which provides additional statistics for diversifying styles. In our implementation, we choose amplitude features from early layer of feature extractors since they encode style information about the inputs. Our results in the ablation studies demonstrate that augmenting amplitude on feature is better than on image, which empirically verify that AAF helps TF-Cal.
4 Experiments
4.1 Datasets and Settings
Our method is evaluated on three DG benchmarks: (1) PACS Li et al. (2017) is an object recognition dataset that contains 4 different domains (i.e., Photo, Art-Painting, Cartoon, Sketch), including 9,991 images and 7 classes. We follow the train-val split provided by Li et al. (2017). The domain shift is primarily contributed by the style difference. (2) Office-Home Venkateswara et al. (2017) is an object recognition dataset that contains 4 different domains (i.e., Art, Clipart, Product, Real-World), including 15,550 images and 65 classes. The dataset is split randomly into 90% training set and 10% validation set. The domain shift is primarily contributed by the style and viewpoint differences, with less diversity than PACS. (3) MICCAI WMH Challenge Kuijf et al. (2019) is a medical image segmentation dataset for White Matter Hyperintensity that contains 3 different domains (i.e., Utrecht, Amsterdam, Singapore), including 20 subjects in each domain. For each subject, there are 47 slices in Utrecht and Singapore, and 82 slices in Amsterdam. The domain shift is mainly due to scanner and protocol differences.
For evaluation, we use the leave-one-domain-out protocol. In other words, the model is trained on source domains, and is evaluated on a held-out target domain. For PACS and OfficeHome, we report the classification accuracy. For WMH Challenge, we report the Dice Similarity Coefficient (DSC) (higher is better). All results are averaged over 3 runs with different random seeds. We use DeepAll as the ERM baseline with no DG mechanisms.
4.2 Implementation Details
(1) For PACS and OfficeHome, we employ the ResNet He et al. (2016) pretrained on the ImageNet as the backbone. The networks are trained with the learning rate of 1e-3 for feature extractors, the learning rate of 1e-4 for classifiers, batch size of 64, and epochs of 50. We use the features from the second residual block as the input for AAF and TF-Cal. We adopt the same data augmentation as Matsuura and Harada (2020). (2) For WMH Challenge, we use the U-Net Ronneberger et al. (2015) trained from scratch as the backbone with the the learning rate of 2e-4, batch size of 30, and epochs of 300. We use the features from the second downsamling layer as the input for our method. We randomly augment each training slices by rotation, scaling, and shearing. All networks are optimized by SGD with weight decay of 5e-4, and all learning rates are decayed by 0.1 after 80% of the epochs. For all experiments: For TF-Cal, we set the calibration strength and to 0.5, and set the to 0.5. For AAF activated with a probability of 0.5, we set . The AAF is applied for all epochs, and TF-Cal is trained after 70% of the epochs.
4.3 Evaluation on PACS
From the results in Table 1, our method outperforms the SOTA methods on both ResNet-18 and ResNet-50, improving the DeepAll by a large margin, which demonstrates the effectiveness of our method on unseen target domains. In comparison to the meta-learning methods MLDG Li et al. (2018a) and MASF Dou et al. (2019), as well as the domain-adversarial method MMLD Matsuura and Harada (2020), our method clearly outperforms them significantly with simple feature transformations and no additional network. With the advantage of alleviating the overfit to source domains, our method also surpasses data augmentation methods DDAIG Zhou et al. (2020) and MixStyle Zhou et al. (2021), as well as a more recent method, RSC Huang et al. (2020). FACT Xu et al. (2021) is most related to our work in terms of Fourier transformation and augmentation. Still, our method outperforms FACT: (1) TF-Cal can calibrate target style at test time, which solves the problem of training with no knowledge of target domains. (2) AAF can encourage more diverse styles of source domains by providing additional signals in higher-level style information via features from early CNN layers. MVRML Zhang et al. (2021) is relevant to our work in terms of test-time augmentation because it relies on augmenting target images with multiple views. Our method beats MVRML since we can calibrate target images with guidance from the source prototype.
4.4 Evaluation on Office-Home
As Table 2 shows, our method has the best performance. For this dataset, we have similar observations and conclusions as the PACS. Our method shows improvement consistently over all target domains. Note that, Clipart has the most style difference compared to other domains. Our method demonstrates its ability to calibrate the target style to offset the source-target shifts, which improves the performance on Clipart by a large margin over the DeepAll.
4.5 Evaluation on WMH Challenge
We evaluate our method on the medical image segmentation task, which is important for real-world applications and is often ignored by other DG works. Our method outperforms other related methods: (1) BigAug Zhang et al. (2020) applies heavy data augmentation. (2) UDA Li et al. (2020) is an unsupervised domain adaptation method. (3) MixDANN Zhao et al. (2021) is a method that combines mixup and DANN. We notice that the DeepAll fails to generalize to the Utrecht due to the significant scanner-difference. Our method significantly improves the overall performance of Utrecht by a considerable margin compared to previous methods.
Method | Art | Cartoon | Photo | Sketch | Avg |
---|---|---|---|---|---|
ResNet-18 | |||||
DeepAll | 77.95 | 77.92 | 95.68 | 69.56 | 80.28 |
MLDG [2017] | 79.50 | 77.30 | 94.30 | 71.50 | 80.65 |
MASF [2019] | 80.29 | 77.17 | 94.99 | 71.69 | 81.04 |
MMLD [2019] | 81.28 | 77.16 | 96.09 | 72.29 | 81.83 |
DDAIG [2020] | 84.20 | 78.10 | 95.30 | 74.70 | 83.08 |
MixStyle [2021] | 84.10 | 78.80 | 96.10 | 75.90 | 83.73 |
FACT [2021] | 85.37 | 78.38 | 95.15 | 79.15 | 84.51 |
RSC [2020] | 83.43 | 80.31 | 95.99 | 80.85 | 85.15 |
MVRML [2021] | 84.59 | 79.22 | 95.38 | 83.55 | 85.69 |
TAF-Cal (Ours) | 85.70 | 82.55 | 96.07 | 82.60 | 86.73 |
ResNet-50 | |||||
DeepAll | 84.91 | 78.58 | 97.78 | 73.89 | 83.79 |
MASF [2019] | 82.89 | 80.49 | 95.01 | 72.29 | 82.67 |
MVRML [2021] | 87.36 | 83.12 | 96.41 | 84.11 | 87.75 |
RSC [2021] | 87.89 | 82.16 | 97.92 | 83.35 | 87.83 |
FACT [2021] | 89.63 | 81.77 | 96.75 | 84.46 | 88.15 |
TAF-Cal (Ours) | 90.53 | 84.91 | 97.84 | 83.83 | 89.28 |
Method | Art | Clipart | Product | Real | Avg |
---|---|---|---|---|---|
DeepAll | 59.93 | 48.80 | 74.13 | 75.42 | 64.57 |
RSC [2020] | 58.42 | 47.90 | 71.63 | 74.54 | 63.12 |
DDAIG [2020] | 59.20 | 52.30 | 74.60 | 76.00 | 65.53 |
MixStyle [2021] | 58.70 | 53.40 | 74.20 | 75.90 | 65.55 |
MVRML [2021] | 61.44 | 51.16 | 74.72 | 77.72 | 66.26 |
FACT [2021] | 60.34 | 54.85 | 74.48 | 76.55 | 66.56 |
TAF-Cal (Ours) | 61.52 | 54.98 | 74.87 | 76.97 | 67.09 |
Method | Utrecht | Amsterdam | Singapore | Avg |
---|---|---|---|---|
DeepAll | 0.183 | 0.619 | 0.781 | 0.528 |
BigAug [2020] | 0.534 | 0.691 | 0.711 | 0.645 |
UDA [2020] | 0.529 | 0.737 | 0.782 | 0.683 |
MixDANN [2021] | 0.694 | 0.700 | 0.839 | 0.744 |
TAF-Cal (Ours) | 0.736 | 0.732 | 0.860 | 0.776 |
AAF | TF-Cal (Train) | TF-Cal (Test) | Art | Cartoon | Photo | Sketch | Avg |
---|---|---|---|---|---|---|---|
- | - | - | 77.95 | 77.92 | 95.68 | 69.56 | 80.28 |
✓ | - | - | 85.35 | 78.63 | 96.23 | 77.58 | 84.45 |
- | ✓ | - | 80.76 | 77.22 | 96.18 | 69.33 | 80.88 |
✓ | ✓ | - | 84.60 | 78.88 | 95.23 | 77.17 | 83.97 |
- | ✓ | ✓ | 83.31 | 81.23 | 95.63 | 77.45 | 84.41 |
✓ | ✓ | ✓ | 85.70 | 82.55 | 96.07 | 82.60 | 86.73 |
TF-Cal with a Random Prototype | |||||||
- | ✓ | ✓ | 71.43 | 76.32 | 85.69 | 68.95 | 75.60 |
4.6 Further Analysis
Ablation Studies of Different Components
We investigate a thorough ablation studies of different components in our method. With only AAF, our method already outperforms most of the methods listed in Table 1. The AAF also achieves similar performance as FACT. As showed in Xu et al. (2021), FACT’s augmentation component obtains accuracy of 83.44 which is lower than our AAF, which empirically verifies the merit of AAF in generating more diverse styles in features. To prove that using TF-Cal does not serve as a way for augmenting source data, we show that only applying TF-Cal at training does not have much difference in performance with the DeepAll. The similar observation is also found when both AAF and TF-Cal (Train) are activated. When we activate TF-Cal at both train-time and test-time, we see the great improvement, which also beats most of the methods in Table 1. Moreover, we see that AAF and TF-Cal can complement each other and further boost the generalizability. Finally, we empirically conduct the experiment on the importance of the source prototype. We randomly sample a prototype from Gaussian distribution for both TF-Cal (Train) and TF-Cal (Test), and we see a significant performance drop.
Visualization of Features
To qualitatively analyze our contribution in calibrating style features, we applied t-SNE to compare the features learned by the DeepAll and our approach on PACS dataset with sketch as the target domain. From the Fig. 2, DeepAll has a satisfactory capacity for categorizing features, but it fails to align the source domain Photo and the target domain Sketch with other source domains. Specifically, AAF can well align the Sketch with the source domains, but with unclear cluster boundaries due to the mixing of source styles. TF-Cal precisely calibrates the target domain to the source domains, with distinct cluster boundaries. With the TF-Cal’s advantage of aligning styles and AAF’s advantage of diversifying styles, we observe a robust representation toward the target domain.
Hyperparameter Sensitivity
We conduct the hyperparameter experiment (Fig.3a) on the calibration stength hyperparameter and for TF-Cal with ResNet-18 and ResNet-50 on PACS. As we define in the method section, and share the same value. We choose and from . The larger and are, the more source prototype and the less target style in the target features. Both ResNet-18 and ResNet-50 have the similar trend. We find that TF-Cal performs the best when the and are equal to 0.5. When the and are smaller, the target style is less calibrated. When the and are larger, class-discriminative information in target styles might be reduced.






Where to Apply the Method
We explore applying TF-Cal+AAF after different residule blocks of ResNet-18 on PACS. From the result in Fig.3b, we see that applying our method after the second residual block performs the best, and the reason is that early layers extract more domain information than the later layers. As we can see that applying our method to the third and the fourth residual blocks, the performance drops. Because the these two blocks are closer to the classification layer, augmenting or calibrating the features will unnecessarily affect the class-discriminative features.
5 Conclusion
In this paper, we propose a test-time calibration function for target styles extracted by Fourier transformation, and a feature style augmentation technique that complements style calibration. With TF-Cal, we alleviate the stress of having no knowledge about target domains during training by calibrating target styles with the source prototype at test-time. Additionally, AAF efficiently diversifies source amplitudes using higher-level style information extracted from features to help TF-Cal, outperforming image-based amplitudes augmentation. While our method TAF-Cal can be easily integrated into any CNN network, we achieve state-of-the-art performance on two primary tasks (i.e., classification and segmentation), and we conduct extensive experiments and qualitative analysis that validate our method’s effectiveness and robustness.
References
- Albuquerque et al. [2019] Isabela Albuquerque, João Monteiro, Mohammad Javad Darvishi Bayazi, Tiago H. Falk, and Ioannis Mitliagkas. Generalizing to unseen domains via distribution matching. arXiv: Learning, 2019.
- Balaji et al. [2018] Yogesh Balaji, Swami Sankaranarayanan, and Rama Chellappa. Metareg: Towards domain generalization using meta-regularization. In NeurIPS, 2018.
- Ben-David et al. [2009] Shai Ben-David, John Blitzer, Koby Crammer, Alex Kulesza, Fernando C Pereira, and Jennifer Wortman Vaughan. A theory of learning from different domains. Machine Learning, 79:151–175, 2009.
- Cha et al. [2021] Junbum Cha, Han-Cheol Cho, Kyungjae Lee, Seunghyun Park, Yunsung Lee, and Sungrae Park. Domain generalization needs stochastic weight averaging for robustness on domain shifts. ArXiv, abs/2102.08604, 2021.
- Dou et al. [2019] Qi Dou, Daniel Coelho de Castro, Konstantinos Kamnitsas, and Ben Glocker. Domain generalization via model-agnostic learning of semantic features. In NeurIPS, 2019.
- Ganin et al. [2016] Yaroslav Ganin, E. Ustinova, Hana Ajakan, Pascal Germain, H. Larochelle, François Laviolette, Mario Marchand, and Victor S. Lempitsky. Domain-adversarial training of neural networks. In J. Mach. Learn. Res., 2016.
- Geirhos et al. [2019] Robert Geirhos, Patricia Rubisch, Claudio Michaelis, Matthias Bethge, Felix Wichmann, and Wieland Brendel. Imagenet-trained cnns are biased towards texture; increasing shape bias improves accuracy and robustness. ArXiv, abs/1811.12231, 2019.
- He et al. [2016] Kaiming He, X. Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 770–778, 2016.
- Huang et al. [2020] Zeyi Huang, Haohan Wang, Eric P. Xing, and Dong Huang. Self-challenging improves cross-domain generalization. In ECCV, 2020.
- Kuijf et al. [2019] Hugo J. Kuijf, Adrià Casamitjana, D. Louis Collins, M. Dadar, Achilleas Georgiou, and et al. Standardized assessment of automatic segmentation of white matter hyperintensities and results of the wmh segmentation challenge. IEEE Transactions on Medical Imaging, 38:2556–2568, 2019.
- Li et al. [2017] Da Li, Yongxin Yang, Yi-Zhe Song, and Timothy M. Hospedales. Deeper, broader and artier domain generalization. 2017 IEEE International Conference on Computer Vision (ICCV), pages 5543–5551, 2017.
- Li et al. [2018a] Da Li, Yongxin Yang, Yi-Zhe Song, and Timothy M. Hospedales. Learning to generalize: Meta-learning for domain generalization. ArXiv, abs/1710.03463, 2018.
- Li et al. [2018b] Haoliang Li, Sinno Jialin Pan, Shiqi Wang, and Alex Chichung Kot. Domain generalization with adversarial feature learning. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5400–5409, 2018.
- Li et al. [2020] Hongwei Li, Timo Loehr, and et al. Domain adaptive medical image segmentation via adversarial learning of disease-specific spatial patterns. arXiv: Computer Vision and Pattern Recognition, 2020.
- Matsuura and Harada [2020] Toshihiko Matsuura and Tatsuya Harada. Domain generalization using a mixture of multiple latent domains. ArXiv, abs/1911.07661, 2020.
- Nussbaumer [1982] Henri J. Nussbaumer. The fast fourier transform. 1982.
- Piotrowski and Campbell [1982] Lisa Piotrowski and Fergus William Campbell. A demonstration of the visual importance and flexibility of spatial-frequency amplitude and phase. Perception, 11:337 – 346, 1982.
- Ronneberger et al. [2015] Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-net: Convolutional networks for biomedical image segmentation. In MICCAI, 2015.
- Sicilia et al. [2021] Anthony Sicilia, Xingchen Zhao, and Seong Jae Hwang. Domain adversarial neural networks for domain generalization: When it works and how to improve. ArXiv, abs/2102.03924, 2021.
- Venkateswara et al. [2017] Hemanth Venkateswara, Jose Eusebio, Shayok Chakraborty, and Sethuraman Panchanathan. Deep hashing network for unsupervised domain adaptation. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 5385–5394, 2017.
- Xu et al. [2021] Qinwei Xu, Ruipeng Zhang, Ya Zhang, Yanfeng Wang, and Qi Tian. A fourier-based framework for domain generalization. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 14378–14387, 2021.
- Zhang et al. [2020] Ling Zhang, Xiaosong Wang, and et al. Generalizing deep learning for medical image segmentation to unseen domains via deep stacked transformation. IEEE Transactions on Medical Imaging, 39:2531–2540, 2020.
- Zhang et al. [2021] Jian Zhang, Lei Qi, Yinghuan Shi, and Yang Gao. More is better: A novel multi-view framework for domain generalization. ArXiv, abs/2112.12329, 2021.
- Zhao et al. [2021] Xingchen Zhao, Anthony Sicilia, and et al. Robust white matter hyperintensity segmentation on unseen domain. 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI), pages 1047–1051, 2021.
- Zhou et al. [2020] Kaiyang Zhou, Yongxin Yang, Timothy M. Hospedales, and Tao Xiang. Deep domain-adversarial image generation for domain generalisation. ArXiv, abs/2003.06054, 2020.
- Zhou et al. [2021] Kaiyang Zhou, Yongxin Yang, Yu Qiao, and Tao Xiang. Domain generalization with mixstyle. ArXiv, abs/2104.02008, 2021.