This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Label-Efficient Domain Generalization via Collaborative Exploration and Generalization

Junkun Yuan 0000-0003-0012-7397 Xu Ma Zhejiang UniversityHangzhouChina [email protected] [email protected] Defang Chen Zhejiang UniversityHangzhouChina [email protected] Kun Kuang Zhejiang UniversityHangzhouChina Shanghai AI LaboratoryShanghaiChina [email protected] Fei Wu Zhejiang UniversityHangzhouChina Shanghai Institute for Advanced Study of Zhejiang UniversityShanghaiChina [email protected]  and  Lanfen Lin Zhejiang UniversityHangzhouChina [email protected]
(2022)
Abstract.

Considerable progress has been made in domain generalization (DG) which aims to learn a generalizable model from multiple well-annotated source domains to unknown target domains. However, it can be prohibitively expensive to obtain sufficient annotation for source datasets in many real scenarios. To escape from the dilemma between domain generalization and annotation costs, in this paper, we introduce a novel task named label-efficient domain generalization (LEDG) to enable model generalization with label-limited source domains. To address this challenging task, we propose a novel framework called Collaborative Exploration and Generalization (CEG) which jointly optimizes active exploration and semi-supervised generalization. Specifically, in active exploration, to explore class and domain discriminability while avoiding information divergence and redundancy, we query the labels of the samples with the highest overall ranking of class uncertainty, domain representativeness, and information diversity. In semi-supervised generalization, we design MixUp-based intra- and inter-domain knowledge augmentation to expand domain knowledge and generalize domain invariance. We unify active exploration and semi-supervised generalization in a collaborative way and promote mutual enhancement between them, boosting model generalization with limited annotation. Extensive experiments show that CEG yields superior generalization performance. In particular, CEG can even use only 5% data annotation budget to achieve competitive results compared to the previous DG methods with fully labeled data on PACS dataset.

domain generalization; image classification; label-efficient learning
journalyear: 2022copyright: acmcopyrightconference: Proceedings of the 30th ACM International Conference on Multimedia; October 10–14, 2022; Lisboa, Portugalbooktitle: Proceedings of the 30th ACM International Conference on Multimedia (MM ’22), October 10–14, 2022, Lisboa, Portugalprice: 15.00doi: 10.1145/3503161.3548059isbn: 978-1-4503-9203-7/22/10ccs: Computing methodologies Computer vision

1. Introduction

Despite the remarkable success achieved by modern machine learning algorithms in visual recognition (He et al., 2016; Voulodimos et al., 2018; Srinivas et al., 2021), it heavily relies on the i.i.d. assumption (Vapnik, 1992) that training and test datasets should have a consistent statistical pattern. Since machine learning systems are usually deployed in a wide range of scenarios where the test data are unknown in advance, it may inevitably result in serious model performance degradation when there exists a distinct distribution/domain shift (Quionero-Candela et al., 2009) between the training and test data.

Refer to caption
Figure 1. Comparison between the conventional DG (a) and the proposed LEDG (b) tasks. DG may need to label all source data. In comparison, LEDG queries the labels of a small quota of data with limited annotation budget and boosts domain generalization by exploiting both labeled and unlabeled data.

With awareness of this problem, domain generalization (DG) (Blanchard et al., 2011) is introduced to extract domain invariance from multiple well-annotated source datasets/domains and train a generalizable model to unknown target domains. Lots of favorable DG algorithms (Shankar et al., 2018; Carlucci et al., 2019; Zhou et al., 2020; Xu et al., 2021; Zhou et al., 2021c; Pandey et al., 2021; Dubey et al., 2021) have been proposed recently, however, these methods may need to be fed with a large amount of labeled multi-source data for identifying domain invariance and improving model generalization. It might impede the realization of the DG approaches in many real-world applications where labeling massive data could be expensive or even infeasible. For example, a highly accurate and robust system for the detection of lung lesion images of COVID-19 patients may demand a large number of labeled medical images from different hospitals as the source data for training (Ettinger et al., 2021), but it could be impractical to require numerous experienced clinicians to complete the annotation. Therefore, a dilemma is encountered: The requirements of obtaining massive labeled source data for training a generalizable model may hard to be met in realistic scenarios due to the limited annotation budget. Meanwhile, without sufficient labeled data to provide adequate information of multi-source distribution, improving model generalization by identifying and learning the domain invariance is at serious risk of being misled.

To escape from the dilemma, we introduce a more practical task named label-efficient domain generalization (LEDG) to enable model generalization with label-limited source domains as shown in Figure 1. Instead of getting fully labeled data, the LEDG task unleashes the potential of budget-limited annotation by querying the labels of a small quota of informative data, and leverages both the labeled and unlabeled data to improve domain generalization. LEDG permits the learning of generalizable models in real scenarios, but it could be much more challenging. The first challenge is from the clear domain distinctions that could exist in the multi-source data, which constitutes enormous obstacles for selecting the most informative samples and learning adequate information of multi-source distribution. Meanwhile, the second challenge is from the discrepant distributions that labeled and unlabeled data may be subject to, making the simultaneous utilization of them for extracting domain invariance and promoting model generalization extremely difficult.

Active learning (AL) (Wang and Shang, 2014; Sener and Savarese, 2018; Ash et al., 2020; Huang et al., 2021b; Kim et al., 2021; Joshi et al., 2009) and semi-supervised learning (SSL) (Tarvainen and Valpola, 2017; Berthelot et al., 2019, 2020; Sohn et al., 2020) provide possible solutions to the introduced LEDG task. AL aims to query the labels of high-quality samples and SSL leverages the unlabeled data to improve performance with limited labeled data. However, the existing AL and SSL methods mostly depend on the i.i.d. assumption, hence may not be favorably extended to the generalization scenarios under distinct domain shifts. Semi-supervised domain generalization (SSDG) (Zhou et al., 2021a; Wang et al., 2021b; Yuan et al., 2021b; Liao et al., 2020; Sharifi-Noghabi et al., 2020) tackles domain shift under the SSL setting. But some data directly assumed to be labeled in this task might not be helpful for generalization improvement but could increase the annotation costs. Thus, it is imperative to find a solution to the challenging LEDG task for getting rid of the raised dilemma between domain generalization and annotation costs, realizing more practical training of generalizable models in real-world scenarios.

To address the LEDG task, in this paper, we propose a novel framework called Collaborative Exploration and Generalization (CEG) which jointly optimizes active exploration and semi-supervised generalization. In active exploration, to unleash the power of the limited annotation, we query the labels of the samples with the highest overall ranking of class uncertainty, domain representativeness, and information diversity, exploring class and domain discriminability while avoiding information divergence and redundancy. In semi-supervised generalization, we augment intra- and inter-domain knowledge with MixUp (Zhang et al., 2018) to expand domain knowledge and generalize domain invariance. An augmentation consistency constraint for unlabeled data and a prediction supervision for labeled data are further included to improve performance. We unify active exploration and semi-supervised generalization in a collaborative way by repeating them alternately, promoting closed-loop mutual enhancement between them for effective learning of domain invariance and label-efficient training of generalizable models.

Our contributions are listed in the following. (1) We introduce a more practical task named label-efficient domain generalization to permit generalization learning in real-world scenarios by tackling the dilemma between domain generalization and annotation costs. (2) To solve this challenging task, we propose a semi-supervised active learning-based framework CEG to unify active query-based distribution exploration and semi-supervised training-based model generalization in a collaborative way, achieving closed-loop mutual enhancement between them. (3) Extensive experiments show the superior generalization performance of CEG, which can even achieve competitive results using 5% annotation budget compared to the previous DG methods with full annotation on PACS dataset.

2. Related Work

Domain Generalization (DG). Different from domain adaptation (DA) (Wang et al., 2021c; Deng et al., 2021; Lv et al., 2021; Yan et al., 2021; Huang et al., 2021c; Ye et al., 2021; Li et al., 2021b; Chen et al., 2021b; Ma et al., 2022; Chen et al., 2022; Chen and Wang, 2021; Chen et al., 2021a) which adapts models from the source domain to the target, DG (Blanchard et al., 2011) assumes that the target domain is unknown during training and aims to train a generalizable model from the source domains. Increasing DG methods (Shankar et al., 2018; Pandey et al., 2021; Dubey et al., 2021; Volpi et al., 2021; Mahajan et al., 2021; Huang et al., 2020; Zhou et al., 2021b; Yuan et al., 2021a, c; Kuang et al., 2018, 2022, 2021, 2020; Shen et al., 2020) are proposed recently, they popularize various strategies via invariant representation learning (Zhao et al., 2020; Dou et al., 2019; Li et al., 2021a, 2018a; Qiao et al., 2020), meta-learning (Shu et al., 2021; Balaji et al., 2018; Li et al., 2018b; Dou et al., 2019; Li et al., 2019), data augmentation (Carlucci et al., 2019; Zhou et al., 2020, 2021c; Xu et al., 2021; Zhang et al., 2021; Huang et al., 2021a; Jeon et al., 2021), and others (Du et al., 2021; Wang et al., 2021a; Liu et al., 2021). But they mostly need the fully labeled data to learn generalization.

Semi-Supervised Domain Generalization (SSDG) (Zhou et al., 2021a; Wang et al., 2021b; Yuan et al., 2021b; Liao et al., 2020; Sharifi-Noghabi et al., 2020) aims to reduce the reliance of DG on annotation via pseudo-labeling (Wang et al., 2021b), consistency learning (Zhou et al., 2021a), or bias filtering (Yuan et al., 2021b). For example, StyleMatch (Zhou et al., 2021a) combines consistency learning, model uncertainty learning, and style augmentation to utilize the annotation for improving model robustness. However, partial samples assumed to be labeled in the SSDG task may not be informative for boosting model generalization but increase the annotation costs.

Semi-Supervised Learning (SSL). SSL (Tarvainen and Valpola, 2017; Berthelot et al., 2019, 2020; Sohn et al., 2020; Jiang et al., 2022) is a practicable way to use both labeled and unlabeled data. For example, MeanTeacher (Tarvainen and Valpola, 2017) achieves significant performance by using the labeled data to optimize a student model, the prediction of which is constrained to be consistent with the prediction of a teacher model. But most of the SSL methods rely on the i.i.d. assumption, which can impair their generalization performance under domain shift.

Refer to caption
Figure 2. Overview of the Collaborative Exploration and Generalization (CEG) framework. In active exploration, samples with the highest overall ranking (RR) of class uncertainty (SuS_{u}), domain representativeness (SrS_{r}), and information diversity (SdS_{d}) are selected to query for learning multi-source distribution. In semi-supervised generalization, expansion and generalization (eg\mathcal{L}_{eg}) of the known knowledge, and augmentation consistency (ac\mathcal{L}_{ac}) for unlabeled data and prediction supervision (ce\mathcal{L}_{ce}) for labeled data are devised to extract domain invariance and improve model generalization via semi-supervised training (ss\mathcal{L}_{ss}).

Active Learning (AL). AL (Ash et al., 2020; Wang and Shang, 2014; Joshi et al., 2009; Sener and Savarese, 2018; Kim et al., 2021; Huang et al., 2021b) aims to select high-quality data for querying the labels. Pool-based AL (Ash et al., 2020; Sener and Savarese, 2018; Huang et al., 2021b; Kim et al., 2021) is the most popular that chooses samples from an unlabeled pool and hand them over to the oracle to label. The labeled samples are then added to a labeled pool as newly acquired knowledge. Some successful uncertainty (Ash et al., 2020; Wang and Shang, 2014; Joshi et al., 2009) and diversity (Ash et al., 2020; Sener and Savarese, 2018) based methods select uncertain and diverse samples for learning task boundary and comprehensive information, respectively. However, the AL algorithms are mainly designed for single-domain data, thus may not be directly extended to the generalization scenarios.

3. Method

3.1. Label-Efficient Domain Generalization

We begin with the task setting of the introduced Label-Efficient Domain Generalization (LEDG). In the LEDG task, we have KK unlabeled source datasets {𝒟1,,𝒟K}\{\mathcal{D}^{1},...,\mathcal{D}^{K}\} sampled from different data distributions {P(X1,Y1),,P(XK,YK)}\{P(X^{1},Y^{1}),...,P(X^{K},Y^{K})\}, respectively. There are NkN^{k} unlabeled data points being sampled for each dataset 𝒟k\mathcal{D}^{k}, i.e., 𝒟k={𝒙ik}i=1Nk\mathcal{D}^{k}=\{\boldsymbol{x}_{i}^{k}\}_{i=1}^{N^{k}}, for k=1,,Kk=1,...,K. We further have an annotation budget BB, i.e., the maximum number of samples that we are allowed to query their class labels. Each sample pair (𝒙,y)(\boldsymbol{x},y) is defined on the image and label joint space 𝒳×𝒴\mathcal{X}\times\mathcal{Y}. Besides, the domain label pikp_{i}^{k} for each sample 𝒙ik\boldsymbol{x}_{i}^{k} is given in our task. We consider a classification model GG composed of a feature extractor FF and a classifier head CC, i.e., G=CFG=C\circ{F}. The goal of LEDG is to train the model GG by utilizing the unlabeled multi-source data {𝒟k}k=1K\{\mathcal{D}^{k}\}_{k=1}^{K} as well as the limited annotation budget BB for improving the generalization performance of the model on the target domains with unknown distributions. For convenience of the statement of our method, we denote the dataset consists of all the labeled (queried) samples (𝒙i(l),yi(l))(\boldsymbol{x}_{i}^{(l)},y_{i}^{(l)}) as 𝒟(l)={(𝒙i(l),yi(l))}i=1N(l)\mathcal{D}^{(l)}=\{(\boldsymbol{x}_{i}^{(l)},y_{i}^{(l)})\}_{i=1}^{N^{(l)}} and the dataset with all the unlabeled (not queried) samples 𝒙i(u)\boldsymbol{x}_{i}^{(u)} as 𝒟(u)={𝒙i(u)}i=1N(u)\mathcal{D}^{(u)}=\{\boldsymbol{x}_{i}^{(u)}\}_{i=1}^{N^{(u)}}, where N(l)N^{(l)} and N(u)N^{(u)} are the data size of the labeled and unlabeled datasets, respectively. It is obvious that the whole data size N(l)+N(u)=N1+N2++NKN^{(l)}+N^{(u)}=N^{1}+N^{2}+...+N^{K}.

Our insight for this challenging task is to consider the labeled and unlabeled samples as the “known” and “unknown” regions of the multi-source distribution, respectively. In view of this, the core idea of our solution is to: (1) explore the key knowledge hidden in the unknown regions via active query for adequate multi-source distribution learning, (2) extract and generalize the domain invariance contained in the obtained knowledge in both the known and unknown regions via semi-supervised training, and (3) make the active query-based exploration and semi-supervised training-based generalization complement and promote each other to train a generalizable model. An overview of our framework, i.e., Collaborative Exploration and Generalization (CEG), is shown in Figure 2.

Refer to caption
Figure 3. Comparisons of different query strategies for the active exploration. (a): Uncertainty criterion selects class-ambiguous samples but can cause information divergence, i.e., it may fail to select the representative samples for each domain. (b): The introduction of representativeness criterion helps to capture domain characteristics but can lead to information redundancy, i.e., it may select adjacent samples and hence reduce the efficiency of the limited annotation budget. (c): The further introduction of information diversity criterion guides the model to learn comprehensive knowledge of the data distribution.

3.2. Active Exploration

There might be distinct domain divergence among the data distributions of the source domains. Meanwhile, each source domain contains the discriminative information of class boundary which is essential for the prediction task. Thus, we take class and domain discriminability as the key knowledge for learning the multi-source distribution in the active exploration. In light of this, we present to select the samples with high class uncertainty and domain representativeness. To avoid information redundancy, we further take information diversity into consideration. Figure 3 provides revealing insights into the elaborate strategy for the active exploration.

To capture the key knowledge of class discriminability, we select the samples with high class uncertainty. Specifically, we adopt the margin of the top two model predictions to choose class-ambiguous samples to query. Let GhG_{h} be the hh-th dimension of the class prediction of the model GG (after softmax\mathrm{softmax} operation), then the class uncertainty score SuS_{u} for each unlabeled sample 𝒙(u)\boldsymbol{x}^{(u)} is defined as

(1) Su(𝒙(u))=(1(maxhGh(𝒙(u))maxh|hhGh(𝒙(u)))).S_{u}(\boldsymbol{x}^{(u)})=\left(1-\left(\max_{h}{G_{h}\left(\boldsymbol{x}^{(u)}\right)}-\max_{h^{\prime}|h^{\prime}\neq{h}}{G_{h^{\prime}}\left(\boldsymbol{x}^{(u)}\right)}\right)\right).

We tend to query the samples with high uncertainty scores for their class-ambiguous. Labeling these samples provides the key knowledge of class discriminability, which helps the model to figure out class boundary and boosts its performance of class prediction.

Different from the single-domain scenario considered in the AL methods (Wang and Shang, 2014; Joshi et al., 2009), multiple source domains may lead to an information divergence problem here, i.e., the selected high-uncertainty samples are scattered at the domain boundary (see Figure 3 (a)). To sufficiently explore and grasp information of the multi-source distribution, the selected samples are required to represent the characteristics of each source domain. Therefore, given domain label p(u)p^{(u)} of each unlabeled sample 𝒙(u)\boldsymbol{x}^{(u)}, we first train a domain discriminator model HH with a domain discriminability loss:

(2) dd=𝔼𝒙(u)𝒟(u)(H(𝒙(u)),p(u)),\mathcal{L}_{dd}=\mathbb{E}_{\boldsymbol{x}^{(u)}\in\mathcal{D}^{(u)}}\ell(H(\boldsymbol{x}^{(u)}),p^{(u)}),

where \ell is the cross-entropy loss. Let HhH_{h} be the hh-th dimension of the domain prediction of the model HH, we then define domain representativeness score SrS_{r} for each unlabeled sample 𝒙(u)\boldsymbol{x}^{(u)} as:

(3) Sr(𝒙(u))=maxhHh(𝒙(u)).S_{r}(\boldsymbol{x}^{(u)})=\max_{h}{H_{h}(\boldsymbol{x}^{(u)})}.

Note that different from the class discriminability learning with class ambiguous data, here, we select the samples with high representativeness score, i.e., high domain confidence. It prevents the model from learning class discriminability in the remote areas of each source domain and hence loosing domain characteristics.

Then an information redundancy problem has arisen, i.e., the selected samples with high class uncertainty and domain representativeness may gather together (see Figure 3 (b)), which wastes the limited annotation budget. To disperse the information, we choose the samples that are far away from the known domain-class knowledge of the labeled data. We make a knowledge dataset 𝒦hk\mathcal{K}_{h}^{k} be composed of the labeled data (𝒙(l),y(l))𝒟(l)(\boldsymbol{x}^{(l)},y^{(l)})\in\mathcal{D}^{(l)} that belongs to domain kk and class hh (if there is no such a sample in 𝒟(l)\mathcal{D}^{(l)}, then 𝒦hk=\mathcal{K}_{h}^{k}=\emptyset). Let |𝒦hk||\mathcal{K}_{h}^{k}| be the number of samples in the knowledge dataset 𝒦hk\mathcal{K}_{h}^{k} and FF be the feature extractor. We generate knowledge centroids 𝝁hk\boldsymbol{\mu}_{h}^{k} for the known regions in the semantic feature space:

(4) 𝝁hk=1|𝒦hk|(𝒙(l),y(l))𝒦hkF(𝒙(l)).\boldsymbol{\mu}_{h}^{k}=\frac{1}{|\mathcal{K}_{h}^{k}|}\sum_{(\boldsymbol{x}^{(l)},y^{(l)})\in\mathcal{K}_{h}^{k}}F(\boldsymbol{x}^{(l)}).

We let a set \mathcal{B} be composed of all the knowledge centroids 𝝁hk\boldsymbol{\mu}_{h}^{k} if 𝒦hk\mathcal{K}_{h}^{k}\neq{\emptyset}. Then, we define information diversity score SdS_{d} as

(5) Sd(𝒙(u))=min𝝁hkdist(𝒙(u),𝝁hk),S_{d}(\boldsymbol{x}^{(u)})=\min_{\boldsymbol{\mu}_{h}^{k}\in\mathcal{B}}\mathrm{dist}(\boldsymbol{x}^{(u)},\boldsymbol{\mu}_{h}^{k}),

where dist(,)\mathrm{dist}(\cdot,\cdot) is a distance metric used as cosine distance in experiments. We tend to choose samples with high diversity score, i.e., far away from the closest centroids, facilitating the unknown exploration for comprehensive learning of multi-source distribution.

To avoid numerical issues, we integrate uncertainty score SuS_{u}, representativeness score SrS_{r}, and diversity score SdS_{d} by adopting their rankings, which we denote as SuS_{u}^{\prime}, SrS_{r}^{\prime}, and SdS_{d}^{\prime}, respectively. Finally, we have an overall query ranking RR for each unlabeled sample 𝒙(u)\boldsymbol{x}^{(u)}:

(6) R(𝒙(u))=\displaystyle R(\boldsymbol{x}^{(u)})= Su(𝒙(u))+γ1Sr(𝒙(u))+γ2Sd(𝒙(u)),\displaystyle S_{u}^{\prime}(\boldsymbol{x}^{(u)})+\gamma_{1}S_{r}^{\prime}(\boldsymbol{x}^{(u)})+\gamma_{2}S_{d}^{\prime}(\boldsymbol{x}^{(u)}),

where γ1\gamma_{1} and γ2\gamma_{2} are trade-off hyper-parameters. Note that we rank each score, i.e., SuS_{u}, SrS_{r}, and SdS_{d}, from high to low, for selecting the most informative samples of the multi-source data distribution.

3.3. Semi-Supervised Generalization

With active query-based exploration, we have a small quota of labeled data and massive unlabeled data, i.e., small range of known regions and large range of unknown regions of the data distribution. In semi-supervised generalization, we aim to expand domain knowledge and learn domain invariance via MixUp-based intra- and inter-domain knowledge augmentation as shown in Figure 4.

We start by defining the unlabeled samples that close to the knowledge centroids 𝝁hk\boldsymbol{\mu}_{h}^{k} as “reliable samples”, and construct a reliable dataset 𝒰\mathcal{U} with expansion threshold TT to tune reliable range:

(7) 𝒰={(𝒙(u),y(u))|min𝝁hkdist(𝒙(u),𝝁hk)<T},\mathcal{U}=\{(\boldsymbol{x}^{(u)},y^{(u)})|\min_{\boldsymbol{\mu}_{h}^{k}\in\mathcal{B}}\mathrm{dist}(\boldsymbol{x}^{(u)},\boldsymbol{\mu}_{h}^{k})<T\},

where pseudo label y(u)y^{(u)} of each unlabeled sample 𝒙(u)\boldsymbol{x}^{(u)} is assigned by the nearest knowledge centroids 𝝁hk\boldsymbol{\mu}_{h}^{k}, that is,

(8) y(u)=argminh:𝝁hkdist(𝒙(u),𝝁hk).y^{(u)}={\arg\min}_{h:\boldsymbol{\mu}_{h}^{k}\in\mathcal{B}}\mathrm{dist}(\boldsymbol{x}^{(u)},\boldsymbol{\mu}_{h}^{k}).

A low threshold value TT leads to few reliable samples but high dependability of its pseudo labels, and vice versa. To arrange learning tasks in the order of difficulty for helping the model gain sufficient basic and easy knowledge before handling more complex data, we let TT increase with epochs and dynamically tune learning difficulty:

(9) T=Tini+TfinTiniEtotEcur,T=T^{ini}+\frac{T^{fin}-T^{ini}}{E_{tot}}*E_{cur},

where TiniT^{ini} and TfinT^{fin} are the initial and final threshold values, EtotE_{tot} and EcurE_{cur} are the total and current epochs, respectively. It makes the model expand knowledge stably with the high dependable samples at the beginning, and break through the hard samples gradually.

Refer to caption
Figure 4. Semi-supervised generalization with knowledge augmentation. (a): Unlabeled samples within reliable range (scaled by expansion threshold), i.e., the reliable samples, are pseudo-labeled by the nearest knowledge centroids, i.e., the centers of the labeled data. (b): The reliable samples are then augmented in the same domain and across domains to expand knowledge and generalize domain invariance.

We expand domain-class knowledge in each domain and across domains with the reliable dataset 𝒰\mathcal{U}, and construct MixUp-based intra- and inter-domain knowledge augmentation datasets, i.e., intra\mathcal{M}^{intra} and inter\mathcal{M}^{inter}, respectively. That is,

(10) intra=\displaystyle\mathcal{M}^{intra}= {(λ𝒙i(u)+(1λ)𝒙j(u),λyi(u)+(1λ)yj(u))|\displaystyle\{(\lambda\boldsymbol{x}^{(u)}_{i}+(1-\lambda)\boldsymbol{x}^{(u)}_{j},\lambda y^{(u)}_{i}+(1-\lambda)y^{(u)}_{j})|
(𝒙i(u),yi(u))𝒟m,(𝒙j(u),yj(u))𝒟n,m=n},\displaystyle(\boldsymbol{x}_{i}^{(u)},y_{i}^{(u)})\in\mathcal{D}^{m},(\boldsymbol{x}_{j}^{(u)},y_{j}^{(u)})\in\mathcal{D}^{n},m=n\},
(11) inter=\displaystyle\mathcal{M}^{inter}= {(λ𝒙i(u)+(1λ)𝒙j(u),λyi(u)+(1λ)yj(u))|\displaystyle\{(\lambda\boldsymbol{x}^{(u)}_{i}+(1-\lambda)\boldsymbol{x}^{(u)}_{j},\lambda y^{(u)}_{i}+(1-\lambda)y^{(u)}_{j})|
(𝒙i(u),yi(u))𝒟m,(𝒙j(u),yj(u))𝒟n,mn},\displaystyle(\boldsymbol{x}_{i}^{(u)},y_{i}^{(u)})\in\mathcal{D}^{m},(\boldsymbol{x}_{j}^{(u)},y_{j}^{(u)})\in\mathcal{D}^{n},m\neq n\},

where λBeta(α,α)\lambda\sim Beta(\alpha,\alpha) with α=0.2\alpha=0.2 as in (Zhang et al., 2018). intra\mathcal{M}^{intra} and inter\mathcal{M}^{inter} open up the association among known regions within and across domains, respectively. To broaden the known regions and learn domain-invariant representations for improving out-of-domain generalization ability, we train the model on the union of the augmented datasets by optimizing an expansion and generalization loss eg\mathcal{L}_{eg}:

(12) eg=𝔼(𝒙(u),y(u))intrainter(G(𝒙(u)),y(u)).\mathcal{L}_{eg}=\mathbb{E}_{(\boldsymbol{x}^{(u)},y^{(u)})\in\mathcal{M}^{intra}\cup\mathcal{M}^{inter}}\ell(G(\boldsymbol{x}^{(u)}),y^{(u)}).
Algorithm 1 Collaborative Exploration and Generalization
1:Source datasets {𝒟k}k=1K\{\mathcal{D}^{k}\}_{k=1}^{K}; Pretraining epochs NpN_{p} and learning epochs NlN_{l}; Annotation budget BB and initial budget BiniB^{ini}.
2:A well trained prediction model G^\hat{G};
3:Initialize labeled dataset 𝒟(l)\mathcal{D}^{(l)} with BiniB^{ini};
4:for n=1n=1 to NlN_{l} do
5:     Get knowledge centroids 𝝁hk\boldsymbol{\mu}_{h}^{k} with 𝒟(l)\mathcal{D}^{(l)} via Equation (4);
6:     if n>Npn>N_{p} then // begin to query after pretraining
7:         Select a subset Δ𝒟(u)\Delta\mathcal{D}^{(u)} from 𝒟(u)\mathcal{D}^{(u)} via Equation (6);
8:         Query labels of Δ𝒟(u)\Delta\mathcal{D}^{(u)} for Δ𝒟(u)Δ𝒟(l)\Delta\mathcal{D}^{(u)}\rightarrow\Delta\mathcal{D}^{(l)};
9:         𝒟(l)𝒟(l)Δ𝒟(l)\mathcal{D}^{(l)}\leftarrow\mathcal{D}^{(l)}\cup\Delta\mathcal{D}^{(l)}, 𝒟(u)𝒟(u)/Δ𝒟(u)\mathcal{D}^{(u)}\leftarrow\mathcal{D}^{(u)}/\Delta\mathcal{D}^{(u)};
10:     end if
11:     Train the discriminator HH with 𝒟(u)\mathcal{D}^{(u)} via Equation (2);
12:     Train the model GG with 𝒟(l)\mathcal{D}^{(l)} and 𝒟(u)\mathcal{D}^{(u)} via Equation (15).
13:end for

We further utilize the unknown and known regions by adopting augmentation consistency constraint (Sohn et al., 2020) for the unlabeled data and prediction supervision for the labeled data, respectively. Let 𝒜w\mathcal{A}^{w} and 𝒜s\mathcal{A}^{s} be weak (flip-and-shift) and strong (Cubuk et al., 2019) augmentation functions, respectively. Pseudo labels q^\hat{q} can be assigned by 𝒜w\mathcal{A}^{w} via q^=argmaxhGh(𝒜w(𝒙(u)))\hat{q}={\arg\max}_{h}G_{h}(\mathcal{A}^{w}(\boldsymbol{x}^{(u)})). The augmentation consistency loss ac\mathcal{L}_{ac} makes the model predictions of strong augmented data, i.e., G(𝒜s(𝒙(u)))G(\mathcal{A}^{s}(\boldsymbol{x}^{(u)})), and weak augmented label, i.e., q^\hat{q}, to be consistent:

(13) ac=𝔼𝒙(u)𝒟(u)I(𝒙(u))(G(𝒜s(𝒙(u))),q^),\mathcal{L}_{ac}=\mathbb{E}_{\boldsymbol{x}^{(u)}\in\mathcal{D}^{(u)}}I(\boldsymbol{x}^{(u)})\ell(G(\mathcal{A}^{s}(\boldsymbol{x}^{(u)})),\hat{q}),

where an indicator I(𝒙(u))=𝟙(maxh(Gh(𝒜w(𝒙(u)))τ)I(\boldsymbol{x}^{(u)})=\mathbbm{1}(\max_{h}{(G_{h}(\mathcal{A}^{w}(\boldsymbol{x}^{(u)}))}\geq\tau) (τ\tau is set to 0.95 as in (Sohn et al., 2020)) selects high dependable data. This constraint helps the model to capture structural knowledge in the unknown regions via unsupervised learning. For prediction supervision, we adopt a cross-entropy classification loss ce\mathcal{L}_{ce} for the labeled data:

(14) ce=𝔼(𝒙(l),y(l))𝒟(l)(G(𝒙(l)),y(l)).\mathcal{L}_{ce}=\mathbb{E}_{(\boldsymbol{x}^{(l)},y^{(l)})\in\mathcal{D}^{(l)}}\ell(G(\boldsymbol{x}^{(l)}),y^{(l)}).

A semi-supervised training loss ss\mathcal{L}_{ss} is then derived as:

(15) ss=ce+ac+δeg,\mathcal{L}_{ss}=\mathcal{L}_{ce}+\mathcal{L}_{ac}+\delta\mathcal{L}_{eg},

where δ\delta is a hyper-parameter of knowledge expansion and generalization. We set the weights of ce\mathcal{L}_{ce} and ac\mathcal{L}_{ac} to 1 as in (Sohn et al., 2020).

Our framework CEG explores informative unlabeled samples for learning key knowledge of multi-source distribution using limited annotation budget, promoting the expansion and generalization of the key knowledge in semi-supervised training. Then a well trained model is continuously utilized to select more effective samples in the next round of query. The active exploration and semi-supervised generalization are unified in a collaborative way by being repeated alternately. They complement and promote each other to enable label-efficient domain generalization. The learning process of CEG is stated in Algorithm 1. Note that we use an initial budget BiniB^{ini} from the annotation budget BB to initialize the labeled dataset D(l)D^{(l)} via uniform sample selection, and pretrain the model before active query to solve a cold start problem, to our empirical experience.

4. Experiments

Table 1. Performance (%) comparisons between CEG and DG methods on PACS dataset with 5% annotation budget. The results with fully labeled source data (100% annotation) are given in parentheses. The best results are emphasized in bold.
Methods Art Cartoon Photo Sketch Average
DeepAll 55.79±\pm2.92 (73.59±\pm2.89) 61.84±\pm1.98 (70.63±\pm2.33) 80.21±\pm1.14 (89.36±\pm1.21) 63.00±\pm1.60 (80.06±\pm0.95) 65.21±\pm0.46 (78.41±\pm0.44)
JiGen (Carlucci et al., 2019) 53.03±\pm5.19 (77.05±\pm1.83) 55.09±\pm1.85 (76.16±\pm2.02) 78.62±\pm5.83 (93.81±\pm1.51) 22.28±\pm4.05 (70.93±\pm0.72) 52.26±\pm0.25 (79.49±\pm0.67)
FACT (Xu et al., 2021) 74.21±\pm0.30 (84.76±\pm0.77) 65.82±\pm2.03 (77.52±\pm0.70) 90.48±\pm1.17 (95.29±\pm0.31) 55.24±\pm3.62 (78.97±\pm0.32) 71.44±\pm1.02 (84.13±\pm0.24)
DDAIG (Zhou et al., 2020) 62.87±\pm3.13 (77.80±\pm1.09) 57.64±\pm6.59 (75.35±\pm3.21) 81.48±\pm3.82 (89.66±\pm1.74) 36.61±\pm2.23 (73.70±\pm2.99) 59.65±\pm2.22 (79.13±\pm0.91)
RSC (Huang et al., 2020) 59.57±\pm3.37 (77.88±\pm0.66) 59.61±\pm3.22 (73.90±\pm2.12) 84.43±\pm3.59 (93.85±\pm0.80) 57.38±\pm6.61 (80.66±\pm0.81) 65.25±\pm2.95 (81.57±\pm0.71)
CrossGrad (Shankar et al., 2018) 56.06±\pm7.04 (75.69±\pm2.25) 52.73±\pm4.17 (76.51±\pm3.24) 80.51±\pm1.97 (91.33±\pm0.50) 41.25±\pm5.21 (70.50±\pm0.97) 57.64±\pm2.13 (78.51±\pm1.40)
DAEL (Zhou et al., 2021b) 66.24±\pm1.86 (83.51±\pm0.83) 61.72±\pm1.89 (72.31±\pm2.67) 89.98±\pm0.37 (95.74±\pm0.08) 32.50±\pm2.20 (78.87±\pm0.59) 62.61±\pm0.46 (82.61±\pm0.98)
CEG (ours) 80.12±\pm0.37 71.11±\pm0.96 92.32±\pm1.68 73.13±\pm2.87 79.17±\pm0.83
Table 2. Performance (%) comparisons between CEG and DG methods on Office-Home dataset with 5% annotation budget. The results with fully labeled source data (100% annotation) are given in parentheses. The best results are emphasized in bold.
Methods Art Clipart Product Real-World Average
DeepAll 34.73±\pm1.13 (47.06±\pm1.35) 34.46±\pm2.11 (47.50±\pm0.91) 46.20±\pm1.51 (64.89±\pm0.65) 48.89±\pm0.87 (65.16±\pm0.62) 41.07±\pm0.93 (56.15±\pm0.59)
JiGen (Carlucci et al., 2019) 29.62±\pm2.25 (52.67±\pm0.95) 25.52±\pm2.12 (50.40±\pm0.97) 37.91±\pm1.33 (71.21±\pm0.12) 39.84±\pm0.61 (72.24±\pm0.15) 33.22±\pm0.93 (61.63±\pm0.25)
FACT (Xu et al., 2021) 40.71±\pm0.08 (58.98±\pm0.29) 32.12±\pm0.17 (53.53±\pm0.35) 48.05±\pm0.14 (74.47±\pm0.56) 49.16±\pm0.17 (75.63±\pm0.67) 42.51±\pm0.09 (65.65±\pm0.41)
DDAIG (Zhou et al., 2020) 35.20±\pm1.06 (55.05±\pm0.69) 29.75±\pm0.50 (52.37±\pm0.58) 42.42±\pm0.58 (72.00±\pm0.58) 43.07±\pm0.12 (73.54±\pm0.19) 37.61±\pm0.16 (63.24±\pm0.35)
RSC (Huang et al., 2020) 31.95±\pm1.24 (56.06±\pm0.71) 28.62±\pm1.53 (52.95±\pm0.31) 40.88±\pm1.87 (72.61±\pm0.39) 42.43±\pm0.69 (73.42±\pm0.38) 35.97±\pm0.61 (63.76±\pm0.25)
CrossGrad (Shankar et al., 2018) 35.05±\pm0.37 (54.42±\pm0.55) 30.86±\pm1.74 (52.63±\pm0.77) 45.10±\pm1.76 (73.00±\pm0.47) 44.41±\pm2.08 (73.42±\pm0.74) 38.86±\pm0.27 (63.37±\pm0.24)
DAEL (Zhou et al., 2021b) 35.93±\pm0.57 (59.20±\pm0.56) 30.71±\pm0.86 (50.97±\pm2.63) 42.79±\pm0.99 (73.53±\pm0.52) 43.95±\pm0.86 (76.56±\pm0.45) 38.35±\pm0.12 (65.06±\pm0.55)
CEG (ours) 47.60±\pm1.32 42.01±\pm1.19 56.20±\pm1.79 57.69±\pm1.18 50.87±\pm0.99
Table 3. Comparisons between CEG and AL, SSL, SSDG methods on PACS and Office-Home datasets with 5% annotation budget.
Methods PACS dataset (%) Office-Home dataset (%)
Art Cartoon Photo Sketch Average Art Clipart Product Real-World Average
Uniform 55.56±\pm2.92 61.61±\pm1.98 79.98±\pm1.14 62.77±\pm1.60 64.98±\pm0.46 34.27±\pm1.13 34.00±\pm2.11 45.74±\pm1.51 48.43±\pm0.87 40.61±\pm0.93
Entropy (Wang and Shang, 2014) 58.79±\pm2.36 63.49±\pm1.78 82.47±\pm0.99 61.67±\pm0.82 66.61±\pm0.58 34.06±\pm1.59 32.02±\pm1.84 46.72±\pm0.99 47.03±\pm0.96 39.96±\pm0.90
BvSB (Joshi et al., 2009) 62.85±\pm1.83 63.17±\pm1.02 79.57±\pm3.46 63.61±\pm5.97 67.30±\pm0.97 35.58±\pm1.44 35.32±\pm2.29 47.66±\pm1.42 50.39±\pm0.54 42.24±\pm0.82
Confidence (Wang and Shang, 2014) 58.02±\pm2.14 59.48±\pm1.61 81.75±\pm3.40 61.04±\pm2.86 65.07±\pm1.60 36.35±\pm1.23 36.21±\pm1.24 47.88±\pm0.75 50.46±\pm2.20 42.73±\pm0.83
CoreSet (Sener and Savarese, 2018) 61.48±\pm4.57 58.74±\pm2.66 79.03±\pm3.71 60.61±\pm2.25 64.96±\pm1.06 37.54±\pm0.76 35.75±\pm2.55 49.44±\pm1.21 51.06±\pm1.75 43.45±\pm0.66
BADGE (Ash et al., 2020) 54.49±\pm1.67 63.10±\pm1.60 80.84±\pm1.19 65.57±\pm5.99 66.50±\pm2.11 37.81±\pm1.07 36.86±\pm2.62 49.90±\pm2.00 51.26±\pm2.77 43.96±\pm0.82
MeanTeacher (Tarvainen and Valpola, 2017) 53.84±\pm6.41 54.86±\pm4.14 78.86±\pm4.63 35.52±\pm4.57 55.77±\pm1.43 32.70±\pm1.56 27.25±\pm2.92 43.01±\pm2.04 42.41±\pm4.03 36.35±\pm1.25
MixMatch (Berthelot et al., 2019) 63.92±\pm1.77 61.37±\pm2.31 81.14±\pm4.12 55.46±\pm0.61 65.47±\pm1.70 25.65±\pm0.66 22.90±\pm2.24 33.80±\pm0.93 28.35±\pm2.24 27.68±\pm0.80
FixMatch (Sohn et al., 2020) 78.60±\pm1.47 71.14±\pm2.49 92.17±\pm1.02 69.16±\pm0.94 77.77±\pm0.91 36.76±\pm1.84 31.09±\pm2.53 44.79±\pm4.20 45.07±\pm5.18 39.43±\pm2.21
StyleMatch (Zhou et al., 2021a) 72.67±\pm1.08 73.07±\pm0.81 89.61±\pm0.74 76.46±\pm0.93 77.95±\pm0.56 42.01±\pm0.68 40.95±\pm0.97 47.65±\pm1.70 51.93±\pm0.26 45.63±\pm0.29
CEG (ours) 80.12±\pm0.37 71.11±\pm0.96 92.32±\pm1.68 73.13±\pm2.87 79.17±\pm0.83 47.60±\pm1.32 42.01±\pm1.19 56.20±\pm1.79 57.69±\pm1.18 50.87±\pm0.99
Refer to caption
Figure 5. Sensitivity analysis of hyper-parameters. From left to right: δ\delta, TT, γ1\gamma_{1} and γ2\gamma_{2} on PACS, γ1\gamma_{1} and γ2\gamma_{2} on Office-Home.
Refer to caption
Figure 6. Comparisons with increasing annotation budget.
Table 4. Ablation studies on PACS and Office-Home datasets (5% annotation). The best results are emphasized in bold.
Strategies Cases PACS dataset (%) Office-Home dataset (%)
Art Cartoon Photo Sketch Average Art Clipart Product Real-World Average
Active Exploration w/ Uniform 76.42±\pm1.27 69.92±\pm3.01 87.09±\pm2.03 71.92±\pm4.97 76.34±\pm2.05 45.25±\pm1.29 40.48±\pm2.18 53.48±\pm2.00 55.95±\pm1.08 48.79±\pm0.96
w/o SuS_{u} 80.00±\pm2.08 67.56±\pm2.34 89.43±\pm0.71 71.11±\pm5.64 77.03±\pm1.28 46.12±\pm2.04 40.84±\pm0.98 52.56±\pm0.75 55.83±\pm1.47 48.84±\pm0.84
w/o SrS_{r} 76.77±\pm2.74 67.91±\pm4.38 90.90±\pm1.19 72.46±\pm1.44 77.01±\pm0.74 46.14±\pm1.05 39.59±\pm1.25 55.99±\pm0.82 57.02±\pm2.06 49.68±\pm0.71
w/o SdS_{d} 78.21±\pm0.97 68.95±\pm2.54 91.03±\pm1.50 72.58±\pm3.24 77.69±\pm0.61 45.78±\pm2.34 40.79±\pm1.78 54.39±\pm1.46 58.24±\pm1.83 49.84±\pm0.82
Semi-Supervised Generalization w/o ac\mathcal{L}_{ac} w/o eg\mathcal{L}_{eg} 62.61±\pm2.81 68.51±\pm2.34 82.84±\pm3.85 48.65±\pm4.32 65.65±\pm1.35 38.51±\pm1.06 33.62±\pm1.50 48.60±\pm2.59 49.89±\pm5.07 42.66±\pm1.44
w/o ac\mathcal{L}_{ac} 76.16±\pm2.11 67.05±\pm3.37 88.07±\pm1.80 60.37±\pm5.36 72.91±\pm1.31 44.65±\pm1.69 37.99±\pm1.25 55.41±\pm2.16 56.42±\pm2.09 48.62±\pm0.59
w/o eg\mathcal{L}_{eg} 72.13±\pm1.43 68.09±\pm3.37 86.03±\pm4.02 75.12±\pm2.31 75.34±\pm0.84 40.38±\pm2.15 34.68±\pm1.29 46.10±\pm2.20 47.94±\pm1.66 42.27±\pm1.18
w/o eg\mathcal{L}_{eg} (intra\mathcal{M}^{intra}) 76.52±\pm2.01 67.69±\pm3.50 89.89±\pm1.72 74.72±\pm2.88 77.20±\pm1.53 45.85±\pm2.72 38.50±\pm1.47 54.12±\pm2.65 55.80±\pm1.66 48.57±\pm1.36
w/o eg\mathcal{L}_{eg} (inter\mathcal{M}^{inter}) 75.36±\pm2.59 68.20±\pm2.66 91.18±\pm1.05 72.49±\pm2.28 76.81±\pm1.32 47.47±\pm3.43 38.97±\pm2.26 52.99±\pm2.77 55.10±\pm1.48 48.63±\pm1.32
w/ static TT 76.86±\pm2.37 69.61±\pm1.63 90.79±\pm1.13 73.96±\pm2.06 77.81±\pm1.09 45.90±\pm1.76 40.46±\pm2.33 55.90±\pm1.50 56.21±\pm2.30 49.62±\pm1.02
CEG 80.12±\pm0.37 71.11±\pm0.96 92.32±\pm1.68 73.13±\pm2.87 79.17±\pm0.83 47.60±\pm1.32 42.01±\pm1.19 56.20±\pm1.79 57.69±\pm1.18 50.87±\pm0.99
Refer to caption
Figure 7. T-SNE visualization (Maaten and Hinton, 2008) of the selected samples. Different colors represent different classes (left) and domains (right). Each dot is the embedding of a sample and the black crosses are the selected samples. Dataset: PACS; Target domain: Art.
Refer to caption
Figure 8. Accuracy of w/ uniform, w/ active exploration, w/ active exploration + semi-supervised training (CEG).

In this section, we first evaluate our framework CEG in label-limited scenarios, and then give sensitivity analysis of hyper-parameters, ablation studies of the components, and in-depth empirical analysis.

Datasets. We adopt two popular pubilc datasets that are PACS (Li et al., 2017) and Office-Home (Venkateswara et al., 2017). PACS contains 7 categories within 4 domains, i.e., Art, Cartoon, Sketch, and Photo. Office-Home has 65 classes in 4 domains, i.e., Art, Clipart, Product, and Real-World.

Baseline methods. We implement four types of baselines. (1) Domain generalization (DG): DeepAll (training with mixed multi-source data), JiGen (Carlucci et al., 2019), CrossGrad (Shankar et al., 2018), DDAIG (Zhou et al., 2020), DAEL (Zhou et al., 2021b), RSC (Huang et al., 2020), and FACT (Xu et al., 2021). (2) Active learning (AL): Uniform (uniform selection), Entropy (Wang and Shang, 2014), BvSB (Joshi et al., 2009), Confidence (Wang and Shang, 2014), CoreSet (Sener and Savarese, 2018), and BADGE (Ash et al., 2020). (3) Semi-supervised learning (SSL): MeanTeacher (Tarvainen and Valpola, 2017), MixMatch (Berthelot et al., 2019), and FixMatch (Sohn et al., 2020). (4) Semi-supervised domain generalization (SSDG): StyleMatch (Zhou et al., 2021a). See Section 2 for details.

Implementation details. Following (Carlucci et al., 2019; Huang et al., 2020; Xu et al., 2021), we use a pre-trained ResNet-18 (He et al., 2016) as the backbone and conduct leave-one-domain-out experiments by choosing one domain to hold out as the target domain. For fair comparisons, we implement all the methods with the same settings, i.e., SGD optimizer with learning rate 0.003 for feature extractor and 0.01 for classifier, pre-training/learning epochs on PACS and Office-Home datasets are 30/30 and 15/15, respectively, and batch-size is 16, et al. In experiments, we directly use “TT” to represent “TfinT^{fin}” in Equation (9) for simplicity. We adopt the percentage of the unlabeled samples for TT instead of a distance value. We set Tini=T2T^{ini}=\frac{T}{2}. The hyper-parameters {δ,T,γ1,γ2}\{\delta,T,\gamma_{1},\gamma_{2}\} are set to {0.3,50%,3,1}\{0.3,50\%,3,1\} and {0.1,70%,0.5,0.5}\{0.1,70\%,0.5,0.5\} for PACS and Office-Home, respectively. Half annotation budget is used as the initial budget to initialize the labeled dataset. We report the results over five runs.

4.1. Main Results of CEG

CEG vs DG methods. Table 1 and 2 report the results with 5% annotation budget on PACS and Office-Home datasets, respectively. We observe that the accuracy of DG methods drops rapidly when only 5% labeled data is given. In comparison, our method CEG can select the informative data to label and utilize both the labeled and unlabeled data to boost generalization performance in this challenging label-limited scenario. Most notably, CEG can even achieve competitive results with only 5% annotation budget compared to the DG methods with full annotation on the PACS dataset. It reveals that CEG generally realizes label-efficient domain generalization by exploiting only a small quota of labeled data and massive unlabeled data. We attribute this success to the effective collaboration mechanism between active exploration and semi-supervised generalization, which unleashes the latent power of the limited annotation budget. Since the DG methods may not be good at tackling the label-limited task as they can only use the labeled data, we further compare our CEG method with AL, SSL, and SSDG methods.

CEG vs AL, SSL, SSDG methods. Table 3 reports the results with 5% annotation budget on PACS and Office-Home datasets. CEG outperforms other methods on half of the tasks and yields the best average accuracy on the PACS dataset. It is probably because the AL and SSL methods rely on the i.i.d. assumption, and the SSDG method does not label and exploit the important source data. In contrast, CEG selects the most informative samples for query via active exploration, and hence captures multi-source distribution and boosts generalization ability more accurately. Besides, the performance of CEG is significantly better than other methods on the Office-Home dataset. We attribute it to the construction of domain-class knowledge centroids, which greatly helps CEG to precisely explore unknown regions during active exploration, and effectively expand knowledge and generalize domain invariance during semi-supervised generalization on the Office-Home dataset (because Office-Home has 65 classes but PACS only has 7 classes).

4.2. Sensitivity Analysis

As shown in Figure 5, CEG is generally robust to the hyper-parameters and outperforms other methods even with the default settings, i.e., δ=1.0\delta=1.0 (78.79% on PACS and 46.34% on Office-Home), T=100%T=100\% (78.09% on PACS and 50.49% on Office-Home), γ1=γ2=1.0\gamma_{1}=\gamma_{2}=1.0 (78.12% on PACS and 50.37% on Office-Home), indicating that exhaustive hyper-parameter fine-tuning is not necessary for CEG to achieve excellent performance in label-efficient generalization learning.

4.3. Results with Increasing Annotation Budget

Figure 6 shows the results with increasing annotation on Office-Home dataset. CEG consistently outperforms other methods by sharp margins on the average accuracy and three of the four tasks. The significant performance achieved by CEG when given a low budget is probably due to the query based-active exploration, but this advantage could be weakened when given a higher budget.

4.4. Why does CEG Work?

Ablation studies are reported in Table 4. The three criteria of active exploration, i.e., uncertainty SuS_{u}, representativeness SrS_{r}, and diversity SdS_{d}, are all important for learning multi-source distribution, and the integration of them further make full use of the limited annotation, compared with uniform selection. For semi-supervised generalization, both knowledge expansion and generalization eg\mathcal{L}_{eg} and augmentation consistency ac\mathcal{L}_{ac} are necessary to yield remarkable results. The intra- and inter-domain knowledge augmentation datasets, i.e., intra\mathcal{M}^{intra} and inter\mathcal{M}^{inter}, both play vital roles in improving generalization performance. It is noteworthy that the proposed eg\mathcal{L}_{eg} significantly improves average accuracy from 42.27% to 50.87% on Office-Home. Besides, the devised dynamic threshold TT shows its effectiveness of learning with increasing difficulty compared to the static one. The above results illustrate that each component is indispensable, and the exploration and generalization complement and promote each other for achieving the excellent performance.

T-SNE visualization is shown in Figure 7. The left figure shows class-ambiguous samples, i.e., the samples distribute on class boundary, are selected for learning class discriminability. The right figure shows that, in general, the selected samples distribute uniformly and representatively in each domain, illustrating the effectiveness of the domain representativeness and information diversity criteria. These three criteria help CEG to select the most informative samples for learning multi-source distribution, which facilitates the generalizable model training in semi-supervised generalization.

Accuracy curve on Office-Home dataset is shown in Figure 8. Active exploration selects the most important samples and grasps the key knowledge of multi-source distribution to effectively improve performance on each domain, compared with uniform sample selection. Semi-supervised generalization further markedly boosts performance by expanding the obtained knowledge and generalizing domain invariance. They promote each other to achieve remarkable generalization performance on the target domain.

5. Conclusion

We introduce a practical task named label-efficient domain generalization, and propose a novel method called CEG for this task via active exploration and semi-supervised generalization. The two modules promote each other to improve model generalization with the limited annotation. In future work, we may extend our method to a more challenging setting that domain labels are unknown.

Acknowledgements.
This work was supported in part by National Key Research and Development Program of China (2021YFC3340300), Young Elite Scientists Sponsorship Program by CAST (2021QNRC001), National Natural Science Foundation of China (No. 62006207, No. 62037001), Project by Shanghai AI Laboratory (P22KS00111), the Starry Night Science Fund of Zhejiang University Shanghai Institute for Advanced Study (SN-ZJU-SIAS-0010), Natural Science Foundation of Zhejiang Province (LZ22F020012), and the Fundamental Research Funds for the Central Universities (226-2022-00142).

References

  • (1)
  • Ash et al. (2020) Jordan T Ash, Chicheng Zhang, Akshay Krishnamurthy, John Langford, and Alekh Agarwal. 2020. Deep batch active learning by diverse, uncertain gradient lower bounds. In International Conference on Learning Representations.
  • Balaji et al. (2018) Yogesh Balaji, Swami Sankaranarayanan, and Rama Chellappa. 2018. Metareg: Towards domain generalization using meta-regularization. Advances in Neural Information Processing Systems 31 (2018), 998–1008.
  • Berthelot et al. (2020) David Berthelot, Nicholas Carlini, Ekin D Cubuk, Alex Kurakin, Kihyuk Sohn, Han Zhang, and Colin Raffel. 2020. Remixmatch: Semi-supervised learning with distribution alignment and augmentation anchoring. International Conference on Learning Representation (2020).
  • Berthelot et al. (2019) David Berthelot, Nicholas Carlini, Ian Goodfellow, Nicolas Papernot, Avital Oliver, and Colin Raffel. 2019. Mixmatch: A holistic approach to semi-supervised learning. Advances in Neural Information Processing Systems (2019).
  • Blanchard et al. (2011) Gilles Blanchard, Gyemin Lee, and Clayton Scott. 2011. Generalizing from several related classification tasks to a new unlabeled sample. Advances in Neural Information Processing Systems 24 (2011), 2178–2186.
  • Carlucci et al. (2019) Fabio Maria Carlucci, Antonio D’Innocente, S. Bucci, B. Caputo, and T. Tommasi. 2019. Domain Generalization by Solving Jigsaw Puzzles. IEEE Conference on Computer Vision and Pattern Recognition (2019), 2224–2233.
  • Chen et al. (2021b) Yang Chen, Yingwei Pan, Yu Wang, Ting Yao, Xinmei Tian, and Tao Mei. 2021b. Transferrable Contrastive Learning for Visual Domain Adaptation. In Proceedings of the 29th ACM International Conference on Multimedia. 3399–3408.
  • Chen et al. (2021a) Zhengyu Chen, Jixie Ge, Heshen Zhan, Siteng Huang, and Donglin Wang. 2021a. Pareto Self-Supervised Training for Few-Shot Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 13663–13672.
  • Chen and Wang (2021) Zhengyu Chen and Donglin Wang. 2021. Multi-Initialization Meta-Learning with Domain Adaptation. In ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 1390–1394.
  • Chen et al. (2022) Zhengyu Chen, Teng Xiao, and Kun Kuang. 2022. BA-GNN: On Learning Bias-Aware Graph Neural Network. In 2022 IEEE 38th International Conference on Data Engineering (ICDE). IEEE.
  • Cubuk et al. (2019) Ekin D Cubuk, Barret Zoph, Dandelion Mane, Vijay Vasudevan, and Quoc V Le. 2019. Autoaugment: Learning augmentation strategies from data. In IEEE/CVF Conference on Computer Vision and Pattern Recognition. 113–123.
  • Deng et al. (2021) Wanxia Deng, Yawen Cui, Zhen Liu, Gangyao Kuang, Dewen Hu, Matti Pietikäinen, and Li Liu. 2021. Informative Class-Conditioned Feature Alignment for Unsupervised Domain Adaptation. In Proceedings of the 29th ACM International Conference on Multimedia. 1303–1312.
  • Dou et al. (2019) Qi Dou, Daniel Coelho de Castro, Konstantinos Kamnitsas, and Ben Glocker. 2019. Domain generalization via model-agnostic learning of semantic features. Advances in Neural Information Processing Systems 32 (2019), 6450–6461.
  • Du et al. (2021) Zhekai Du, Jingjing Li, Ke Lu, Lei Zhu, and Zi Huang. 2021. Learning Transferrable and Interpretable Representations for Domain Generalization. In Proceedings of the 29th ACM International Conference on Multimedia. 3340–3349.
  • Dubey et al. (2021) Abhimanyu Dubey, Vignesh Ramanathan, Alex Pentland, and Dhruv Mahajan. 2021. Adaptive Methods for Real-World Domain Generalization. In IEEE/CVF Conference on Computer Vision and Pattern Recognition. 14340–14349.
  • Ettinger et al. (2021) Scott Ettinger, Shuyang Cheng, Benjamin Caine, Chenxi Liu, Hang Zhao, Sabeek Pradhan, Yuning Chai, Ben Sapp, Charles R Qi, Yin Zhou, et al. 2021. Large scale interactive motion forecasting for autonomous driving: The waymo open motion dataset. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 9710–9719.
  • He et al. (2016) Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In IEEE Conference on Computer Vision and Pattern Recognition. 770–778.
  • Huang et al. (2021a) Jiaxing Huang, Dayan Guan, Aoran Xiao, and Shijian Lu. 2021a. Fsdr: Frequency space domain randomization for domain generalization. In IEEE/CVF Conference on Computer Vision and Pattern Recognition. 6891–6902.
  • Huang et al. (2021b) Siyu Huang, Tianyang Wang, Haoyi Xiong, Jun Huan, and Dejing Dou. 2021b. Semi-supervised active learning with temporal output discrepancy. In IEEE/CVF International Conference on Computer Vision. 3447–3456.
  • Huang et al. (2021c) Shengqi Huang, Wanqi Yang, Lei Wang, Luping Zhou, and Ming Yang. 2021c. Few-shot Unsupervised Domain Adaptation with Image-to-class Sparse Similarity Encoding. In Proceedings of the 29th ACM International Conference on Multimedia. 677–685.
  • Huang et al. (2020) Zeyi Huang, Haohan Wang, Eric P. Xing, and Dong Huang. 2020. Self-challenging Improves Cross-Domain Generalization. In European Conference on Computer Vision. 124–140.
  • Jeon et al. (2021) Seogkyu Jeon, Kibeom Hong, Pilhyeon Lee, Jewook Lee, and Hyeran Byun. 2021. Feature stylization and domain-aware contrastive learning for domain generalization. In Proceedings of the 29th ACM International Conference on Multimedia. 22–31.
  • Jiang et al. (2022) Ziqi Jiang, Shengyu Zhang, Siyuan Yao, Wenqiao Zhang, Sihan Zhang, Juncheng Li, Zhou Zhao, and Fei Wu. 2022. Weakly-supervised Disentanglement Network for Video Fingerspelling Detection. In ACM MM.
  • Joshi et al. (2009) Ajay J Joshi, Fatih Porikli, and Nikolaos Papanikolopoulos. 2009. Multi-class active learning for image classification. In IEEE Conference on Computer Vision and Pattern Recognition. IEEE, 2372–2379.
  • Kim et al. (2021) Kwanyoung Kim, Dongwon Park, Kwang In Kim, and Se Young Chun. 2021. Task-aware variational adversarial active learning. In IEEE/CVF Conference on Computer Vision and Pattern Recognition. 8166–8175.
  • Kuang et al. (2018) Kun Kuang, Peng Cui, Susan Athey, Ruoxuan Xiong, and Bo Li. 2018. Stable prediction across unknown environments. In proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining. 1617–1626.
  • Kuang et al. (2022) Kun Kuang, Haotian Wang, Yue Liu, Ruoxuan Xiong, Runze Wu, Weiming Lu, Yue Ting Zhuang, Fei Wu, Peng Cui, and Bo Li. 2022. Stable Prediction with Leveraging Seed Variable. IEEE Transactions on Knowledge and Data Engineering (2022).
  • Kuang et al. (2020) Kun Kuang, Ruoxuan Xiong, Peng Cui, Susan Athey, and Bo Li. 2020. Stable prediction with model misspecification and agnostic distribution shift. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 34. 4485–4492.
  • Kuang et al. (2021) Kun Kuang, Hengtao Zhang, Runze Wu, Fei Wu, Yueting Zhuang, and Aijun Zhang. 2021. Balance-Subsampled stable prediction across unknown test data. ACM Transactions on Knowledge Discovery from Data (TKDD) 16, 3 (2021), 1–21.
  • Li et al. (2017) Da Li, Yongxin Yang, Yi-Zhe Song, and Timothy M Hospedales. 2017. Deeper, broader and artier domain generalization. In Proceedings of the IEEE International Conference on Computer Vision. 5542–5550.
  • Li et al. (2018b) Da Li, Yongxin Yang, Yi-Zhe Song, and Timothy M Hospedales. 2018b. Learning to generalize: Meta-learning for domain generalization. In AAAI Conference on Artificial Intelligence.
  • Li et al. (2018a) Haoliang Li, Sinno Jialin Pan, Shiqi Wang, and Alex C Kot. 2018a. Domain generalization with adversarial feature learning. In IEEE Conference on Computer Vision and Pattern Recognition. 5400–5409.
  • Li et al. (2021a) Lei Li, Ke Gao, Juan Cao, Ziyao Huang, Yepeng Weng, Xiaoyue Mi, Zhengze Yu, Xiaoya Li, and Boyang Xia. 2021a. Progressive Domain Expansion Network for Single Domain Generalization. In IEEE/CVF Conference on Computer Vision and Pattern Recognition. 224–233.
  • Li et al. (2021b) Xinhao Li, Jingjing Li, Lei Zhu, Guoqing Wang, and Zi Huang. 2021b. Imbalanced Source-free Domain Adaptation. In Proceedings of the 29th ACM International Conference on Multimedia. 3330–3339.
  • Li et al. (2019) Yiying Li, Yongxin Yang, Wei Zhou, and Timothy Hospedales. 2019. Feature-critic networks for heterogeneous domain generalization. In International Conference on Machine Learning. PMLR, 3915–3924.
  • Liao et al. (2020) Yixiao Liao, Ruyi Huang, Jipu Li, Zhuyun Chen, and Weihua Li. 2020. Deep semisupervised domain generalization network for rotary machinery fault diagnosis under variable speed. IEEE Transactions on Instrumentation and Measurement 69, 10 (2020), 8064–8075.
  • Liu et al. (2021) Chang Liu, Lichen Wang, Kai Li, and Yun Fu. 2021. Domain Generalization via Feature Variation Decorrelation. In Proceedings of the 29th ACM International Conference on Multimedia. 1683–1691.
  • Lv et al. (2021) Jianming Lv, Kaijie Liu, and Shengfeng He. 2021. Differentiated Learning for Multi-Modal Domain Adaptation. In Proceedings of the 29th ACM International Conference on Multimedia. 1322–1330.
  • Ma et al. (2022) Xu Ma, Junkun Yuan, Yen-wei Chen, Ruofeng Tong, and Lanfen Lin. 2022. Attention-based cross-layer domain alignment for unsupervised domain adaptation. Neurocomputing 499 (2022), 1–10.
  • Maaten and Hinton (2008) L. V. D. Maaten and Geoffrey E. Hinton. 2008. Visualizing Data using t-SNE. Journal of Machine Learning Research 9 (2008), 2579–2605.
  • Mahajan et al. (2021) Divyat Mahajan, Shruti Tople, and Amit Sharma. 2021. Domain generalization using causal matching. In International Conference on Machine Learning. PMLR, 7313–7324.
  • Pandey et al. (2021) Prashant Pandey, Mrigank Raman, Sumanth Varambally, and Prathosh AP. 2021. Generalization on Unseen Domains via Inference-Time Label-Preserving Target Projections. In IEEE/CVF Conference on Computer Vision and Pattern Recognition. 12924–12933.
  • Qiao et al. (2020) Fengchun Qiao, Long Zhao, and Xi Peng. 2020. Learning to learn single domain generalization. In IEEE/CVF Conference on Computer Vision and Pattern Recognition. 12556–12565.
  • Quionero-Candela et al. (2009) Joaquin Quionero-Candela, Masashi Sugiyama, Anton Schwaighofer, and Neil D Lawrence. 2009. Dataset shift in machine learning. The MIT Press.
  • Sener and Savarese (2018) Ozan Sener and Silvio Savarese. 2018. Active learning for convolutional neural networks: A core-set approach. In International Conference on Learning Representations.
  • Shankar et al. (2018) Shiv Shankar, Vihari Piratla, Soumen Chakrabarti, Siddhartha Chaudhuri, Preethi Jyothi, and Sunita Sarawagi. 2018. Generalizing across domains via cross-gradient training. International Conference on Learning Representation (2018).
  • Sharifi-Noghabi et al. (2020) Hossein Sharifi-Noghabi, Hossein Asghari, Nazanin Mehrasa, and Martin Ester. 2020. Domain generalization via semi-supervised meta learning. arXiv preprint arXiv:2009.12658 (2020).
  • Shen et al. (2020) Zheyan Shen, Peng Cui, Tong Zhang, and Kun Kunag. 2020. Stable learning via sample reweighting. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 34. 5692–5699.
  • Shu et al. (2021) Yang Shu, Zhangjie Cao, Chenyu Wang, Jianmin Wang, and Mingsheng Long. 2021. Open Domain Generalization with Domain-Augmented Meta-Learning. In IEEE/CVF Conference on Computer Vision and Pattern Recognition. 9624–9633.
  • Sohn et al. (2020) Kihyuk Sohn, David Berthelot, Chun-Liang Li, Zizhao Zhang, Nicholas Carlini, Ekin Dogus Cubuk, Alexey Kurakin, Han Zhang, and Colin Raffel. 2020. FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence. Advances in Neural Information Processing Systems 33 (2020).
  • Srinivas et al. (2021) Aravind Srinivas, Tsung-Yi Lin, Niki Parmar, Jonathon Shlens, Pieter Abbeel, and Ashish Vaswani. 2021. Bottleneck transformers for visual recognition. In IEEE/CVF Conference on Computer Vision and Pattern Recognition. 16519–16529.
  • Tarvainen and Valpola (2017) Antti Tarvainen and Harri Valpola. 2017. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. Advances in Neural Information Processing Systems (2017).
  • Vapnik (1992) Vladimir Vapnik. 1992. Principles of risk minimization for learning theory. In Advances in neural information processing systems. 831–838.
  • Venkateswara et al. (2017) Hemanth Venkateswara, Jose Eusebio, Shayok Chakraborty, and Sethuraman Panchanathan. 2017. Deep hashing network for unsupervised domain adaptation. In IEEE conference on computer vision and pattern recognition. 5018–5027.
  • Volpi et al. (2021) Riccardo Volpi, Diane Larlus, and Grégory Rogez. 2021. Continual Adaptation of Visual Representations via Domain Randomization and Meta-learning. In IEEE/CVF Conference on Computer Vision and Pattern Recognition. 4443–4453.
  • Voulodimos et al. (2018) Athanasios Voulodimos, Nikolaos Doulamis, Anastasios Doulamis, and Eftychios Protopapadakis. 2018. Deep learning for computer vision: A brief review. Computational intelligence and neuroscience 2018 (2018).
  • Wang and Shang (2014) Dan Wang and Yi Shang. 2014. A new active labeling method for deep learning. In International joint conference on neural networks. IEEE, 112–119.
  • Wang et al. (2021c) Mengzhu Wang, Wei Wang, Baopu Li, Xiang Zhang, Long Lan, Huibin Tan, Tianyi Liang, Wei Yu, and Zhigang Luo. 2021c. InterBN: Channel Fusion for Adversarial Unsupervised Domain Adaptation. In Proceedings of the 29th ACM International Conference on Multimedia. 3691–3700.
  • Wang et al. (2021b) Ruiqi Wang, Lei Qi, Yinghuan Shi, and Yang Gao. 2021b. Better Pseudo-label: Joint Domain-aware Label and Dual-classifier for Semi-supervised Domain Generalization. arXiv preprint arXiv:2110.04820 (2021).
  • Wang et al. (2021a) Yufei Wang, Haoliang Li, Lap-pui Chau, and Alex C Kot. 2021a. Embracing the Dark Knowledge: Domain Generalization Using Regularized Knowledge Distillation. In Proceedings of the 29th ACM International Conference on Multimedia. 2595–2604.
  • Xu et al. (2021) Qinwei Xu, Ruipeng Zhang, Ya Zhang, Yanfeng Wang, and Qi Tian. 2021. A Fourier-based Framework for Domain Generalization. In IEEE/CVF Conference on Computer Vision and Pattern Recognition. 14383–14392.
  • Yan et al. (2021) Zizheng Yan, Xianggang Yu, Yipeng Qin, Yushuang Wu, Xiaoguang Han, and Shuguang Cui. 2021. Pixel-level intra-domain adaptation for semantic segmentation. In Proceedings of the 29th ACM International Conference on Multimedia. 404–413.
  • Ye et al. (2021) Mucong Ye, Jing Zhang, Jinpeng Ouyang, and Ding Yuan. 2021. Source Data-free Unsupervised Domain Adaptation for Semantic Segmentation. In Proceedings of the 29th ACM International Conference on Multimedia. 2233–2242.
  • Yuan et al. (2021a) Junkun Yuan, Xu Ma, Defang Chen, Kun Kuang, Fei Wu, and Lanfen Lin. 2021a. Collaborative Semantic Aggregation and Calibration for Separated Domain Generalization. arXiv e-prints (2021), arXiv–2110.
  • Yuan et al. (2021b) Junkun Yuan, Xu Ma, Defang Chen, Kun Kuang, Fei Wu, and Lanfen Lin. 2021b. Domain-Specific Bias Filtering for Single Labeled Domain Generalization. arXiv preprint arXiv:2110.00726 (2021).
  • Yuan et al. (2021c) Junkun Yuan, Xu Ma, Kun Kuang, Ruoxuan Xiong, Mingming Gong, and Lanfen Lin. 2021c. Learning domain-invariant relationship with instrumental variable for domain generalization. arXiv preprint arXiv:2110.01438 (2021).
  • Zhang et al. (2018) Hongyi Zhang, Moustapha Cisse, Yann N Dauphin, and David Lopez-Paz. 2018. mixup: Beyond empirical risk minimization. International Conference on Learning Representation (2018).
  • Zhang et al. (2021) Xingxuan Zhang, Peng Cui, Renzhe Xu, Linjun Zhou, Yue He, and Zheyan Shen. 2021. Deep Stable Learning for Out-Of-Distribution Generalization. In IEEE/CVF Conference on Computer Vision and Pattern Recognition. 5372–5382.
  • Zhao et al. (2020) Shanshan Zhao, Mingming Gong, Tongliang Liu, Huan Fu, and Dacheng Tao. 2020. Domain generalization via entropy regularization. Advances in Neural Information Processing Systems 33 (2020).
  • Zhou et al. (2021a) Kaiyang Zhou, Chen Change Loy, and Ziwei Liu. 2021a. Semi-Supervised Domain Generalization with Stochastic StyleMatch. arXiv preprint arXiv:2106.00592 (2021).
  • Zhou et al. (2020) Kaiyang Zhou, Yongxin Yang, Timothy Hospedales, and Tao Xiang. 2020. Deep domain-adversarial image generation for domain generalisation. In AAAI Conference on Artificial Intelligence.
  • Zhou et al. (2021b) Kaiyang Zhou, Yongxin Yang, Yu Qiao, and Tao Xiang. 2021b. Domain adaptive ensemble learning. IEEE Transactions on Image Processing 30 (2021), 8008–8018.
  • Zhou et al. (2021c) Kaiyang Zhou, Yongxin Yang, Yu Qiao, and Tao Xiang. 2021c. Domain Generalization with Mixstyle. In International Conference on Learning Representation.