This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Self-training Guided Adversarial Domain Adaptation For Thermal Imagery

Ibrahim Batuhan Akkaya1,2     Fazil Altinel111footnotemark: 1     Ugur Halici2,3
1Research Center
indicates equal contribution
   Aselsan Inc    Turkey
2Department of Electrical and Electronics Engineering
   Middle East Technical University    Turkey
3NOROM Neuroscience and Neurotechnology Excellency Center
   Turkey
{ibakkaya, faltinel}@aselsan.com.tr
   [email protected]
Abstract

Deep models trained on large-scale RGB image datasets have shown tremendous success. It is important to apply such deep models to real-world problems. However, these models suffer from a performance bottleneck under illumination changes. Thermal IR cameras are more robust against such changes, and thus can be very useful for the real-world problems. In order to investigate efficacy of combining feature-rich visible spectrum and thermal image modalities, we propose an unsupervised domain adaptation method which does not require RGB-to-thermal image pairs. We employ large-scale RGB dataset MS-COCO as source domain and thermal dataset FLIR ADAS as target domain to demonstrate results of our method. Although adversarial domain adaptation methods aim to align the distributions of source and target domains, simply aligning the distributions cannot guarantee perfect generalization to the target domain. To this end, we propose a self-training guided adversarial domain adaptation method to promote generalization capabilities of adversarial domain adaptation methods. To perform self-training, pseudo labels are assigned to the samples on the target thermal domain to learn more generalized representations for the target domain. Extensive experimental analyses show that our proposed method achieves better results than the state-of-the-art adversarial domain adaptation methods. The code and models are publicly available.111https://github.com/avaapm/SGADA

1 Introduction

Refer to caption
Figure 1: Visualization of class activation maps on a target domain image using occlusion sensitivity [35]. Given a person image (a), our proposed model (d) activates semantically more meaningful parts of the image compared to our base method (c) [32], while the model trained on only source domain images (b) misclassifies the image as bicycle and activates wrong regions. Best viewed in color.

Recently, significant improvements have been made by using RGB images for image classification and detection problems [9, 14, 20, 27, 28]. The state-of-the-art methods have been trained on large-scale RGB datasets such as MS-COCO [22], ImageNet [5], Pascal-VOC [7], etc. However, low-lighting conditions hinder current state-of-the-art deep learning methods trained on visible spectrum images from performing well on computer vision tasks such as image classification, object detection, etc. Since thermal IR cameras are more robust against these conditions, exploiting them is useful for real-world applications. Therefore, usage of thermal IR cameras has become more common in the tasks related to autonomous driving, military operations, security surveillance, etc. Since such large-scale thermal datasets are not publicly available, it still remains an important challenge to achieve same level of performance on thermal image datasets. Therefore, exploiting complementary information offered by visible spectrum images is a straightforward technique to improve performance of the methods which work on thermal images for classification and detection problems. Unfortunately, recent studies demonstrated that performance of a deep model well-trained on visible spectrum images may significantly drop when applying to thermal images [6, 12, 13, 16, 17].

Since deep networks are sensitive to domain shift, a deep model trained on a large amount of labeled source domain data may fail at generalizing to unlabeled target domain data which are not similar to source domain data. To overcome these issues, unsupervised domain adaptation (UDA) aims to learn a model which maps both domains into a common feature space without requiring image pairs. Among the recent UDA methods, adversarial domain adaptation methods have become popular [8, 23, 25, 32]. These approaches incorporate adversarial learning as a two-player game similar to generative adversarial networks (GANs) [10]. Adversarial domain adaptation methods utilize a domain discriminator to distinguish source domain from target domain and a feature extractor to learn domain invariant representations to fool the domain discriminator. By learning domain invariant feature representations, adversarial domain adaptation methods assume that a classifier trained on source domain features is able to successfully classify target domain samples as well.

In this paper, we propose an unsupervised adversarial domain adaptation method to align source and target domain distributions as described in Section 3. We employ Adversarial Discriminative Domain Adaptation (ADDA) [32] method as our base method. Although ADDA and other adversarial domain adaptation methods have achieved successful results, these methods face a major generalization limitation. The limitation is that even though the distributions are aligned by learning domain invariant representations with a feature extractor, theoretically, the classifier may not work well on the target domain as shown in [1]. Therefore, learning discriminative representations for the unlabeled target domain is a difficult problem.

Based on the assumption of self-training, a classifiers’ own high-confidence predictions are correct [38]. Since we assume that the predictions are mostly correct, exploiting the samples with high confidence values and retraining the classifier further improves the performance of the classifier. To this end, recent adversarial domain adaptation methods proposed to use pseudo-labels obtained from a classifier and retrain the model using the pseudo-labeled samples [29, 34, 36]. With these in mind, in this study, we propose a self-training guided adversarial domain adaptation (SGADA) method to overcome the generalization problems of adversarial domain adaptation methods (Figure 2). To perform self-training, pseudo labels obtained after warm-up phase of our method are assigned to the samples on the target domain to learn more generalized representations for the target domain. Pseudo-labels are assigned if confidences of the classifier trained on source domain and the domain discriminator reaches to threshold values for a target domain sample.

Our proposed method makes use of features obtained from visual spectrum images to improve classification performance on thermal domain. Moreover, our method does not need paired samples of RGB and thermal datasets. In order to train and test our proposed method, we use large-scale RGB dataset MS-COCO [22] and thermal imagery dataset FLIR ADAS [11]. We evaluate the proposed method quantitatively and qualitatively. We demonstrate our methods’ success compared to the state-of-the-art unsupervised domain adaptation methods in Section 4. The results show that our method improves the performance of our base model and outperforms the state-of-the-art methods. Moreover, Figure 1 depicts that given a thermal image, our method classifies the image correctly by activating semantically more meaningful regions compared to our base method ADDA [32] and the model trained only on source domain data.

Effective classification for imbalanced data is an important field of research since class imbalance exists in many real-world applications [3, 18]. Therefore, it is important to address the class imbalance problem. In our experimental studies (Section 4), we show that class imbalanced datasets cause UDA methods to over-classify the majority category. We show that our proposed method achieves better results compared to the state-of-the-art UDA methods when class imbalance exists.

Our contributions are summarized as follows:

  • We demonstrate the efficacy of combining visible spectrum and thermal image modalities by using unsupervised domain adaptation without requiring RGB-to-thermal image pairs.

  • We propose a self-training guided adversarial domain adaptation method for thermal imagery. In order to learn more generalized feature representations for target thermal domain, we employ pseudo-labels generated by the classifier trained on RGB images and the discriminator, and train our model with these pseudo-labels.

  • In order to demonstrate results of our method, we employ the large-scale RGB dataset MS-COCO as source domain and the thermal dataset FLIR ADAS as target domain. Extensive experimental analyses show that our proposed method outperforms the state-of-the-art unsupervised domain adaptation methods.

2 Related Work

Refer to caption
Figure 2: An overview of our proposed self-training guided adversarial domain adaptation (SGADA) method. Pseudo-labels generated after our methods’ warm-up phase are assigned to target thermal images. Then, the target CNN 𝐅t\mathbf{F}_{t} is trained using the pseudo-labels. The classifier 𝐂\mathbf{C} and the source CNN 𝐅s\mathbf{F}_{s} are reused from our base model, and thus they are fixed. Target feature representations are learned by updating parameters of the target CNN 𝐅t\mathbf{F}_{t} with respect to losses generated by the discriminator 𝐃\mathbf{D} and the classifier 𝐂\mathbf{C}. Blue boxes indicate fixed network parameters while red boxes indicate trainable network parameters. Best viewed in color.

By using RGB images, deep neural networks have gained popularity on computer vision tasks such as object detection, classification etc. Although significant improvements have been accomplished by using visible spectrum images, it still remains a critical problem to train a deep model robust to real-world problems, e.g. low-lighting conditions. To overcome these problems, thermal imagery has been used for object detection and classification problems [13, 16, 31].

Some of recent studies investigated the effects of combining RGB and thermal images to the performance on object detection problem [6, 13, 16, 17]. Since large-scale thermal datasets are not publicly available, in this study, we exploit complementary information offered by visible spectrum images to improve classification performance on thermal imagery without requiring RGB-to-thermal image pairs. We propose an unsupervised domain adaptation method in order to investigate efficacy of combining visible spectrum and thermal image modalities.

Numerous recent methods attempted to address domain adaptation problem. Recently, Generative Adversarial Networks (GANs) [10] inspired the field of domain adaptation, and thus deep adversarial domain adaptation methods have become popular [2, 8, 23, 25, 32].

Feature-level adversarial domain adaptation methods incorporate a domain discriminator to distinguish source and target domains while feature extractor learns features to fool the discriminator. Ganin et al. [8] proposed a gradient reversal layer to learn a feature extractor which generates features that maximize domain discriminator loss while minimizing label prediction loss. More recently, Tzeng et al. [32] proposed a method to learn a discriminative mapping of target images to source feature space by fooling a domain discriminator which distinguishes the encoded target images from source samples. Many recent works employ adversarial training paradigm in their domain adaptation procedure [23, 25]. Although feature-level adversarial domain adaptation methods have accomplished successful empirical results, these methods suffer from a major limitation. As shown in [1], even if a feature extractor is well learned to generate domain invariant features, theoretically, the classifier may not work well on the target domain. Therefore, learning discriminative representations for the unlabeled target domain is considered difficult.

On the other hand, pixel-level adversarial domain adaptation methods translate source domain data into target domain data or vice versa by using image-to-image translation [24]. Bousmalis et al. [2] proposed an approach to learn a transformation in pixel-level from one domain to the other. Inspired by CycleGAN [37], Hoffman et al. [15] proposed CyCADA to increase semantic consistency of the image translation to improve the pixel-level methods. Even though pixel-level adversarial domain adaptation studies present remarkable results, image-to-image translation sometimes performs poorly on the datasets which have objects with many complex structures.

To overcome the limitations of adversarial domain adaptation methods, recent studies propose to directly deal with relationship between decision boundary and learned feature representations [21, 30]. Saito et al. [30] introduced to use a minimax training method to push target feature distributions away. Lee et al. [21] proposed to exploit adversarial dropout mechanism to learn more discriminative features by enforcing cluster assumption [4]. However, our experimental studies show that these methods and aforementioned adversarial domain adaptation methods have a drawback: Class imbalanced datasets [3, 18] lead to a performance drop for these methods.

We employ a self-training guided adversarial domain adaptation method to deal with the generalization problems of adversarial domain adaptation methods for thermal imagery. To the best of our knowledge, there is no self-training guided domain adaptation study in the literature of thermal image classification. Self-training is a technique to assign pseudo-labels to unlabeled samples using predictions of a classifier and retrain the model including the pseudo-labeled samples. Based on the assumption of self-training, a classifiers’ own high-confidence predictions are correct [38]. Recent adversarial domain adaptation methods proposed to use pseudo-labels [29, 34, 36]. In our experiments, we show that our self-training guided method performs better than previous domain adaptation methods under class imbalance problem by learning more generalized representations for target thermal domain.

3 Proposed Method

Our proposed self-training guided adversarial domain adaptation method is illustrated in Figure 2. Before performing self-training, we extract pseudo-labels for the target domain samples. The pseudo-label extraction mechanism is depicted in Figure 3.

First, a feature extractor 𝐅s\mathbf{F}_{s} and a classifier 𝐂\mathbf{C} are trained on source domain using labeled source domain RGB images (Figure 3-(a)). This step is named as pre-training. After this step, the classifier network 𝐂\mathbf{C} can successfully classify the source domain images by exploiting the features which are extracted by the source Convolutional Neural Network (CNN) 𝐅s\mathbf{F}_{s}. After the training on the source domain, we perform the second step: the warm-up phase for pseudo-label generation (Figure 3-(b)). In this step, we fix the parameters of the feature extractor 𝐅s\mathbf{F}_{s} trained on the source domain. A target specific feature extractor 𝐅t\mathbf{F}_{t} is learned in an unsupervised manner. By performing this step, features extracted from the source domain and the target domain are aligned with adversarial training. Therefore, we can use the classifier 𝐂\mathbf{C} trained on the source domain to classify target domain samples. We perform aforementioned two steps by following the training process of our base method ADDA [32]. In the last step, we fix the parameters of the feature extractor 𝐅t\mathbf{F}_{t} trained on the target domain, the classifier 𝐂\mathbf{C} trained on the source domain and the discriminator 𝐃\mathbf{D}. Then, we obtain predictions from the classifier and confidences from both the classifier and the discriminator for the target domain samples.

Once the predictions and the confidences are obtained, we utilize the predictions to give pseudo-labels for target domain samples using the confidences obtained from the classifier 𝐂\mathbf{C} and the discriminator 𝐃\mathbf{D} as shown in Figure 3-(c). We use prediction of the classifier for a target sample if classifiers’ confidence value is higher than a threshold and domain label prediction of discriminator is close to source domain. That is, we can use the prediction of the classifier if the discriminator incorrectly classifies target samples. By using this pseudo-label selection mechanism, intuitively, we select samples with feature representations which are close to data with known labels. Next, as illustrated in Figure 2, we train our proposed method using extracted pseudo-labels.

Refer to caption
Figure 3: Illustration of our pseudo-label generation mechanism.

A general definition of unsupervised domain adaptation, and self-training guided adversarial domain adaptation procedures of our proposed method are described in Section 3.1 and 3.2, respectively.

3.1 Unsupervised Domain Adaptation

In the general definition of unsupervised domain adaptation (UDA) problem, we are given nsn_{s} labeled samples from a source domain 𝒟s={(xis,yis)}i=1ns\mathcal{D}_{s}=\{(x^{s}_{i},y^{s}_{i})\}_{i=1}^{n_{s}} and ntn_{t} unlabeled samples from a target domain 𝒟t={(xjt)}j=1nt\mathcal{D}_{t}=\{(x^{t}_{j})\}_{j=1}^{n_{t}}. The goal of UDA problem is to learn a feature extractor 𝐅t\mathbf{F}_{t} for the target domain and a classifier 𝐂t\mathbf{C}_{t} which correctly classifies the features. It is not possible to perform supervised training since there is no labeled samples in the target domain. Therefore, UDA learns to adapt the source feature extractor 𝐅s\mathbf{F}_{s} and the source classifier 𝐂s\mathbf{C}_{s} to be able to use them on target domain.

3.2 Self-training Guided Adversarial Domain Adaptation (SGADA)

The task of adversarial domain adaptation methods is to adversarially align source and target domain representations. For this purpose, adversarial domain adaptation methods propose to reduce the gap between 𝐅s(xs)\mathbf{F}_{s}(x^{s}) and 𝐅t(xt)\mathbf{F}_{t}(x^{t}). Thus, the classifier 𝐂s\mathbf{C}_{s} trained on source domain can be applied to the representations on target domain, and necessity to train a separate 𝐂t\mathbf{C}_{t} can be eliminated. As a result, we obtain 𝐂=𝐂s=𝐂t\mathbf{C}=\mathbf{C}_{s}=\mathbf{C}_{t} [32]. We employ the feature extractor 𝐅s\mathbf{F}_{s} and the classifier 𝐂\mathbf{C} which are learned during the warm-up phase. In this subsection, we elaborate our training scheme shown in Figure 2.

We use the following loss function for the domain discriminator 𝐃\mathbf{D} which distinguishes source domain from target domain:

advD(xs,xt,𝐅s,𝐅t)=1nsi=1nslog[𝐃(𝐅s(xis))]\displaystyle\mathcal{L}_{advD}(x^{s},x^{t},\mathbf{F}_{s},\mathbf{F}_{t})=-\frac{1}{n_{s}}\sum_{i=1}^{n_{s}}\log[\mathbf{D}(\mathbf{F}_{s}(x_{i}^{s}))] (1)
1nti=1ntlog[1𝐃(𝐅t(xit))].\displaystyle-\frac{1}{n_{t}}\sum_{i=1}^{n_{t}}\log[1-\mathbf{D}(\mathbf{F}_{t}(x_{i}^{t}))].

Given source images xsx^{s} and target images xtx^{t}, we update the parameters of the domain discriminator 𝐃\mathbf{D} with respect to outputs of the feature extractors 𝐅s\mathbf{F}_{s} and 𝐅t\mathbf{F}_{t}. While updating the parameters, we fix and reuse the source feature extractor 𝐅s\mathbf{F}_{s} which is trained in the pre-training step of our pseudo-label generation.

We employ two loss functions to train the target CNN 𝐅t\mathbf{F}_{t}: adversarial loss advF\mathcal{L}_{advF} and self-training loss clsP\mathcal{L}_{clsP}. The adversarial loss is formulated as follows:

advF(xs,ys,𝐃)=1nti=1ntlog[𝐃(𝐅t(xit))].\mathcal{L}_{advF}(x^{s},y^{s},\mathbf{D})=\frac{1}{n_{t}}\sum_{i=1}^{n_{t}}\log[\mathbf{D}(\mathbf{F}_{t}(x_{i}^{t}))]. (2)

Note that we reuse the parameters of the source feature extractor 𝐅s\mathbf{F}_{s} from the previous step to initialize 𝐅t\mathbf{F}_{t}.

We exploit pseudo-labeled target domain samples to perform self-training guided adversarial learning. After the learning of the classifier 𝐂\mathbf{C} and the domain discriminator 𝐃\mathbf{D} is completed during the warm-up phase, we obtain predictions from the classifier and confidences for these predictions. Given an unlabeled target domain sample, if confidence of the classifier 𝐂\mathbf{C} is higher than pre-defined threshold and the domain discriminator 𝐃\mathbf{D} classifies the sample as source domain, we include the sample during our self-training guided adversarial domain adaptation step. Also, if the domain discriminator 𝐃\mathbf{D} classifies the sample as target domain with a confidence lower than a pre-defined threshold, we assign the pseudo-label y^t\hat{y}^{t} generated by the classifier 𝐂\mathbf{C} to the sample as well. By using n^t\hat{n}_{t} pseudo-labeled samples on target domain, we aim to train a target specific feature extractor (Figure 2). We use the new self-training loss function for our proposed method:

clsP(xt,y^t)=1n^ti=1n^tce(𝐂(𝐅t(xit)),y^it).\mathcal{L}_{clsP}(x^{t},\hat{y}^{t})=\frac{1}{\hat{n}^{t}}\sum_{i=1}^{\hat{n}^{t}}\ell_{ce}(\mathbf{C}(\mathbf{F}_{t}(x_{i}^{t})),\hat{y}_{i}^{t}). (3)

The overall objective function to train our proposed method SGADA is defined as:

min𝐃advD(xs,xt,𝐅s,𝐅t)\displaystyle\min_{\mathbf{D}}\mathcal{L}_{advD}(x^{s},x^{t},\mathbf{F}_{s},\mathbf{F}_{t}) (4)
min𝐅tadvF(xs,ys,𝐃)\displaystyle\min_{\mathbf{F}_{t}}\mathcal{L}_{advF}(x^{s},y^{s},\mathbf{D})
+λclsP(xt,y^t),\displaystyle+\lambda\mathcal{L}_{clsP}(x^{t},\hat{y}^{t}),

where λ\lambda is a trade-off parameter. We set the trade-off parameter λ\lambda and thresholds based on validation split (see Section 4.2 for further details).

4 Experiments

We perform extensive evaluations and compare our proposed method with several state-of-the-art unsupervised domain adaptation methods.

4.1 Datasets

We prepare a new RGB-to-thermal domain adaptation setting for classification using FLIR ADAS [11] as thermal dataset and MS-COCO [22] as visible spectrum dataset for our experimental studies.

Refer to caption
Figure 4: The network architectures used for our experimental analyses.

FLIR ADAS [11] consists of 9,214 thermal images with bounding box annotations. Each image has a resolution of 640 ×\times 512 and obtained from FLIR Tau2 camera. 60% of the images are captured during daytime and the remaining 40% of the images are captured during night. The dataset provides both visible spectrum (RGB) images and thermal images. We consider only the thermal images of the dataset for our experiments. We use the training and test splits as suggested in the dataset documentation for our experiments. The objects in the dataset are classified into four categories i.e. bicycle, car, dog, and person. However, the dog class has very few annotations. Therefore, the dog class is not considered in our experimental studies. We crop square images using bounding box annotations for objects. After objects are extracted, we resize the images to 224 ×\times 224. Finally, our thermal dataset consists of 4,137 samples of bicycle, 43,734 samples of car, and 26,294 samples of person images. Example images from FLIR ADAS dataset are shown in the second row of Figure 5.

Refer to caption
Figure 5: Example images from MS-COCO dataset [22] and FLIR ADAS thermal dataset [11]. Best viewed in color.
Refer to caption
Figure 6: The t-SNE visualization of network activations on target thermal domain generated by source only model (a), our base method ADDA [32] (b), our proposed method SGADA with classifier confidences only (c), and our proposed method SGADA with classifier and discriminator confidences (d). Best viewed in color.

Our proposed method incorporates publicly available large-scale visible spectrum datasets to improve classification performance on thermal dataset. Therefore, we consider using an RGB dataset which includes the same classes as FLIR dataset [11] (bicycle, car, and person). For this purpose, we use MS-COCO dataset [22] as our visible spectrum dataset. In the first row of Figure 5, we show some example images from MS-COCO dataset. MS-COCO dataset contains 91 object categories (airplane, bicycle, bird, car, person, etc.). In total, there are 123,287 images and around 886,000 bounding boxes. 118,287 of the images are for training while 5,000 of the images are for validation split. We apply standard training and test splits as provided in the dataset documentation for our experiments. We use only bicycle, car, and person classes to match with our thermal dataset. We cropped the annotated objects with the same procedure as applied to FLIR dataset. Once objects are extracted, we resize the images to 224 ×\times 224. Our visible spectrum image dataset extracted from MS-COCO consists of 5,732 samples of bicycle, 38,453 samples of car, and 209,162 samples of person images.

4.2 Implementation Details

For our experiments, we used same training procedure with ADDA [32]. For a fair comparison with other methods, we employed ResNet-50 [14] pre-trained on ImageNet [5] as backbone for all methods. Our network architectures are given in Figure 4. The architecture of our feature extractors (source CNN 𝐅s\mathbf{F}_{s} and target CNN 𝐅t\mathbf{F}_{t}) is the ResNet-50 without the last fully connected (FC) layer. In the figure, each convolutional residual unit is depicted with the size of filters at the top and the outputs of each convolutional layer at the bottom. The notation k ×\times k, n in the convolutional layer blocks represents a filter of size k and n channel. The number on the top of the convolutional layer blocks denotes the number of repetition for the unit. The domain discriminator 𝐃\mathbf{D} consists of three FC layers: two consecutive hidden units of 500 neurons and the discriminator output. The classifier 𝐂\mathbf{C} has only one FC layer.

To train our method, we set batch size to 32. Number of epochs was set to 15 in our experiments. Parameters were updated using ADAM optimization algorithm [19]. For pre-training of our method on source domain only, we set learning rate to 5e-4. For adversarial adaptation step of our method, we set learning rate as 1e-5 and discriminator learning rate as 1e-3. We used same learning rates as used in the previous step for self-training guided adversarial adaptation step of our method. We set λ\lambda value as 0.7 and threshold value as 0.87 for our method with classifier confidences only (SGADA-Cls). For our method with classifier and discriminator confidences (SGADA-Cls+Disc), we used same learning rates as used in the previous step, and λ\lambda value of 0.25, classifier threshold value of 0.79, and discriminator threshold value of 0.87. We used the same experimental settings for training and testing. We exploit classification accuracy to compare our proposed method with other methods.

We implemented our proposed method using PyTorch framework [26]. Implementation details, models, and the code were made publicly available at https://github.com/avaapm/SGADA.

4.3 Evaluation of SGADA

In our experiments, we select visual spectrum (RGB) domain as the source domain, and thermal domain as the target domain. As a general practice in the field of domain adaptation, we denote source only as the target domain performance of a model trained using only source domain images, and target only as that of a model trained on the target domain. Performances of source only and target only models serve as baselines for the lower and upper bound performances.

Table 1: Per-class classification performance comparison.
Method

Bicycle

Car

Person

Average

Source only 69.89 83.89 86.52 80.10
Pixel-DA [2] 62.53 89.99 76.73 76.42
DTA [21] 75.45 97.65 92.45 88.52
MCD-DA [30] 81.71 94.90 91.83 89.48
DANN [8] 78.16 95.07 96.24 89.82
CDAN [25] 78.16 97.10 94.82 90.03
ADDA [32] 86.67 96.95 89.10 90.90
SGADA (ours) 87.13 94.44 92.03 91.20
Target only 87.59 98.78 96.35 94.24

Quantitative Analysis.

We compare our proposed method SGADA with several state-of-the-art unsupervised domain adaptation methods: Unsupervised Pixel-Level Domain Adaptation with Generative Adversarial Networks (Pixel-DA) [2], Drop to Adapt (DTA) [21], Maximum Classifier Discrepancy for Unsupervised Domain Adaptation (MCD-DA) [30], Domain Adversarial Neural Network (DANN) [8], Conditional Domain Adversarial Adaptation (CDAN) [25], and Adversarial Discriminative Domain Adaptation (ADDA) [32]. Since these methods do not consider domain adaptation problem for thermal datasets, there exist no reported results on their paper for our dataset. Therefore, we trained and evaluated all these methods for our dataset.

Table 2: Ablation studies of different pseudo-label selection scenarios for self-training guided adversarial domain adaptation.
Bicycle Car Person
Number of samples 3702 38657 21081
Classifier confidences only Number of selected samples 3995 35494 20323
Number of correctly selected samples 2901 34905 18911
Accuracy of selected samples (%) 72.62 98.34 93.05
Discriminator confidences only Number of selected samples 3598 36024 3800
Number of correctly selected samples 2874 35251 3549
Accuracy of selected samples (%) 79.88 97.85 93.39
Classifier and discriminator confidences together Number of selected samples 3557 35123 3558
Number of correctly selected samples 2873 34558 3454
Accuracy of selected samples (%) 80.77 98.39 97.08
Table 3: Ablation experiments.
Method

Bicycle

Car

Person

Average

SGADA-Cls 87.36 95.27 90.62 91.08
SGADA-Cls+Disc 87.13 94.44 92.03 91.20

Per-class domain adaptation performances are reported in Table 1. The results show that our proposed method which uses classifier and discriminator confidences together for self-training outperforms the state-of-the-art methods. Although DTA [21] and DANN [8] perform well for car and person classes respectively, their performance for bicycle class cannot reach the top performances since the number of samples for bicycle class is much less than the other classes. On the other hand, our proposed method achieves more balanced performance scores for all classes and outperforms other methods. It is important to address this problem since datasets for real-world problems usually include imbalanced classes [3, 18]. As shown in the table, our proposed method achieves more balanced class-wise accuracies compared to our base method ADDA [32], and furthermore our method increases the average accuracy over our base method.

Qualitative Analysis.

We visualize the feature representations on target thermal domain with t-SNE [33] for qualitative analysis in Figure 6. The features of source only model on target domain can not be discriminated very well while ADDA [32] discriminate some overlapping points in the feature space. Our proposed model which uses only classifier confidences for self-training (SGADA-Cls) learns more discriminative representations. As shown in the figure, our proposed model which uses classifier and discriminator confidences together for self-training (SGADA-Cls+Disc) further enlarges inter-class distances, especially for car and person classes.

Ablation Study.

To evaluate the contributions of our proposed method, we perform ablation studies. We examine effects of using classifier and/or discriminator confidences for pseudo-label selection. As described in Section 3, we select samples using our base method. Table 2 shows three cases where we select target domain samples using only the confidences of the classifier, using only the confidences of the discriminator, and using the confidences of both the classifier and the discriminator. As shown in the table, if we utilize the discriminator confidences, the number of selected samples for the person class decreases. Moreover, when we use both discriminator and classifier confidences to select target domain samples, accuracy of the pseudo-labels increases significantly for all classes. This results in better separation of feature representations as depicted in Figure 6 (c)-(d). Furthermore, since accuracy of selected samples for all classes using classifier and discriminator confidences is higher than the other cases, class imbalance of our proposed method (SGADA-Cls+Disc) reduces compared to SGADA-Cls as shown in Table 3. And thus, the overall accuracy of our proposed method surpasses SGADA-Cls, resulting in the best overall performance.

5 Conclusion

In this paper, we propose a self-training guided adversarial domain adaptation method in order to investigate the efficacy of combining visible spectrum and thermal image modalities by using unsupervised domain adaptation. To overcome the generalization problems of the current adversarial domain adaptation methods, we employ pseudo-labels obtained from a classifier trained on RGB images, and train our method with these pseudo-labels. In order to demonstrate results of our method, we use large-scale RGB dataset MS-COCO as the source domain and thermal dataset FLIR ADAS as the target domain. Quantitative and qualitative results show that our proposed method achieves better results than the state-of-the-art adversarial domain adaptation methods by learning more generalized feature representations for target thermal domain.

References

  • [1] Shai Ben-David, John Blitzer, Koby Crammer, Alex Kulesza, Fernando Pereira, and Jennifer Wortman Vaughan. A theory of learning from different domains. Machine learning, 2010.
  • [2] Konstantinos Bousmalis, Nathan Silberman, David Dohan, Dumitru Erhan, and Dilip Krishnan. Unsupervised pixel-level domain adaptation with generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
  • [3] Mateusz Buda, Atsuto Maki, and Maciej A Mazurowski. A systematic study of the class imbalance problem in convolutional neural networks. Neural Networks, 2018.
  • [4] Olivier Chapelle and Alexander Zien. Semi-supervised classification by low density separation. In Proceedings of the International Conference on Artificial Intelligence and Statistics (AISTATS), 2005.
  • [5] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei. ImageNet: A Large-Scale Hierarchical Image Database. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2009.
  • [6] Chaitanya Devaguptapu, Ninad Akolekar, Manuj M Sharma, and Vineeth N Balasubramanian. Borrow from anywhere: Pseudo multi-modal object detection in thermal imagery. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 2019.
  • [7] Mark Everingham, Luc Gool, Christopher K. Williams, John Winn, and Andrew Zisserman. The pascal visual object classes (voc) challenge. International Journal of Computer Vision (IJCV), 2010.
  • [8] Yaroslav Ganin and Victor Lempitsky. Unsupervised domain adaptation by backpropagation. In Proceedings of the International Conference on Machine Learning (ICML), 2015.
  • [9] Ross Girshick, Jeff Donahue, Trevor Darrell, and Jitendra Malik. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2014.
  • [10] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Advances in Neural Information Processing Systems (NeurIPS), 2014.
  • [11] F. A. Group. Flir thermal dataset for algorithm training. https://www.flir.com/oem/adas/adas-dataset-form/.
  • [12] Dayan Guan, Xing Luo, Yanpeng Cao, Jiangxin Yang, Yanlong Cao, George Vosselman, and Michael Ying Yang. Unsupervised domain adaptation for multispectral pedestrian detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 2019.
  • [13] Tiantong Guo, Cong Phuoc Huynh, and Mashhour Solh. Domain-adaptive pedestrian detection in thermal images. In Proceedings of the IEEE International Conference on Image Processing (ICIP), 2019.
  • [14] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
  • [15] Judy Hoffman, Eric Tzeng, Taesung Park, Jun-Yan Zhu, Phillip Isola, Kate Saenko, Alexei Efros, and Trevor Darrell. Cycada: Cycle-consistent adversarial domain adaptation. In Proceedings of the International Conference on Machine Learning (ICML), 2018.
  • [16] Soonmin Hwang, Jaesik Park, Namil Kim, Yukyung Choi, and In So Kweon. Multispectral pedestrian detection: Benchmark dataset and baseline. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2015.
  • [17] Shu Wang Jingjing Liu, Shaoting Zhang and Dimitris Metaxas. Multispectral deep neural networks for pedestrian detection. In Proceedings of the British Machine Vision Conference (BMVC), 2016.
  • [18] Justin M Johnson and Taghi M Khoshgoftaar. Survey on deep learning with class imbalance. Journal of Big Data, 2019.
  • [19] Diederik P Kingma and Jimmy Lei Ba. Adam: A method for stochastic gradient descent. In Proceedings of the International Conference on Learning Representations (ICLR), 2015.
  • [20] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. Advances in Neural Information Processing Systems (NeurIPS), 2012.
  • [21] Seungmin Lee, Dongwan Kim, Namil Kim, and Seong-Gyun Jeong. Drop to adapt: Learning discriminative features for unsupervised domain adaptation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2019.
  • [22] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollar, and Larry Zitnick. Microsoft coco: Common objects in context. In Proceedings of the IEEE/CVF European Conference on Computer Vision (ECCV), 2014.
  • [23] Hong Liu, Mingsheng Long, Jianmin Wang, and Michael Jordan. Transferable adversarial training: A general approach to adapting deep classifiers. In Proceedings of the International Conference on Machine Learning (ICML), 2019.
  • [24] Ming-Yu Liu, Thomas Breuel, and Jan Kautz. Unsupervised image-to-image translation networks. In Advances in Neural Information Processing Systems (NeurIPS), 2017.
  • [25] Mingsheng Long, Zhangjie Cao, Jianmin Wang, and Michael I Jordan. Conditional adversarial domain adaptation. In Advances in Neural Information Processing Systems (NeurIPS), 2018.
  • [26] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. Pytorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems (NeurIPS), 2019.
  • [27] Joseph Redmon, Santosh Divvala, Ross Girshick, and Ali Farhadi. You only look once: Unified, real-time object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
  • [28] Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster r-cnn: Towards real-time object detection with region proposal networks. In Advances in Neural Information Processing Systems (NeurIPS), 2015.
  • [29] Kuniaki Saito, Yoshitaka Ushiku, and Tatsuya Harada. Asymmetric tri-training for unsupervised domain adaptation. In Proceedings of the International Conference on Machine Learning (ICML), 2017.
  • [30] Kuniaki Saito, Kohei Watanabe, Yoshitaka Ushiku, and Tatsuya Harada. Maximum classifier discrepancy for unsupervised domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2018.
  • [31] Philip Saponaro, Scott Sorensen, Abhishek Kolagunda, and Chandra Kambhamettu. Material classification with thermal imagery. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2015.
  • [32] Eric Tzeng, Judy Hoffman, Kate Saenko, and Trevor Darrell. Adversarial discriminative domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
  • [33] Laurens van der Maaten and Geoffrey Hinton. Visualizing data using t-sne. Journal of Machine Learning Research, 2008.
  • [34] Shaoan Xie, Zibin Zheng, Liang Chen, and Chuan Chen. Learning semantic representations for unsupervised domain adaptation. In Proceedings of the International Conference on Machine Learning (ICML), 2018.
  • [35] Matthew D. Zeiler and Rob Fergus. Visualizing and understanding convolutional networks. In Proceedings of the IEEE/CVF European Conference on Computer Vision (ECCV), 2014.
  • [36] Weichen Zhang, Wanli Ouyang, Wen Li, and Dong Xu. Collaborative and adversarial network for unsupervised domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2018.
  • [37] J. Zhu, T. Park, P. Isola, and A. A. Efros. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2017.
  • [38] Xiaojin Jerry Zhu. Semi-supervised learning literature survey. University of Wisconsin-Madison Department of Computer Sciences, 2005.