This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Stagewise Unsupervised Domain Adaptation with Adversarial Self-Training for Road Segmentation of Remote Sensing Images

Lefei Zhang,  Meng Lan, Jing Zhang,  and Dacheng Tao Manuscript received 27th February 2021, revised 19th May 2021 and 17th July 2021, accepted 7th august 2021. This work was supported by the National Natural Science Foundation of China under Grant 62076188, the Science and Technology Major Project of Hubei Province (Next-Generation AI Technologies) under Grant 2019AEA170, and the Fundamental Research Funds for the Central Universities under Grant 2042021kf0196. The numerical calculations in this paper had been supported by the supercomputing system in the Supercomputing Center of Wuhan University. Dr. Jing Zhang is supported by ARC project FL-170100117. This work was done during M. Lan’s internship at JD Explore Academy.L. Zhang and M. Lan are with the School of Computer Science, Institute of Artificial Intelligence, Wuhan University, Wuhan China. (e-mail:{zhanglefei, menglan}@whu.edu.cn).J. Zhang is with the School of Computer Science, in the Faculty of Engineering, at the University of Sydney, Sydney, Australia. (e-mail: [email protected]). D. Tao is with JD Explore Academy, China (e-mail: [email protected])
Abstract

Road segmentation from remote sensing images is a challenging task with wide ranges of application potentials. Deep neural networks have advanced this field by leveraging the power of large-scale labeled data, which, however, are extremely expensive and time-consuming to acquire. One solution is to use cheap available data to train a model and deploy it to directly process the data from a specific application domain. Nevertheless, the well-known domain shift (DS) issue prevents the trained model from generalizing well on the target domain. In this paper, we propose a novel stagewise domain adaptation model called RoadDA to address the DS issue in this field. In the first stage, RoadDA adapts the target domain features to align with the source ones via generative adversarial networks (GAN) based inter-domain adaptation. Specifically, a feature pyramid fusion module is devised to avoid information loss of long and thin roads and learn discriminative and robust features. Besides, to address the intra-domain discrepancy in the target domain, in the second stage, we propose an adversarial self-training method. We generate the pseudo labels of target domain using the trained generator and divide it to labeled easy split and unlabeled hard split based on the road confidence scores. The features of hard split are adapted to align with the easy ones using adversarial learning and the intra-domain adaptation process is repeated to progressively improve the segmentation performance. Experiment results on two benchmarks demonstrate that RoadDA can efficiently reduce the domain gap and outperforms state-of-the-art methods. The code is available at https://github.com/LANMNG/RoadDA.

Index Terms:
remote sensing, road segmentation, unsupervised domain adaptation, self-training.

I Introduction

Road segmentation from remote sensing (RS) images is a crucial research topic in remote sensing field. It aims to separate the road areas from the complex background and assign the right label to each pixel of the whole image. Road segmentation has many important applications, such as vehicle navigation [1], urban planning [2, 3], disaster assistance [4] and so on. Various methods have been proposed to effectively address the road segmentation task. However, the unsupervised learning based methods usually meet low accuracy for depending on the predefined features [1, 5]. As the rapid development of deep learning technique, supervised learning based convolutional neural networks have shown excellent ability of feature extraction and have been applied to many remote sensing image processing tasks [6, 7, 8, 9, 10]. Especially, with the aid of deep learning, road segmentation of remote sensing images has made great progress [11, 12].

Refer to caption
Figure 1: How to exploit the available labeled road data (source domain) to improve the segmentation performance of unlabeled road data (target domain) are of practical importance. (a1) and (b1) are the results obtained by directly applying the segmentation model trained on source domain to two target domain images; (a2) and (b2) are the results of our model.

Many CNNs based road segmentation approaches achive good performance on the public remote sensing benchmark datasets. Cheng et al. [13] proposed a cascaded encoder-decoder network for road segmentation and evaluated it on the collected road dataset which contains 224 manually labeled Google Earth RGB images. [3] designed a RCNN unit with shared weights of convolutional filters to build a deeper network. The model was test on the RoadTracer dataset which obtains the satellite images from Google Map labeled with OpenStreeMap [14]. Zhou et al. [15] introduced the D-LinkNet and won the first prize of Road Extraction Challenge in DeepGlobe 2018, in which the released road dataset has 6226 DigitalGlobe RGB training images. It is exciting to witness various approaches to achieve better performance on different road scenarios, whereas these supervised learning based road segmentation methods always require a large amount of labeled data to train the model and it could be extremely expensive and time-consuming to manually label the road areas from a lot of unlabeled high-resolution remote sensing images in the practical applications.

Fortunately, we have access to some public road datasets released by previous researchers and abundant digital map resources, such as Google Maps and OpenStreetMap, which can be used to create labeled road dataset with minimum efforts. These data may help us to complete the road segmentation task on unlabeled target domain. However, as shown in Fig. 1 (a1) and (b1), directly applying the road segmentation model, which is trained on the source domain, to infer the target road images may encounter significant performance drop. It is caused by the domain shift between training images and test images, since these images may have different types of road surfaces (unpaved, paved, dirt roads), rural and urban background areas, etc. Hence, it is important and meaningful to find a suitable technique to address the domain shift issue and take advantage of the available labeled road data to conduct automated road segmentation on target RS images without manual labeling.

Recently, unsupervised domain adaptation (UDA) [16, 17] is introduced to solve this problem. Unsupervised domain adaptation utilizes labeled data in one or more relevant source domains to execute new tasks in a target domain with unlabeled data and aims to mitigate the domain shift. Although UDA has been widely studied in remote sensing image classification task [18, 19], few efforts have been made in the context of road segmentation. The recent work ASPN [20] introduces the random noise to the source domain as synthetic data and performs the adversarial domain adaptation on the output space of source and target domain. However, the existing UDA methods for road segmentation neither fine-tune the segmentation model on the target data with pseudo labels for better performance nor tackle the possible intra-domain discrepancy within the target domain caused during data collection. As a specific technique of semi-supervised learning, self-training [21] could be an useful strategy for UDA. Specifically, it generates pseudo labels of unlabeled target data from the classifier and uses them to fine-tune the model, therefore adapting the model to the target domain. In addition, the adapted model can be used to update the pseudo labels and this procedure can be repeated to improve the performance.

Inspired by these techniques, we propose a novel stagewise unsupervised domain adaptation framework called RoadDA for road segmentation of remote sensing images. RoadDA consists of two stages, namely, the inter-domain adaptation and the adversarial self-training, as illustrated in Fig. 2. In the first stage, given the labeled source data and unlabeled target data, a segmentation model serves as the generator to produce the predictions while an inter-domain discriminator predicts the domain labels of these predictions. The target domain are adapted to align with the source domain at the output-level by optimizing the segmentation loss of source domain and the inter-domain adversarial loss. Besides, a feature pyramid fusion module is devised to avoid information loss of long and thin roads and learn discriminative and robust features. In the second stage, to address the intra-domain discrepancy in target domain and further improve the segmentation performance, we design an adversarial self-training scheme. The segmentation model trained in the previous stage is utilized to generate pseudo labels of target domain images. Since road information is the most important prior knowledge in road segmentation task, we pay more attention to the quality of road prediction of the pseudo labels. Therefore, these pseudo labels are checked by a quality estimator according to the confidence score of the road pixels. The retained pseudo labels along with their corresponding images are regarded as the easy split, while the left images in the target domain are regarded as the unlabeled hard split. Similarly, we use the same technique in the first stage to mitigate the intra-domain gap by adapting the distribution of the hard split to that of the easy split. Moreover, we devise a progressive training scheme to iteratively update the pseudo labels from the trained segmentation model and then use them to retrain the segmentation model. During the inference phase, only the adapted segmentation model of the second stage is used for road segmentation without any extra computational requirements.

In summary, we conclude our contributions as follows:

  • 1)

    We propose a novel stagewise UDA framework for road segmentation of remote sensing images, which exploits available labeled road data to achieve promising road segmentation on unlabeled target images. This work could be useful in practical application scenarios.

  • 2)

    We devise a new feature pyramid fusion module to avoid the information loss of long and thin roads in high-level feature maps and learn discriminative and robust features for better performance.

  • 3)

    In addition to addressing the inter-domain discrepancy between source and target domains, we propose an adversarial self-training stage to mitigate the intra-domain discrepancy in the target domain. It can progressively achieve intra-domain adaptation and improve road segmentation performance on target domain.

The rest of the paper is organized as follows. Section II briefly reviews the related work. In Section III, we describe the details of our proposed RoadDA model. Section IV provides extensive comparison experiments and ablation studies. Finally, we conclude the paper in Section V.

II Related work

In this section, we review the related methods of road segmentation and unsupervised domain adaptation in the context of remote sensing.

II-A Supervised Road Segmentation

Various applications, such as vehicle navigation and urban planning, require a constantly updated road database, which could be created and updated using road segmentation model on high-resolution remote sensing images. Before the era of deep learning, most of the road segmentation methods are derived from pixel-level classification strategy. Yuan et al. [5] proposed an automatic road extraction approach, where a three-step method including segmentation, medial axis points selection, and road grouping was adopted for road segmentation. Similarly, Das et al. [22] introduced a multistage algorithm in which probabilistic SVM and salient features were exploited to extract road regions from high-resolution multispectral satellite images.

As the rapid development of deep learning techniques especially the deep convolutional neural networks, they have been widely applied in various kinds of computer vision tasks [23, 24, 25, 26, 27], owing to their strong capability of feature learning and end-to-end modeling. Some researchers attempt to introduce deep learning techniques to the field of road segmentation of remote sensing images, which advance the development of this field[13, 28, 29, 30]. Cheng et al.[13] designed a cascaded end-to-end CNN for both road segmentation and centerline extraction tasks, where a symmetric encoder-decoder structure was employed for road segmentation. Bruzzone et al. [12] proposed a direction-aware residual network for enhancing the structural completeness of the road networks, in which a direction supervision is introduced to strengthen the detection of linear features. Wei et al. [29] proposed a multistage framework to accurately extract the road surface and road centerline simultaneously. Li et al. [31] proposed to combine multi-scale global context with adversarial networks. Specifically, the images with ground truth labels and images with prediction maps are fed into the discriminator network, which plays a min-max game with the segmentation model by back-propagating the classification error reversely. Although these approaches achieve remarkable performance, deep supervised learning based methods require a large amount of labeled data, which are expensive and time-consuming to collect and annotate, especially for various unseen scenarios. Since we could get access to some public road datasets released by previous researchers, these data may be useful to help extract road information from unlabeled data by using domain adaptation techniques.

II-B Unsupervised Domain Adaptation

Unsupervised Domain Adaptation (UDA) can be regarded as a particular case of domain adaptation that exploits labeled data in a relevant source domain to conduct similar task in the target domain. The purpose of UDA methodologies is to reduce the domain shift, which typically deteriorates the performance of the models [32, 33, 34, 35]. In recent years, various UDA methods have been proposed to address the domain shift problem in the semantic segmentation field. Tsai et al. [36] proposed a multi-level adversarial learning framework for UDA semantic segmentation, which employed adversarial learning in the structured output space at different feature levels. Vu et al. Li et al. [37] proposed the collaborative class conditional GAN to bypass the dependence on the source data and improved the performance of the predicted model through generated target-style data. Pan et al. [38] proposed the intra-domain adaptation of the target domain and designed the division criterion for target domain based on the mean of multi-class entropy map. Nevertheless, this criterion may be not suitable for binary road segmentation, since mean entropy based criterion will misclassify all predicted background pixels as easy samples and mislead the intra-domain adaptation. Therefore, we pay more attention to road pixels in the road segmentation task and design a new intra-domain division strategy to better divide the target domain. Furthermore, we propose to combine the intra-domain adaptation with self-training to further improve the segmentation performance.

UDA has also been introduced to the hyperspectral image classification. Tasar et al. [16] designed the color mapping GAN to generate fake training images with the same semantics as training images and used them to fine-tune the trained classifier. [39] used multiple classifiers and variational autoencoders to construct a GAN [40] to address the domain discrepancy. CaGAN [41] proposed a class-aware GAN for UDA multisource remote sensing image classification. CaGAN selects the reliable per-category feature centers based on the clustering and reduces the intra-class and the inter-class discrepancies across domains by optimizing the class-level l1l_{1}-norm loss between the source and target category feature centers. Different from CaGAN, our RoadDA adopts the pseudo labels of easy split and address the intra-class discrepancy within target domain by adversarial self-training. Compared with the multi-class classification task of hyperspectral image, road segmentation can be regarded as a binary classification task of RGB remote sensing images, where more attention should be paid to the road information than background pixels. Moreover, the road shape characteristics and the proportion of road and background pixels should also be considered to improve the model performance. In this sense, UDA methods of hyperspectral image classification that require considerations of spatial-spectral information and interclass relationships may not be suitable for road segmentation. Recently, ASPN [20] designed an adversarial spatial pyramid network for the domain adaptation in road segmentation of RS images, which focuses on extracting multi-level effective features and enhancing the feature representation. Few studies have been done on road segmentation of RS images and there are still many issues to be explored and settled.

Refer to caption
Figure 2: The overview of our proposed RoadDA method, which consists of two stages. In the inter-domain adaptation stage, given the source and unlabeled target data, a segmentation model equipped with a specifically devised feature pyramid fusion module (FPFM) acts as the generator to predict the segmentation results. The discriminator is trained to distinguish the domain label of the input while the generator aims to generate similar distribution for both source and target domain to fool the discriminator. In the adversarial self-training stage, we predict the segmentation maps of target domain images using the trained model. The predictions are used to mine the easy target samples by calculating and ranking the road confidence scores and assigned to them as pseudo labels. Theses pseudo labels are used to help adapt the unlabeled hard split to the easy one and therefore reduce the intra-domain discrepancy. The pseudo labels generation and intra-domain adaptation are repeated to gradually improve the performance of the segmentation model.
Refer to caption
Figure 3: Illustration of the proposed feature pyramid fusion module. 8×\times denotes a downsampling rate of 8.

III proposed method

III-A Overview

Our method aims to improve the performance of road segmentation on unlabeled target data using the unsupervised domain adaptation and self-training techniques. As depicted in Fig. 2, RoadDA consists of two stages: inter-domain adaptation stage and adversarial self-training stage. In the first stage, a generative adversarial network (GAN) equipped with a specifically devised feature pyramid fusion module is constructed to reduce the domain shift between the labeled source domain and unlabeled target domain. In the second stage, we use the trained generator to predict the pseudo labels of target domain and then divide the target domain into easy and hard split based on the road confidence scores. An adversarial learning based intra-domain adaptation is performed to fine-tune the trained generator on the target domain and mitigate the intra-domain discrepancy. We adopt the fine-tuned generator to get more accurate pseudo labels and use them for further intra-domain adaptation. This procedure is iterated to progressively improve the segmentation performance until saturation. During the inference phase, only the adaptive generator is used for road segmentation.

III-B Inter-domain Adaptation

Inter-domain adaptation is achieved by the GAN architecture. The input are the labeled source data (𝒳s\mathcal{X}_{s}, 𝒴s\mathcal{Y}_{s}) and unlabeled target data (𝒳t\mathcal{X}_{t}), where 𝒳s,𝒳tH×W×3\mathcal{X}_{s},\mathcal{X}_{t}\in\mathbb{R}^{H\times W\times 3} and 𝒴s(0,1)H×W\mathcal{Y}_{s}\in(0,1)^{H\times W} (0 for background pixel, 1 for road pixel). A segmentation model acts as the generator GinterG_{inter} and predicts the segmentation probability map Ps=Ginter(𝒳s)P_{s}=G_{inter}(\mathcal{X}_{s}), Pt=Ginter(𝒳t)P_{t}=G_{inter}(\mathcal{X}_{t}) (Ps,PtH×W×2P_{s},P_{t}\in\mathbb{R}^{H\times W\times 2}) for source images and target images, respectively. As the intermediate representation learned in the segmentation process, the high-level semantic feature of high dimension actually has complex implicit semantics. Therefore, it may be less effective to perform domain alignment at the feature level and there is no guarantee that the joint image-label distributions are aligned between domains. While the binary segmentation predictions of both source images and target images have strong similarity in context and layout. Therefore, we adopt an adversarial learning scheme to align the target domain to the source domain at the output level. We send the predictions PsP_{s}, PtP_{t} to the discriminator DinterD_{inter} as input to correctly predict their domain labels, while the generator GinterG_{inter} is trained to fool DinterD_{inter}, i.e., the generator GinterG_{inter} is encouraged to generate similar prediction distributions on both the target and source domains.

Here, we describe the training process of GinterG_{inter} and DinterD_{inter}. For the discriminator, given the segmentation predictions on the source domain and target domain Ps,PtP_{s},P_{t}, the fully convolutional discriminator is trained to classify the domain labels of the input samples. Here we artificially set the labels of source domain samples to 1 and the labels of target domain samples to 0, thus the discriminator could be optimized with the binary cross-entropy (CE) domain classification loss. The original CE loss for discriminator could be formulated as follows:

interD(Ps,Pt)=\displaystyle\mathcal{L}_{inter}^{D}\left(P_{s},P_{t}\right)= h,w[slog(𝐃inter(Ps)(h,w))\displaystyle-\sum_{h,w}\left[s\log\left(\mathbf{D}_{inter}\left(P_{s}\right)^{(h,w)}\right)\right. (1)
+(1s)log(1𝐃inter(Ps)(h,w))]\displaystyle\left.+(1-s)\log\left(1-\mathbf{D}_{inter}\left(P_{s}\right)^{(h,w)}\right)\right]
h,w[tlog(𝐃inter(Pt)(h,w))\displaystyle-\sum_{h,w}\left[t\log\left(\mathbf{D}_{inter}\left(P_{t}\right)^{(h,w)}\right)\right.
+(1t)log(1𝐃inter(Pt)(h,w))].\displaystyle\left.+(1-t)\log\left(1-\mathbf{D}_{inter}\left(P_{t}\right)^{(h,w)}\right)\right].

when s=1 and t=0, it is rewritten as:

interD(Ps,Pt)=\displaystyle\mathcal{L}_{inter}^{D}\left(P_{s},P_{t}\right)= h,w[log(𝐃inter(Ps)(h,w))\displaystyle-\sum_{h,w}\left[\log\left(\mathbf{D}_{inter}\left(P_{s}\right)^{(h,w)}\right)\right. (2)
+log(1𝐃inter(Pt)(h,w))].\displaystyle\left.+\log\left(1-\mathbf{D}_{inter}\left(P_{t}\right)^{(h,w)}\right)\right].

For the training of generator GinterG_{inter}, it contains two parts. Firstly, the segmentation loss of images from source domain, which is the cross-entropy loss:

sourceseg(𝒳s,𝒴s)=h,w,c𝒴s(h,w,c)log(𝐆inter(𝒳s)(h,w,c)),\displaystyle\mathcal{L}_{source}^{seg}\left(\mathcal{X}_{s},\mathcal{Y}_{s}\right)=-\sum_{h,w,c}\mathcal{Y}_{s}^{(h,w,c)}\log\left(\mathbf{G}_{inter}\left(\mathcal{X}_{s}\right)^{(h,w,c)}\right), (3)

where the source domain labels 𝒴s\mathcal{Y}_{s} have been converted to the one-hot vector form and cc denotes the number of classes. Secondly, the adversarial loss for target images is defined as:

interadv(𝒳t)=\displaystyle\mathcal{L}_{inter}^{adv}\left(\mathcal{X}_{t}\right)= h,w[tlog(𝐃inter(Pt)(h,w))\displaystyle-\sum_{h,w}\left[t\log\left(\mathbf{D}_{inter}\left(P_{t}\right)^{(h,w)}\right)\right. (4)
+(1t)log(1𝐃inter(Pt)(h,w))].\displaystyle\left.+(1-t)\log\left(1-\mathbf{D}_{inter}\left(P_{t}\right)^{(h,w)}\right)\right].

According to the adversarial strategy, we set t = 1, which is opposite of the discriminator, then the adversarial loss could be rewritten as:

interadv(𝒳t)=h,wlog(𝐃inter(𝐆inter(𝒳t))(h,w)).\mathcal{L}_{inter}^{adv}\left(\mathcal{X}_{t}\right)=-\sum_{h,w}\log\left(\mathbf{D}_{inter}\left(\mathbf{G}_{inter}(\mathcal{X}_{t})\right)^{(h,w)}\right). (5)

This loss forces the generator to produce a source-like distribution on the target domain to fool the discriminator and finally the model reaches a balance state. In summary, the final training objective for the generator is:

interG(𝒳s,𝒳t)=sourceseg(𝒳s)+αadvinteradv(𝒳t),\mathcal{L}_{inter}^{G}\left(\mathcal{X}_{s},\mathcal{X}_{t}\right)=\mathcal{L}_{source}^{seg}\left(\mathcal{X}_{s}\right)+\alpha_{adv}\mathcal{L}_{inter}^{adv}\left(\mathcal{X}_{t}\right), (6)

where αadv\alpha_{adv} is the loss weight to balance the two losses.

During training, We alternatively optimize discriminator DinterD_{inter} and generator GinterG_{inter} using the loss function in Eq. (2) and Eq. (6), while in inference, only the segmentation model (generator) is used for road segmentation.

III-B1 Feature Pyramid Fusion Module

In the GAN framework, a fully convolutional segmentation model acts as the generator and the backbone network is ResNet-101 [42]. Let the hierarchical features extracted by ResNet-101 be denoted as [C1,C2,C3,C4,C5C_{1},C_{2},C_{3},C_{4},C_{5}], where the corresponding downsampling rate are [2, 4, 8, 16, 32]. High downsampling rate is originally designed to reduce the feature size and the computation cost and increase the feature robustness. However, in the different scenarios of road segmentation from remote sensing images, high downsampling rate may result in the spatial information loss of long and thin roads, which will affect the final road segmentation performance in dealing with different domain data. In contrast, if we keep high resolution in the deep layer, such as using the downsampling rate set of [2, 4, 8, 8, 8], it will reduce the effective receptive field of the model as well as increase the computation cost in the deep layer. To better handle the various scenarios in both source and target domains, we design the feature pyramid fusion module to fuse multi-level and and multi-scale road features. As shown in Fig. 3, we employ the downsampling rate set of [2, 4, 8, 8, 16] by setting the stride of fourth stage in ResNet101 to 1, and select the features C4C_{4} and C5C_{5} as input to the ASPP module [43] to generate the global context features C4C_{4}^{*} and C5C_{5}^{*} at different level. We then upsample C5C_{5}^{*} to the same size as C4C_{4}^{*} and add them together to explicitly enhance the feature representation ability and generate the robust fused feature at the downsampling rate of 8.

III-C Adversarial Self-training

After the training of the inter-domain adaptation stage, we obtain the adapted segmentation model which can achieve considerable performance improvement compared with model trained only on the source domain (source-only). However, according to our observation, there may be intra-domain discrepancy in the target domain due to different illumination conditions and background context. Thus we propose the adversarial learning based self-training stage to further improve the segmentation performance on the target domain. This stage is composed of three parts: intra-domain division, intra-domain adaptation, and self-training. We will describe each part in detail in the following subsection.

III-C1 Intra-domain Division

To address the intra-domain discrepancy in the target domain, it should be decomposed into two parts and reduce their discrepancy. To this end, we propose a quality estimator to accomplish this task. Inspired by the observation that the performance of segmentation model trained on the source domain will deteriorate on the target domain, we use the adapted segmentation model GinterG_{inter} to predict segmentation maps for target domain and accordingly divide the target domain into an easy split and a hard split. Specifically, we design a quality estimator to estimate the road confidence scores of prediction results and rank them based on the scores. Finally, the target images are separated into an easy split 𝒳te\mathcal{X}_{te} with pseudo labels and an unlabeled hard split 𝒳th\mathcal{X}_{th}.

Let 𝒫t=Ginter(𝒳t)\mathcal{P}_{t}=G_{inter}(\mathcal{X}_{t}) denote the predicted segmentation map of target images, and 𝒫i,jt\mathcal{P}_{i,j}^{t} denotes the probability of road at location (i,j)(i,j). t\mathcal{M}_{t} represents the predicted binary mask (pseudo labels). i,jt\mathcal{M}_{i,j}^{t} equals 1 when the probability of pixel at (i,j)(i,j) is higher than a threshold of 0.5 and the pixel is predicted as road, otherwise it equals 0. Since we focus on the road pixels rather than the large numbers of background pixels which may bias the estimation, we filter out the background and calculate the confidence score of the prediction as follows:

𝒮cf=i,j𝒫i,jti,jti,ji,jt,\mathcal{S}_{cf}=\frac{\sum_{i,j}\mathcal{P}_{i,j}^{t}\cdot\mathcal{M}_{i,j}^{t}}{\sum_{i,j}\mathcal{M}_{i,j}^{t}}, (7)

which is the mean value of the confidence of all predicted road pixels in each image. Then we rank the target images in a descending order according to their 𝒮cf\mathcal{S}_{cf} scores and employ a hyperparameter λ(0,1)\lambda\in(0,1) as a ratio to divide the ranked target images into an easy split 𝒳te\mathcal{X}_{te} with high-confidence pseudo labels te\mathcal{M}_{te} and an unlabeled hard split 𝒳th\mathcal{X}_{th}, where |Xte|=λ|Xt|\left|X_{te}\right|=\lambda\left|X_{t}\right|, |Xth|=(1λ)|Xt|\left|X_{th}\right|=(1-\lambda)\left|X_{t}\right|. In the experiment section, we conducted a hyperparameter analysis experiment to investigate the influence of λ\lambda.

III-C2 Intra-domain Adaptation

In this part, the same GANs architecture as the first stage is constructed to align the hard split with the easy split at the output level and reduce the intra-domain gap. The generator GintraG_{intra} is initialized with the generator adapted in the first stage and takes 𝒳te\mathcal{X}_{te} and 𝒳th\mathcal{X}_{th} as input, and the output are fed into the discriminator DintraD_{intra}, which aims to predict the domain labels. DintraD_{intra} is trained to minimize the binary cross-entropy classification loss while GintraG_{intra} is trained to minimize the segmentation loss and the adversarial loss. They are defined as follows:

intraD(𝒳te,𝒳th)=\displaystyle\mathcal{L}_{intra}^{D}\left(\mathcal{X}_{te},\mathcal{X}_{th}\right)= h,w[log(𝐃intra(𝐆intra(𝒳te))(h,w))\displaystyle-\sum_{h,w}\left[\log\left(\mathbf{D}_{intra}\left(\mathbf{G}_{intra}(\mathcal{X}_{te})\right)^{(h,w)}\right)\right. (8)
+log(1𝐃intra(𝐆intra(𝒳th))(h,w))],\displaystyle\left.+\log\left(1-\mathbf{D}_{intra}\left(\mathbf{G}_{intra}(\mathcal{X}_{th})\right)^{(h,w)}\right)\right],
interG(𝒳te,𝒳th)=targetseg(𝒳te)+βadvintraadv(𝒳th),\mathcal{L}_{inter}^{G}\left(\mathcal{X}_{te},\mathcal{X}_{th}\right)=\mathcal{L}_{target}^{seg}\left(\mathcal{X}_{te}\right)+\beta_{adv}\mathcal{L}_{intra}^{adv}\left(\mathcal{X}_{th}\right), (9)

where βadv\beta_{adv} is the loss weight to balance these two losses and

targetseg(𝒳te,te)=h,w,cte(h,w,c)log(𝐆intra(𝒳te)(h,w,c)),\displaystyle\mathcal{L}_{target}^{seg}\left(\mathcal{X}_{te},\mathcal{M}_{te}\right)=-\sum_{h,w,c}\mathcal{M}_{te}^{(h,w,c)}\log\left(\mathbf{G}_{intra}\left(\mathcal{X}_{te}\right)^{(h,w,c)}\right), (10)
intraadv(𝒳th)=h,wlog(𝐃intra(𝐆intra(𝒳th))(h,w)).\mathcal{L}_{intra}^{adv}\left(\mathcal{X}_{th}\right)=-\sum_{h,w}\log\left(\mathbf{D}_{intra}\left(\mathbf{G}_{intra}(\mathcal{X}_{th})\right)^{(h,w)}\right). (11)

As in the first stage, DintraD_{intra} and GintraG_{intra} are alternatively optimized using the loss functions in Eq. (8) and Eq. (9).

III-C3 Self-training

After the intra-domain adaptation training on the target domain, the adapted segmentation model GintraG_{intra} achieves better performance and predicts more accurate road segmentation maps. Therefore, we update the pseudo labels of target images using this adapted generator and repeat the intra-domain division and adaptation process to progressively improve the performance of the road segmentation model until saturation, as shown in Fig. 2. The training algorithm of RoadDA is summarized in Algorithm 1.

Algorithm 1 The Training Algorithm of RoadDA

Stage1: Inter-domain adaptation
   Input: source domain data: Ds=(𝒳s,𝒴s)D_{s}=(\mathcal{X}_{s},\mathcal{Y}_{s});
             target domain data: Dt=𝒳tD_{t}=\mathcal{X}_{t}.
   Initialize: GinterG_{inter} and DinterD_{inter}
   while epoch<epochmaxepoch<epoch_{max} do
      epoch=epoch+1epoch=epoch+1
      for i<iterationmaxi<iteration_{max}; i++ do
        Derive BsB_{s} and BtB_{t} sampled from 𝒳s\mathcal{X}_{s} and 𝒳t\mathcal{X}_{t}
        Train GinterG_{inter} on BsB_{s} and BtB_{t} by optimizing interG\mathcal{L}_{inter}^{G}
        Train DinterD_{inter} on BsB_{s} and BtB_{t} by optimizing interD\mathcal{L}_{inter}^{D}
      end for
   end while
   Output: Adapted segmentation model GinterG_{inter}
Stage2: Adversarial self-training
   Input: target data: 𝒳t\mathcal{X}_{t}, adapted GinterG_{inter}, hyperparameter λ\lambda.
   Initialize: Initialize GintraG_{intra} using GinterG_{inter}; DintraD_{intra}
   Step1: generate pseudo labels t\mathcal{M}_{t} for 𝒳t\mathcal{X}_{t} using GintraG_{intra}
   Step2: divide 𝒳t\mathcal{X}_{t} into 𝒳te\mathcal{X}_{te} and 𝒳th\mathcal{X}_{th} s.t. |𝒳te|=λ|Xt|\left|\mathcal{X}_{te}\right|=\lambda\left|X_{t}\right|
   Step3: train GintraG_{intra} on 𝒳te\mathcal{X}_{te} and 𝒳th\mathcal{X}_{th} by optimizing intraG\mathcal{L}_{intra}^{G}
             train DintraD_{intra} on 𝒳te\mathcal{X}_{te} and 𝒳th\mathcal{X}_{th} by optimizing intraD\mathcal{L}_{intra}^{D}
   Step4: evaluate the segmentation performance of GintraG_{intra}
   Repeat: Step1-Step3
   Until: performance saturation of GintraG_{intra}
   Output: GintraG_{intra}

III-D Implementation Details

For the generator, ResNet-101 is pre-trained on ImageNet. The discriminative fused feature from FPFM is directly upsampled to predict the final segmentaion probability map. For the discriminator, we adopt the fully convolutional network similar to [44]. The network contains five convolution layers with 4 ×\times 4 kernel and stride of 2. Leaky ReLU with slope of 0.2 follows each convolution layer as activation function except the last layer.

We implemented our model with PyTorch on a single NVIDIA V100 GPU. The generators (Ginter,GintraG_{inter},G_{intra}) were optimized using the SGD optimizer with the momentum of 0.9 and a weight decay of 10410^{-4}. We used the Adam optimizer to optimize the discriminator with the momentum of 0.9 and 0.99. The initial learning rates are 4×1044\times 10^{-4} and 1×1041\times 10^{-4} for generators and discriminators, respectively, and both of which are decreased using the polynomial decay policy with a power of 0.9. The loss weights αadv\alpha_{adv} and βadv\beta_{adv} are set to 0.1 and 0.01, respectively. We set the batch size to 4 and λ\lambda to 0.7.

IV experiments

In this section, we conduct extensive experiments to demonstrate the effectiveness of RoadDA in terms of both objective evaluation metrics and subjective visual comparisons.

IV-A Datasets

In the UDA experiments of road segmentation, we consider the adaptation setting that includes different spatial resolution, background region, and different types of road surfaces. Here, we employ the Roadtracer dataset [45] and DeepGlobe dataset [46] as the source domains, and the CasNet dataset [13] as the target domain. The segmentation model is trained on the labeled source data and unlabeled target data, and is evaluated on the test set of CasNet dataset.

  • 1)

    DeepGlobe dataset (DeepGlobe) This dataset is proposed in the road extraction challenge of DeepGlobe 2018 [46]. The images of this dataset are captured over Thailand, Indonesia, and India with the spatial resolution of 0.5 m/pixel. Images are sampled uniformly between rural and urban areas and cropped to extract useful region by GIS experts. The dataset consists of 8,570 RGB images with size of 1024x1024, where 6,226 images are randomly sampled for training, 1,243 images and 1,101 images are chosen as the validation and test sets, respectively. Specially, the official available dataset only provides labels for the training set. The dataset is available at https://competitions.codalab.org/competitions/18467.

  • 2)

    Roadtracer dataset (RTDS): RTDS is collected and first used in [45]. RTDS is a large corpus of high-resolutional satellite images and ground-truth road network graphs, in which images are obtained from Google Map with 0.6 m/pixel resolution and ground-truth road network are from the OpenStreeMap (OSM). RTDS covers the urban core of 40 cities in 6 countries. In each city, the center area of approximately 24 km2km^{2} is selected as the sample of the dataset for a total of 300 RGB images having a size of 4096x4096 pixels. Following [45], images from 25 cities are randomly selected as the training set and images of the 15 remaining cities serve as the test set. The dataset is available at https://github.com/mitroadmaps/roadtracer.

  • 3)

    CasNet dataset (CNDS): the CasNet dataset, built by Cheng et al [13], consists of 224 RGB remote sensing images collected from Google Earth with manual annotations. The images in this dataset are at least 600x600 pixels with the spatial resolution of 1.2 m/pixel. Moreover, these images contain complex background and diversified road shapes. Following [13], we randomly select 180 images as the training set, 14 images for validation, and the rest 30 images as the test set. The dataset is available at https://shibiaoxu.github.io/TGRS2017.html.

Some image examples are depicted in Fig. 4.

IV-B Data Augmentation

Because of the limitation of the GPU resource. we cannot directly train the model using the high-resolution remote sensing images. Besides, the number of samples in the original dataset is insufficient to train a model with good feature representation and generalization. Therefore, data preprocessing and augmentation are carried out before training.

For each image in these datasets, we randomly crop 20 patches having a size of 512×512512\times 512 and then filter out the patches where the number of road pixels is less than 4,000. Since the total number of training samples in CNDS is still insufficient, we flip each cropped training sample in the horizontal direction and rotate the original and flipped samples at the step of 9090^{\circ} for 3 times. The detailed information of the final processed and augmented datasets is reported in Table I.

TABLE I: Detailed information of the datasets.
Datasets Resolution Size Training Data Test Data
DeepGlobe 0.5 m/pixel 512×512512\times 512 44,165 -
RTDS 0.6 m/pixel 512×512512\times 512 51,106 -
CNDS 1.2 m/pixel 512×512512\times 512 28,420 589
Refer to caption
Figure 4: Some visual examples from different datasets. (a) DeepGlobe; (b) RTDS; (c) CNDS.

IV-C Evaluation Metrics

Four common metrics are used to evaluated the quantitative performance of different road segmentation models, including intersection-over-union (IoU), completeness (COM), correctness (COR), and F1 score. COM measures the proportion of matched road pixels in the ground truth map, COR reports the percentage of matched road areas in the predicted segmentation map. The F1 score is a harmonic average between COM and COR which can measure the robustness of methods. The four metrics are defined as follows:

IoU=TPTP+FN+FP\displaystyle\mathrm{IoU}=\frac{\mathrm{TP}}{\mathrm{TP}+\mathrm{FN}+\mathrm{FP}} COM=TPTP+FN\displaystyle\mathrm{COM}=\frac{\mathrm{TP}}{\mathrm{TP}+\mathrm{FN}} (12)
COR=TPTP+FP\displaystyle\mathrm{COR}=\frac{\mathrm{TP}}{\mathrm{TP}+\mathrm{FP}} F1=2×COM×CORCOM+COR,\displaystyle\mathrm{F1}=\frac{2\times\mathrm{COM}\times\mathrm{COR}}{\mathrm{COM}+\mathrm{COR}},

where TP, FP, and FN denote the number of true positive pixels, false positive pixels, and false negative pixels, respectively. A larger metric value indicates better performance.

IV-D Comparative Methods

To demonstrate the superiority of the proposed method, we compare it with several state-of-the-art methods, including AdaptSegNet [36], ADVENT [47], BDL [48], and IntraDA [38]. Here, we briefly describe these approaches as follows:

  • 1)

    Source-only: This is a baseline model which is only trained on the labeled source domain data while directly tested on the target domain.

  • 2)

    Target-only: Target-only represents the typical supervised road segmentation method which is trained on the target domain with labels. It is the oracle model to provide the upper bound of performance and indicate the effectiveness and reliability of those UDA methods.

  • 3)

    AdaptSegNet: AdaptSegNet firstly adopts adversarial learning in the output space for UDA semantic segmentation. Compared with the adaptation in the feature-level, structured output spaces of source and target domain contain the global context and spatial similarities, which is beneficial to the adaptation. Moreover, a multi-level adversarial network is designed to enhance the model.

  • 4)

    ADVENT: Based on AdaptSegNet, ADVENT calculates the entropy of output space to avoid low-confident predictions on the target domain. The entropy map is fed into the discriminator for adversarial training.

  • 5)

    BDL: BDL adopts a bidirectional learning framework which first carries out image-to-image translation and then uses the translated images to complete the adversarial domain adaptation. The image translation and adversarial adaptation promote each other to progressively improve performance.

  • 6)

    IntraDA: In addition to the adaptation between source and target domain, IntraDA tries to address the intra-domain shift, although the entropy based criterion designed to separate the multi-class target domain is not suitable for binary road segmentation tasks.

IV-E UDA Results on different adaptation setting

DeepGlobe \longrightarrow CNDS. In Table II, we report the road segmentation performance of our method and other comparison methods on the CNDS test set. For a fair comparison, the baseline model and target-only model adopt the same segmentation model as RoadDA and all the methods also use the same ResNet-101 backbone. Overall, RoadDA achieves the best performance of 74.92% IoU and 85.81% F1 score compared with the state-of-the-art UDA methods, which shows the reliability and robustness of our model. Source-only and target-only models produce the worst and best results, respectively, as expected. Compared with the methods with only inter-domain adaptation, such as AdaptSegNet and ADVENT, the intra-domain adaptation and self-training adopted in our model bring considerable performance improvement. For instance, RoadDA is 15.52% higher than ADVENT in terms of IoU. Moreover, our method also outperforms IntraDA, which also uses intra-domain adaptation. We argue that the performance gain owes to the employment of more effective division criterion for target domain and self-training.

Fig.5 presents the visual road segmentation results of different methods on four representative test images from the CNDS test set. It can be observed that RoadDA obtains the best segmentation results close to the ground-truth and could handle the remote sensing images with different complex backgrounds. Moreover, most of the road contours in the images can be captured by RoadDA, demonstrating the great potential of using domain adaptation techniques to solve the unsupervised road segmentation problem.

RTDS \longrightarrow CNDS. Here, we use the RTDS as the source domain and report the evaluation results on the CNDS test set in Table III. All the methods employ the same configuration as the previous experiment. As shown in Table III, our proposed method achieves 61.76% IoU and 77.48% F1 score, which is superior to all the UDA comparison methods. Specially, RoadDA is 15.8% higher than source-only and 7.32% higher than ADVENT in terms of IoU. Compared with IntraDA equipped with intra-domain adaptation, our model also brings 5.7% improvement in IoU metric.

Comparing the experimental results of the two adaptation settings, we can find that the adaptation performance of DeepGlobe is better than RTDS. This may be caused by the inherent properties of the domain itself. The images in DeepGlobe dataset have more similar spatial information to the target dataset, such as the background context and road shape, while the images in RTDS are captured in the big city centers where roads are too wide or indistinguishable from surrounding buildings. This may give us some hints for choosing the suitable source domain in practice.

TABLE II: Quantitative results on the setting of DeepGlobe \longrightarrow CNDS. Adv. learning: adversarial learning.
DeepGlobe \longrightarrow CNDS
Methods backbone Input size Strategy IoU COM COR F1
Source-only ResNet-101 512×512512\times 512 Segmentation 52.84 57.47 89.57 70.01
AdaptSegNet [36] ResNet-101 512×512512\times 512 Adv. learning 54.13 59.04 91.86\mathbf{91.86} 71.88
ADVENT [47] ResNet-101 512×512512\times 512 Adv. learning 59.40 72.76 90.66 80.72
BDL [48] ResNet-101 512×512512\times 512 Style transfer & Adv. learning 72.35 79.91 89.97 84.64
IntraDA [38] ResNet-101 512×512512\times 512 Adv. learning 72.73 80.06 90.21 84.83
RoadDA ResNet-101 512×512512\times 512 Adv. learning & self-training 74.92\mathbf{74.92} 82.11\mathbf{82.11} 89.88 85.81\mathbf{85.81}
Target-only ResNet-101 512×512512\times 512 Segmentation 85.35 91.23 93.51 92.35
TABLE III: quantitative results of adapting RTDS to CNDS. Adv. learning: adversarial learning
RTDS \longrightarrow CNDS
Methods backbone Input size Strategy IoU COM COR F1
Source-only ResNet-101 512×512512\times 512 Segmentation 45.96 48.39 91.36\mathbf{91.36} 63.26
AdaptSegNet [36] ResNet-101 512×512512\times 512 Adv. learning 53.34 58.03 89.48 70.40
ADVENT [47] ResNet-101 512×512512\times 512 Adv. learning 54.44 58.03 90.71 71.26
BDL [48] ResNet-101 512×512512\times 512 Style transfer & Adv. learning 55.47 59.65 90.75 71.98
IntraDA [38] ResNet-101 512×512512\times 512 Adv. learning 56.06 60.26 91.16 72.55
RoadDA ResNet-101 512×512512\times 512 Adv. learning & self-training 61.76\mathbf{61.76} 67.56\mathbf{67.56} 90.81 77.48\mathbf{77.48}
Target-only ResNet-101 512×512512\times 512 Segmentation 85.35 91.23 93.51 92.35
Refer to caption
Figure 5: Visual results of typical methods on CNDS. (a) the target images. (b) The ground-truth. (c) Our RoadDA. (d) IntraDA. (e) ADVENT. (f) Source-only.

IV-F Model Analysis

1) Ablation study: In this part, we evaluate the effects of the proposed modules in RoadDA on the setting of DeepGlobe \longrightarrow CNDS. The input image size is 512 ×\times 512.

In Table IV, we summarize the performance of RoadDA using different variants of feature pyramid fusion module on the setting of DeepGlobe \longrightarrow CNDS. RoadDA with 32×\times means that we only use the deep feature of a downsampling rate of 32 as input for ASPP module and then directly generate the segmentation map with out the fusion operation. Among different variants of RoadDA only using the single-level feature, the setting of 8×\times performs best by obtaining a 74.17% IoU which shows that the high-resolution features preserve more road spatial information and beneficial to the performance. RoadDA with 32×\times, 16×\times, and Fusion denotes using the FPFM which takes the two levels of features of the downsampling rate of 32 and 16 as input. Overall, RoadDA using the proposed setting of FPFM achieves the best performance of 74.92% IoU compared with all other variants.

We also conducted an ablation study on the second adversarial self-training stage. The results are summarized in Table V. As can be seen, RoadDA without intra-domain adaptation and self-training, which only employs the inter-domain adaptation, achieves 61.74% IoU and 81.13% F1 score. After using the intra-domain adaptation and self-training strategies, the performance is improved. The intra-domain adaptation provides 11.9% IoU improvement and the self-training brings 12.34% IoU gains. RoadDA using both intra-domain adaptation and self-training achieves a gain of 13.18% IoU over the vanilla baseline, which confirms the effectiveness of the proposed techniques in the adversarial self-training stage.

2) Division ratio λ\lambda: For intra-domain division, λ\lambda controls the number of samples in the easy and hard splits and also has a impact on the distribution of simple samples and difficult samples. To investigate its influence, we conducted the experiments using λ\lambda from 0.6 to 0.9 on both adaptation settings. As reported in Table VI, RoadDA achieves the best performance for both settings when λ\lambda = 0.7.

TABLE IV: Ablation study results of the proposed FPFM on the setting of DeepGlobe \longrightarrow CNDS.
Model 32×\times 16×\times 8×\times Fusion IoU
RoadDA \checkmark ×\times ×\times ×\times 73.24
RoadDA ×\times \checkmark ×\times ×\times 73.65
RoadDA ×\times ×\times \checkmark ×\times 74.17
RoadDA \checkmark \checkmark ×\times \checkmark 73.98
RoadDA ×\times \checkmark \checkmark \checkmark 74.92\mathbf{74.92}
TABLE V: Ablation study results of the proposed techniques in the adversarial self-training stage on the setting of DeepGlobe \longrightarrow CNDS. IA: intra-domain adaptation. ST: self-training.
Model IA ST IoU COM COR F1
RoadDA ×\times ×\times 61.74 73.48 90.57 81.13
RoadDA \checkmark ×\times 73.64 80.72 90.03 85.12
RoadDA ×\times \checkmark 74.08 81.15 89.97 85.33
RoadDA \checkmark \checkmark 74.92\mathbf{74.92} 82.11\mathbf{82.11} 89.88 85.81\mathbf{85.81}
TABLE VI: IoU results of different hyperparameter λ\lambda on both adaptation settings. DPG: DeepGlobe dataset.
λ\lambda 0.6 0.7 0.8 0.9
DPG \longrightarrow CNDS 73.54 74.92\mathbf{74.92} 74.01 72.87
RTDS \longrightarrow CNDS 60.25 61.76\mathbf{61.76} 61.33 60.62
Refer to caption
Figure 6: IoU results versus the iterations of self-training.
Refer to caption
Figure 7: Segmentation loss and adversarial loss versus the iterations while training RoadDA on DeepGlobe and CNDS.
Refer to caption
Figure 8: T-SNE visualization of the features obtained by different models on the setting of DeepGlobe \longrightarrow CNDS. (a) Source-only. (b) RoadDA using inter-domain adaptation only. (c) RoadDA.
TABLE VII: The results of two choices of domain adaptation space on the setting of DeepGlobe \longrightarrow CNDS.
Methods IoU COM COR F1
Feature-level adaptation 51.42 56.36 87.65 68.61
RoadDA (output space) 74.92 82.11 89.88 85.81

3) Iterarion for self-training: In the adversarial self-training stage, we iterate the intra-domain adaptation process to progressively improve model performance, while it is unknown how many iterations it will take to achieve performance saturation. Therefore, we analyze the relationship between the iteration and the IoU on two adaptation setting. As depicted in Fig. 6, the performance gradually increase and tends to be saturated after three or four iterations. Thus, three or four iterations may be a proper choice to save the training time while achieve a decent performance.

4) Domain adaptation space: Here, we investigate the influence of domain adaptation space on the segmentation results. As shown in Table VII, when RoadDA performs the feature-level domain adaptation, the model suffers significant performance deterioration and only achieves 51.42% IoU which is 23.5% lower than RoadDA using output space domain adaptation. It indicates that the structured prediction space is more appropriate for domain adaptation in the binary road segmentation task since remote sensing images always contain complex structures and consequently complex feature distributions may obtain for different domains.

5) Model training: Fig. 7 presents the curves of the segmentation loss and adversarial loss against the number of iterations on the DeepGlobe \longrightarrow CNDS setting. The loss curves in the inter-domain adaptation stage and the intra-domain adaptation are listed in the left side and right side, respectively. As can be seen, with the increasing of iterations, the segmentation loss and discriminator loss become smaller until convergence while the adversarial loss increases gradually until convergence. Finally, the adversarial learning reaches a balanced state where the model generates similar distributions on both target and source domains. In addition, we show the t-SNE visualization [49] of the features before the prediction layer of two variants of our RoadDA and the source-only model in Fig. 8. The results demonstrate that using the two-stage adaptation in the output-level space, the features in target domain is gradually aligned with the those in the source domain.

V Conclusion

In this paper, we propose a novel two-stage unsupervised domain adaptation framework for road segmentation of remote sensing images. In the first inter-domain adaptation stage, our model equipped with a specifically designed feature pyramid fusion module efficiently mitigates the domain discrepancy between the labeled source domain and the unlabeled target domain, resulting in an adapted segmentation model that can generalize well to the target domain. In the second adversarial self-training stage, our model effectively mine easy target samples and assign pseudo labels to them, which are used to guide the intra-domain adaptation to mitigate the discrepancy within the target domain. Our model outperforms state-of-the-art domain adaptation methods on two benchmark settings for remote sensing road segmentation, showing a promising application potential in real-world scenarios. In the future works, we will attempt to use the remote sensing image dataset of river, which has similar shape with road, as the source domain to guide the segmentation of road in target domain.

References

  • [1] C. Unsalan and B. Sirmacek, “Road network detection using probabilistic and graph theoretical methods,” IEEE Transactions on Geoscience and Remote Sensing, vol. 50, no. 11, pp. 4441–4453, 2012.
  • [2] R. Cao, W. Tu, C. Yang, Q. Li, J. Liu, J. Zhu, Q. Zhang, Q. Li, and G. Qiu, “Deep learning-based remote and social sensing data fusion for urban region function recognition,” ISPRS Journal of Photogrammetry and Remote Sensing, vol. 163, pp. 82–97, 2020.
  • [3] X. Yang, X. Li, Y. Ye, R. Y. K. Lau, X. Zhang, and X. Huang, “Road detection and centerline extraction via deep recurrent convolutional neural network u-net,” IEEE Transactions on Geoscience and Remote Sensing, vol. 57, no. 9, pp. 7209–7220, 2019.
  • [4] M. O. Sghaier and R. Lepage, “Road damage detection from vhr remote sensing images based on multiscale texture analysis and dempster shafer theory,” in International Geoscience and Remote Sensing Symposium, 2015, pp. 4224–4227.
  • [5] J. Yuan, D. Wang, B. Wu, L. Yan, and R. Li, “Legion-based automatic road extraction from satellite imagery,” IEEE Transactions on Geoscience and Remote Sensing, vol. 49, no. 11, pp. 4528–4538, 2011.
  • [6] G. Cheng, J. Han, P. Zhou, and D. Xu, “Learning rotation-invariant and fisher discriminative convolutional neural networks for object detection,” IEEE Transactions on Image Processing, vol. 28, no. 1, pp. 265–278, 2018.
  • [7] M. E. Paoletti, J. M. Haut, N. S. Pereira, J. Plaza, and A. Plaza, “Ghostnet for hyperspectral image classification,” IEEE Transactions on Geoscience and Remote Sensing, 2021.
  • [8] G. Cheng and J. Han, “A survey on object detection in optical remote sensing images,” ISPRS Journal of Photogrammetry and Remote Sensing, vol. 117, pp. 11–28, 2016.
  • [9] L. Zhang, L. Zhang, and B. Du, “Deep learning for remote sensing data: A technical tutorial on the state of the art,” IEEE Geoscience and Remote Sensing Magazine, vol. 4, no. 2, pp. 22–40, 2016.
  • [10] L. Zhang, J. Zhang, W. Wei, and Y. Zhang, “Learning discriminative compact representation for hyperspectral imagery classification,” IEEE Transactions on Geoscience and Remote Sensing, vol. 57, no. 10, pp. 8276–8289, 2019.
  • [11] Y. Liu, J. Yao, X. Lu, M. Xia, X. Wang, and Y. Liu, “Roadnet: Learning to comprehensively analyze road networks in complex urban scenes from high-resolution remotely sensed images,” IEEE Transactions on Geoscience and Remote Sensing, vol. 57, no. 4, pp. 2043–2056, 2019.
  • [12] L. Ding and L. Bruzzone, “Diresnet: Direction-aware residual network for road extraction in vhr remote sensing images,” IEEE Transactions on Geoscience and Remote Sensing, 2021.
  • [13] G. Cheng, Y. Wang, S. Xu, H. Wang, S. Xiang, and C. Pan, “Automatic road detection and centerline extraction via cascaded end-to-end convolutional neural network,” IEEE Transactions on Geoscience and Remote Sensing, vol. 55, no. 6, pp. 3322–3337, 2017.
  • [14] J. E. Vargas Munoz, S. Srivastava, D. Tuia, and A. X. Falcao, “Openstreetmap: Challenges and opportunities in machine learning and remote sensing,” IEEE Geoscience and Remote Sensing Magazine, 2021.
  • [15] L. Zhou, C. Zhang, and M. Wu, “D-linknet: Linknet with pretrained encoder and dilated convolution for high resolution satellite imagery road extraction,” in Proceedings of Conference on Computer Vision and Pattern Recognition Workshop, 2018, pp. 182–186.
  • [16] O. Tasar, S. Happy, Y. Tarabalka, and P. Alliez, “Colormapgan: Unsupervised domain adaptation for semantic segmentation using color mapping generative adversarial networks,” IEEE Transactions on Geoscience and Remote Sensing, vol. 58, no. 10, pp. 7178–7193, 2020.
  • [17] Q. Zhang, J. Zhang, W. Liu, and D. Tao, “Category anchor-guided unsupervised domain adaptation for semantic segmentation,” in Advances in Neural Information Processing Systems, 2019, pp. 433–443.
  • [18] A. Elshamli, G. W. Taylor, A. Berg, and S. Areibi, “Domain adaptation using representation learning for the classification of remote sensing images,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 10, no. 9, pp. 4198–4209, 2017.
  • [19] L. Ma, M. M. Crawford, L. Zhu, and Y. Liu, “Centroid and covariance alignment-based domain adaptation for unsupervised classification of remote sensing images,” IEEE Transactions on Geoscience and Remote Sensing, vol. 57, no. 4, pp. 2305–2323, 2019.
  • [20] P. Shamsolmoali, M. Zareapoor, H. Zhou, R. Wang, and J. Yang, “Road segmentation for remote sensing images using adversarial spatial pyramid networks,” IEEE Transactions on Geoscience and Remote Sensing, 2021.
  • [21] Q. Xie, M.-T. Luong, E. Hovy, and Q. V. Le, “Self-training with noisy student improves imagenet classification,” in Proceedings of Conference on Computer Vision and Pattern Recognition, 2020, pp. 10 687–10 698.
  • [22] S. Das, T. T. Mirnalinee, and K. Varghese, “Use of salient features for the design of a multistage framework to extract roads from high-resolution multispectral satellite images,” IEEE Transactions on Geoscience and Remote Sensing, vol. 49, no. 10, pp. 3906–3931, 2011.
  • [23] J. Xie, N. He, L. Fang, and A. Plaza, “Scale-free convolutional neural network for remote sensing scene classification,” IEEE Transactions on Geoscience and Remote Sensing, vol. 57, no. 9, pp. 6916–6928, 2019.
  • [24] B. Du, L. Ru, C. Wu, and L. Zhang, “Unsupervised deep slow feature analysis for change detection in multi-temporal remote sensing images,” IEEE Transactions on Geoscience and Remote Sensing, vol. 57, no. 12, pp. 9976–9992, 2019.
  • [25] J. Zhang and D. Tao, “Empowering things with intelligence: A survey of the progress, challenges, and opportunities in artificial intelligence of things,” IEEE Internet of Things Journal, 2021.
  • [26] W. Wang, W. Zhai, and Y. Cao, “Deep inhomogeneous regularization for transfer learning,” in 2020 IEEE International Conference on Image Processing (ICIP).   IEEE, 2020, pp. 221–225.
  • [27] G. Cheng, C. Yang, X. Yao, L. Guo, and J. Han, “When deep learning meets metric learning: Remote sensing image scene classification via learning discriminative cnns,” IEEE transactions on geoscience and remote sensing, vol. 56, no. 5, pp. 2811–2821, 2018.
  • [28] M. Lan, Y. Zhang, L. Zhang, and B. Du, “Global context based automatic road segmentation via dilated convolutional neural network,” Information Sciences, vol. 535, pp. 156–171, 2020.
  • [29] Y. Wei, K. Zhang, and S. Ji, “Simultaneous road surface and centerline extraction from large-scale remote sensing images using cnn-based segmentation and tracing,” IEEE Transactions on Geoscience and Remote Sensing, vol. 58, no. 12, pp. 8919–8931, 2020.
  • [30] L. Zhang, L. Song, B. Du, and Y. Zhang, “Nonlocal low-rank tensor completion for visual data,” IEEE transactions on cybernetics, vol. 51, no. 2, pp. 673–685, 2021.
  • [31] Y. Li, B. Peng, L. He, K. Fan, and L. Tong, “Road segmentation of unmanned aerial vehicle remote sensing images using adversarial network with multiscale context aggregation,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 12, no. 7, pp. 2279–2287, 2019.
  • [32] M. Li, Y.-M. Zhai, Y.-W. Luo, P.-F. Ge, and C.-X. Ren, “Enhanced transport distance for unsupervised domain adaptation,” in Proceedings of Conference on Computer Vision and Pattern Recognition, 2020, pp. 13 936–13 944.
  • [33] J. Dong, Y. Cong, G. Sun, Y. Liu, and X. Xu, “Cscl: Critical semantic-consistent learning for unsupervised domain adaptation,” in European Conference on Computer Vision, 2020, pp. 745–762.
  • [34] L. Song, C. Wang, L. Zhang, B. Du, Q. Zhang, C. Huang, and X. Wang, “Unsupervised domain adaptive re-identification: Theory and practice,” Pattern Recognition, vol. 102, p. 107173, 2020.
  • [35] Z. Chen, J. Zhang, and D. Tao, “Progressive lidar adaptation for road detection,” IEEE/CAA Journal of Automatica Sinica, vol. 6, no. 3, pp. 693–702, 2019.
  • [36] Y.-H. Tsai, W.-C. Hung, S. Schulter, K. Sohn, M.-H. Yang, and M. Chandraker, “Learning to adapt structured output space for semantic segmentation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 7472–7481.
  • [37] R. Li, Q. Jiao, W. Cao, H.-S. Wong, and S. Wu, “Model adaptation: Unsupervised domain adaptation without source data,” in Proceedings of Conference on Computer Vision and Pattern Recognition, 2020, pp. 9641–9650.
  • [38] F. Pan, I. Shin, F. Rameau, S. Lee, and I. S. Kweon, “Unsupervised intra-domain adaptation for semantic segmentation through self-supervision,” in Proceedings of Conference on Computer Vision and Pattern Recognition, 2020, pp. 3764–3773.
  • [39] X. Ma, X. Mou, J. Wang, X. Liu, J. Geng, and H. Wang, “Cross-dataset hyperspectral image classification based on adversarial domain adaptation,” IEEE Transactions on Geoscience and Remote Sensing, 2021.
  • [40] I. J. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial networks,” arXiv:1406.2661, 2014.
  • [41] Q. Xu, X. Yuan, and C. Ouyang, “Class-aware domain adaptation for semantic segmentation of remote sensing images,” IEEE Transactions on Geoscience and Remote Sensing, 2021.
  • [42] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of Conference on Computer Vision and Pattern Recognition, 2016, pp. 770–778.
  • [43] L.-C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille, “Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 40, no. 4, pp. 834–848, 2017.
  • [44] A. Radford, L. Metz, and S. Chintala, “Unsupervised representation learning with deep convolutional generative adversarial networks,” arXiv:1511.06434, 2015.
  • [45] F. Bastani, S. He, S. Abbar, M. Alizadeh, and H. Balakrishnan, “Roadtracer: Automatic extraction of road networks from aerial images,” in Proceedings of Conference on Computer Vision and Pattern Recognition, 2018, pp. 4720–4728.
  • [46] I. Demir, K. Koperski, D. Lindenbaum, G. Pang, J. Huang, S. Basu, F. Hughes, D. Tuia, and R. Raskar, “Deepglobe 2018: A challenge to parse the earth through satellite images,” in Proceedings of Conference on Computer Vision and Pattern Recognition Workshops, 2018, pp. 172–181.
  • [47] T.-H. Vu, H. Jain, M. Bucher, M. Cord, and P. Pérez, “Advent: Adversarial entropy minimization for domain adaptation in semantic segmentation,” in Proceedings of Conference on Computer Vision and Pattern Recognition, 2019, pp. 2517–2526.
  • [48] Y. Li, L. Yuan, and N. Vasconcelos, “Bidirectional learning for domain adaptation of semantic segmentation,” in Proceedings of Conference on Computer Vision and Pattern Recognition, 2019, pp. 6936–6945.
  • [49] L. van der Maaten and G. Hinton, “Visualizing data using t-sne.” Journal of Machine Learning Research, vol. 9, no. 11, 2008.
[Uncaptioned image] Lefei Zhang (S’11-M’14-SM’20) received the B.S. and Ph.D. degrees from Wuhan University, Wuhan, China. He is currently a professor with the School of Computer Science, Wuhan University. He was a Big Data Institute Visitor with the Department of Statistical Science, University College London, and a Hong Kong Scholar with the Department of Computing, The Hong Kong Polytechnic University. His research interests include pattern recognition, image processing, and remote sensing. Dr. Zhang serves as an associate editor for Pattern Recognition and IEEE Geoscience and Remote Sensing Letters.
[Uncaptioned image] Meng Lan received the B.S. degree from the School of Computer Science, Wuhan University, China, in 2018. He is currently pursuing the Ph.D degree in the School of Computer Science, Wuhan University, China. His research interests include deep learning and computer vision.
[Uncaptioned image] Jing Zhang is a research fellow at the School of Computer Science of the University of Sydney. His research interests include computer vision and deep learning. He has published more than 30 papers in prestigious journals and proceedings at leading conferences. He serves as a reviewer for many journals and conferences and a senior program committee member of the AAAI Conference on Artificial Intelligence and the International Joint Conference on Artificial Intelligence. He is a member of IEEE and ACM.
[Uncaptioned image] Dacheng Tao (Fellow, IEEE) is the President of the JD Explore Academy and a Senior Vice President of JD.com. He is also an advisor and chief scientist of the digital science institute in the University of Sydney. He mainly applies statistics and mathematics to artificial intelligence and data science, and his research is detailed in one monograph and over 200 publications in prestigious journals and proceedings at leading conferences. He received the 2015 Australian Scopus-Eureka Prize, the 2018 IEEE ICDM Research Contributions Award, and the 2021 IEEE Computer Society McCluskey Technical Achievement Award. He is a fellow of the Australian Academy of Science, AAAS, ACM and IEEE.