This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Building Damage Mapping with Self-Positive Unlabeled Learning

Junshi Xia
RIKEN
[email protected]
Naoto Yokoya
The University of Tokyo and RIKEN
[email protected]
Bruno Adriano
RIKEN
[email protected]
Abstract

Humanitarian organizations must have fast and reliable data to respond to disasters. Deep learning approaches are difficult to implement in real-world disasters because it might be challenging to collect ground truth data of the damage situation (training data) soon after the event. The implementation of recent self-paced positive-unlabeled learning (PU) is demonstrated in this work by successfully applying to building damage assessment with very limited labeled data and a large amount of unlabeled data. The self-PU learning is compared with the supervised baselines and traditional PU learning using different datasets collected from the 2011 Tohoku earthquake, the 2018 Palu tsunami, and the 2018 Hurricane Michael. By utilizing only a portion of labeled damaged samples, we show how models trained with self-PU techniques may achieve comparable performance as supervised learning.

1 Introduction

Natural disasters, such as earthquakes, tsunami, and wildfires, are extremely rare occurrences that can wreak havoc on urban environments, resulting in enormous human and economic losses [19]. When a new natural disaster strikes, accurate and rapid actions are necessary in Humanitarian Assistance and Disaster Response (HADR) in order to save thousands of lives and prevent further damage [20]. Thus, a high-precision disaster damage mapping product needs to meet this demand. However, using very high-resolution optical and SAR images, the disaster damage areas are manually annotated by specialists, which might take a long time if the disaster area is huge [18].

Deep learning is an effective method to automatically recognize damages in remote sensing datasets  [2, 16]. However, deep learning models necessitate a huge number of training images as well as a large amount of high-quality labeled data  [8]. This is not possible in the case of new disasters. One possible solution is to use the datasets and labels of past disasters to identify the damages of new disasters. However, due to domain gaps caused by many factors, such as various geographical locations, different damage characteristics, and seasonal variation, models that have only been trained on previous disasters would not work well for future disasters.

Since human-expert labeling is time-consuming, only a small number of labels are used when annotations obtained after a disaster are used to train models. The availability of labeled data is extremely restricted right after a new disaster, yet an enormous amount of unlabeled satellite images can be acquired over the affected area. Lee et al. [14] applied two semi-supervised learning (SSL) techniques, MixMatch and FixMatch, for the building damages with very limited labeled samples (damaged and undamaged). However, it is very time-consuming to collect both damaged and undamaged samples. Positive-unlabeled learning (PU), which integrates only positive and unlabeled training data, has recently made significant progress [3]. By using PU techniques, we can train accurate damage assessment models for new disasters by spending less time on gathering fewer positive labels of damaged class.

In this work, the recent Self-PU learning [6] is applied to train building damage detection models using very limited labeled positive damage samples. The method is applied to three different datasets to validate the performance. We demonstrate that by utilizing only a fraction of the labeled positive damage, models may obtain outcomes that are comparable to those obtained from fully supervised methods.

2 Related works

Deep learning in building damage mapping Previous studies conducted in fully supervised settings have effectively utilized deep learning algorithms for building damage assessment. The xView2 Challenge [12] developed the first large-scale pre- and post-disaster datasets, namely xBD, for mapping building damage, which incorporates VHR optical imagery taken during 19 disasters, including floods, wildfires, and earthquakes. The winning solution adopted a Siamese neural network which is initialized by the building detection from pre-event dataset. Hao et al. [13] combined the features from pre- and post-disaster imagery with a Siamese attention-based Unet model. Gupta and Shah [11] proposed an end-to-end RescueNet for both building segmentation and individual building damage level assessment. Shen et al. [17] developed a cross-directional fusion strategy to investigate the relationships between the datasets taken before and after the disaster. Boin et al. [5] proposed the techniques to mitigate the imbalanced class problem for different damage levels. Adriano et al. [1] construct high-resolution optical imagery and high-to-moderate-resolution SAR data. The multi- and cross-modality deep neural networks are applied to the new dataset.

In the above studies, the testing datasets are drawn from the same distribution as the training dataset. They do not address the difficulty of conducting inference for a disaster that just happened with very limited labeled datasets. Xu et al. [22] and Benson and Ecker [4] investigated the generalization experiments among the disasters in different regions and indicated that a significant generalization gap existed in current state-of-the-art models. The findings of models trained on previous disasters did not perform well when applied to a new disaster that did not have any labeled datasets.

Positive-unlabeled (PU) learning PU learning is a subdivision of SSL. However, when compared to SSL, there are no negative samples in PU learning, making the task more difficult and necessitating the invention of an algorithm to compensate for the lack of negative samples [3]. There are currently two solutions: 1) find credible negative samples in unlabeled data using heuristics or semi-supervised learning approaches [15], 2) treat unlabeled samples as weighted negative ones, including non-negative PU (nnPU) [7]. To further improve the performance of PU learning, the self-paced learning [6], active learning is considered in the framework of PU learning. In the remote sensing applications, Gui et al. [10] investigated eigenvalue statistical components-based PU-Learning to extract built-up from PolSAR datasets. Wu et al. [21] produced a landslide susceptibility map via the PU-bagging technique.

3 Data

ABCD. ABCD (AIST Building Change Detection) dataset 111https://github.com/gistairc/ABCDdataset [9] is a new labeled dataset to identify whether buildings have been washed-away by tsunami of Tohoku region of Japan in 2011. The class label assigned to each patch pair (i.e. "damaged" or "survived") represents whether or not a building at the center of the before-tsunami patch got washed-away by tsunami. In this work, the number of images (the size is 160×160160\times 160) for "damaged" and "survived" are 4253/4253.

xBD. In additional to ABCD dataset for 2011 Tohoku earthquake, we generated the patch-based datasets for the 2018 Palu tsunami and 2018 Hurricane Michael from xBD. There are four damage levels, including no damage, minor damage, major damage and destroyed. To formulate the binary classification problem and make the class label same as ABCD, major damage and destroyed are categorized into the class Damaged, while no damage, and minor damage are combined as the class Survived. The patches of each building were cropped from the original WorldView images of xBD with a fixed size of 128 ×\times 128 pixels. Thus, the total number of samples for the 2018 Palu tsunami and 2018 Hurricane Michael are 7645 (27% as the damaged) and 9214 (29% as the damaged), respectively.

4 Proposed approach

Refer to caption
Figure 1: The proposed approach.

The objective is formulated as a binary classification with two classes (𝑿6\bm{X}\in\mathcal{R}^{6}, Y{1,+1}Y\in\{-1,+1\}): damaged and survived, where 𝑿\bm{X} is the input that contains six bands of pre- and post-event RGB and YY is the class label. The damaged and survived classes are treated as the positive (+1+1) and negative (1-1). In the disaster damage mapping, we only have few numbers of damaged samples (P). The remaining datasets are unlabeled (U), which include positive or negative. Assume that we have the training dataset DD, which include positive (DPD_{P}) and unlabeled datasets (DUD_{U}). Then, we have D=DP+DUD=D_{P}+D_{U}, where DPD_{P} and DUD_{U} contain nPn_{P} and nUn_{U} samples, respectively. Let p(x,y)p(x,y) be the underlying joint distribution of (𝑿,Y)(\bm{X},Y), p(x|Y=+1)p(x|Y=+1) and p(x|Y=1)p(x|Y=-1) are the distributions of P and N datasets. πp=p(Y=+1)\pi_{p}=p(Y=+1) and πn=p(Y=1)=1πp\pi_{n}=p(Y=-1)=1-\pi_{p} are the positive and negative class-prior probabilities. As shown in Fig. 1, we firstly train a classification model to create a trusted pseudo PN dataset for U (DtrustD_{trust}), and the remaining is treated as a new PU dataset (DUDtrustD_{U}-D_{trust}). Then, knowledge distillation with two student models and their teacher models is used as a regularization to enhance performance while overcoming the inconsistent between the two models.

Due to the fact that deep neural networks are prone to memorizing simple patterns first, followed by irregular patterns, We might start by looking at simple instances and labeling them with confidence. Given an input xx with the label yy to the given model gg, the probability of xx being positive are calculated. A larger probability implies greater confidence that xx belongs to the positive class, and vice versa. Based on this assumption, we selected nn most confident positive and negative samples from DUD_{U} to form DtrustD_{trust}. Then, the loss function that integrates the supervised one of DtrustD_{trust} and PU loss of DUDtrustD_{U}-D_{trust}, is given by

Lhybrid=xDtrustLCE(x,y)+xDUDtrustLnnPU(x)L_{hybrid}=\sum_{x\in D_{trust}}L_{CE}(x,y)+\sum_{x\in D_{U}-D_{trust}}L_{nnPU}(x) (1)

where, the cross-entropy loss are used for DtrustD_{trust} and non-negative PU (nnPU) [7] are used for DUDtrustD_{U}-D_{trust}. To make the the selected confident samples as accurate as possible, We start by training 10 epochs to “warm up” the model. Two ways are used to select the trusted pseudo PN datasets: 1) fixed size: the model choose the fixed number of samples while removing lower-confidence samples, 2) without replacement: the model choose a fixed number of samples without removing lower-confidence samples. In this work, we adapt the similar dynamic selection strategy [6]: linearly increase the size of DtrustD_{trust} from 0 to 20% of DuD_{u} with the 75% bootstrap size of training set for the model.

Then, knowledge distillation with two student models (g1g_{1} and g2g_{2}) and two teacher models (G1G_{1} and G2G_{2}) are further applied [6]. The two student networks shared the same D=DP+DUD=D_{P}+D_{U} and same networks architectures. However, two different trusted pseudo PN datasets may be extracted by two student networks, thus, the consistently between two student networks should be considered as the distillation via MSE loss. The distillations from g1g_{1} to g2g_{2} and from g2g_{2} to g1g_{1} are given by:

LMSE(g1,g2,x)=f(g1(x))f(g2(x))2,xDDtrust1L_{MSE}(g_{1},g_{2},x)=\|f(g_{1}(x))-f(g_{2}(x))\|^{2},x\in D-D_{trust1} (2)
LMSE(g2,g1,x)=f(g2(x))f(g1(x))2,xDDtrust2L_{MSE}(g_{2},g_{1},x)=\|f(g_{2}(x))-f(g_{1}(x))\|^{2},x\in D-D_{trust2} (3)

where, ff is a monotonic function of mapping (e.g., sigmoid function). The loss function of the student networks is given by (we only consider the MSE over hard unlabeled samples with larger nnPU):

Lstudent=xDUDtrust1LS(g1,g2,x)+xDUDtrust2LS(g2,g1,x)L_{student}=\sum_{x\in D_{U}-D_{trust1}}L_{S}(g_{1},g_{2},x)+\sum_{x\in D_{U}-D_{trust2}}L_{S}(g_{2},g_{1},x) (4)

where, LS(g1,g2,x)=LMSE(g1,g2,x)L_{S}(g_{1},g_{2},x)=L_{MSE}(g_{1},g_{2},x) when LnnPU(x)>αLMSE(g1,g2,x)L_{nnPU}(x)>\alpha L_{MSE}(g_{1},g_{2},x), otherwise LS(g1,g2,x)=0L_{S}(g_{1},g_{2},x)=0, α\alpha is the control parameter.

Beside the distillation between two student networks, two teacher networks are included to make sure the consistency with the weighted moving average trajectory of the student networks. The weights of G1G_{1} and G2G_{2} are updated as

Θ1,t=θ1,t1+βθ1,t\Theta_{1,t}=\theta_{1,t-1}+\beta\theta_{1,t} (5)
Θ2,t=θ2,t1+βθ2,t\Theta_{2,t}=\theta_{2,t-1}+\beta\theta_{2,t} (6)

where, Θ1,t\Theta_{1,t} and θ1,t\theta_{1,t} are the parameters of G1G_{1} and g1g_{1} at time tt, respectively. β\beta is the user-defined parameter. The distillations from G1G_{1} to g1g_{1} and from G2G_{2} to g2g_{2} are given by:

Lteacher=xDUf(G1(x))f(g1(x))2+xDUf(G2(x))f(g2(x))2L_{teacher}=\sum_{x\in D_{U}}\|f(G_{1}(x))-f(g_{1}(x))\|^{2}+\sum_{x\in D_{U}}\|f(G_{2}(x))-f(g_{2}(x))\|^{2} (7)

The total loss function is given by:

L=Lhybrid+Lstudent+LteacherL=L_{hybrid}+L_{student}+L_{teacher} (8)

5 Experimental results and analysis

Settings. The efficacy of the proposed methods on classifiy damages are tested on ABCD and xView2 datasets. We split the datasets with the ratios of 4/3/3 for train/validation/test. To formulate the PU setting, we select r% (r ranges from 10 to 30) damaged samples as P, and the remaining (1-r)% damaged and 100% survived class are treated as U. The fully-supervised learning with the following setting: 1) Full PN; 2) Standard PN (r% P+100% N); 3) Small PN (r% P+r% N); 4) PU (U is treated as the N) and nnPU are used as the baselines. In this work, we adopt the simple VGG11 that contains eight convolutional layers and three full-connected layer. The pre- and post-disaster RGB images are treated as the input of the models. We set the SGD with momentum of 0.9 and a batch size of 16 for 100 epochs on a single NVIDIA V100 GPU with 16GB memory.

Table 1: Classification accuracy of the proposed method and the baselines averaged over 5 runs
Datasets Ratio (%) Full PN Standard PN Small PN PU nnPU Self-PU
ABCD 10 0.94 0.72 0.64 0.53 0.74 0.79
20 0.75 0.67 0.54 0.78 0.81
30 0.77 0.69 0.55 0.80 0.87
Palu tsunami 10 0.95 0.74 0.71 0.64 0.79 0.82
20 0.76 0.73 0.66 0.82 0.87
30 0.81 0.75 0.66 0.85 0.89
Hurricane 10 0.92 0.76 0.67 0.59 0.79 0.82
Michael 20 0.78 0.69 0.60 0.84 0.84
30 0.80 0.72 0.59 0.86 0.86

Results. Table 1 reports the accuracies of different methods. The fully-supervised model on all labeled data (Full PN), which is not available in real scenario, can be treated as the upper bound of the performance. The fully-supervised model with PU setting obtained the worst results because the training label is very noisy, especially for the negative labels. The clean training sets with small amount of P (standard PN and small PN) improved the results. nnPU makes further improvements in all cases. Self-PU improves the performance on all three disaster datasets. It should be noticed that we can obtain nearly the accuracies of 90% using only 30% positive damaged class as the input.

Table 2: Classification accuracies with different components on ABCD datasets.
Methods Accuracy Methods Accuracy
nnPU 0.74 LhybirdL_{hybird} 0.76
LhybirdL_{hybird} (fixed size) 0.75 Lhybird+LstudentL_{hybird}+L_{student} 0.77
LhybirdL_{hybird} (w.o. replacement) 0.74 Self-PU 0.79
Table 3: Sensitive analysis of α\alpha and β\beta on the performance of ABCD dataset.
α\alpha Accuracy β\beta Accuracy
10 0.76 0.3 0.77
20 0.79 0.4 0.77
30 0.77 0.5 0.79
40 0.75 0.5 0.78

Ablation studies. From Table 2, it can be clearly seen that the proposed Self-PU methods gain the better performance than the ones which only consider the LhybirdL_{hybird} and Lhybird+LstudentL_{hybird}+L_{student}. Table 3 investigates three different values of α\alpha and β\beta. α\alpha indicates the quality of consistency between two students and β\beta controls the updates of teachers from the students. Both parameters favors a reasonably moderate values. Larger or smaller values will decrease the accuracies.

6 Conclusion

In this paper, we introduced a novel application of PU learning to detect damaged buildings in satellite imagery with limited labeled samples. We experimented with the recent PU techniques with the combination of self-paced learning and knowledge distillation, and showed how they are able to achieve strong performance using very limited samples of damaged buildings by leveraging unlabeled samples. They consistently outperformed fully supervised models and even achieved performance close to that of a fully supervised setting with no data constraints. The results empirically showed how self-PU learning approaches can be useful to train models when a new disaster is unfolding in an unseen region. For future work, we will work towards the combination of the labeled datasets from past disasters and very limited damaged samples to further improve the performance. .

7 Application Context

In the HADR domain, the most urgent need is to provide the building damage mapping that requires a large number of labeled datasets as the input of current SOTA deep learning models, which are already explored in the past AI for HADR workshops [17, 5]. However, such labeled datasets are not available for a new disaster. Previous studies in [22] and [4] proved that the result of a new disaster is not accurate by training the model on the past disasters. Lee et al. [14] attempted to apply SSL methods for the building damage mapping. In our work, we provided the building damage mapping using very limited positive damaged images, requiring less time for the annotation before the response.

Acknowledgments and Disclosure of Funding

This work was supported by KAKENHI (Grant Number 19K20309 and 19H02408), JSPS Bilateral Joint Research Project (Grant number JPJSBP 120203211).

References

  • Adriano et al. [2021] B. Adriano, N. Yokoya, J. Xia, H. Miura, W. Liu, M. Matsuoka, and S. Koshimura. Learning from multimodal and multitemporal earth observation data for building damage mapping. ISPRS Journal of Photogrammetry and Remote Sensing, 175:132–143, 2021.
  • Bai et al. [2018] Y. Bai, C. Gao, S. Singh, M. Koch, B. Adriano, E. Mas, and S. Koshimura. A Framework of Rapid Regional Tsunami Damage Recognition From Post-event TerraSAR-X Imagery Using Deep Neural Networks. IEEE Geoscience and Remote Sensing Letters, 15(1):43–47, 2018.
  • Bekker and Davis [2020] J. Bekker and J. Davis. Learning from positive and unlabeled data: a survey. Mach. Learn., 109(4):719–760, 2020.
  • Benson and Ecker [2020] V. Benson and A. Ecker. Assessing out-of-domain generalization for robust building damage detection, 2020. URL https://arxiv.org/abs/2011.10328.
  • Boin et al. [2020] J.-B. Boin, N. Roth, J. Doshi, P. Llueca, and N. Borensztein. Multi-class segmentation under severe class imbalance: A case study in roof damage assessment, 2020. URL https://arxiv.org/abs/2010.07151.
  • Chen et al. [2020] X. Chen, W. Chen, T. Chen, Y. Yuan, C. Gong, K. Chen, and Z. Wang. Self-PU: Self boosted and calibrated positive-unlabeled training. In Proceedings of the 37th International Conference on Machine Learning (ICML), volume 119, pages 1510–1519, 2020.
  • du Plessis et al. [2015] M. C. du Plessis, G. Niu, and M. Sugiyama. Convex formulation for learning from positive and unlabeled data. In ICML, volume 37, pages 1386–1394, 2015.
  • Everingham et al. [2015] M. Everingham, S. M. A. Eslami, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman. The Pascal Visual Object Classes Challenge: A Retrospective. International Journal of Computer Vision, 111(1):98–136, 2015.
  • Fujita et al. [2017] A. Fujita, K. Sakurada, T. Imaizumi, R. Ito, S. Hikosaka, and R. Nakamura. Damage detection from aerial images via convolutional neural networks. In 2017 Fifteenth IAPR International Conference on Machine Vision Applications (MVA), pages 5–8, 2017.
  • Gui et al. [2020] R. Gui, X. Xu, L. Wang, R. Yang, and F. Pu. Eigenvalue statistical components-based pu-learning for polsar built-up areas extraction and cross-domain analysis. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 13:3192–3203, 2020.
  • Gupta and Shah [2020] R. Gupta and M. Shah. Rescuenet: Joint building segmentation and damage assessment from satellite imagery, 2020. URL https://arxiv.org/abs/2004.07312.
  • Gupta et al. [2019] R. Gupta, R. Hosfelt, S. Sajeev, N. Patel, B. Goodman, J. Doshi, E. Heim, H. Choset, and M. Gaston. Creating xBD: A Dataset for Assessing Building Damage from Satellite Imagery. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, June 2019.
  • Hao et al. [2020] H. Hao, S. Baireddy, E. R. Bartusiak, L. Konz, K. LaTourette, M. Gribbons, M. Chan, M. L. Comer, and E. J. Delp. An attention-based system for damage assessment using satellite imagery, 2020. URL https://arxiv.org/abs/2004.06643.
  • Lee et al. [2020] J. Lee, J. Z. Xu, K. Sohn, W. Lu, D. Berthelot, I. Gur, P. Khaitan, Ke-Wei, Huang, K. Koupparis, and B. Kowatsch. Assessing post-disaster damage from satellite imagery using semi-supervised learning techniques, 2020. URL https://arxiv.org/abs/2011.14004.
  • Li and Liu [2003] X. Li and B. Liu. Learning to classify texts using positive and unlabeled data. In Proceedings of the Eighteenth International Joint Conference on Artificial Intelligence (IJCAI), pages 587–594, 2003.
  • Ma et al. [2019] H. Ma, Y. Liu, Y. Ren, and J. Yu. Detection of Collapsed Buildings in Post-Earthquake Remote Sensing Images Based on the Improved YOLOv3. Remote Sensing, 12(1):44, dec 2019. ISSN 2072-4292.
  • Shen et al. [2020] Y. Shen, S. Zhu, T. Yang, and C. Chen. Cross-directional feature fusion network for building damage assessment from satellite imagery. 2020. URL https://arxiv.org/abs/2010.14014.
  • Valentijn et al. [2020] T. Valentijn, J. Margutti, M. van den Homberg, and J. Laaksonen. Multi-hazard and spatial transferability of a cnn for automated building damage assessment. Remote Sensing, 12(17), 2020.
  • Wallemacq and Below [2018] P. Wallemacq and R. Below. Economic losses, poverty and disasters: 1998-2017. Technical report, United Nations Office for Disaster Risk Reduction, 9-11 Rue de Varembee CH 1202, Geneva Switzerland, 2018.
  • Weber et al. [2020] E. Weber, N. Marzo, D. P. Papadopoulos, A. Biswas, A. Lapedriza, F. Ofli, M. Imran, and A. Torralba. Detecting natural disasters, damage, and incidents in the wild. In ECCV, Glasgow, UK, August 23-28, volume 12364, pages 331–350, 2020.
  • Wu et al. [2021] B. Wu, W. Qiu, J. Jia, and N. Liu. Landslide susceptibility modeling using bagging-based positive-unlabeled learning. IEEE Geoscience and Remote Sensing Letters, 18(5):766–770, 2021.
  • Xu et al. [2019] J. Z. Xu, W. Lu, Z. Li, P. Khaitan, and V. Zaytseva. Building damage detection in satellite imagery using convolutional neural networks, 2019. URL https://arxiv.org/abs/1910.06444.