This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\cftpagenumbersoff

figure \cftpagenumbersofftable

A Light-Weighted Convolutional Neural Network for Bitemporal SAR Image Change Detection

Rongfang Wang Key Laboratory of Intelligent Perception and Image Understanding of Ministry of Education, School of Artificial Intelligence, Xidian University, Xi’an, China, 710071 Fan Ding Key Laboratory of Intelligent Perception and Image Understanding of Ministry of Education, School of Artificial Intelligence, Xidian University, Xi’an, China, 710071 Licheng Jiao Key Laboratory of Intelligent Perception and Image Understanding of Ministry of Education, School of Artificial Intelligence, Xidian University, Xi’an, China, 710071 Jia-Wei Chen Key Laboratory of Intelligent Perception and Image Understanding of Ministry of Education, School of Artificial Intelligence, Xidian University, Xi’an, China, 710071 Bo Liu Key Laboratory of Intelligent Perception and Image Understanding of Ministry of Education, School of Artificial Intelligence, Xidian University, Xi’an, China, 710071 Wenping Ma Key Laboratory of Intelligent Perception and Image Understanding of Ministry of Education, School of Artificial Intelligence, Xidian University, Xi’an, China, 710071 Mi Wang State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan 430079, China.
Abstract

Recently, many Convolution Neural Networks (CNN) have been successfully employed in bitemporal SAR image change detection. However, most of the existing networks are too heavy and occupy a large volume of memory for storage and calculation. Motivated by this, in this paper, we propose a lightweight neural network to reduce the computational and spatial complexity and facilitate the change detection on an edge device. In the proposed network, we replace normal convolutional layers with bottleneck layers that keep the same number of channels between input and output. Next, we employ dilated convolutional kernels with a few non-zero entries that reduce the running time in convolutional operators. Comparing with the conventional convolutional neural network, our light-weighted neural network will be more efficient with fewer parameters. We verify our light-weighted neural network on four sets of bitemporal SAR images. The experimental results show that the proposed network can obtain better performance than the conventional CNN and has better model generalization, especially on the challenging datasets with complex scenes.

keywords:
Light-weighted neural network, Bitemporal SAR images, Change detection

*Corresponding Author: Jia-Wei Chen  \linkable[email protected]

1 Introduction

Over past decades, much attentions focus on multitemporal Synthetic Aperture Radar (SAR) image change detection, since a SAR is capable of working in all-time and all-weather without the influence of extremely bad weather and the cloud. In a past decade, most traditional SAR image change detection methods are developed how to extract changed areas from a difference image (DI), which suppose to include the information of changed regions. The DI calculated by the log-ratio (LR) [1] is usually subject to the speckle and it is challenging to extract the accurate and clear information on the changed region. To tackle this issue, sparse learning[2] was recently proposed learning robust features from the noisy DI. Wang et al.[3] analyzed the affects of the SAR image speckle on change detection and proposed a sparse learning-based method for SAR image change detection.

Recently, deep neural networks have been successfully employed to computer vision and remote sensing image analysis due to its ability to exploiting essential and robust structural features on categories of objects. It also has been introduced into the field of change detection. Gong et al[4] proposed a deep neural network for SAR image change detection. Gao et al. [5] proposed a simple convolutional neural network, known as PCA-Net, exploring robust features on the changed regions from the noisy DI. However, the performance of these two unsupervised methods are limited without the correct guidance. To tackle this issue, Wang et al. [6] proposed a supervised PCA-Net to improve the performance by carefully collecting typical training samples, which obtains state-of-arts performance of bitemporal SAR image change detection. However, this two-layer convolutional neural network is low efficient, since the convolutional kernels are trained or generated by a Principle Component Analysis (PCA) decomposition. Recently, Li et al. [7] proposed a convolutional neural network (CNN) for SAR image change detection based on both unsupervised and supervised learning. Zhao et al. [8] proposed a bitemporal PolSAR image change detection by a joint classification and a similarity measurement. Currently, it is still an open problem to extract the changed regions from the noisy DI. Nowadays, a large volume of SAR images are acquired by satellites and it is imperative to develop an efficient model that can produce promising results of SAR image interpretation. Most above networks are too heavy and cost much computational burden. It is strongly required to develop a lightweight convolutional neural network.

Recently, several lite networks are proposed to improve the inference efficiency. Howard et al. and Sandler et al. proposed two lite networks MobileNetV1 [9] and MobileNetV2 [10] for visual category. Recently, Howard et al. proposed to search for MobileNetV3 [11]. Tan et al. [12] proposed an efficient network for visual category. These lite networks have been extensively employed to visual category and the experimental results show that they can achieve comparable performance with heavy networks, but with low latency and network capacity. It can be potentially performed on edge devices with low power. Most recently, Chen et al.[13] proposed a lightweight multiscale spatial pooling network for bitemporal SAR image change detection.

Following the idea of the lightweight neural network, in this letter, we focus on the application of lite networks in SAR image change detection. To achieve this, we propose a lite CNN for SAR image change detection. In the proposed network, bottleneck layers are introduced to reduce the number of output channel. Furthermore, the dilated convolutional layers[14] are introduced to enlarge receptive field with a few of non-zero entries in the kernel, which reduces the number of network parameters. We verify the proposed network by comparing other conventional CNN. Compared with the lightweight network in [13], the proposed network is more robust with the residual and bottleneck structure. Experimental results on four sets of bitemporal SAR image show that our proposed method obtain comparable performance with CNN, while being much more efficient than CNNs.

The rest of paper will be organized as follows. We will introduce our proposed method in Section 2. Then the proposed method will be verified on four datasets in Section 3. Finally, we will draw a conclusion in Section 4.

2 Proposed Method

Given bitemporal SAR images 𝐈1{\bf I}_{1} and 𝐈2{\bf I}_{2}, the DI can be generated as follows

𝐈DI=𝐈1𝐈2{\bf I}_{DI}={\bf I}_{1}\ominus{\bf I}_{2} (1)

where \ominus denote the difference operator. However, most existing difference operator is subject to the speckle. Then we will propose a lightweight convolutional neural network to exploit the changed regions from the noisy DI.

Refer to caption
Figure 1: The framework of the proposed network.(a)Network Architecture. (b) Bottleneck for encoder. (c)Bottleneck for decoder.
Table 1: The tensors of all the layers.
Layer Name Tensor Size Layer Name Tensor Size Layer Name Tensor Size
Initial Block
Input (32,32,1)
Conv (16,16,13) Max-pooling (16,16,1) Concatenation (16,16,14)
Group 1
BottleNeck 1.0 (8,8,64) BottleNeck 1.1 (8,8,64) BottleNeck 1.2 (8,8,64)
BottleNeck 1.3 (8,8,64) BottleNeck 1.4 (8,8,64)
Group 2
BottleNeck 2.0 (4,4,128) BottleNeck 2.1 (4,4,128) BottleNeck 2.2 (4,4,128)
BottleNeck 2.3 (4,4,128) BottleNeck 2.4 (4,4,128) BottleNeck 2.5 (4,4,128)
BottleNeck 2.6 (4,4,128) BottleNeck 2.7 (4,4,128) BottleNeck 2.8 (4,4,128)
Group 3
BottleNeck 3.0 (4,4,128) BottleNeck 3.1 (4,4,128) BottleNeck 3.2 (4,4,128)
BottleNeck 3.3 (4,4,128) BottleNeck 3.4 (4,4,128) BottleNeck 3.5 (4,4,128)
BottleNeck 3.6 (4,4,128) BottleNeck 3.7 (4,4,128)
Group 4
BottleNeck 4.0 (8,8,64) BottleNeck 4.1 (8,8,64) BottleNeck 4.2 (8,8,64)
Group 5
BottleNeck 5.0 (16,16,16) BottleNeck 5.1 (16,16,16)
Output
Conv (32,32,2)

The whole framework of the proposed network can illustrated in Fig.1(a). It is shown that the network consists of five groups of bottleneck layers [15] with an 1×\times1 kernels, illustrated by variety of colorful bars, among which the former three ones work as the encoder, and the latter two ones as the decoder. The tensors of all layers are listed in Table 1.

In the forward process, the network takes a patch of DI as the input. Firstly, the input data go through a normal convolutional layer and a max-pooling (MP) layer, respectively and then the outputs are concatenated. Next, the contact activations go through the decoder with three groups bottleneck layers. The essential structure of a bottleneck of encoder can be illustrated in Fig.1(b). It is shown that a bottleneck layer is constructed by a small residual block, including a maxpooling path and a convolutional path. More specifically, the convolutional path consists of two 1×\times1 convolutions and one main convolution. The main convolution will vary with the various function of the bottleneck. It can be a normal convolution, a dilated convolution [14] or an asymmetrical convolution[16]. The tensors inside the encoding bottleneck are listed in Table 2.

Table 2: The tensors inside an encoding bottleneck layer.
Layer Name Tensor Size Layer Name Tensor Size
Input (16,16,14)
Branch 1
Conv1 (8,8,16) Conv2 (8,8,16)
Conv3 (8,8,64) Dropout (8,8,64)
Branch 2
Max-pooling (8,8,14) Padding (8,8,64)
Output
Addition (8,8,64)
Table 3: The tensors inside a decoding bottleneck layer.
Layer Name Tensor Size Layer Name Tensor Size
Input (4,4,128)
Branch 1
Conv1 (4,4,16) Conv2 (8,8,16)
Conv3 (8,8,64)
Branch 2
Conv (4,4,64) Up-Sampling (8,8,64)
Output
Addition (8,8,64)

The first group consist a down-sampling bottleneck and four normal convolutional bottleneck layers. In the normal bottleneck layers, the main convolution component is default set as the main normal convolutional layer. Especially, in the down-sampling bottleneck, the main convolutional component is set as a normal convolution kernel with 3×\times3 and the 1×\times1 convolution component is replaced by a 2×\times2 one. In the next two groups, to exploit the spatial context, we insert the bottleneck layers with asymmetrical and dilated convolution layers with various kernel sizes among the normal bottleneck, where the kernel sizes are set 2, 4, 8 and 16, respectively, as shown the digits below the green bars. In these bottleneck layers, the main convolutional layers are replaced by the dilated convolutional layers, where the kernels are sparse and most entries are zeros. Furthermore, the bottleneck layers with asymmetric convolution layers are also insert between the dilated and normal convolutional layer, illustrated by the blue bars. After encoding, the sizes of feature maps decrease as the one fourth as the original image.

In the decoding part, the context information collected by the encoder will be propagated to the pixel level. To achieve this, inspired by the idea of U-Net [17], we schedule a upsampling layer and two bottleneck layers. The essential structure of bottleneck for decoder can be illustrated in Fig.1(c). Similar to the bottleneck for the encoder, the bottleneck for the decoder contains a pooling path and a convolution path. The former includes a maxpooling layer and a 1×\times1 convolution, while the latter includes two 1×\times1 convolutional layers and a 3×\times3 convolutional layer. Especially, when the bottleneck layer is used for upsampling, the 3×\times3 convolution component, illustrated by the yellow bar, will be replaced by a 3×\times3 transpose convolution [18]. Then the bottleneck will do the 2x upsampling. Trough two groups of decoding bottleneck layers, the feature map will be recover the same size as the input image. The tensors inside the decoding bottleneck are listed in Table 3. Finally, we put a 2×\times2 convolutional layer to get the probability map of two categories.

Refer to caption
Figure 2: Four sets of bitemporal SAR images. The first two rows are bitemporal images and the last row is the DIs. (a) YR-A. (b) YR-B. (c) Sendai-A. (d) Sendai-B.

3 Experimental Results

3.1 Experiment Datasets

In this paper, the proposed method is verified on four sets of bitemporal SAR images. Two scenes (YR-A and YR-B) are from bitemporal Yellow River SAR images [4] acquired by the Radarsat-2 satellite in 2008 and 2009, respectively. Their image sizes are 306 ×\times 291 and 400 ×\times 350, respectively. Other two are parts of TerraSAR-X images acquired prior to (on Oct. 20, 2010) and after (on May 6, 2011) the Sendai earthquake in Japan [19]. Their sizes (Sendai-A and Sendai-B) are 590 ×\times687 and 689 ×\times 734, respectively. These four datasets are shown in Fig.2). These four datasets are quite challenging, such as the linear-shape changed regions in YR-B dataset and complex scene in both Sendai-A and Sendai-B datasets.

Table 4: The number of training samples.
Dataset YR-A YR-B Sendai-A Sendai-B
No. Samples 3596 6205 15375 20294
Refer to caption
Figure 3: The variations of loss and accuracies.

3.2 Implementations

We have introduced a lite CNN for change detection for bitemporal SAR images. We first generate the DI by the Eq.(1) and the difference operator is implemented by the neighborhood-based LR operator [20]. To train the network, we collect a training dataset according to the method in [6]. The numbers of training samples for each dataset are listed in Table 4.

More specifically, the patchsize of each sample is set as 32 ×\times 32 and 8 samples are fed in each training step. Additionally, the network is trained by an end-to-end back-propagation manner and the loss is set as the binary entropy function defined in [21]. The training is optimized by the Adam algorithm [22] in the training stage, where the initial learning rate is set as 0.005. The training is performed on the PyTorch platform built on the Ubuntu 16.04 installed in a PC with a 16 GB DDR memory and an NVIDIA TITAN Xp Graphics Processing Unit of 11 GB memory. The training process will converge at around 15 epochs. We show the variations of loss values and accuracies of training process in Fig.3.

3.3 Comparison Experiments

To verify the benefits of the proposed method, it is compared with the unsupervised PCA-Net (U-PCA-Net)[5], the supervised PCA-Net (S-PCA-Net) [6] which achieves the state-of-arts performance on SAR image change detection. We also compare the proposed method with the deep neural network (DNN) method [4] and CNN [7]. Among these methods, DNN and U-PCA-Net are unsupervised methods,while S-PCA-Net, CNN and the proposed method are supervised ones.

The performance of the compared methods is evaluated by probabilistic Missed Alarm (pMA), probabilistic False Alarm (pFA) and kappa coefficient, where pFA (pMA) are calculated by the ratios between FA (MA) and the number of Non-Changed pixels (NC)[3].

Refer to caption
Figure 4: The visual comparison results. (a)U-PCA-Net. (b)S-PCA-Net. (c)DNN. (d) CNN. (e)Lite CNN. (f)Reference.
Refer to caption
Figure 5: The quantitative evaluations of compared methods.(a) MA. (b)FA. (c) Kappa.

3.4 Experiment Results on Individual Dataset Change Detection

In this experiment, for the supervised learning methods, we collect the training samples from an individual dataset, which covers 30% areas of the whole image frame. The rest part is employed for testing.

The visual comparison results are shown in Fig.4. It is shown that for the YR-A dataset, S-PCA-Net, DNN and Lite CNN get less noisy but more completed changed regions. Lite CNN can get more clear boundary of changed regions than DNN. For the YR-B dataset, Lite CNN can get more completed changed regions, especially the line at the bottom of the image. Other methods can not get the completed changed regions. For the Sendai-A dataset with complex scene, S-PCA-Net and the Lite CNN are less subject to the speckle and the background and get more clear changed regions, while other compared methods are subjected to the speckle and background and they are almost failed. Moreover, compared with S-PCA-Net, Lite CNN get better inner regional consistence. For the Sendai-B dataset, Lite CNN gets more accurate changed regions than other methods.

Moreover, we show the quantitative evaluations in Fig.5. It is shown in Fig.5 (a) that the proposed Lite CNN gets a lower pMA on YR-A, Sendai-A and Sendai-B dataset, while the CNN gets a lower pMA on YR-B dataset. It is shown in Fig.5 (b) that the proposed Lite CNN gets a lower pFA on Sendai-B dataset, while S-PCA-Net gets a lower pFA on YR-A, YR-B and Sendai-A dataset. It is shown in Fig.5 (c) that on the YR-A dataset, S-PCA-Net gets the best kappa among all the methods, while the Lite CNN gets the comparable kappas with other methods, except S-PCA-Net. On the YR-B dataset, Lite CNN gets the comparable kappas with other methods. However, on both Sendai-A and Sendai-B datasets, the Lite CNN performs better than other methods in terms of pMA and kappas.

Refer to caption
Figure 6: The visual comparison results. (a) S-PCA-Net. (b) CNN. (c) Lite CNN. (d) Reference.
Refer to caption
Figure 7: The quantitative evaluations of compared methods.(a) MA. (b)FA. (c) Kappa.

3.5 Experiment Results on Cross-dataset Change Detection

To further compare the proposed method with other supervised learning methods, S-PCA-Net and CNN, we perform the comparisons on the cross-dataset change detection, where the network trained on several datasets is applied to an unknown testing dataset. More specifically, to achieve this, this experiment is conducted through the leave-one-out manner, i.e. each dataset alternative is selected as the testing dataset and others as training datasets.

The visual comparisons are shown in Fig.6. It is shown that on the YR-A dataset Lite CNN gets more clear visual result with less noisy spots. On the YR-B dataset, Lite CNN performs better than CNN, but not better than S-PCA-Net. There is many miss alarms in the results of Lite CNN. On both Sendai-A and Sendai-B datasets, Lite CNN gets better results than other two methods.

The quantitative evaluations in terms of pFA, pMA and Kappa are shown in Fig.7. It is shown in Fig.7(a) that the S-PCA-Net gets a lower pMA on all datasets. It is shown in Fig.7(b) that the Lite CNN gets a lower pFA on the Sendai-A dataset, while S-PCA-Net gets a lower pFA on the YR-A dataset and CNN gets a lower pFA on the YR-B and Sendai-B dataset. It is shown in Fig.7(c) that Lite CNN performs better than CNN but comparable with S-PCA-Net on YR-A and YR-B datasets. However, Lite CNN shows great advantages over other two methods on Sendai-A and Sendai-B datasets. It indicates that the Lite CNN performs better than other two methods in model generalization, especially on challenging datasets with complex scenes.

Table 5: The training times of compared methods.
Methods S-PCA-Net CNN Lite CNN
Times  3 h 30 mins. 15 mins.

3.6 Discussion

From the above comparisons, it is shown that Lite CNN can obtain comparable performance with other method on the YR-A and YR-B datasets. On the challenging datasets, e.g. Sendai-A and Sendai-B, Lite CNN outperforms other methods. Moreover, it has better ability to model generalization. Moreover, Lite CNN is more computationally efficient than S-PCA-Net and CNN. The training times of three supervised learning methods are compared in Table 5. It is shown that Lite CNN is easy to train and take less time than S-PCA-Net and CNN, while S-PCA-Net takes longer time, since the convolutional kernel is generated by the principle component analysis decomposition.

Overall, Lite CNN can obtain comparable or even better performance than S-PCA-Net and CNN. It is also more computationally efficient than other two methods. It is expected that Lite CNN is more practical in change detection, especially for the requirement of real-time detection.

4 Conclusion

In this paper, we develop a lightweight convolutional neural network for bitemporal SAR image change detection. The proposed network consists of groups of bottlenecks layers which exploit the image feature. To verify the benefits of our proposed method, we compare it with several traditional neural networks and the comparisons are performed on fours sets of bitemporal SAR images. The experimental results show that our proposed method Lite CNN performs better than other two methods on cross-dataset change detection, especially when the scene is complex. Furthermore, Lite CNN is a lightweight network, which is more computationally efficient than CNN and S-PCA-Net. In the future, we will further optimize Lite CNN and make it more efficient on the edge device.

Acknowledgment

This work was supported by the State Key Program of National Natural Science of China (No. 61836009), the National Natural Science Foundation of China (No. 61701361, 61806154), the Major Research Plan of National Natural Science Foundation of China (No. 91838303), the Open Fund of Key Laboratory of Intelligent Perception and Image Understanding of Ministry of Education, Xidian University (Grant No. IPIU2019006), the Natural Science Basic Research Plan in Shaanxi Province of China (No.2018JM6083) and the Open Fund of State Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University (No. 17E02).

References

  • [1] Y. Bazi, L. Bruzzone, and F. Melgani, “Automatic identification of the number and values of decision thresholds in the log-ratio image for change detection in sar images,” IEEE Geoscience and Remote Sensing Letters 3(3), 349–353 (2006).
  • [2] S. Wang, L. Jiao, and S. Yang, “Sar images change detection based on spatial coding and nonlocal similarity pooling,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 9(8), 3452–3466 (2016).
  • [3] R. Wang, J.-W. Chen, L. Jiao, et al., “How can despeckling and structural features benefit to change detection on bitemporal sar images?,” Remote Sensing 11(4), 421 (2019).
  • [4] M. Gong, J. Zhao, J. Liu, et al., “Change detection in synthetic aperture radar images based on deep neural networks.,” IEEE Transactions on Neural Networks and Learning Systems 27(1), 125–138 (2017).
  • [5] F. Gao, J. Dong, B. Li, et al., “Automatic change detection in synthetic aperture radar images based on pcanet,” IEEE Geoscience and Remote Sensing Letters 13(12), 1792–1796 (2016).
  • [6] R. Wang, J. Zhang, J. Chen, et al., “Imbalanced learning-based automatic sar images change detection by morphologically supervised pca-net,” IEEE Geoscience and Remote Sensing Letters 16(4), 554–558 (2018).
  • [7] Y. Li, C. Peng, Y. Chen, et al., “A deep learning method for change detection in synthetic aperture radar images,” IEEE Transactions on Geoscience and Remote Sensing (2019).
  • [8] J. Zhao, J. Yang, Z. Lu, et al., “A novel method of change detection in bi-temporal polsar data using a joint-classification classifier based on a similarity measure,” Remote Sensing 9(8), 846 (2017).
  • [9] A. G. Howard, M. Zhu, B. Chen, et al., “Mobilenets: Efficient convolutional neural networks for mobile vision applications,” arXiv preprint arXiv:1704.04861 (2017).
  • [10] M. Sandler, A. Howard, M. Zhu, et al., “Mobilenetv2: Inverted residuals and linear bottlenecks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 4510–4520 (2018).
  • [11] A. Howard, M. Sandler, G. Chu, et al., “Searching for mobilenetv3,” arXiv preprint arXiv:1905.02244 (2019).
  • [12] M. Tan and Q. V. Le, “Efficientnet: Rethinking model scaling for convolutional neural networks,” arXiv preprint arXiv:1905.11946 (2019).
  • [13] J.-W. Chen, R. Wang, F. Ding, et al., “A convolutional neural network with parallel multi-scale spatial pooling to detect temporal changes in sar images,” Remote Sensing 12(10), 1619 (2020).
  • [14] Y. Li, X. Zhang, and D. Chen, “Csrnet: Dilated convolutional neural networks for understanding the highly congested scenes,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 1091–1100 (2018).
  • [15] M. Lin, Q. Chen, and S. Yan, “Network in network,” arXiv preprint arXiv:1312.4400 (2013).
  • [16] C. Szegedy, V. Vanhoucke, S. Ioffe, et al., “Rethinking the inception architecture for computer vision,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2818–2826 (2016).
  • [17] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical image computing and computer-assisted intervention, 234–241, Springer (2015).
  • [18] V. Dumoulin and F. Visin, “A guide to convolution arithmetic for deep learning,” arXiv preprint arXiv:1603.07285 (2016).
  • [19] S. Cui, G. Schwarz, and M. Datcu, “A benchmark evaluation of similarity measures for multitemporal sar image change detection,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 9(3), 1101–1118 (2016).
  • [20] M. Gong, Y. Cao, and Q. Wu, “A neighborhood-based ratio approach for change detection in sar images,” IEEE Geoscience and Remote Sensing Letters 9(2), 307–311 (2011).
  • [21] P. Sadowski, “Notes on backpropagation,” homepage: https://www. ics. uci. edu/pjsadows/notes. pdf (online) (2016).
  • [22] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980 (2014).