This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

1st Place Solution to MultiEarth 2023 Challenge on
Multimodal SAR-to-EO Image Translation

Jingi Ju  Hyeoncheol Noh11footnotemark: 1  Minwoo Kim11footnotemark: 1  Dong-Geol Choi
Hanbat National University
{jingi.ju, hyeoncheol.noh, minwoo.kim, dgchoi}@hanbat.ac.kr
equal contributioncorresponding author
Abstract

The Multimodal Learning for Earth and Environment Workshop (MultiEarth 2023) aims to harness the substantial amount of remote sensing data gathered over extensive periods for the monitoring and analysis of Earth’s ecosystems’ health. The subtask, Multimodal SAR-to-EO Image Translation, involves the use of robust SAR data, even under adverse weather and lighting conditions, transforming it into high-quality, clear, and visually appealing EO data. In the context of the SAR2EO task, the presence of clouds or obstructions in EO data can potentially pose a challenge. To address this issue, we propose the Clean Collector Algorithm (CCA), designed to take full advantage of this cloudless SAR data and eliminate factors that may hinder the data learning process. Subsequently, we applied pix2pixHD for the SAR-to-EO translation and Restormer for image enhancement. In the final evaluation, the team ‘CDRL’ achieved an MAE of 0.07313, securing the top rank on the leaderboard.

1 Introduction

Satellite remote sensing data possesses remarkable capabilities in capturing a wide range of situations, making it widely utilized in various tasks such as environmental monitoring  [1, 2], change detection  [10, 11], and disaster assessment  [5, 15]. However, the susceptibility to noise, caused by factors like weather, clouds, smoke, and lighting conditions, imposes significant limitations on the usage of electro-optical (EO) data. In particular, the presence of noise poses a significant challenge in environmental monitoring tasks that require continuous EO imagery. To overcome these limitations, a methodology utilizing synthetic aperture radar (SAR) is being proposed. SAR provides advantages such as the ability to characterize surface properties and collect data during both day and night, as well as in adverse weather conditions. Nevertheless, SAR data presents challenges in human interpretation, restricting its utilization. To overcome these challenges, the task of converting SAR data into EO data, which can be more easily understood and interpreted by humans, has been proposed.

Refer to caption
Figure 1: Our Clean Collector Algorithm (CCA) example images.

The SAR to EO translation track at The Multimodal Learning for Earth and Environment Workshop (MultiEarth 2023) provides SAR (Sentinel-1) and EO (Sentinel-2) data for the Amazon rainforest. Due to the year-round high temperatures and heavy rainfall in the Amazon rainforest, frequent cloud cover poses limitations on continuous monitoring. Therefore, the objective of this task is to transform SAR data into cloud-free, high-quality EO data. To achieve this goal, it is necessary to construct a training dataset without noise such as clouds, lighting differences, weather conditions, and no-data. However, the provided QA60 band alone has its limitations. To effectively remove noise, we propose the Clean Collector Algorithm (CCA), which consists of three stages. Firstly, utilizing the QA60 band and experimentally obtained pixel values, we eliminate images with a significant amount of clouds. Then, brightness and the ratio of pixels with a value of 0 are calculated to eliminate nighttime and no-data. Finally, we remove data that still contains clouds by using a manually created subset of cloud EO and the frechet inception distance (FID) scores of each EO data.  Fig. 1 shows example images of each stage. We utilize a pix2pixHD  [12] model trained on data constructed with CCA to perform the SAR to EO translation. Furthermore, we employ the Restormer  [14] model to enhance the quality of the transformed images.

2 Methodology

2.1 Data preparation

The noise present in EO data (such as clouds, cloudy weather, night, and other no-data) makes it challenging to train the shared terrain information between SAR and EO. In order to generate clear and high-quality EO data, we need to remove observable noise and proceed with the training. However, the provided QA60 band alone has limitations in noise removal.

We propose the Clean Collector Algorithm (CCA), which consists of three stages, to effectively remove various types of noise. In the first stage, we remove data containing clouds based on the QA60 band and pixel threshold. We have discovered that a significant number of images within the cloud data exceed a certain threshold value, and we consider data points surpassing the pixel threshold α\alpha as being indicative of cloud data. We empirically used α\alpha as 4,096. In the second stage, we removed nighttime data and no-data. Specifically, we remove images with an average brightness lower than 30 after conversion to HSV. Also, remove images where the number of pixels with a pixel value less than 10 exceeds 10% of the total pixels. We determined that nighttime data and no-data, similar to cloud images, are difficult to interpret and would have a negative impact on the training process. Lastly, we select sharp EO data through FID score calculation. This utilizes a feature-based similarity measure, FID, to remove cloud data that could not be processed based on pixel values. Specifically, we constructed a handcrafted subset of 1,000 cloud images from the provided EO data, which contains a significant amount of clouds. We calculate the FID score between cloud subset and each EO data and construct a set of scores, S={Si|i=1,2,,N}S=\{S_{i}|i=1,2,...,N\}, where NN is the number of preprocessed EO data from the previous two stages. After that, we removed EO data with scores lower than the score threshold FthF_{th} which is formulated as Eq. 1.

Fth=(min(S)+max(S)min(S))×βF_{th}=(\mathrm{min}(S)+\mathrm{max}(S)-\mathrm{min}(S))\times\beta (1)

Here, β\beta is a hyperparameter that determines the quality of the training dataset, and we used 0.4. Finally, we were able to select approximately 19,000 clean EO data samples, and we generated around 23,000 SAR-EO pairs within ±30 days.

2.2 SAR-to-EO Translation

There are prominent models in the image-to-image translation, such as pix2pix  [6], pix2pixHD, and SPADE  [13]. Recently, there has been significant interest in diffusion-based generative models, with notable examples being I2SB  [9], BBDM  [7], and DPGAN  [8]. These diffusion models have demonstrated exceptional performance in terms of semantic interpretability and generation quality across various benchmark datasets. Inspired by this, we also utilized the I2SB model. However, we encountered challenges of excessively long training time and the generation of excessive cloud images. We determined that this issue arose due to the slight inclusion of clouds in our training data. Consequently, we opted for the stable and efficient training offered by pix2pixHD.

We have determined that it is difficult to establish a clear criterion for cutting out outliers due to the scattered pixels in SAR data. These outliers tend to be disregarded during training, resulting in the generation of values unrelated to the ground truth when inputted. Therefore, to enhance image quality and performance, we have employed the Restormer, an enhancement model.

Submission dataset Method Score
Submission1 dataset1 CCA + pix2pixHD 0.077576
Submission2 dataset1 CCA + pix2pixHD + Restormer 0.073135
Submission3 dataset2 CCA + pix2pixHD + L1 loss 0.073675
Table 1: Test Data Submission Result.

3 Experiments

We use the provided Challenge Data only and trained approximately 23,000 SAR-EO pair data with a minute amount of clouds through our preprocessing algorithm. We generated 3-channel SAR data described in  [3] and employed median blur to reduce the outlier pixel values. We conducted experiments with two versions of datasets. In dataset1, we performed min-max normalization with a range of (0, 1) followed by normalization to the range of (-1, 1) in order to input SAR and EO data into the translation model. For dataset2, we applied the Tanh function to mitigate the impact of outliers in SAR data, while keeping the other settings the same as dataset1. After training the translation model, we used its output as input for the enhancement model. The performance is measured using the metric proposed in  [4], and the metric is formalized using Eq. 2.

jmin𝑖f(x)iyj\sum_{j}\underset{i}{\mathrm{min}}\left\|f(\mathrm{x})_{i}-\mathrm{y}_{j}\right\| (2)

Where f(x)if(x)_{i} represents the output image of the model, while yjy_{j} represents the reference image.

Table 1 presents the performance on the test dataset. Submission1 achieved a performance of 0.07757 when using the single model, pix2pixHD. Combining the enhancement model, Restormer, in Submission2 resulted in a performance of 0.07313. Additionally, Submission3 utilized dataset2 and experimented with the addition of L1 loss, achieving a performance of 0.07367.  Fig. 2 shows test image examples for each submission.

Refer to caption
Figure 2: (a), (b), and (c) are test example images for submission1, 2, and 3, respectively

4 Analysis Result

We engaged in the competition employing three distinct methodologies, all of which yielded exceptional performance. However, in  Fig. 2, even though the image in (c) appears to have the highest visual quality, it is (b) that achieved the best performance in terms of score, despite being a challenging image to interpret. This result can be attributed to the characteristics of Eq. 2, indicating that blurred images tend to achieve higher performance.

5 Conclusion

In this report, we propose an effective noise removal algorithm called Clean Collector Algorithm (CCA) for converting SAR data from the Amazon rainforest into interpretable EO data. CCA effectively removes various noise data, enabling the construction of a clean training dataset. Leveraging the dataset created by CCA, we trained the pix2pixHD and Restormer models, achieving a score of 0.07313 under the team name “CDRL” in the SAR to EO translation track of the Multimodal Learning for Earth and Environment Workshop (MultiEarth 2023). The outcome suggests that, for the SAR to EO translation, it is not only crucial to employ high-performance models but also essential to effectively eliminate noise data from the training dataset.

Acknowledgement

The authors would like to thank Minseok Seo of SI Analytics and Jongchan Park of Lunit for their technical discussions.

References

  • [1] Katherine Anderson, Barbara Ryan, William Sonntag, Argyro Kavvada, and Lawrence Friedl. Earth observation in service of the 2030 agenda for sustainable development. Geo-spatial Information Science, 20(2):77–96, 2017.
  • [2] Daniela Ballari, Luis M Vilches-Blázquez, María Lorena Orellana-Samaniego, Francisco Salgado-Castillo, Ana Elizabeth Ochoa-Sánchez, Valerie Graw, Nazli Turini, and Jörg Bendix. Satellite earth observation for essential climate variables supporting sustainable development goals: A review on applications. Remote Sensing, 15(11):2716, 2023.
  • [3] Armando Cabrera, Miriam Cha, Prafull Sharma, and Michael Newey. Sar-to-eo image translation with multi-conditional adversarial networks. In 2021 55th Asilomar Conference on Signals, Systems, and Computers, pages 1710–1714. IEEE, 2021.
  • [4] Miriam Cha, Gregory Angelides, Mark Hamilton, Andy Soszynski, Brandon Swenson, Nathaniel Maidel, Phillip Isola, Taylor Perron, and Bill Freeman. Multiearth 2023 – multimodal learning for earth and environment workshop and challenge, 2023.
  • [5] Masroor Hussain, Dongmei Chen, Angela Cheng, Hui Wei, and David Stanley. Change detection from remotely sensed images: From pixel-based to object-based approaches. ISPRS Journal of photogrammetry and remote sensing, 80:91–106, 2013.
  • [6] Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A Efros. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1125–1134, 2017.
  • [7] Bo Li, Kaitao Xue, Bin Liu, and Yu-Kun Lai. Bbdm: Image-to-image translation with brownian bridge diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1952–1961, 2023.
  • [8] Shijie Li, Ming-Ming Cheng, and Juergen Gall. Dual pyramid generative adversarial networks for semantic image synthesis. arXiv preprint arXiv:2210.04085, 2022.
  • [9] Guan-Horng Liu, Arash Vahdat, De-An Huang, Evangelos A Theodorou, Weili Nie, and Anima Anandkumar. I2sb: Image-to-image schr\\backslash” odinger bridge. arXiv preprint arXiv:2302.05872, 2023.
  • [10] Hyeoncheol Noh, Jingi Ju, Minseok Seo, Jongchan Park, and Dong-Geol Choi. Unsupervised change detection based on image reconstruction loss. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1352–1361, 2022.
  • [11] Minseok Seo, Hakjin Lee, Yongjin Jeon, and Junghoon Seo. Self-pair: Synthesizing changes from single source for object change detection in remote sensing imagery. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 6374–6383, 2023.
  • [12] Ting-Chun Wang, Ming-Yu Liu, Jun-Yan Zhu, Andrew Tao, Jan Kautz, and Bryan Catanzaro. High-resolution image synthesis and semantic manipulation with conditional gans. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 8798–8807, 2018.
  • [13] Mohammed J Zaki. Spade: An efficient algorithm for mining frequent sequences. Machine learning, 42:31–60, 2001.
  • [14] Syed Waqas Zamir, Aditya Arora, Salman Khan, Munawar Hayat, Fahad Shahbaz Khan, and Ming-Hsuan Yang. Restormer: Efficient transformer for high-resolution image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5728–5739, 2022.
  • [15] Xueliang Zhang, Pengfeng Xiao, Xuezhi Feng, and Min Yuan. Separate segmentation of multi-temporal high-resolution remote sensing images for object-based change detection in urban area. Remote Sensing of Environment, 201:243–255, 2017.