11email: [email protected]
† Corresponding Author
Masked Angle-Aware Autoencoder for Remote Sensing Images
Abstract
To overcome the inherent domain gap between remote sensing (RS) images and natural images, some self-supervised representation learning methods have made promising progress. However, they have overlooked the diverse angles present in RS objects. This paper proposes the Masked Angle-Aware Autoencoder (MA3E) to perceive and learn angles during pre-training. We design a scaling center crop operation to create the rotated crop with random orientation on each original image, introducing the explicit angle variation. MA3E inputs this composite image while reconstruct the original image, aiming to effectively learn rotation-invariant representations by restoring the angle variation introduced on the rotated crop. To avoid biases caused by directly reconstructing the rotated crop, we propose an Optimal Transport (OT) loss that automatically assigns similar original image patches to each rotated crop patch for reconstruction. MA3E111Our code will be released at: https://github.com/benesakitam/MA3E demonstrates more competitive performance than existing pre-training methods on seven different RS image datasets in three downstream tasks.
Keywords:
Masked Autoencoder Optimal transport Angle restoration Remote sensing image1 Introduction
Nowadays, deep learning-based interpretation of remote sensing (RS) images has been widely applied in fields related to national defense [69, 49, 48] and people’s well-being [24, 45, 30]. The increasing number of Earth observation satellites makes it possible to acquire a massive amount of unlabeled RS images. Despite the abundance of data, many RS models still initialize with the ImageNet [15] pre-trained weights. Inherent domain gap between natural images and RS images limit the performance of these models. Therefore, exploring self-supervised representation learning on RS images is highly necessary.


Self-supervised representation learning [5, 26, 1, 25] for natural images has emerged as a new paradigm for pre-training models on large-scale datasets. Among these, Masked Image Modeling (MIM) [1, 25, 64, 58] learns visual representations by reconstructing masked portions of the input. With its concise architecture and outstanding performance on downstream tasks [15, 33, 71], it has attracted widespread attention. Recently, several noteworthy MIM studies have surfaced in the RS image field [11, 46, 54], offering excellent initialization for the vision transformer [16, 70] and achieving good results across various downstream tasks. This demonstrates the potential of MIM in representation learning for RS images.
Although existing customized MIM methods for RS images take various factors into account, such as different resolutions [46], multi-scale objects and the complex background [50], and imaging from multiple spectral bands [11], they are not effective learners in the face of angles of RS objects. Objects in natural images typically have fixed orientations due to gravity, whereas in RS images, objects captured from an overhead perspective often exhibit a wide range of angles. The same RS object presents diverse shapes and appearances when viewed from different angles. Properly perceiving and considering angle information aligns with the nature of how objects are captured in the RS community, which is conducive to accurate image interpretation. The above methods only focus on reconstructing the pixel values of RS objects, the learning for angles is often implicitly accompanied within reconstruction. Therefore, we propose the Masked Angle-Aware Autoencoder (MA3E), which perceives and learns angle information by restoring the preset angle variations during original pixel reconstruction. As an illustrative example, when fine-tuning a pre-trained model for rotated object detection, we count the angles of all correctly detected objects in two datasets and report AP50 for objects falling within different angle ranges in Fig. 1. [54] only obtains the higher AP50 for objects with angles close to horizontal (e.g., 0∘ or 90∘). In contrast, MA3E significantly improves AP50 for objects with large inclinations. This indicates that MA3E effectively becomes aware of diverse angles of objects and learn robust rotation-invariant representations.
MA3E follows an asymmetric encoder-decoder architecture and the similar pre-training principle of the classical Masked Autoencoder (MAE) [25]. Firstly, MA3E creates the rotated crop at arbitrary position within each RS image, introducing an explicit angle variation. We propose a scaling center crop operation to construct diverse rotated crops while preserving primary scenes. Each rotated crop is center-rotated at a random angle and replaces the original scene. Then, MA3E takes this composite image as the training input and reconstructs the original image. Before feeding into the encoder, an additional angle embedding is added to the rotated crop to implicitly prompt the model to perceive the angle of this region. Masking is applied respectively to the rotated crop and the remaining background to avoid losing too much or all patches of rotated crop. Due to the scene offset in the rotated crop, direct reconstruction using the original image patches of corresponding positions would result in obvious biases. Therefore, the reconstruction for the rotated crop is treated as an Optimal Transport (OT) problem. We propose an OT loss that automatically allocates similar original image patches as reconstruction targets for each patch of the rotated crop.
By simultaneously restoring the angle variations while reconstructing the original pixels. MA3E exhibits an awareness of diverse angles, enabling it to effectively learn rotation-invariant visual representations. MA3E demonstrates outstanding performance in several downstream tasks, including scene classification on NWPU-RESISC45 [7], AID [60], and UC Merced [65], rotated object detection on DOTA1.0 [59] and DIOR-R [8], as well as semantic segmentation on iSAID [57] and Potsdam.
2 Related Works
The development of vision transformers [16, 37, 56, 70] has advanced the masked image modeling [1, 25, 18]. MIM has gradually replaced contrastive learning [26, 5, 23], becoming the currently prominent pre-training paradigm in the computer vision field.
Model Image Modeling. MIM aims to reconstruct masked parts using the visible input. BEiT [1] masks 60% of the image and relies on tokens extracted from these masked regions by dVAE [47] for reconstruction. SimMIM [64] encodes visible patches and mask tokens, directly predicting the original pixel values. MAE [25] improves reconstruction efficiency by feeding only visible patches into the encoder. GreenMIM [29] proposes an optimal grouping algorithm, deploying MAE on a hierarchical transformer [37] by dividing each window into multiple groups. Some studies design diverse reconstruction targets, such as advanced CLIP [27] or DINO features [19], HOG features [58], frequencies [62, 35], and multi-level features [55, 36]. Moreover, some works pay attention to process input images; [52] recovers masked patches with five different learning targets, LoMaR [3] reconstructs multiple local regions of an image, and MixMAE [34] takes a mixed image as input and simultaneously reconstructs multiple original images before mixing. These methods significantly advance self-supervised representation learning based on natural images.
MIM in RS Images. Imaging sources for RS images are diverse, covering complex scenes with uneven scales and distributions of foreground objects or land cover. Current works have gradually transitioned from contrastive learning utilizing land cover information such as seasonal changes [42] and temporal differences [41] to the customized MIM methods. Wang et al. [54] pre-trains with MAE [25] on the MillionAID dataset [38] and fine-tunes by replacing the original transformer’s global attention with rotated varied-size window attention in downstream tasks. CMID [44] introduces contrastive learning into the MIM branch to learn consistency. RingMo [50] collects two millions images from satellite and aerial platforms and designs a patch incomplete masking strategy for reconstruction. GFM [43] pre-trains on GeoPile, a constructed dataset with multiple sources, and continual learns valuable in-domain representations under the guidance of the ImageNet-22k models. SatMAE [11] encodes temporal and multi-spectral information in position embeddings to extend spatio-temporal relationships in the fMoW dataset [9]. ScaleMAE [46] leverages the inherent ground sample distance to reconstruct multi-scale resolution images. It is regrettable that the above methods do not explore angles during pre-training. In this paper, we propose MA3E, which simultaneously performs pixel reconstruction and angle restoration, thus perceiving angles and learning rotation-invariant representations.
3 Method
3.1 Preliminary: MAE
MAE [25] employs an asymmetric encoder-decoder architecture for efficient masked image modeling. An input image, , is first reshaped into a series of non-overlapping image patches of size , denoted as , where is the number of patches. Then, is linearly mapped into patch embeddings. MAE adds positional encoding information to these embeddings and randomly masks them at a certain ratio, such as 75%. The masked patches, , are discarded, and only the remaining visible patches, , are fed into the encoder to extract latent features. These latent features, along with the shared and learnable mask tokens representing the substituted masked patches, constitute the input to the decoder, with positional embeddings also added. After obtaining the decoder output , MAE predicts only the pixel values of the masked patches, using the original image patches as the reconstruction targets. This is achieved by computing the mean squared error (MSE) loss:
(1) |
where denotes the output of the decoder for masked patches. The decoder is adopts for pre-training only, while the encoder is further fine-tuned for downstream tasks. This often results in the common practice of using a lightweight decoder and a complete transformer encoder. The proposed MA3E shares the similar principle with MAE, described next.
3.2 Masked Angle-Aware Autoencoder (MA3E)

MA3E aims to be aware of diverse angles and learn rotation-invariant visual representations. Fig. 2(a) illustrates our pipeline. MA3E constructs rotated crops at arbitrary positions on the RS images by deploying the designed scaling center crop operation to introduce explicit angle variations. These rotated crops have random angles and replace the scenes at their original locations. We add an angle embedding in each rotated crop, and mask the rotated crop and remaining background separately. For the reconstruction of the rotated crop, MA3E automatically allocates similar image patches as reconstruction targets for each rotated crop patch based on the transportation plan solved by the Sinkhorn-Knopp [14] algorithm. This avoids biases introduced by the crop operation.
Rotated crop. Using a simple random rotation operation to construct the rotated crop of side length for each RS image would lead to adverse results as shown in Fig. 3(b). The model struggles to learn high-quality representations from these regions, resulting in wasted computational resources. Therefore, we propose a scaling center crop operation to create diverse rotated crops that preserve scenes to a substantial extent, as shown in Fig. 3(a). For a square region (blue) of side length at arbitrary positions on an image, rotating this region at random angle would lead to loss of edge scenes (gray). However, the scenes within its largest inscribed circle (red) are fully preserved. Hence, we perform center cropping to extract the largest inscribed square region with side length from the red circle as the rotated crop. This region holds an arbitrary orientation, replacing the original scene and introducing the explicit angle variation to the composite image. To ensure that each rotated crop can be entirely patchified, the side length needs to be divisible by the patch size , with the starting position being a multiple of .

Angle embedding. In addition to adding positional embeddings for the divided patches of composite images, MA3E also includes the angle embeddings for the rotated crops. Each angle embedding is a learnable vector shared across each patch within a rotated crop. They serve as implicit cues for the model to perceive the angle variation in the rotated crop, while also distinguishing them from the remaining background.
Random masking. Given patches from the rotated crop (denoted as ) and patches from the background (denoted as ), to avoid the random masking strategy that removes too much or all patches from the rotated crop, we separately mask and at a certain ratio, e.g., 75%. Thus, visible patches and masked patches from the background are denoted as and respectively. These definition holds similarly for the rotated crop patches .
Reconstruction. MA3E uses MSE loss to predict the pixel values of masked patches in the background. For the rotated crop, an OT loss is proposed to minimize the distance in pixel space between each patch and its matched image patches. The overall loss can be written as follows:
(2) |
where means predictions for the masked patches of the background, and denotes all predicted patches of the rotated crop.
3.3 Reconstruction for Rotated Crop
After the scaling center crop operation, the scenes in all rotated crop patches are offset compared to the original image patches at the same positions. Directly calculating MSE between the masked patches and image patches not only results in biases but also overlooks the changes in angles and scenes on the visible patches. Inspired by [20], this paper treats the reconstruction for rotated crops as an OT problem, allowing each predicted patch to automatically match similar image patches for reconstruction.
Optimal transport. Supposing there are suppliers and demanders, where the -th supplier holds units of goods, and the -th demander needs units of goods. The transportation cost for transporting one unit of goods from the -th supplier to the -th demander is denoted as . OT aims to find a transportation plan, denoted as , that minimizes the total transportation costs, ensuring that all goods are transported from suppliers to demanders:
(3) | ||||
s.t. | ||||
OT for reconstruction. The context of reconstructing the rotated crop is described in Fig. 2(b), considering original image patches and predicted patches of the rotated crop, each image patch is treated as a supplier, holding units of pixel values (i.e., ), and each predicted patch as a demander, with units of channels (i.e., ) needing units of pixel values for reconstruction. The similarity between the each unit of pixel value in the image patch and the any unit of channel in the predicted patch represents the transportation cost . This is extended to matrix-wise MSE computation for GPU acceleration. Thus, the transportation cost from the -th image patch to the -th predicted patch is given by:
(4) |
where image patches closer in L2 distance to the predicted patch have higher similarity, tending to lower costs. A fast iterative algorithm, Sinkhorn-Knopp [14], is employed to calculate the transportation plan in Eq. 3. According to the solved , as shown in Fig. 2(b), the OT loss automatically allocates similar multiple image patches for -th predicted patch as the reconstruction targets, defined as follows:
(5) |
The proposed , during the pixel reconstruction and angle restoration, guides our model to perceive the angle variations of rotated crops. As the iterations progress, MA3E can effectively learn rotation-invariant visual representations. Supplementary material provides more details about .
4 Experiments
Following the evaluation protocol of [1, 25], MA3E is first pre-trained on MillionAID [38]. Then, only the encoder is fine-tuned on downstream tasks, including scene classification, rotated object detection, and semantic segmentation.
4.1 Experimental Setups
Methods | Backbone | PT | Ep. | FLOPs | GPU | NU45 | AID | UCM |
Data. | (G) | H. / Ep. | (2:8) | (5:5) | (5:5) | |||
Pre-training Methods for Natural Images | ||||||||
MoCo v3 [6] | ViT-B | IN1k | 300 | 17.5 | - | 80.69 | 85.74 | 81.43 |
MoCo v3 † [6] | ViT-B | MA | 300 | 17.5 | 1.9 | 92.40 | 93.99 | 97.70 |
DINO [2] | ViT-B | IN1k | 400 | 17.5 | - | 78.71 | 83.14 | 80.14 |
DINO † [2] | ViT-B | MA | 300 | 17.5 | 3.6 | 90.88 | 93.36 | 97.66 |
MAE [25] | ViT-B | IN1k | 1600 | 17.5 | - | 95.20 | 97.64 | 99.05 |
MAE † [25] | ViT-B | MA | 300 | 17.5 | 1.2 | 95.31 | 98.16 | 99.05 |
MAE † [25] | ViT-B | MA | 1600 | 17.5 | 1.2 | 95.40 | 98.36 | 99.44 |
SimMIM † [64] | ViT-B | MA | 400 | 17.5 | 2.7 | 95.54 | 98.19 | 99.06 |
LoMaR † [3] | ViT-B | MA | 300 | 17.5 | 1.9 | 95.47 | 98.11 | 98.91 |
MixMAE † [34] | Swin-B/W14 | MA | 300 | 16.3 | 1.8 | 95.45 | 98.22 | 99.04 |
Pre-training Methods for RS Images | ||||||||
SeCo [42] | RN-50 | S-2 | 200 | 4.1 | - | 92.91 | 95.99 | 97.81 |
CACo [41] | RN-50 | S-2 | 200 | 4.1 | - | 91.94 | 95.05 | 97.05 |
RingMo [50] | Swin-B | 2M | 200 | 15.6 | - | 95.67 | 98.34 | - |
CMID [44] | Swin-B | MA | 200 | 15.6 | - | 95.16 | 96.98 | 98.21 |
GFM [43] | Swin-B | G.I. | 100 | 15.6 | - | 96.06 | 97.09 | 99.14 |
MAE [54] | ViT-B+RVSA | MA | 1600 | 33.6 | - | 95.49 | 98.33 | 99.70 |
MAE [54] | ViTAE-B+RVSA | MA | 1600 | 26.3 | - | 95.69 | 98.48 | 99.56 |
SatMAE [11] | ViT-B | f-S | 200 | 17.5 | - | 76.04 | 83.84 | 81.05 |
SatMAE [11] | ViT-L | f-R | 800 | 61.3 | - | 93.78 | 98.70 | 97.14 |
ScaleMAE [46] | ViT-L | f-R | 800 | 61.3 | - | 88.54 | 97.42 | 93.28 |
MA3E | ViT-B | MA | 300 | 17.5 | 1.4 | 95.77 | 98.44 | 99.05 |
MA3E | ViT-B | MA | 1600 | 17.5 | 1.4 | 96.23 | 99.04 | 99.81 |
Methods | Backbone | Pre-training | Epoch | NU45 | AID | UCM |
Dataset | (2:8) | (5:5) | (5:5) | |||
Pre-training Methods for Natural Images | ||||||
MoCo v3 [6] | ViT-B | ImageNet-1k | 300 | 48.80 | 68.51 | 25.40 |
MoCo v3 † [6] | ViT-B | MillionAID | 300 | 61.45 | 78.72 | 38.34 |
DINO [2] | ViT-B | ImageNet-1k | 400 | 51.42 | 72.10 | 31.14 |
DINO † [2] | ViT-B | MillionAID | 300 | 63.67 | 78.51 | 40.04 |
MAE [25] | ViT-B | ImageNet-1k | 1600 | 66.09 | 74.60 | 49.81 |
MAE † [25] | ViT-B | MillionAID | 300 | 73.94 | 83.12 | 51.52 |
MAE † [25] | ViT-B | MillionAID | 1600 | 75.98 | 84.21 | 52.75 |
SimMIM † [64] | ViT-B | MillionAID | 400 | 74.86 | 83.19 | 51.48 |
LoMaR † [3] | ViT-B | MillionAID | 300 | 74.30 | 82.26 | 51.89 |
MixMAE † [34] | Swin-B/W14 | MillionAID | 300 | 73.95 | 81.53 | 50.63 |
Pre-training Methods for RS Images | ||||||
SeCo [42] | RN-50 | Sentinel-2 | 200 | 65.02 | 78.26 | 47.45 |
CACo [41] | RN-50 | Sentinel-2 | 200 | 63.24 | 77.81 | 40.53 |
CMID [44] | Swin-B | MillionAID | 200 | 65.63 | 79.05 | 47.43 |
GFM [43] | Swin-B | GeoPile+ImageNet-22k | 100 | 76.09 | 80.58 | 49.73 |
MAE [54] | ViT-B+RVSA | MillionAID | 1600 | 75.72 | 84.06 | 50.86 |
SatMAE [11] | ViT-B | fMoW-Sentinel | 200 | 20.60 | 33.72 | 19.14 |
SatMAE [11] | ViT-L | fMoW-RGB | 800 | 37.15 | 55.10 | 34.28 |
ScaleMAE [46] | ViT-L | fMoW-RGB | 800 | 33.03 | 48.46 | 28.19 |
MA3E | ViT-B | MillionAID | 300 | 74.61 | 84.21 | 52.24 |
MA3E | ViT-B | MillionAID | 1600 | 76.41 | 85.86 | 55.69 |
Unless otherwise stated, all experiments are implemented using PyTorch and conducted on a machine equipped with eight 24GB RTX 3090 GPUs. More experimental setups and datasets are detailed in the supplementary material.
Pre-training details. The testing set of MillionAID [38], consisting of 990,848 RS images, is used for pre-training. Each image are resized to pixels. The patch size is 16. For each rotated crop, the side length is set to 96, and the rotation range is . We randomly mask the rotated crop and background patches with a ratio of 75% respectively. MA3E employs a plain ViT-B [16] as the encoder and 8 ViT blocks with 512-D as the decoder. Except for the batchsize of 1024, all other pre-training configurations follow [25].
Scene classification. All fine-tuning and linear probing experiments are conducted on NWPU-RESISC45 [7] (NU45), AID [60], and UC Merced [65] (UCM) datasets. For NU45, 20% of images from each class are randomly sampled as the training set, and the remaining 80% are used for testing. For AID and UCM, these two ratios are both 50%. Fine-tuning is performed with a batchsize of 512 for 200 epochs, and linear probing is trained with a batchsize of 2048 for 100 epochs. We follow the other default fine-tuning and linear probing settings outlined in [25] and report the Top-1 accuracy on each testing set.
Rotated object detection. Experiments for detection are conducted on DOTA1.0 [59] and DIOR-R [8] using the Oriented R-CNN [63] detector. MA3E pre-trained models serves as the backbone of the detector and undergoes end-to-end fine-tuning. The detector uses the batchsize of 2 for DOTA1.0 and 4 for DIOR-R. We train for 12 epochs, with other hyper-parameters following default settings of the detector. Mean Average Precision (mAP) on each testing set is reported, where results on DOTA1.0 are obtained from the official evaluation server. Due to the limited GPU memory, above experiments are deployed on two GPUs and implemented by the OBBDetection and the ViTDet [32] codebases.
Semantic segmentation. Similarly, segmentation experiments are conducted using the UperNet [61] framework for end-to-end supervised fine-tuning on iSAID [57] and Potsdam. The UperNet is trained for iterations with a batchsize of 4, while other hyper-parameters remain at the default settings. Mean Intersection over Union (mIoU) on the iSAID validation set and mean F1 score (mF1) on the Potsdam testing set are reported. These experiments are implemented using the mmsegmentation [12] library and also run on two GPUs.
Methods | Backbone | PT | Ep. | DOTA1.0 | DIOR-R | iSAID | Potsdam | |
Data. | mAP | mIoU | mF1 | |||||
Pre-training Methods for Natural Images | ||||||||
MoCo v3 [6] | ViT-B | IN1k | 300 | 59.35 | 44.22 | 40.18 | 83.59 | |
MoCo v3 † [6] | ViT-B | MA | 300 | 71.46 | 59.41 | 58.72 | 90.13 | |
DINO [2] | ViT-B | IN1k | 400 | 73.53 | 62.67 | 50.40 | 86.29 | |
DINO † [2] | ViT-B | MA | 300 | 74.91 | 64.87 | 54.61 | 88.56 | |
MAE [25] | ViT-B | IN1k | 1600 | 76.04 | 64.84 | 61.08 | 90.14 | |
MAE † [25] | ViT-B | MA | 300 | 75.85 | 64.54 | 60.96 | 90.08 | |
MAE † [25] | ViT-B | MA | 1600 | 77.53 | 67.72 | 61.38 | 90.49 | |
SimMIM † [64] | ViT-B | MA | 400 | 76.17 | 65.24 | 60.92 | 90.20 | |
LoMaR † [3] | ViT-B | MA | 300 | 75.76 | 64.55 | 60.86 | 90.21 | |
MixMAE † [34] | Swin-B/W14 | MA | 300 | 75.87 | 64.67 | 60.64 | 90.13 | |
Pre-training Methods for RS Images | ||||||||
SeCo [42] | RN-50 | S-2 | 200 | 69.95 | 62.74 | 57.45 | 89.83 | |
CACo [41] | RN-50 | S-2 | 200 | 75.35 | 65.10 | 61.32 | 90.35 | |
CMID [44] | Swin-B | MA | 200 | 77.36 | - | - | - | |
CMID [44] | Swin-B | MA | 200 | 77.29 | 66.13 | 62.42 | 90.71 | |
GFM [43] | Swin-B | G.I. | 100 | 77.81 | 67.67 | 62.54 | 90.62 | |
MAE [54] | ViT-B+RVSA | MA | 1600 | 78.75 | 70.67 | 63.76 | 90.60 | |
MAE [54] | ViTAE-B+RVSA | MA | 1600 | 78.96 | 70.95 | 63.48 | 91.22 | |
SatMAE [11] | ViT-B | f-S | 200 | 68.54 | 48.55 | 53.55 | 86.43 | |
MA3E | ViT-B | MA | 300 | 77.93 | 68.41 | 62.74 | 90.67 | |
MA3E | ViT-B | MA | 1600 | 79.47 | 71.82 | 64.06 | 91.50 |
4.2 Main Results
MA3E is compared with eight state-of-the-art pre-training methods for RS images, including seasonal-contrasted SeCo [42], change-aware contrasted CACo [41], GFM [43] with continual pre-training, RingMo [50] adopting incomplete masking, CMID [44] combining contrastive learning and MIM, [54] using rotated varied-size window attention (RVSA) to replace the original global attention during downstream fine-tuning, sptaio-temporal encoded SatMAE [11], and scale-aware ScaleMAE [46]. However, these methods adopt different datasets and fine-tuning settings for downstream tasks. For fairness, we normalize the experimental setups and further compared with six pre-training methods for natural images, including the popular MoCo v3 [6], DINO [2], MAE [25], SimMIM [64], region-reconstructed LoMaR [3], and input-mixed MixMAE [34].
Scene classification. The fine-tuning and linear probing results on three datasets are shown in Table 1 and Table 2, repectively. MA3E pre-trained for 300 epochs achieves competitive results. Although the fine-tuning accuracy on UCM is lower than ViT+RVSA [54], MA3E requires only 52% of the latter’s FLOPs. The fine-tuning and linear probing results on three datasets continually improve as training progresses. MA3E pre-trained for 1600 epochs comprehensively leads, demonstrating that MA3E effectively learns the discriminative rotation-invariant representations of RS objects. In addition, the training time per epoch on a single GPU is calculated. Compared to MAE [25], we achieve a significant improvement in accuracy with only about 0.2 hours of extra training time.
MAE | SCC | AE | Mask. | OT | ft | det | seg |
---|---|---|---|---|---|---|---|
✓ | 95.31 | 75.85 | 60.96 | ||||
✓ | ✓ | 95.43 | 76.12 | 61.24 | |||
✓ | ✓ | ✓ | 95.47 | 76.41 | 61.86 | ||
✓ | ✓ | ✓ | 95.36 | 76.46 | 61.88 | ||
✓ | ✓ | ✓ | 95.06 | 77.23 | 62.17 | ||
✓ | ✓ | ✓ | ✓ | 95.53 | 76.70 | 61.93 | |
✓ | ✓ | ✓ | ✓ | ✓ | 95.77 | 77.93 | 62.74 |
Num. | ft | det | seg | |
---|---|---|---|---|
32 | 1 | 95.21 | 76.13 | 61.89 |
32 | 4 | 94.16 | 75.59 | 62.14 |
64 | 1 | 95.36 | 77.53 | 62.51 |
64 | 2 | 94.90 | 76.21 | 62.43 |
80 | 1 | 95.40 | 77.38 | 62.83 |
96 | 1 | 95.77 | 77.93 | 62.74 |
128 | 1 | 95.74 | 76.92 | 62.01 |
Rotated object detection. Table 3 presents the fine-tuning results on DOTA1.0 and DIOR-R for different methods, where MA3E, using a simple backbone, obtains superior detection performance. The version pre-trained for 300 epochs outperforms other methods with a similar number of epochs. MA3E pre-trained for 1600 epochs surpasses all methods. Compared to [54] with ViTAE+RVSA, the mAP on DOTA1.0 and DIOR-R has increased by 0.51 and 0.87, respectively. This significant improvement in detection performance demonstrates the effectiveness of MA3E in angle perception during pre-training.
Semantic segmentation. The fine-tunning results on iSAID and Potsdam are also reported in Table 3. With fewer pre-training epochs, MA3E demonstrates competitive performance, achieving mF1 only 0.04 lower than CMID [44] on Potsdam. When pre-trained for 1600 epochs, MA3E again achieves the best results, outperforming the second-best [54] by 0.3 mIoU and 0.28 mF1 on iSAID and Potsdam, respectively. It is indicated that the significance of the rotation-invariant representations learned by our model in semantic segmentation.
4.3 Ablation Study
Position | ft | det | seg |
---|---|---|---|
fixed | 95.48 | 76.44 | 62.10 |
random | 95.77 | 77.93 | 62.74 |
selective search [53] | 95.96 | 78.08 | 62.68 |
Operation | ft | det | seg |
---|---|---|---|
random rotation | 93.82 | 76.14 | 61.23 |
scaling center crop | 95.77 | 77.93 | 62.74 |
Range | ft | det | seg |
---|---|---|---|
95.78 | 77.68 | 62.49 | |
95.77 | 77.93 | 62.74 | |
95.32 | 77.22 | 62.55 | |
94.89 | 76.45 | 61.90 |
Strategy | ft | det | seg |
---|---|---|---|
random masking | 95.77 | 77.93 | 62.74 |
block-wise masking | 94.98 | 77.69 | 62.39 |
uniform sampling [31] | 95.33 | 77.57 | 62.58 |
In this section, a series of ablation studies are conducted to analyze how each key design is available and demonstrate the effectiveness of each component in MA3E. By default, MA3E is pre-trained for 300 epochs. We report the Top-1 accuracy on NU45 [7] after fine-tuning, mAP on DOTA1.0 [59], and mIoU on iSAID [57]. The results of MA3E with default settings are marked in green. Our supplementary material provides additional ablation results.
Each component. The ablation results on different components are shown in Table 5. Each key design of MA3E improves the performance of the baseline MAE across the three RS tasks, and the combination of all components in MA3E yields the best results. This demonstrates the effectiveness of our proposal.
Side length and crop numbers. We study the effects of different side lengths for the rotated crop and the number rotated crops in each image, as shown in Table 5. As increases, MA3E’s performance gradually improves until . However, for detection and segmentation tasks, there is a sudden decrease in the performance when . Furthermore, an increase in the number of rotated crops also negatively impacts performance. Excessive large and a more number of crops make angle restoration challenging, and this decline is more obvious in detection where angles are crucial.
Position of rotated crops. We ablate three schemes for selecting the position of the rotated crop: fixed positions at the center of the images, random positions, and positions determined by the selective search [53]. The corresponding results are shown in Table 7. The selective search algorithm generates candidate bounding boxes for potential objects in an unsupervised manner. Although determining the position of the rotated crop through this method results in an improvement of 0.19 in classification accuracy and 0.15 mAP in detection compared to randomly choosing positions, the limited performance gain comes at the cost of at least 10% extra training time per epoch. Therefore, it is not employed in the default settings.
Method to create rotated crops. We compare MA3E with a model using the simple random rotation operation to generate rotated crops with . The results in Table 7 demonstrate that rotated crops constructed using the proposed scaling center crop operation significantly enhance the performance in all tasks. Note that our method may still incur minor losses in edge scenes. But this corresponds to deliberately increasing the difficulty of reconstructing the complete original image, thereby improving the quality of learned rotation-invariant representations.
Rotation ranges. Table 9 shows the results for different rotation ranges of rotated crops. An reasonable range maximizes the ability of to restore the angle variations introduced on rotated crops, promoting MA3E to learn rotation-invariant visual representations and achieving good performance.
Masking strategy. Table 9 report the effect of different masking strategies on MA3E. The random masking strategy [25] performs better. Block-wise masking [1] increases the reconstruction difficulty. Meanwhile, uniform sampling [31] with masking in adjacent four patches makes the reconstruction of rotated crops easier but leads to lower quality representations learned by the model.
4.4 Visualization

Fig. 4 quantitatively visualizes the reconstruction performance of MA3E on RS images. Example images are randomly sampled from the training set of MillionAID [38], which contains images. We resize each image to (196 patches, ) and set the rotated crop of (36 patches) for reconstruction. The rotated crop and background are masked at a 75% ratio, corresponding to 9 and 40 visible patches, respectively. It can be seen that MA3E effectively restores the preset angle variations of the rotated crops during reconstructing original pixels. In addition, some uneven color patches may appear in the reconstructed rotated crop, such as the water surface and playground in Fig. 4. This phenomenon is attributed to the proposed OT loss, which calculates the mean square error between each predicted patch and multiple target patches. The supplementary material shows more visualizations.
5 Conclusion and Discussion
This paper proposes Masked Angle-Aware Autoencoder (MA3E) for self-supervised representation learning on RS images. The scaling center crop operation is designed to construct the rotated crop within each original image, introducing the explicit angle variation. MA3E takes the created composite image as input, with the goal of simultaneously achieving original pixel reconstruction and angle restoration. The reconstruction for the rotated crop is treated as an optimal transport problem, and we propose an OT loss to automatically assign similar original image patches for each rotated crop patch. Finally, MA3E can effectively perceive angles and learn rotation-invariant representations, achieving competitive performance in various downstream tasks. We hope MA3E can contribute to the advancement of foundational models in RS research.
Limitation. Despite MA3E demonstrates the potential of angle awareness, in many RS scenarios, the angles of only man-made objects require more emphasis. For extensive land cover, the model may not benefit significantly from angle information. In future work, we will consider the scale that exists in any RS scene and further explore the combination of angles for man-made objects and scale for land cover.
Acknowledgements
This work is supported in part by the National Natural Science Foundation of China under Grant 62171347, 62101405, 62371373, 62271377; 111 Project; the Postdoctoral Fellowship Program of China Postdoctoral Science Foundation under Grant GZC20232036, GZC20232032; the Shaanxi Province postdoctoral research project under Grant 2023BSHEDZZ168, 2023BSHYDZZ96; the Fundamental Research Funds for the Central Universities.
References
- [1] Bao, H., Dong, L., Piao, S., Wei, F.: Beit: Bert pre-training of image transformers. arXiv preprint arXiv:2106.08254 (2021)
- [2] Caron, M., Touvron, H., Misra, I., Jégou, H., Mairal, J., Bojanowski, P., Joulin, A.: Emerging properties in self-supervised vision transformers. In: Proceedings of the IEEE/CVF international conference on computer vision. pp. 9650–9660 (2021)
- [3] Chen, J., Hu, M., Li, B., Elhoseiny, M.: Efficient self-supervised vision pretraining with local masked reconstruction. arXiv preprint arXiv:2206.00790 (2022)
- [4] Chen, M., Radford, A., Child, R., Wu, J., Jun, H., Luan, D., Sutskever, I.: Generative pretraining from pixels. In: International conference on machine learning. pp. 1691–1703. PMLR (2020)
- [5] Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A simple framework for contrastive learning of visual representations. In: International conference on machine learning. pp. 1597–1607. PMLR (2020)
- [6] Chen, X., Xie, S., He, K.: An empirical study of training self-supervised vision transformers. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 9620–9629 (2021)
- [7] Cheng, G., Han, J., Lu, X.: Remote sensing image scene classification: Benchmark and state of the art. Proceedings of the IEEE 105(10), 1865–1883 (2017)
- [8] Cheng, G., Wang, J., Li, K., Xie, X., Lang, C., Yao, Y., Han, J.: Anchor-free oriented proposal generator for object detection. IEEE Transactions on Geoscience and Remote Sensing 60, 1–11 (2022)
- [9] Christie, G., Fendley, N., Wilson, J., Mukherjee, R.: Functional map of the world. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 6172–6180 (2018)
- [10] Clark, K., Luong, M.T., Le, Q.V., Manning, C.D.: Electra: Pre-training text encoders as discriminators rather than generators. arXiv preprint arXiv:2003.10555 (2020)
- [11] Cong, Y., Khanna, S., Meng, C., Liu, P., Rozi, E., He, Y., Burke, M., Lobell, D., Ermon, S.: Satmae: Pre-training transformers for temporal and multi-spectral satellite imagery. Advances in Neural Information Processing Systems 35, 197–211 (2022)
- [12] Contributors, M.: Mmsegmentation: Openmmlab semantic segmentation toolbox and benchmark (2020)
- [13] Cubuk, E.D., Zoph, B., Shlens, J., Le, Q.V.: Randaugment: Practical automated data augmentation with a reduced search space. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops. pp. 702–703 (2020)
- [14] Cuturi, M.: Sinkhorn distances: Lightspeed computation of optimal transport. Advances in neural information processing systems 26 (2013)
- [15] Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: A large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition. pp. 248–255. Ieee (2009)
- [16] Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et al.: An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)
- [17] Drusch, M., Del Bello, U., Carlier, S., Colin, O., Fernandez, V., Gascon, F., Hoersch, B., Isola, C., Laberinti, P., Martimort, P., et al.: Sentinel-2: Esa’s optical high-resolution mission for gmes operational services. Remote sensing of Environment 120, 25–36 (2012)
- [18] Gao, P., Ma, T., Li, H., Lin, Z., Dai, J., Qiao, Y.: Convmae: Masked convolution meets masked autoencoders. arXiv preprint arXiv:2205.03892 (2022)
- [19] Gao, P., Zhang, R., Fang, R., Lin, Z., Li, H., Li, H., Yu, Q.: Mimic before reconstruct: Enhancing masked autoencoders with feature mimicking. arXiv preprint arXiv:2303.05475 (2023)
- [20] Ge, Z., Liu, S., Li, Z., Yoshie, O., Sun, J.: Ota: Optimal transport assignment for object detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 303–312 (2021)
- [21] Glorot, X., Bengio, Y.: Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of the thirteenth international conference on artificial intelligence and statistics. pp. 249–256. JMLR Workshop and Conference Proceedings (2010)
- [22] Goyal, P., Dollár, P., Girshick, R., Noordhuis, P., Wesolowski, L., Kyrola, A., Tulloch, A., Jia, Y., He, K.: Accurate, large minibatch sgd: Training imagenet in 1 hour. arXiv preprint arXiv:1706.02677 (2017)
- [23] Grill, J.B., Strub, F., Altché, F., Tallec, C., Richemond, P., Buchatskaya, E., Doersch, C., Avila Pires, B., Guo, Z., Gheshlaghi Azar, M., et al.: Bootstrap your own latent-a new approach to self-supervised learning. Advances in neural information processing systems 33, 21271–21284 (2020)
- [24] Gu, Y., Wang, Y., Li, Y.: A survey on deep learning-driven remote sensing image scene understanding: Scene classification, scene retrieval and scene-guided object detection. Applied Sciences 9(10), 2110 (2019)
- [25] He, K., Chen, X., Xie, S., Li, Y., Dollár, P., Girshick, R.: Masked autoencoders are scalable vision learners. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 16000–16009 (2022)
- [26] He, K., Fan, H., Wu, Y., Xie, S., Girshick, R.: Momentum contrast for unsupervised visual representation learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 9729–9738 (2020)
- [27] Hou, Z., Sun, F., Chen, Y.K., Xie, Y., Kung, S.Y.: Milan: Masked image pretraining on language assisted representation. arXiv preprint arXiv:2208.06049 (2022)
- [28] Huang, G., Sun, Y., Liu, Z., Sedra, D., Weinberger, K.Q.: Deep networks with stochastic depth. In: Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11–14, 2016, Proceedings, Part IV 14. pp. 646–661. Springer (2016)
- [29] Huang, L., You, S., Zheng, M., Wang, F., Qian, C., Yamasaki, T.: Green hierarchical vision transformer for masked image modeling. Advances in Neural Information Processing Systems 35, 19997–20010 (2022)
- [30] Ippoliti, C., Candeloro, L., Gilbert, M., Goffredo, M., Mancini, G., Curci, G., Falasca, S., Tora, S., Di Lorenzo, A., Quaglia, M., et al.: Defining ecological regions in italy based on a multivariate clustering approach: A first step towards a targeted vector borne disease surveillance. PloS one 14(7), e0219072 (2019)
- [31] Li, X., Wang, W., Yang, L., Yang, J.: Uniform masking: Enabling mae pre-training for pyramid-based vision transformers with locality. arXiv preprint arXiv:2205.10063 (2022)
- [32] Li, Y., Mao, H., Girshick, R., He, K.: Exploring plain vision transformer backbones for object detection. In: European Conference on Computer Vision. pp. 280–296. Springer (2022)
- [33] Lin, T.Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Dollár, P., Zitnick, C.L.: Microsoft coco: Common objects in context. In: Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13. pp. 740–755. Springer (2014)
- [34] Liu, J., Huang, X., Zheng, J., Liu, Y., Li, H.: Mixmae: Mixed and masked autoencoder for efficient pretraining of hierarchical vision transformers. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 6252–6261 (2023)
- [35] Liu, Y., Zhang, S., Chen, J., Chen, K., Lin, D.: Pixmim: Rethinking pixel reconstruction in masked image modeling. arXiv preprint arXiv:2303.02416 (2023)
- [36] Liu, Y., Zhang, S., Chen, J., Yu, Z., Chen, K., Lin, D.: Improving pixel-based mim by reducing wasted modeling capability. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 5361–5372 (2023)
- [37] Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., Guo, B.: Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE/CVF international conference on computer vision. pp. 10012–10022 (2021)
- [38] Long, Y., Xia, G.S., Li, S., Yang, W., Yang, M.Y., Zhu, X.X., Zhang, L., Li, D.: On creating benchmark dataset for aerial image interpretation: Reviews, guidances, and million-aid. IEEE Journal of selected topics in applied earth observations and remote sensing 14, 4205–4230 (2021)
- [39] Loshchilov, I., Hutter, F.: Sgdr: Stochastic gradient descent with warm restarts. arXiv preprint arXiv:1608.03983 (2016)
- [40] Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)
- [41] Mall, U., Hariharan, B., Bala, K.: Change-aware sampling and contrastive learning for satellite images. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 5261–5270 (2023)
- [42] Manas, O., Lacoste, A., Giró-i Nieto, X., Vazquez, D., Rodriguez, P.: Seasonal contrast: Unsupervised pre-training from uncurated remote sensing data. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 9414–9423 (2021)
- [43] Mendieta, M., Han, B., Shi, X., Zhu, Y., Chen, C.: Towards geospatial foundation models via continual pretraining. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 16806–16816 (2023)
- [44] Muhtar, D., Zhang, X., Xiao, P., Li, Z., Gu, F.: Cmid: A unified self-supervised learning framework for remote sensing image understanding. IEEE Transactions on Geoscience and Remote Sensing (2023)
- [45] Mulla, D.J.: Twenty five years of remote sensing in precision agriculture: Key advances and remaining knowledge gaps. Biosystems engineering 114(4), 358–371 (2013)
- [46] Reed, C.J., Gupta, R., Li, S., Brockman, S., Funk, C., Clipp, B., Keutzer, K., Candido, S., Uyttendaele, M., Darrell, T.: Scale-mae: A scale-aware masked autoencoder for multiscale geospatial representation learning. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 4088–4099 (2023)
- [47] Rolfe, J.T.: Discrete variational autoencoders. arXiv preprint arXiv:1609.02200 (2016)
- [48] Rolnick, D., Donti, P.L., Kaack, L.H., Kochanski, K., Lacoste, A., Sankaran, K., Ross, A.S., Milojevic-Dupont, N., Jaques, N., Waldman-Brown, A., et al.: Tackling climate change with machine learning. ACM Computing Surveys (CSUR) 55(2), 1–96 (2022)
- [49] Schumann, G.J., Brakenridge, G.R., Kettner, A.J., Kashif, R., Niebuhr, E.: Assisting flood disaster response with earth observation data and products: A critical assessment. Remote sensing 10(8), 1230 (2018)
- [50] Sun, X., Wang, P., Lu, W., Zhu, Z., Lu, X., He, Q., Li, J., Rong, X., Yang, Z., Chang, H., et al.: Ringmo: A remote sensing foundation model with masked image modeling. IEEE Transactions on Geoscience and Remote Sensing (2022)
- [51] Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 2818–2826 (2016)
- [52] Tian, Y., Xie, L., Fang, J., Shi, M., Peng, J., Zhang, X., Jiao, J., Tian, Q., Ye, Q.: Beyond masking: Demystifying token-based pre-training for vision transformers. arXiv preprint arXiv:2203.14313 (2022)
- [53] Uijlings, J.R., Van De Sande, K.E., Gevers, T., Smeulders, A.W.: Selective search for object recognition. International journal of computer vision 104, 154–171 (2013)
- [54] Wang, D., Zhang, Q., Xu, Y., Zhang, J., Du, B., Tao, D., Zhang, L.: Advancing plain vision transformer toward remote sensing foundation model. IEEE Transactions on Geoscience and Remote Sensing 61, 1–15 (2022)
- [55] Wang, H., Tang, Y., Wang, Y., Guo, J., Deng, Z.H., Han, K.: Masked image modeling with local multi-scale reconstruction. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 2122–2131 (2023)
- [56] Wang, W., Xie, E., Li, X., Fan, D.P., Song, K., Liang, D., Lu, T., Luo, P., Shao, L.: Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In: Proceedings of the IEEE/CVF international conference on computer vision. pp. 568–578 (2021)
- [57] Waqas Zamir, S., Arora, A., Gupta, A., Khan, S., Sun, G., Shahbaz Khan, F., Zhu, F., Shao, L., Xia, G.S., Bai, X.: isaid: A large-scale dataset for instance segmentation in aerial images. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops. pp. 28–37 (2019)
- [58] Wei, C., Fan, H., Xie, S., Wu, C.Y., Yuille, A., Feichtenhofer, C.: Masked feature prediction for self-supervised visual pre-training. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 14668–14678 (2022)
- [59] Xia, G.S., Bai, X., Ding, J., Zhu, Z., Belongie, S., Luo, J., Datcu, M., Pelillo, M., Zhang, L.: Dota: A large-scale dataset for object detection in aerial images. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 3974–3983 (2018)
- [60] Xia, G.S., Hu, J., Hu, F., Shi, B., Bai, X., Zhong, Y., Zhang, L., Lu, X.: Aid: A benchmark data set for performance evaluation of aerial scene classification. IEEE Transactions on Geoscience and Remote Sensing 55(7), 3965–3981 (2017)
- [61] Xiao, T., Liu, Y., Zhou, B., Jiang, Y., Sun, J.: Unified perceptual parsing for scene understanding. In: Proceedings of the European conference on computer vision (ECCV). pp. 418–434 (2018)
- [62] Xie, J., Li, W., Zhan, X., Liu, Z., Ong, Y.S., Loy, C.C.: Masked frequency modeling for self-supervised visual pre-training. arXiv preprint arXiv:2206.07706 (2022)
- [63] Xie, X., Cheng, G., Wang, J., Yao, X., Han, J.: Oriented r-cnn for object detection. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 3520–3529 (2021)
- [64] Xie, Z., Zhang, Z., Cao, Y., Lin, Y., Bao, J., Yao, Z., Dai, Q., Hu, H.: Simmim: A simple framework for masked image modeling. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 9653–9663 (2022)
- [65] Yang, Y., Newsam, S.: Bag-of-visual-words and spatial extensions for land-use classification. In: Proceedings of the 18th SIGSPATIAL international conference on advances in geographic information systems. pp. 270–279 (2010)
- [66] You, Y., Gitman, I., Ginsburg, B.: Large batch training of convolutional networks. arXiv preprint arXiv:1708.03888 (2017)
- [67] Yun, S., Han, D., Oh, S.J., Chun, S., Choe, J., Yoo, Y.: Cutmix: Regularization strategy to train strong classifiers with localizable features. In: Proceedings of the IEEE/CVF international conference on computer vision. pp. 6023–6032 (2019)
- [68] Zhang, H., Cisse, M., Dauphin, Y.N., Lopez-Paz, D.: mixup: Beyond empirical risk minimization. arXiv preprint arXiv:1710.09412 (2017)
- [69] Zhang, J., Liu, J., Pan, B., Shi, Z.: Domain adaptation based on correlation subspace dynamic distribution alignment for remote sensing image scene classification. IEEE Transactions on Geoscience and Remote Sensing 58(11), 7920–7930 (2020)
- [70] Zhang, Q., Xu, Y., Zhang, J., Tao, D.: Vitaev2: Vision transformer advanced by exploring inductive bias for image recognition and beyond. International Journal of Computer Vision pp. 1–22 (2023)
- [71] Zhou, B., Zhao, H., Puig, X., Xiao, T., Fidler, S., Barriuso, A., Torralba, A.: Semantic understanding of scenes through the ade20k dataset. International Journal of Computer Vision 127, 302–321 (2019)
Masked Angle-Aware Autoencoder for Remote Sensing Images
Zhihao Li Biao Hou† Siteng Ma Zitong Wu Xianpeng Guo Bo Ren Licheng Jiao
Appendix 0.A Full Implementation Details
0.A.1 Experimental Setups
Pre-training. The pre-training settings of MA3E is in Table. 10. All ViT blocks are initialized by xavier_uniform [21]. We use the batchisze of 1024 and the linear lr scaling rule [22]: . Simple data augmentations such as random cropping and horizontal flipping are also applied before creating rotated crops.
configs | pre-training | linear probing |
---|---|---|
optimizer | AdamW [40] | LARS [66] |
base learning rate | 1.5e-4 | 0.1 |
weight decay | 0.05 | 0 |
optimizer momentum | [4] | 0.9 |
learning rate schedule | cosine decay [39] | cosine decay |
warmup epochs [22] | 40 | 10 |
augmentation | ScalingCenterCrop | RandomResizedCrop |
Fine-tuning and linear probing. We fine-tune for 200 epochs with a batchsize of 512 and linear probe for 100 epochs with a batchsize of 2048. Other default linear probing and fine-tuning settings are respectively shown in Table. 10 and Table. 11, which also follow that of MAE [25].
configs | value |
---|---|
optimizer | AdamW |
base learning rate | 1e-3 |
weight decay | 0.05 |
optimizer momentum | |
layer-wise lr decay [1, 10] | 0.75 |
learning rate schedule | cosine decay |
warmup epochs | 5 |
augmentation | RandAug (9, 0.5) [13] |
label smoothing [51] | 0.1 |
mixup [68] | 0.8 |
cutmix [67] | 1.0 |
drop path [28] | 0.1 |
Fine-tuning on DOTA1.0 and DIOR-R. The fine-tuning details for rotated object detection is shown in Table. 13. We adopt a multi-step scheduler to adjust the learning rate, which is reduced by at the and epoch.
Fine-tuning on iSAID and Potsdam. The implementation details for fine-tuning on semantic segmentation datatsets is in Table. 13. The learning rate schedule adopts the polynomial decay policy with a power of 1.0 and min_lr of 0, following [54].
configs | value |
---|---|
optimizer | AdamW |
base learning rate | 1e-4 |
weight decay | 0.05 |
optimizer momentum | |
learning rate schedule | multi-step scheduler |
drop path | 0.15 |
configs | value |
---|---|
optimizer | AdamW |
base learning rate | 6e-5 |
weight decay | 0.05 |
optimizer momentum | |
layer-wise lr decay | 0.9 |
learning rate schedule | polynomial scheduler |
warmup iters | 1500 |
drop path | 0.1 |
0.A.2 Dataset Preparations
MillionAID [38] is a large-scale RS scene dataset containing 1,000,848 RGB images collected from Google Earth. The training set consists of 10,000 images categorized into 51 classes, while the testing set includes the remaining 990,848 images without labels. These images are captured by various sensors and therefore have different resolutions, ranging from to pixels.
NWPU-RESISC45 [7] is a common RS image benchmark collected by Northwestern Polytechnical University from Google Earth. It contains 31,500 images in RGB color space, which are equally divided into 45 classes, each with 700 images of pixels.
AID [60] has images from different countries on Google Earth. These images are extracted at different times and seasons under different imaging conditions. The dataset contains 10,000 images with pixels in 30 classes.
UC Merced [65] contains 21 land-use classes, and each category has 100 images with the size of pixels. There are a total of 2,100 RGB images from the United States Geological Survey (USGS) National Map.
DOTA1.0 [59] is a large-scale rotated object detection dataset. It contains 2,806 images ranging from to pixels and has 188,282 instances with rotated bounding box annotations belonging to 15 object classes. The training, validation, and testing sets have 1,411, 458, and 937 images, respectively. Each image is cropped to patches with a stride of 824 and both training and validation sets are used for training. Note that the testing set does not have published labels, evaluation metrics are obtained by submitting predictions on the testing set to the official evaluation server.
DIOR-R [8] is a rotated object detection dataset consisting of 20 classes. It comprises 23,463 images with a total of 192,518 instances. The training set and the testing set consist of 11,725 images with 68,073 instances and 11,738 images with 124,445 instances, respectively. All images are cropped to pixels, with resolutions ranging from 0.5 to 30 m.
iSAID [57] is a large-scale instance segmentation dataset. Note that this dataset and DOTA1.0 share the same scenes, the difference is that iSAID is labeled with a semantic mask containing one background and 15 foregrounds class. It also consists of 2,806 high-resolution images with pixel dimension ranging from to . We crop all images to patches with a stride of 512 and use only the validation set for evaluation since the testing set is unavailable.
Potsdam is released by ISPRS Commission WG II/4. It contains 38 images with an average size of pixels. These images cover a 3.42 km2 area of Potsdam city and include six scenes, i.e., Impervious surface, Building, Low vegetation, Tree, Car, and Clutter. The training and testing sets have 24 and 14 images, respectively. Each image is cropped to patches with a stride of 384. We exclude the clutter class from the dataset when calculating evaluation metrics.

ft | det | seg | ||
---|---|---|---|---|
67% | 67% | 95.72 | 77.56 | 62.50 |
67% | 75% | 95.70 | 77.81 | 62.66 |
75% | 75% | 95.77 | 77.93 | 62.74 |
80% | 75% | 95.62 | 77.65 | 62.48 |
80% | 80% | 95.41 | 77.31 | 62.28 |
Type | ft | det | seg |
---|---|---|---|
masked | 94.89 | 77.11 | 62.33 |
all | 95.77 | 77.93 | 62.74 |
Methods | Backbone | Dataset | Epoch | GPU H. / Ep. | ft | det | seg |
---|---|---|---|---|---|---|---|
MAE [25] | ViT-S | MillionAID | 300 | 0.6 | 93.54 | 72.01 | 57.86 |
MA3E | ViT-S | MillionAID | 300 | 0.7 | 93.89 | 74.23 | 58.46 |
Appendix 0.B Additional Ablation Results
We conduct additional ablation studies, maintaining the same experimental settings as described in the main paper.

Masking ratio. The results of randomly masking the rotated crop and the background with different ratios are in Table 15. [25] indicates that the optimal masking ratio is 75%, so we explore suitable ratios around this value. It can be seen that our model achieves the best performance when the masking ratios of both the rotated crop and the background are 75%. In addition, we visualize the reconstruction performance at different masking ratios in Fig. 5.
Reconstruction for rotated crops. Table 15 presents the ablation results of MA3E using to reconstruct only masked patches or reconstruct all patches within the rotated crop. The performance is better when reconstructing all patches because it also takes into account offset angles and scenes in the visible patches from the rotated crop.
The effect of pre-training epochs. Fig. 6 illustrates the performance of MA3E under different pre-training epochs. The performance on three tasks improves gradually with the increase in epochs, and MA3E pre-trained for 1600 epochs has not reached saturation. It can be observed that longer training times may still have the potential for performance improvement, especially in rotated object detection.
Scalability. Table 16 shows various downstream results of MA3E and MAE [25] when using ViT-S as the encoder. With only an extra 0.1 hours of training time per epoch, MA3E achieves better results than MAE, showcasing its good scalability. Due to limited computational resources, we do not experiment with larger backbones such as ViT-L/H.
Unified Pre-training Dataset. Considering the typical MIM methods for RS images, SatMAE [11] and ScaleMAE [46], use fMoW [9] for pre-training, we follow their default settings and pre-train on MillionAID for a fair comparison. Note that MillionAID lacks multi-spectral bands, so we pre-train SatMAE according to it facing fMoW-RGB. In Table 17, MA3E achieves excellent downstream results with relative shorter training time per epoch.
Appendix 0.C Reconsidering OT

During the reconstruction of rotated crops, we already define original image patches as suppliers and predicted rotated crop patches as demanders, with the L2 similarity between patch pairs serving as the transportation cost. The transportation plan of this OT problem can be solved via the Sinkhorn-Knopp Iteration [14], which transforms the complex marginal linear programming problem into a solution process over a smooth feasible domain by introducing an entropic regularization term. Note that this classic algorithm is textbook knowledge and not a contribution of this paper. For further details, please refer to prior works [14, 20].
MA3E leverages to allocate similar original image patches as reconstruction targets for each predicted patch. Fig. 7 displays some heatmaps of the sovled transportation plan. Essentially, can be seen as the weight used in computing the mean squared error between the -th target patch and -th predicted patch in , where higher weights are assigned to more similar target-prediction pairs. When reconstructing the -th predicted patch, the model computes a weighted sum of MSE between it and multiple target patches (-th column in the heatmap of Fig. 7). It’s evident that each predicted patch matches similar target patches distributed across multiple positions rather than just at the same location. This demonstrates the effectiveness of MA3E in reconstructing rotated crops by solving the OT problem.
Appendix 0.D More Visualizations
In this section, we show more MA3E reconstructed images in Fig. 8. Rotated crops are marked with red bounding boxes. For the downstream tasks, parts of semantic segmentation results and rotated object detection results are further presented in Fig. 9 and Fig. 10, respectively.


