Promptable Representation Distribution Learning and Data Augmentation for Gigapixel Histopathology WSI Analysis
Abstract
Gigapixel image analysis, particularly for whole slide images (WSIs), often relies on multiple instance learning (MIL). Under the paradigm of MIL, patch image representations are extracted and then fixed during the training of the MIL classifiers for efficiency consideration. However, the invariance of representations makes it difficult to perform data augmentation for WSI-level model training, which significantly limits the performance of the downstream WSI analysis. The current data augmentation methods for gigapixel images either introduce additional computational costs or result in a loss of semantic information, which is hard to meet the requirements for efficiency and stability needed for WSI model training. In this paper, we propose a Promptable Representation Distribution Learning framework (PRDL) for both patch-level representation learning and WSI-level data augmentation. Meanwhile, we explore the use of prompts to guide data augmentation in feature space, which achieves promptable data augmentation for training robust WSI-level models. The experimental results have demonstrated that the proposed method stably outperforms state-of-the-art methods.
Code — https://github.com/lazytkm/PRDL
Introduction
Histopathology whole slide image (WSI) classification (Lu et al. 2021; Shao et al. 2021; Yang et al. 2022; Zheng et al. 2023) is increasingly popular in computer-aided pathological diagnosis, presenting unique challenges to the field of computer vision (Chen et al. 2022; Nakhli et al. 2023; Zhang et al. 2022b). Unlike conventional natural images, WSIs have massive image resolutions, often reaching up to billions of pixels. To tackle the gigapixel problem, a variety of methods within this domain employ multiple instance learning (MIL) frameworks (Ilse, Tomczak, and Welling 2018; Li, Li, and Eliceiri 2021; Campanella et al. 2019; Zhang et al. 2022a) to address the specific needs of WSI analysis.
Within the MIL framework, WSI analysis is usually divided into three phases: 1) Divide a WSI into patches; 2) Extract features for these patches; and 3) Aggregate these features to make a prediction for the entire WSI. Data augmentation serves as an effective expanding data strategy for training deep models (Chen et al. 2020a). It is also important in the domain of histopathology WSI analysis. As shown in Figure 1(a), if we draw a parallel to the process used for natural images, data augmentation for WSI should ideally be performed continuously after the WSI patching stage and before the patch-level feature extraction stage. However, for efficiency, the first two stages are generally conducted only once during the entire training process (Lu et al. 2021; Shao et al. 2021). This leads to the inapplicability of traditional image augmentation techniques, thereby inspiring the shift towards performing data augmentations directly in feature space.


To achieve data augmentation for WSI, several studies have proposed using generative methods (Dai et al. 2024; Zaffar et al. 2022) or Mixup techniques (Yang et al. 2022; Chen and Lu 2023; Gadermayr et al. 2023) to create data augmentations in feature space, as illustrated in Figure 1(b). However, these methods either introduce additional computational costs due to the need for training another parameterized model or cause a loss of semantic information. Additionally, the augmented results generated by these methods often lack control, in contrast to data augmentations applied directly in image space where changes are more visually intuitive and easier to manage. Currently, there lacks data augmentation methods to meet the unique demands of gigapixel image analysis that are not only computationally efficient but also preserve control.
In this paper, we introduce a novel approach named promptable representation distribution learning (PRDL) to address the challenges associated with patch-level representation learning and WSI-level data augmentation in histopathological WSI analysis. Within this framework, a representation distribution estimator is designed and trained during self-supervised representation learning. As shown in Figure 1(c), this estimator is capable of predicting a distribution of potential representation augmentations for each patch. After patch-level feature extraction, each patch within a WSI is represented as an individual distribution rather than a static point in feature space. Additionally, the representation distribution of each dimension is confined within a range by specific augmentation prompts, to simulate different augmentation operations that are used in image space. Finally, we implement a non-parameterized representation augmentation process through online sampling patch representations from these prompted distributions, to efficiently achieve the data augmentation for WSI-level model training. The proposed method was evaluated on a lung dataset with 754 WSIs and two public lung datasets with 696 WSIs and 3064 WSIs. The experimental results have demonstrated that the proposed method stably outperforms state-of-the-art methods. The main contributions of this paper can be summarized as follows:
(1) We proposed promptable representation distribution learning (PRDL), a novel representation learning framework with prompted representation distribution estimation for WSI classification. A promptable distribution estimator is designed to incorporate representation augmentation into representation learning. Compared with the traditional image-level data augmentation, the proposed method can provide more expansive augmentation. This significantly improves the discrimination of the patch representations and thereby enhances the performance of the subsequent WSI analysis model.
(2) We designed a promptable representation sampling (PRS) module based on PRDL. Utilizing PRS, we successfully facilitated the interchange between data augmentation and patch encoding processes, and achieved promptable and flexible data augmentation in the feature space for gigapixel histopathology image analysis. Furthermore, we leveraged the augmentation prompts in image space to guide the training of the learnable augmentation masks in feature space. This strategy enables us to conduct representation augmentation with greater control, enhancing the flexibility of the augmentation process.
Related Work
Data Augmentation for WSI analysis
The conventional method of data augmentation for WSI, similar to that used for natural images, involves continuously extracting representations of augmented patches from each ”bag” (a set of patches) throughout the training process. However, this is obviously inefficient for WSI model training due to the huge amount of time required for feature extraction. Consequently, augmentation is primarily performed during WSI preprocessing, as seen in methods like ABMIL (Ilse, Tomczak, and Welling 2018), to enhance patch diversity before training.
Data augmentation strategies developed for WSI analysis fall into two main categories. The former is realized with generative models (Zaffar et al. 2022; Dai et al. 2024), while the latter generates new subsets from bags through various mixing approaches (Yang et al. 2022; Chen and Lu 2023; Gadermayr et al. 2023). Specifically, DAGAN (Zaffar et al. 2022) involves training networks like generative adversarial networks to create synthetic data augmentations within the feature space. However, these generative models require an extra training phase separate from the representation model, as well as additional computational resources during the inference stage. ReMix (Yang et al. 2022) mixes class-specific prototypes determined by K-means clustering. Although Mixup methods can enhance data diversity, they may sometimes generate representations that deviate from the distribution of real-world data, which potentially compromises the model’s performance on actual datasets. The shared shortcoming of these data augmentation methods is their inability to direct the augmentation process, leading to a lack of control.
Self-Supervised Representation Learning
In MIL methods, it is essential to select an appropriate patch encoder. CNN models pre-trained on the ImageNet (Russakovsky et al. 2015) often serve as patch encoders due to their robust feature extraction capabilities (Lu et al. 2021; Shao et al. 2021). However, there are inevitably semantic differences between pathological and natural images. Self-supervised learning (SSL) has been widely used in representation learning (Jaiswal et al. 2020; Bao et al. 2022; Huang et al. 2023; Assran et al. 2023; Zhou et al. 2022). Chen et al. (Chen et al. 2020a) propose an end-to-end model SimCLR and systematically study the impact of data augmentations. They observe that single augmentation is not sufficient to learn good representations. MoCov2 (Chen et al. 2020b) expands upon MoCo (He et al. 2020) by incorporating blur augmentation, thereby improving the baseline on ImageNet. Grill et al. (Grill et al. 2020) develop BYOL, a unique metric-learning approach that learns representations by predicting one view from another, demonstrating the importance of color diversity in augmentations. Most recently, DINO (Caron et al. 2021) is proposed to utilize self-supervised ViT (Dosovitskiy et al. 2021) for representation learning. DINO (Caron et al. 2021) follows the data augmentations of BYOL (Grill et al. 2020) and multi-crop (Caron et al. 2020), which have also proven the advantage for patch representation pretraining in histopathology WSI analysis (Chen et al. 2022; Wu et al. 2024) . Our approach leverages the DINO framework and ViT architecture, setting this combination as the baseline for our representation learning. Unlike previous methods that focus on image space, we investigate the potential of applying data augmentation in feature space to enhance representation learning.
Method
Our goal is to develop a data augmentation strategy that operates within the feature space after the encoding of image patches, rather than applying traditional image augmentation techniques before patch encoding. Additionally, it is crucial that the representation augmentation remains as promptable as traditional image augmentations, avoiding that the outcomes are unpredictable and harm the performance.
As illustrated in Figure 2a, the architecture of our proposed model is constructed on the foundation of DINO (Caron et al. 2021). We extend the model by introducing an additional student branch specifically for representation augmentation. This branch shares weights with the image augmentation student branch and incorporates a promptable representation augmentation module. Within this framework, we can employ prompts aligned with image augmentation to guide the augmentation in feature space.
Following the training of the model, as depicted in Figure 2d, we can continuously obtain diverse WSI data in feature space through online sampling representations from the distributions. These distributions can be adjusted by different combinations of the trained augmentation masks. In this way, we can perform data augmentations in feature space that are as promptable as image augmentations.
Promptable Representation Distribution Learning
In self-supervised learning, the typical approach involves generating two different sets of augmented data from the same original data to help the model learn useful features without labeled data. In this approach, we modify traditional image augmentation by integrating a set of prompts that specify the type of augmentation. Given an image, we obtain by pre-augmentation, including flip, color distortion and gray scaling. Assuming that we have an augmentation set consisting of augmentation operators , we produce two augmentation prompts and by sampling random compositions of augmentation operators, where a value of appears in the -th bin of the prompt indicates the -th augmentation operator is active.
Promptable Image Augmentation
Image augmentation is the basis of traditional self-distillation. We first define the augmentations required in our image augmentation student branch and teacher branch. Guided by the augmentation prompts and , we produce two different views and . The views and are further encoded by and to obtain their representations and . Subsequently, these representations are transformed into embeddings through projector heads and . Referring to the knowledge distillation paradigm, we train the student network to match the output of the given teacher network, parameterized by and , respectively.
Representation Distribution Estimation
The core of the representation augmentation student branch is representation distribution estimation. In our method, we construct a neural network that can be trained to act as an estimator, using Gaussian prior to characterizing the distribution of patch representations (Zang, Huang, and Loy 2021). This architecture includes two main components: a mean head and a variance head , both composed of fully connected layers. Given the inherent non-negativity of the variance, we compute its logarithm rather than the variance itself directly. The corresponding representation mean and standard deviation of each image that together define a Gaussian distribution , which can be calculated by equations:
(1) |
Promptable Representation Augmentation
The estimated distribution handles the potential outcomes of representation augmentations for the patch. To make the distribution reflect representations from specific augmentation operators, we introduce a set of masks denoted by , where each row corresponds to a specific augmentation operator from the set of transformations . To ensure , we adopt a strategy where M is obtained by applying the sigmoid function to another random initialized matrix . By incorporating the augmentation prompt , we constrain this distribution to a more specific augmentation space. Preciously,
(2) |
where represents Hadamard product. With the narrowed distribution for a patch, we can obtain variable representations of the patch by sampling process , which is computationally efficient. Furthermore, by configuring , we can control the specific type of augmentation to be executed. Here, we adopt the reparameterization trick (Doersch 2021) to enable backpropagation for training. Specifically, we sample representations under the representation independence assumption
(3) |
where denotes -dimensional identity matrix.

Objective and Optimization
Knowledge Distillation
Then, we adapt the augmentation paradigm to self-supervised learning. Following the structure of DINO (Caron et al. 2021), we obtain the probability distribution over dimensions denoted by by a softmax operation on the branch outputs:
(4) |
where is a temperature parameter that controls the sharpness of the probability distribution. Similar formulas hold for and from the two student branches with the mapping head . We minimize the basic loss function
(5) |
where . It is important to note that we assign the same random prompt for a patch to obtain and . It guides the distillation architecture to align representations from both the image augmentation and representation augmentation for the same combination of augmentation operators. This is the basis on which we can decouple the augmentation operators and thereby control the process of the representation augmentation. Moreover, we follow the DINO (Caron et al. 2021) framework to adopt the multi-crop strategy by using 2 global views and several local views. All crops are fed into the image augmentation student branch while only the global views are fed into the teacher branch.
Representation Distribution Constraint
As we model the representation distribution based on Gaussian prior, We add a Kullback-Leibler (KL) divergence constraint to the distribution estimator, which is represented as
(6) |
where is the Gaussian distribution and is the K-L divergence.
Promptable Mask Constraint
The prompted masks aim to identify the most specific feature dimensions associated with different augmentations. Therefore, we employ L1 normalization to induce sparsity in the augmentation masks:
(7) |
Additionally, we introduce a variance regularization term on the standard deviation of the embeddings across the feature dimension (Bardes, Ponce, and LeCun 2022), to mitigate the issue of augmentation masks trending towards zero:
(8) |
where is a small scalar preventing numerical instabilities.
Overall Objective
The final object function is composed as follows
(9) |
where and controls the weights of each term in the loss. The parameters of the student network and the distribution estimator are optimized by the gradient descent algorithm, and the teacher network is updated by the exponential moving average (EMA) mechanism
(10) |
with the update rule of following a cosine schedule during training.
Methods | CLAM (Lu et al. 2021) | TransMIL (Shao et al. 2021) | DTFD-MIL (Zhang et al. 2022a) | ||||||
---|---|---|---|---|---|---|---|---|---|
AUC | F1-Score | ACC | AUC | F1-Score | ACC | AUC | F1-Score | ACC | |
DINO+Random Perturbation | 74.9 | 35.0 | 40.8 | 74.1 | 39.2 | 39.9 | 75.6 | 42.8 | 43.9 |
DINO+MC Sampling | 75.9 | 37.2 | 39.5 | 75.2 | 36.7 | 38.6 | 81.3 | 45.0 | 47.1 |
PRDL w/o PRA (DINO) | 79.4 | 49.8 | 50.2 | 75.2 | 41.3 | 44.0 | 76.1 | 41.5 | 43.1 |
PRDL w/o | 83.8 | 51.2 | 51.6 | 81.0 | 43.1 | 46.6 | 80.0 | 39.3 | 41.7 |
PRDL w/o | 83.8 | 52.7 | 54.3 | 78.4 | 43.5 | 45.3 | 79.6 | 41.7 | 44.0 |
PRDL w/o | 84.6 | 49.9 | 51.6 | 80.8 | 43.1 | 49.8 | 80.9 | 39.6 | 47.3 |
PRDL | 86.4 | 52.5 | 57.9 | 82.0 | 47.1 | 48.4 | 81.3 | 48.7 | 49.8 |
PRDL+PRS | 89.4 | 65.1 | 65.9 | 84.5 | 55.1 | 56.5 | 83.4 | 51.3 | 52.9 |
WSI Analysis with Representation Augmentation
Given a WSI sized by , consisting of patches , a common MIL model for WSI classification can be formulated as:
(11) |
where is the WSI-level prediction, is the patch encoder and is the WSI-level representation aggregator.
Promptable Representation Sampling
Through knowledge distillation in the promptable representation distribution learning, we achieved the goal of procedure exchange of data augmentation and patch encoding. Then we proposed a representation augmentation strategy named promptable representation sampling (PRS) for WSI augmentation. As illustrated in Figure 2d, patch-level representation distributions are obtained after patch encoding. During WSI classifier training, patch representations within X can be replaced by representations online sampled from the corresponding distributions with random combined augmentation prompts, which can be formulated as
(12) |
Then, we feed instead of Z into the WSI model to achieve data augmentation for training, which is represented as
(13) |
As in natural images, WSI augmentation is only used in the training phase to improve the generalization of the model. The image augmentation student branch and the representation augmentation student branch share weights in patch-level representation learning, determining that the WSI augmentation module can be skipped in the inference phase. Therefore, we can directly feed representations extracted from the original patches into the WSI classifier for prediction:
(14) |
Experiments
The proposed method was evaluated based on WSI classification tasks in three datasets: 1) USTC-EGFR: a private dataset with 754 WSIs categorized into 5 types. 2) TCGA-EGFR: a public dataset with 696 WSIs categorized into 2 types. 3) TCGA-LUNG-3K: a public dataset with 3064 WSIs categorized into 3 types. Please refer to the supplemental material for detailed information about the datasets.
The WSIs were segmented into non-overlapping patches in size of 224 × 224 under 20× lenses for representation learning and extraction. Each dataset was split into training, validation, and testing subsets with the ratio of 6:1:3 at the patient-level.
We followed the implementation in DINO (Caron et al. 2021) to use ViT-S/16 (Dosovitskiy et al. 2021) as the backbone and the class token of ViT as the final feature embedding. Our evaluation metrics include average accuracy, micro-average area under the curve (AUC), and macro-average F1 score. All the methods were implemented in Python 3.8 with torch 1.8.1 and run on a computer cluster with 10 Xeon 2.66GHz CPUs and 10 GPUs of Nvidia Geforce 2080Ti.
Comparison with SOTA Methods
We compared our method with 5 different WSI representing strategies, including ImageNet (Russakovsky et al. 2015), SimCLR (Chen et al. 2020a), BYOL (Grill et al. 2020), MoCov3 (Chen, Xie, and He 2021) and DINO (Caron et al. 2021). The first three take ResNet50 (He et al. 2016) as the backbone, while the others utilize ViT-S (Dosovitskiy et al. 2021). Moreover, we compared our method with 4 WSI data augmentation frameworks, including DAGAN (Zaffar et al. 2022), ReMix (Yang et al. 2022), RankMix (Chen and Lu 2023), Intra-Mixup (Gadermayr et al. 2023). Given DINO is the basis of our representation learning framework, the comparative experiments with WSI augmentations are performed on the representations extracted by DINO.
As depicted in Figure 3, our method achieves the best performance on the USTC-EGFR dataset, TCGA-EGFR dataset and TCGA-LUNG-3K dataset under CLAM (Lu et al. 2021), TransMIL (Shao et al. 2021) and DTFD-MIL (Zhang et al. 2022a) benchmarks.
Comparison with Representation Learning Methods
SimCLR (Chen et al. 2020a) is the self-supervised representation learning framework widely applied in feature extraction. The results on the TCGA-EGFR dataset show that it achieves 12.5%, 2.3%, 1.9% AUC better than ImageNet, which showcases the importance of addressing the semantic gap between natural and pathological images. MoCov3 (Chen, Xie, and He 2021) and DINO (Caron et al. 2021) utilize the ViT architecture as the backbone, demonstrating superior performance on the USTC-EGFR and TCGA-LUNG-3K datasets compared to SimCLR (Chen et al. 2020a) and BYOL (Grill et al. 2020). However, the observed performance gaps of MoCov3 and DINO on the TCGA-EGFR dataset compared to top-tier results indicate that current image augmentation strategies may not be sufficiently robust for generating discriminative features needed for effective WSI classification.
Our proposed PRDL additionally introduces representation augmentation into the process of self-supervised learning and employs augmentation prompts to control the representation augmentations, which improves the performance by effectively increasing the diversity of representations. Compared with our baseline model DINO, PRDL achieves increase in AUC of 7.0%, 18.4%, 1.3% under the CLAM (Lu et al. 2021) benchmark, 6.8%, 19.2%, 2.7% under the TransMIL (Shao et al. 2021) benchmark and 5.2%, 24.4%, 1.7% under the DTFD-MIL (Zhang et al. 2022a) benchmark on the three datasets, respectively. Moreover, the promptable distribution estimator and the augmentation masks trained in representation learning can be utilized in the downstream task.
Comparison with WSI Augmentation Methods
Mixup-based methods generate varied representations by mixing features at different levels. ReMix (Yang et al. 2022) proposes to mix instance prototypes formed by clustering. The performance of ReMix in AUC on the TCGA-EGFR dataset has improved by 8.8% under the CLAM benchmark when compared to DINO. Nevertheless, the performance degradation observed across all Mixup-based methods on the TCGA-LUNG dataset under the CLAM benchmark suggests that Mixup-based strategies may compromise semantic integrity, leading to noisy training samples and a subsequent decline in model performance. DAGAN (Zaffar et al. 2022) incorporates a Generative Adversarial Network (GAN) (Goodfellow et al. 2020) model to create new representations, adding another training process. It improves performance in AUC by 0.6%, 3.8%, 0.3% on the USTC-EGFR dataset. However, the generative model is trained after representation learning so that the performance is limited by the quality of the original representations. Moreover, DAGAN introduces extra computation costs in both training and inference phases.
Our WSI augmentation strategy PRS, which samples from prompted distributions, offers flexible augmentation, enhancing WSI classification without additional parameters. This can also be viewed as a reasonable perturbation to representations, which breaks the invariance of representations during training. With the guide of the augmentation prompts, the noise generated during sampling is much less than Mixup. Moreover, the distribution estimator is concurrently trained with representation learning, avoiding additional training overhead and continuously adapting to the representation change over the training procedure. With PRS, there are observed improvements in performance across three datasets, with increases in AUC of 9.4%, 15.2%, 2.1% under the CLAM (Lu et al. 2021) benchmark, 4.8%, 9.8%, 1.3% under the TransMIL (Shao et al. 2021) benchmark, 6.6%, 18.4%, 1.2% under the DTFD-MIL (Zhang et al. 2022a) benchmark compared to the second-best data augmentation methods.
Methods | AUC | F1-Score | ACC |
---|---|---|---|
DINO | 79.4 | 49.8 | 50.2 |
PRDL | 86.4 | 52.5 | 57.9 |
PRDL+RS (w/o M) | 86.0 | 59.6 | 59.6 |
PRDL+ | 89.3 | 61.2 | 61.9 |
PRDL+ | 88.8 | 61.0 | 61.0 |
PRDL+ | 88.9 | 61.9 | 62.3 |
PRDL+ | 89.3 | 62.1 | 63.2 |
PRDL+ | 89.0 | 62.7 | 62.8 |
PRDL+ | 89.1 | 63.0 | 63.2 |
PRDL+PRS | 89.4 | 65.1 | 65.9 |


Ablation Study
We first conduct ablation study to verify the necessity of model components. The results are detailed in Table 1. Here, we first investigate WSI augmentation within the feature space through random perturbation using a standard Gaussian distribution. It shows that the performance on the USTC-EGFR dataset is decreased in F1-Score by 14.8% compared to DINO under the CLAM benchmark, while it is increased by 1.3% under the DTFD-MIL benchmark. This suggests that while perturbations affect the model’s performance, they do not completely ruin the representations. Meanwhile, we utilize Monte Carlo Sampling (Zheng et al. 2023) where patch representations are randomly discarded during the training of the WSI models. As a result, the performance under the DTFD-MIL benchmark is increased in AUC by 5.2% compared to DINO, which indicates that it works in some cases. However, this strategy has a high risk of losing important information for classification, thus leading to significant degradation of performance under the CLAM benchmark. The results of these two strategies show that the impact of purposeless perturbation on the patch representations is unstable, but opens the possibility of guiding perturbations to enhance the performance of WSI classification.
PRDL w/o PRA denotes PRDL without promptable representation augmentation related operations, where data augmentations are merely carried out in image space. The decrease in the evaluation metrics demonstrates that the PRA module plays a crucial role in representation learning. Considering that the representation distribution is estimated based on a Gaussian prior, we constrain it with KL divergence. The results of PRDL w/o show that KL divergence is essential for the training of the distribution estimator. Additionally, components like and within the PRDL framework are crucial for effective learning of augmentation masks. Finally, we verify the impact of the PRS strategy on WSI classification. We observe that PRS can improve the performance of the WSI classifier even further building on PRDL, where the Micro-AUC under the CLAM benchmark increases from 0.864 to 0.894. This suggests that the representations generated by PRS are effective to the model and contain useful semantic information.
The impact of augmentation prompts
As shown in Table 2, we study the impact of different augmentation prompts. The results of PRDL+RS(w/o M) indicate that an overly broad range of representation augmentations can result in limited improvement. Individual augmentation prompts produce better but varied results, yet they still fall short compared to PRDL+PRS that employs a random combination of the augmentation prompts. This demonstrates the importance of a guided and variable representation augmentation for improving model performance. Figure 4 shows that different augmentation prompts have different effects on the representation dimensions, and some augmentations cause similar impacts in feature space. ReisizedCrop and HorizontalFlip are both operated in spatial space and thus their corresponding prompts are similar. However, HorizontalFlip has less improvement in performance compared with other prompts, which indicates that the prompt obtained from this simple image transformation also has relatively limited effects on the augmented representations. ColorJitter and GaussianBlur have minimal effects on structural attributes of images, leading to similar impacts in feature space. Conversely, Grayscale and Solarization significantly alter the image, which directs focus towards key dimensions in representation, hence their prompts perform better.
Conclusion
We proposed a novel promptable representation distribution learning (PRDL) framework with a promptable representation sampling (PRS) strategy for promptable and efficient data augmentation in feature space for histopathology WSI classification. The approach involves a promptable distribution estimator and augmentation prompts for generating diverse representation augmentations, thereby improving the representation quality. This is complemented by a PRS strategy, tailored to leverage the trained estimator and prompts for effective WSI augmentation. The prompted representation augmentations significantly enhance image representations and preserve control, thus avoiding the loss of important semantic information, all of which are beneficial for WSI analysis.
Acknowledgments
This work was partly supported by Beijing Natural Science Foundation (Grant No. 7242270), partly supported by the National Natural Science Foundation of China (Grant No. 62171007, 61901018, and 61906058), partly supported by the Fundamental Research Fund for the Central Universities of China (grant No. YWF-23-Q-1075), partly supported by the Anhui Provincial Natural Science Foundation (Grant No. 2408085MF162), partly supported by Emergency Key Program of Guangzhou Laboratory (Grant No. EKPG21-32), partly supported by Joint Fund for Medical Artificial Intelligence (Grant No. MAI2023C014), and partly supported by National Key Research and Development Program of China (Grant No. 2021YFF1201004).
References
- Assran et al. (2023) Assran, M.; Duval, Q.; Misra, I.; Bojanowski, P.; Vincent, P.; Rabbat, M.; LeCun, Y.; and Ballas, N. 2023. Self-supervised learning from images with a joint-embedding predictive architecture. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 15619–15629.
- Bao et al. (2022) Bao, H.; Dong, L.; Piao, S.; and Wei, F. 2022. BEiT: BERT Pre-Training of Image Transformers. arXiv:2106.08254.
- Bardes, Ponce, and LeCun (2022) Bardes, A.; Ponce, J.; and LeCun, Y. 2022. VICReg: Variance-Invariance-Covariance Regularization for Self-Supervised Learning. arXiv:2105.04906.
- Campanella et al. (2019) Campanella, G.; Hanna, M. G.; Geneslaw, L.; Miraflor, A.; Werneck Krauss Silva, V.; Busam, K. J.; Brogi, E.; Reuter, V. E.; Klimstra, D. S.; and Fuchs, T. J. 2019. Clinical-grade computational pathology using weakly supervised deep learning on whole slide images. Nature medicine, 25(8): 1301–1309.
- Caron et al. (2020) Caron, M.; Misra, I.; Mairal, J.; Goyal, P.; Bojanowski, P.; and Joulin, A. 2020. Unsupervised learning of visual features by contrasting cluster assignments. Advances in neural information processing systems, 33: 9912–9924.
- Caron et al. (2021) Caron, M.; Touvron, H.; Misra, I.; Jégou, H.; Mairal, J.; Bojanowski, P.; and Joulin, A. 2021. Emerging properties in self-supervised vision transformers. In Proceedings of the IEEE/CVF international conference on computer vision, 9650–9660.
- Chen et al. (2022) Chen, R. J.; Chen, C.; Li, Y.; Chen, T. Y.; Trister, A. D.; Krishnan, R. G.; and Mahmood, F. 2022. Scaling vision transformers to gigapixel images via hierarchical self-supervised learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 16144–16155.
- Chen et al. (2020a) Chen, T.; Kornblith, S.; Norouzi, M.; and Hinton, G. 2020a. A simple framework for contrastive learning of visual representations. In International conference on machine learning, 1597–1607. PMLR.
- Chen et al. (2020b) Chen, X.; Fan, H.; Girshick, R.; and He, K. 2020b. Improved Baselines with Momentum Contrastive Learning. arXiv:2003.04297.
- Chen, Xie, and He (2021) Chen, X.; Xie, S.; and He, K. 2021. An Empirical Study of Training Self-Supervised Vision Transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 9640–9649.
- Chen and Lu (2023) Chen, Y.-C.; and Lu, C.-S. 2023. RankMix: Data Augmentation for Weakly Supervised Learning of Classifying Whole Slide Images With Diverse Sizes and Imbalanced Categories. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 23936–23945.
- Dai et al. (2024) Dai, L.; Wang, Y.; Wang, H.; Zhang, Y.; et al. 2024. AugDiff: Diffusion based feature augmentation for multiple instance learning in whole slide image. IEEE Transactions on Artificial Intelligence.
- Doersch (2021) Doersch, C. 2021. Tutorial on Variational Autoencoders. arXiv:1606.05908.
- Dosovitskiy et al. (2021) Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; Uszkoreit, J.; and Houlsby, N. 2021. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv:2010.11929.
- Gadermayr et al. (2023) Gadermayr, M.; Koller, L.; Tschuchnig, M.; Stangassinger, L. M.; Kreutzer, C.; Couillard-Despres, S.; Oostingh, G. J.; and Hittmair, A. 2023. Mixup-mil: Novel data augmentation for multiple instance learning and a study on thyroid cancer diagnosis. In International Conference on Medical Image Computing and Computer-Assisted Intervention, 477–486. Springer.
- Goodfellow et al. (2020) Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; and Bengio, Y. 2020. Generative adversarial networks. Communications of the ACM, 63(11): 139–144.
- Grill et al. (2020) Grill, J.-B.; Strub, F.; Altché, F.; Tallec, C.; Richemond, P.; Buchatskaya, E.; Doersch, C.; Avila Pires, B.; Guo, Z.; Gheshlaghi Azar, M.; et al. 2020. Bootstrap your own latent-a new approach to self-supervised learning. Advances in neural information processing systems, 33: 21271–21284.
- He et al. (2020) He, K.; Fan, H.; Wu, Y.; Xie, S.; and Girshick, R. 2020. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 9729–9738.
- He et al. (2016) He, K.; Zhang, X.; Ren, S.; and Sun, J. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, 770–778.
- Huang et al. (2023) Huang, C.; Goh, H.; Gu, J.; and Susskind, J. 2023. MAST: Masked Augmentation Subspace Training for Generalizable Self-Supervised Priors. arXiv:2303.03679.
- Ilse, Tomczak, and Welling (2018) Ilse, M.; Tomczak, J.; and Welling, M. 2018. Attention-based deep multiple instance learning. In International conference on machine learning, 2127–2136. PMLR.
- Jaiswal et al. (2020) Jaiswal, A.; Babu, A. R.; Zadeh, M. Z.; Banerjee, D.; and Makedon, F. 2020. A survey on contrastive self-supervised learning. Technologies, 9(1): 2.
- Li, Li, and Eliceiri (2021) Li, B.; Li, Y.; and Eliceiri, K. W. 2021. Dual-stream multiple instance learning network for whole slide image classification with self-supervised contrastive learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 14318–14328.
- Lu et al. (2024) Lu, M. Y.; Chen, B.; Williamson, D. F.; Chen, R. J.; Liang, I.; Ding, T.; Jaume, G.; Odintsov, I.; Le, L. P.; Gerber, G.; et al. 2024. A visual-language foundation model for computational pathology. Nature Medicine, 30: 863–874.
- Lu et al. (2021) Lu, M. Y.; Williamson, D. F.; Chen, T. Y.; Chen, R. J.; Barbieri, M.; and Mahmood, F. 2021. Data-efficient and weakly supervised computational pathology on whole-slide images. Nature Biomedical Engineering, 5(6): 555–570.
- Nakhli et al. (2023) Nakhli, R.; Moghadam, P. A.; Mi, H.; Farahani, H.; Baras, A.; Gilks, B.; and Bashashati, A. 2023. Sparse Multi-Modal Graph Transformer With Shared-Context Processing for Representation Learning of Giga-Pixel Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 11547–11557.
- Russakovsky et al. (2015) Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. 2015. Imagenet large scale visual recognition challenge. International journal of computer vision, 115: 211–252.
- Shao et al. (2021) Shao, Z.; Bian, H.; Chen, Y.; Wang, Y.; Zhang, J.; Ji, X.; et al. 2021. Transmil: Transformer based correlated multiple instance learning for whole slide image classification. Advances in neural information processing systems, 34: 2136–2147.
- Wu et al. (2024) Wu, K.; Jiang, Z.; Tang, K.; Shi, J.; Xie, F.; Wang, W.; Wu, H.; and Zheng, Y. 2024. Pan-cancer Histopathology WSI Pre-training with Position-aware Masked Autoencoder. IEEE Transactions on Medical Imaging.
- Yang et al. (2022) Yang, J.; Chen, H.; Zhao, Y.; Yang, F.; Zhang, Y.; He, L.; and Yao, J. 2022. ReMix: A General and Efficient Framework for Multiple Instance Learning Based Whole Slide Image Classification. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2022: 25th International Conference, Singapore, September 18–22, 2022, Proceedings, Part II, 35–45. Springer.
- Zaffar et al. (2022) Zaffar, I.; Jaume, G.; Rajpoot, N.; and Mahmood, F. 2022. Embedding Space Augmentation for Weakly Supervised Learning in Whole-Slide Images. arXiv:2210.17013.
- Zang, Huang, and Loy (2021) Zang, Y.; Huang, C.; and Loy, C. C. 2021. Fasa: Feature augmentation and sampling adaptation for long-tailed instance segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 3457–3466.
- Zhang et al. (2022a) Zhang, H.; Meng, Y.; Zhao, Y.; Qiao, Y.; Yang, X.; Coupland, S. E.; and Zheng, Y. 2022a. Dtfd-mil: Double-tier feature distillation multiple instance learning for histopathology whole slide image classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 18802–18812.
- Zhang et al. (2022b) Zhang, J.; Zhang, X.; Ma, K.; Gupta, R.; Saltz, J.; Vakalopoulou, M.; and Samaras, D. 2022b. Gigapixel whole-slide images classification using locally supervised learning. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2022: 25th International Conference, Singapore, September 18–22, 2022, Proceedings, Part II, 192–201. Springer.
- Zheng et al. (2023) Zheng, Y.; Li, J.; Shi, J.; Xie, F.; Huai, J.; Cao, M.; and Jiang, Z. 2023. Kernel Attention Transformer for Histopathology Whole Slide Image Analysis and Assistant Cancer Diagnosis. IEEE Transactions on Medical Imaging.
- Zhou et al. (2022) Zhou, J.; Wei, C.; Wang, H.; Shen, W.; Xie, C.; Yuille, A.; and Kong, T. 2022. iBOT: Image BERT Pre-Training with Online Tokenizer. arXiv:2111.07832.
Supplementary Material
Dataset Details
The evaluation of the proposed method involved a private dataset and two public datasets, detailed as below:
USTC-EGFR is a private lung adenocarcinoma dataset that contains 754 WSIs for epidermal growth factor receptor (EGFR) gene mutation identification. These WSIs are categorized into 5 classes, including EGFR 19del mutation (19del), EGFR L858R mutation (L858R), none common driver mutations (Wild), other driver gene mutation (Other), and cancer-free tissue (Normal).
TCGA-EGFR is a public lung adenocarcinoma dataset for EGFR mutation classification that contains 696 WSIs collected from the cancer genome atlas (TCGA) program of NCI. These WSIs are categorized into 2 classes, including wild type and mutant type.
TCGA-LUNG-3K is a public lung dataset that contains 3064 WSIs collected from the cancer genome atlas (TCGA) program of NCI. These WSIs are categorized into 3 classes, including lung adenocarcinoma (LUAD), lung squamous cell carcinoma (LUSC), and cancer-free tissue (Normal).
The detailed distributions of these datasets are shown in Table 3.
Training Details
For the representation learning phase, 1000 patches are sampled from each WSI in the training subset to ensure a comprehensive and diverse set of samples for model learning. This strategy generates 457,000, 422,000, and 1,839,000 patches, respectively for USTC-EGFR, TCGA-EGFR, and TCGA-LUNG-3K dataset, for the training of the self-supervised models. Before the training of the WSI classifier begins, each patch is encoded as a Gaussian distribution characterized by mean () and standard deviation (). This process prepares the patches for subsequent classification tasks.
We follow the implementation details of DINO (Caron et al. 2021), to train a backbone of ViT-S/16. The learning rate is linearly ramped up during warmup with the following linear scaling rule with denoting the batch size. After warmup, the learning rate and the weight decay follow a cosine schedule. The temperature is set to 0.1 while is linearly ramped up from 0.04 to 0.07. We train the SSL model by 20 epochs on the training set with the batch size as . Our data augmentation strategy, detailed in Table 4, includes three distinct phases. Initially, pre-augmentation involves basic spatial and color modifications. The subsequent image augmentations add more complex transformations like cropping. The student branch mirrors the teacher’s operations but incorporates 8 local views.
Hyper-parameter Verification

USTC-EGFR | Normal | 19del | L858R | Wild | Other |
---|---|---|---|---|---|
Train | 101 | 69 | 118 | 85 | 84 |
Val | 16 | 11 | 19 | 14 | 14 |
Test | 48 | 38 | 47 | 47 | 43 |
TCGA-EGFR | Wild | Mutant | |||
Train | 357 | 65 | |||
Val | 58 | 9 | |||
Test | 175 | 32 | |||
TCGA-LUNG-3K | Normal | LUAD | LUSC | ||
Train | 339 | 740 | 760 | ||
Val | 56 | 123 | 126 | ||
Test | 158 | 394 | 368 |
Pre-Augmentation | RandomHorizontalFlip |
---|---|
ColorJitter (0.4, 0.4, 0.2, 0.1) | |
RandomGrayscale | |
Image Augmentation Teacher | RandomResizedCrop (224) |
RandomHorizontalFlip | |
ColorJitter (0.4, 0.4, 0.2, 0.1) | |
GaussianBlur | |
Solarization | |
Image Augmentation Student | RandomResizedCrop (224) |
RandomResizedCrop (96) | |
RandomHorizontalFlip | |
ColorJitter (0.4, 0.4, 0.2, 0.1) | |
GaussianBlur | |
Solarization |
The main factors that decide the capacity of the PRDL include 4 hyper-parameters . We recorded Micro-AUC, F1 score, and accuracy metrics for the validation subset to assess the impact of each hyper-parameter while keeping other hyper-parameters constant. The detailed outcomes of this tuning process are illustrated in Figure 5.
The temperature of representation augmentation
The sharpness of representations from the representation augmentation student branch is regulated by adjusting the softmax temperature parameter . Optimal performance is observed with lower temperatures. As the temperature increases, the targets become smoother and the evaluation metrics decrease. Consequently, we have set for subsequent experiments.
The weight of KL divergence
The hyper-parameter controls the coefficient of , which constrains the distribution estimator in the PRDL framework. We observe that from the performance on the validation subset is improved with a decrease in . Therefore, we have set for PRDL training to optimize performance.
The weight of L1 normalization
The hyper-parameter sets the weight of , impacting the sparsity of augmentation masks in PRDL training. The optimal performance on the validation subset is achieved at , hence this value has been chosen for training the PRDL to enhance performance.
The weight of variance regularization
The hyper-parameter determines the weight of that affects the variance of augmentation masks across the feature dimension during PRDL training. is found to achieve the best performance on the validation subset, thus we have select this value in PRDL training.
Complete Numerical Results
Methods | USTC-EGFR | TCGA-EGFR | TCGA-LUNG-3K |
---|---|---|---|
CONCH (Lu et al. 2024) | 85.0 | 65.4 | 96.0 |
PRDL | 86.4 | 81.4 | 97.9 |
PRDL+PRS | 89.4 | 87.0 | 97.5 |
Methods | USTC-EGFR | TCGA-EGFR | TCGA-LUNG-3K |
---|---|---|---|
DINO+PAMA (Wu et al. 2024) | 79.8 | 70.7 | 96.6 |
PRDL+PRS+PAMA | 85.9 | 85.8 | 98.0 |
Comparison with the Foundation Model
Given the large foundation models in histopathology (Lu et al. 2024) have demonstrated robust representational capabilities, we evaluated the performance of the foundation models across the datasets under the CLAM benchmark. The AUC results are detailed in Table 5. It shows that CONCH (Lu et al. 2024) performs well in the TCGA-LUNG-3K dataset, which demonstrates its good generalization in common cancer subtyping task. However, the AUC of CONCH is 21.6% lower than our method in the TCGA-EGFR dataset. It indicates that well-designed representation pre-training and augmentation methodologies are still necessary for challenging prediction tasks. Moreover, our method is not specifically designed for certain cancers and is general for histopathology patch representation learning, which makes it adequate to build histopathology foundation models.
Integration with WSI-level Pretraining
WSI-level pretraining is one of the downstream tasks that benefits from data augmentation techniques (Wu et al. 2024). Our method can enhance the variability of the data augmentation strategies used in WSI-level pertaining, thereby improving the generalization of the model. We experiment with a fine-tuning training strategy that involves training all network parameters after pretraining. The results in Table 6 show that our method achieves increase in AUC of 6.1%, 15.1% and 1.4% on the three datasets. This demonstrates that our method can work effectively in conjunction with the state-of-the-art WSI-level pretraining methods.
Methods | CLAM (Lu et al. 2021) | TransMIL (Shao et al. 2021) | DTFD-MIL (Zhang et al. 2022a) | ||||||
---|---|---|---|---|---|---|---|---|---|
AUC | F1-Score | ACC | AUC | F1-Score | ACC | AUC | F1-Score | ACC | |
ImageNet (Russakovsky et al. 2015) | 77.9 | 38.4 | 43.1 | 73.4 | 30.4 | 34.5 | 73.1 | 38.3 | 39.0 |
SimCLR (Chen et al. 2020a) | 69.4 | 38.1 | 39.5 | 62.5 | 22.8 | 30.9 | 70.3 | 37.8 | 40.4 |
BYOL (Grill et al. 2020) | 61.5 | 25.6 | 31.8 | 54.8 | 26.0 | 20.7 | 61.7 | 28.7 | 34.1 |
MoCov3 (Chen, Xie, and He 2021) | 77.6 | 40.7 | 43.5 | 63.9 | 23.1 | 29.6 | 74.7 | 31.9 | 38.1 |
DINO (Caron et al. 2021) | 79.4 | 49.8 | 50.2 | 75.2 | 41.3 | 44.0 | 76.1 | 41.5 | 43.1 |
DINO+Random Perturbation | 74.9 | 35.0 | 40.8 | 74.1 | 39.2 | 39.9 | 75.6 | 42.8 | 43.9 |
DINO+MC Sampling (Zheng et al. 2023) | 75.9 | 37.2 | 39.5 | 75.2 | 36.7 | 38.6 | 81.3 | 45.0 | 47.1 |
DINO+ReMix (Yang et al. 2022) | 78.2 | 37.1 | 43.5 | 75.6 | 36.3 | 39.9 | 73.7 | 36.7 | 41.7 |
DINO+RankMix (Chen and Lu 2023) | 73.3 | 29.7 | 32.7 | 75.5 | 36.7 | 43.1 | 73.9 | 30.4 | 35.9 |
DINO+Intra-Mixup (Gadermayr et al. 2023) | 79.0 | 38.7 | 40.8 | 79.7 | 42.1 | 43.9 | 76.8 | 41.8 | 42.6 |
DINO+DAGAN (Zaffar et al. 2022) | 80.0 | 49.0 | 51.1 | 79.0 | 42.6 | 44.4 | 76.4 | 44.9 | 46.6 |
DINO+AugDiff (Dai et al. 2024) | 81.7 | 48.8 | 52.3 | 77.1 | 35.2 | 39.8 | 75.5 | 42.5 | 43.8 |
PRDL | 86.4 | 52.5 | 57.9 | 82.0 | 47.1 | 48.4 | 81.3 | 48.7 | 49.8 |
PRDL+PRS | 89.4 | 65.1 | 65.9 | 84.5 | 55.1 | 56.5 | 83.4 | 51.3 | 52.9 |
Methods | CLAM (Lu et al. 2021) | TransMIL (Shao et al. 2021) | DTFD-MIL (Zhang et al. 2022a) | ||||||
---|---|---|---|---|---|---|---|---|---|
AUC | F1-Score | ACC | AUC | F1-Score | ACC | AUC | F1-Score | ACC | |
ImageNet (Russakovsky et al. 2015) | 63.8 | 61.3 | 76.8 | 65.5 | 54.6 | 75.4 | 73.4 | 62.4 | 76.3 |
SimCLR (Chen et al. 2020a) | 76.3 | 61.7 | 73.9 | 67.8 | 60.1 | 70.5 | 75.3 | 60.1 | 79.6 |
BYOL (Grill et al. 2020) | 60.9 | 51.3 | 78.3 | 64.0 | 56.0 | 70.1 | 58.9 | 55.6 | 76.8 |
MoCov3 (Chen, Xie, and He 2021) | 48.1 | 45.7 | 68.1 | 58.2 | 49.0 | 70.1 | 63.6 | 55.9 | 78.7 |
DINO (Caron et al. 2021) | 63.0 | 54.6 | 75.4 | 59.1 | 55.1 | 82.2 | 59.4 | 53.1 | 74.9 |
DINO+Random Perturbation | 61.5 | 53.4 | 64.7 | 63.4 | 57.4 | 71.0 | 67.6 | 52.4 | 59.9 |
DINO+MC Sampling (Zheng et al. 2023) | 70.7 | 60.5 | 77.8 | 66.9 | 60.5 | 76.8 | 69.0 | 54.2 | 80.2 |
DINO+ReMix (Yang et al. 2022) | 71.8 | 47.4 | 81.6 | 61.2 | 45.8 | 84.5 | 58.7 | 50.7 | 74.9 |
DINO+RankMix (Chen and Lu 2023) | 61.8 | 36.5 | 57.4 | 62.9 | 36.5 | 57.4 | 63.9 | 45.8 | 84.5 |
DINO+Intra-Mixup (Gadermayr et al. 2023) | 67.9 | 63.1 | 64.7 | 73.6 | 68.5 | 69.1 | 69.5 | 59.3 | 74.3 |
DINO+DAGAN (Zaffar et al. 2022) | 71.0 | 39.6 | 52.9 | 71.7 | 64.4 | 64.7 | 61.6 | 57.8 | 76.8 |
DINO+AugDiff (Dai et al. 2024) | 67.6 | 60.6 | 77.3 | 67.8 | 54.5 | 78.8 | 73.4 | 61.8 | 80.8 |
PRDL | 81.4 | 70.6 | 84.1 | 78.3 | 67.9 | 80.2 | 83.8 | 71.6 | 85.5 |
PRDL+PRS | 87.0 | 75.1 | 83.6 | 83.4 | 77.0 | 86.5 | 87.9 | 77.2 | 85.5 |
Methods | CLAM (Lu et al. 2021) | TransMIL (Shao et al. 2021) | DTFD-MIL (Zhang et al. 2022a) | ||||||
---|---|---|---|---|---|---|---|---|---|
AUC | F1-Score | ACC | AUC | F1-Score | ACC | AUC | F1-Score | ACC | |
ImageNet (Russakovsky et al. 2015) | 94.6 | 84.1 | 82.6 | 91.9 | 80.2 | 78.2 | 93.7 | 82.8 | 80.9 |
SimCLR (Chen et al. 2020a) | 93.5 | 80.3 | 78.8 | 91.9 | 81.2 | 79.0 | 93.0 | 82.5 | 81.0 |
BYOL (Grill et al. 2020) | 92.6 | 78.5 | 77.1 | 89.6 | 75.9 | 74.6 | 90.9 | 78.3 | 76.9 |
MoCov3 (Chen, Xie, and He 2021) | 94.1 | 82.6 | 81.4 | 93.1 | 79.7 | 77.7 | 93.9 | 82.3 | 80.4 |
DINO (Caron et al. 2021) | 96.6 | 87.2 | 85.8 | 94.7 | 85.5 | 83.6 | 95.7 | 87.4 | 85.5 |
DINO+Random Perturbation | 96.4 | 86.7 | 85.2 | 95.0 | 84.9 | 83.3 | 96.1 | 87.2 | 85.4 |
DINO+MC Sampling (Zheng et al. 2023) | 96.1 | 87.2 | 85.7 | 95.2 | 84.1 | 83.0 | 96.3 | 87.9 | 86.6 |
DINO+ReMix (Yang et al. 2022) | 95.4 | 83.8 | 82.8 | 93.7 | 83.0 | 81.3 | 96.1 | 84.4 | 81.8 |
DINO+RankMix (Chen and Lu 2023) | 94.1 | 81.6 | 79.5 | 93.4 | 75.3 | 73.2 | 94.2 | 75.7 | 73.8 |
DINO+Intra-Mixup (Gadermayr et al. 2023) | 94.6 | 80.9 | 80.6 | 96.0 | 87.7 | 86.1 | 96.6 | 89.4 | 88.0 |
DINO+DAGAN (Zaffar et al. 2022) | 94.1 | 81.6 | 79.5 | 94.7 | 85.4 | 83.6 | 95.5 | 86.0 | 84.2 |
DINO+AugDiff (Dai et al. 2024) | 95.9 | 86.4 | 85.1 | 94.5 | 84.2 | 82.7 | 95.8 | 87.4 | 85.8 |
PRDL | 97.9 | 90.5 | 89.1 | 97.4 | 89.8 | 88.5 | 97.4 | 90.3 | 89.0 |
PRDL+PRS | 97.5 | 90.6 | 89.2 | 97.3 | 90.5 | 89.4 | 97.8 | 91.1 | 89.9 |