This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

ParamNet: A Dynamic Parameter Network for Fast Multi-to-One Stain Normalization

Hongtao Kang Die Luo Li Chen Junbo Hu Tingwei Quan Shaoqun Zeng Shenghua Cheng [email protected] Xiuli Liu [email protected] School of Biomedical Engineering and Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, Guangdong 510515, China Britton Chance Center and MoE Key Laboratory for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics-Huazhong University of Science and Technology, Wuhan, Hubei 430074, China Department of Clinical Laboratory, Tongji Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, Hubei 430030, China Department of Pathology, Hubei Maternal and Child Health Hospital, Wuhan, Hubei 430072, China
Abstract

In practice, digital pathology images are often affected by various factors, resulting in very large differences in color and brightness. Stain normalization can effectively reduce the differences in color and brightness of digital pathology images, thus improving the performance of computer-aided diagnostic systems. Conventional stain normalization methods rely on one or several reference images, but one or several images may not adequately represent the entire dataset. Although learning-based stain normalization methods are a general approach, they use complex deep networks, which not only greatly reduce computational efficiency, but also risk introducing artifacts. Some studies use specialized network structures to enhance computational efficiency and reliability, but these methods are difficult to apply to multi-to-one stain normalization due to insufficient network capacity. In this study, we introduced dynamic-parameter network and proposed a novel method for stain normalization, called ParamNet. ParamNet addresses the challenges of limited network capacity and computational efficiency by introducing dynamic parameters (weights and biases of convolutional layers) into the network design. By effectively leveraging these parameters, ParamNet achieves superior performance in stain normalization while maintaining computational efficiency. Results show ParamNet can normalize one whole slide image (WSI) of 100,000×100,000 within 25s. The code is available at: https://github.com/khtao/ParamNet.

keywords:
Stain normalization, Cytopathology, Histopathology, Convolutional neural network, Generative adversarial network.
journal: Expert Systems with Applications

1 Introduction

With the development of computer technology, computer-aided diagnosis (CAD) has received more and more attention, and many automated pathology diagnosis tools have been developed [1, 2, 3, 4]. In practice, however, the pathology images from different medical centers are often very different due to the differences in staining reagent manufacturers, staff operations, and imaging systems [5, 6]. Even within one medical center, there is some variability over time. For pathologists, small color differences will not affect the diagnosis, but larger color differences will affect even pathologists [7]. For CAD systems, even small color differences can affect their reliability [8]. Stain normalization can effectively reduce these differences by normalizing different color distributions to a uniform color distribution, thereby improving the performance of the CAD system [9, 10, 11].

Conventional stain normalization methods rely on one or several reference images to extract color mapping relationships. However, one or several images may not adequately represent the entire dataset [12, 13]. Conventional methods can be mainly divided into two categories, stain matching based methods [14] and stain-separation methods [15, 16, 5, 17, 18]. Stain matching based methods usually match the mean and standard deviation of the source image with the target image in the color space, e.g. LAB [14]. Stain-separation methods try to separate and normalize each staining channel independently in Optical Density (OD) space [15, 16, 5, 17, 18]. It is very difficult to separate each staining channel for some pathology images using multiple stains, such as cytopathology images [13, 19].

With the rise of deep learning, more and more stain normalization methods have adopted deep learning-based methods [20, 21, 22, 23, 24, 25, 26, 27, 28]. These methods can be roughly divided into two categories, one is the pix2pix-based methods [20, 21, 22, 23, 24], another is CycleGAN-based methods [25, 26, 27, 28].

The pix2pix-based methods try to reconstruct the original image from the color transformed image. These color transformation methods include grayscale [20, 21, 22], Hue-Saturation-Value (HSV) [23] and CIEL*a*b* color space [24]. However, the color variation of the source image still affects the reconstructed image. The CycleGAN-based methods generally adopt cycle consistency loss and adversarial loss to learn the color mapping from the real source and target image [25, 26, 27, 28]. By using real source and target images, the image normalized by these methods is very close to the real target image. However, in real data, due to the mismatch in the distribution of the source and target images and the instability of the GAN network, artifacts, abnormal brightness, or distortion may occur [12, 13, 29]. These phenomena are unacceptable in stain normalization. Some researchers have tried to solve this problem by using additional structures to make the source image and the normalized image consistent in content [30, 31]. These methods include using the segmentation branch to ensure semantic consistency [30] and using additional branches to ensure the content consistency[31]. However, these methods add extra branches, complicate training, and cannot really guarantee against anomalies.

Because pathology images often contain a large number of pixels, researchers hope to improve the computational efficiency of stain normalization algorithms. Anand et al. [32] optimized the Vahadane [5]. and Anghel et al. [10] optimized the Macenko [16] method. Zheng et al. [12] proposed to use 1×1 convolution and sparse routing to achieve fast stain normalization. However, the style of the normalized image by this method is formed automatically and cannot be normalized to a specific color distribution. StainNet [13] used distillation learning and 1×1 convolutional neural network to reduce the amount of normalization computation. Benefiting from its fully 1×1 convolutional network structure, StainNet [13] can retain as much information from the source image as possible. However, its overly simplified network structure limits its color mapping capabilities. When images of multiple styles need to be normalized into one style, StainNet [13] must be trained multiple times to obtain different mapping relationships, which undoubtedly increases its application cost.

In this study, we present ParamNet, a novel method to address the challenges of limited network capacity and computational efficiency by introducing dynamic parameters into the network design. These parameters are carefully designed to enhance the network’s representation capabilities, enabling it to effectively handle stain variations in histopathological and cytopathological images. ParamNet achieves superior performance in stain normalization while maintaining computational efficiency, making it a promising solution for practical applications in histopathology and cytopathology.

Our ParamNet contains two sub-networks, a modified ResNet18 [33] as the prediction sub-network and a fully 1×1 convolutional network as the color mapping sub-network, where the prediction sub-network can predict all parameters (weights and biases of convolutional layers) of the color mapping sub-network at low resolution, and the color mapping sub-network can directly perform stain normalization at the original resolution. The fully 1×1 convolutional color mapping sub-network contains only two convolutional layers with a total of 59 parameters (weights and biases of convolutional layers), which allows our network to be computationally efficient and structurally consistent. The prediction sub-network running at low resolution can automatically determine the parameters (weights and biases of convolutional layers) of the color mapping sub-network according to each input image. So, our ParamNet can perform various stain normalization tasks, including one-to-one stain normalization and multi-to-one stain normalization, in a fast and robust manner. This allows our ParamNet to normalize an 100,000×100,000 whole slide image (WSI) within 25s. Our adversarial training framework introduces a texture module between ParamNet and the discriminator, which solves the misfit between weak generators and strong discriminators. Further, we validate the effectiveness of the method on datasets containing multiple staining styles.

Refer to caption
Figure 1: Network structure of ParamNet. First, the parameter prediction sub-network predicts the parameters (weights and biases of convolutional layers) of the color mapping sub-network at low resolution. Then, the color mapping sub-network uses these parameters to normalize the input image at original resolution.

2 Methods

In this section, we introduce the network structure of ParamNet and its adversarial training framework.

2.1 Network Structure of ParamNet

Stain normalization is an important preprocessing step in CAD system. A practical stain normalization method should have the following features: 1. Fast: Pathological full-slide images usually contain a large number of pixels, so the stain normalization method needs to be computationally efficient. 2. Structural consistency: The stain normalization method should have the same structure as the source image and only change the color style. 3. Multi-domain: The stain normalization method should be able to normalize the images of multiple color styles at the same time.

StainNet [13] is a fast and robust stain normalization method, but only for one-to-one stain normalization. When normalizing multiple color styles, StainNet [13] has to be trained multiple times to obtain different network parameters (weights and biases of convolutional layers). To solve this problem, we propose ParamNet, which can automatically determine the appropriate network parameters (weights and biases of convolutional layers) for each input image.

As shown in Fig. 1, ParamNet consists of two parts, one is the parameter prediction sub-network, and another is the color mapping sub-network. The parameter prediction sub-network can predict all the parameters (weights and biases of convolutional layers) of the color map sub-network according to the color style of the input image at low resolution. The color mapping sub-network utilizes these parameters to directly normalize the source image at original resolution. The feature of dynamic parameter ensures that our network has a sufficient capability for multi-domain normalization.

In our implementation, the parameter prediction sub-network uses a modified ResNet18 [33], in which all BatchNorm layers are removed and all convolutional layers are set as biased convolutional layers. For the stability of the network, the output of the prediction sub-network will be normalized by the tanh function, and then multiplied by a learnable parameter. The learnable parameter represents the numerical range of the parameters (weights and biases of convolutional layers) in the color mapping sub-network and its initial value is set to 4.5. The input image of the parameter prediction sub-network are downsampled to 128×128 by default.

The color mapping sub-network is a fully 1×1 convolutional network, which is similar to StainNet [13]. Since only 1×1 convolutional layer is used, our color mapping sub-network only maps a single pixel without interference from the input image’s local neighborhood, thus ensuring the structural consistency of the output image. The rectified linear unit (ReLU) is used as the activation function of the convolutional layer, and tanh is used as the activation function of the output layer. Due to the dynamic network, the color mapping sub-network can use a more concise network structure. The color mapping sub-network uses two 1×1 convolutional layers by default, and the number of channels is set to 8. So, the color mapping sub-network only contains 59 parameters, which allows our network to be computationally efficient.

Refer to caption
Figure 2: Adversarial training framework of ParamNet. Images from source domain are mapped to the target domain, after which texture details are added, and the result is mapped back to the source domain. The same reverse process is also performed for the images from the target domain. The difference is that in the inverse process, the texture module uses TBT_{B} instead of TAT_{A}, and the discriminator uses DBD_{B} instead of DAD_{A}.

2.2 Adversarial Training Framework of ParamNet

Since ParamNet only transforms the color value of the source image, it retains a lot of information of the source image. In adversarial training, the discriminator can easily distinguish the difference between the image normalized by ParamNet and the real image, resulting in the failure of generating adversarial training. To solve this problem, we introduce a texture module to transform the texture of the output image of ParamNet to obtain an image that can deceive the discriminator. Further, during training, we randomly transform image resolutions (0.5\sim1.0) to stabilize learning.

As shown in Fig. 2, there are two generators (ParamNet A (GAG_{A}) and ParamNet B (GBG_{B})), two texture modules (Texture Module A (TAT_{A}) and Texture Module B (TBT_{B})) and two discriminators (Discriminator A (DAD_{A}) and Discriminator B (DBD_{B})) included in our framework. GAG_{A} is used to map the images from source domain to target domain, and GBG_{B} is used to map the images from target domain to source domain. TAT_{A} is used to add some details to the normalized image by GAG_{A}. TBT_{B} is used to add some details to the normalized image by GBG_{B}. DAD_{A} is used to distinguish the difference between the real target image and the output of TAT_{A}, and DBD_{B} is used to distinguish the difference between the real source image and the output of TBT_{B}.

There are four losses included in our framework, adversarial loss, cycle consistency loss, domain consistency loss and identity loss.

The adversarial loss is used to ensure that the outputs of the two texture modules have the same distribution as the real source and target images. The objective function is defined as:

GAN(GA,TA,DA)=𝔼tpdata(t)[DA(t)2]+𝔼spdata(s)[1DA(TA(GA(s)))2]\begin{split}\mathcal{L}_{GAN}(G_{A},T_{A},D_{A})=\mathbb{E}_{t\sim p_{data}(t)}[\|D_{A}(t)\|_{2}]\\ +\mathbb{E}_{s\sim p_{data}(s)}[\|1-D_{A}(T_{A}(G_{A}(s)))\|_{2}]\end{split} (2)
GAN(GB,TB,DB)=𝔼spdata(s)[DB(s)2]+𝔼tpdata(t)[1DB(TB(GB(t)))2]\begin{split}\mathcal{L}_{GAN}(G_{B},T_{B},D_{B})=\mathbb{E}_{s\sim p_{data}(s)}[\|D_{B}(s)\|_{2}]\\ +\mathbb{E}_{t\sim p_{data}(t)}[\|1-D_{B}(T_{B}(G_{B}(t)))\|_{2}]\end{split} (3)

where 2\|\cdot\|_{2} is the 2\ell_{2}-norm, ss is the real source image, and tt is the real target image.

The cycle consistency loss constrains the consistency of image content by reconstructing the original image from the transformed image. The objective function can be expressed as:

cyc(GA,GB,TA,TB)=𝔼spdata(s)[sGB(TA(GA(s)))1]+𝔼tpdata(t)[tGA(TB(GB(t)))1]\begin{split}\mathcal{L}_{cyc}(G_{A},G_{B},T_{A},T_{B})=\\ \mathbb{E}_{s\sim p_{data}(s)}[\|s-G_{B}(T_{A}(G_{A}(s)))\|_{1}]\\ +\mathbb{E}_{t\sim p_{data}(t)}[\|t-G_{A}(T_{B}(G_{B}(t)))\|_{1}]\end{split} (4)

where 1\|\cdot\|_{1} is the 1\ell_{1}-norm.

Domain consistency loss aims to make the texture module change the image content as little as possible, and only adds some details that can deceive the discriminator. The objective function can be expressed as:

dom(GA,GB,TA,TB)=𝔼spdata(s)[GA(s)TA(GA(s))1]+𝔼tpdata(t)[GB(t)TB(GB(t))1]\begin{split}\mathcal{L}_{dom}(G_{A},G_{B},T_{A},T_{B})=\\ \mathbb{E}_{s\sim p_{data}(s)}[\|G_{A}(s)-T_{A}(G_{A}(s))\|_{1}]\\ +\mathbb{E}_{t\sim p_{data}(t)}[\|G_{B}(t)-T_{B}(G_{B}(t))\|_{1}]\end{split} (5)

where 1\|\cdot\|_{1} is the 1\ell_{1}-norm.

Identity loss is designed such that when the input image comes from the corresponding target domain, the generator can maintain the identity map without making any changes to the image. Specifically, in this paper, we hope that when the inputs of GAG_{A} and TAT_{A} are from the target domain, GAG_{A} and TAT_{A} do not make any changes to the input. Similarly, when the inputs of GBG_{B} and TBT_{B} are from the source domain, GBG_{B} and TBT_{B} do not make any changes to the input. The objective function can be expressed as:

ide(GA,GB,TA,TB)=ide(GA,GB)+ide(TA,TB)\begin{split}\mathcal{L}_{ide}(G_{A},G_{B},T_{A},T_{B})=\\ \mathcal{L}_{ide}(G_{A},G_{B})+\mathcal{L}_{ide}(T_{A},T_{B})\end{split} (6)
ide(GA,GB)=𝔼spdata(s)[sGB(s)1]+𝔼tpdata(t)[tGA(t)1]\begin{split}\mathcal{L}_{ide}(G_{A},G_{B})=\mathbb{E}_{s\sim p_{data}(s)}[\|s-G_{B}(s)\|_{1}]\\ +\mathbb{E}_{t\sim p_{data}(t)}[\|t-G_{A}(t)\|_{1}]\end{split} (7)
ide(TA,TB)=𝔼spdata(s)[sTB(s)1]+𝔼tpdata(t)[tTA(t)1]\begin{split}\mathcal{L}_{ide}(T_{A},T_{B})=\mathbb{E}_{s\sim p_{data}(s)}[\|s-T_{B}(s)\|_{1}]\\ +\mathbb{E}_{t\sim p_{data}(t)}[\|t-T_{A}(t)\|_{1}]\end{split} (8)

In our implementation, the network structure of the texture module and discriminator uses the generator and discriminator in StainGAN [25] respectively.

3 Experiments and Results

3.1 Datasets

In this study, we used four different datasets to verify the performance of different methods, including two public datasets and two private datasets. This study was approved by the Ethics Committee of Tongji Medical College, Huazhong University of Science and Technology. The related descriptions are below:

3.1.1 Aligned Cytopathology Dataset

The aligned cytopathology dataset from StainNet [13] was used to verify the similarity between the normalized and target images. In this dataset, the same slides (Thinprep cytologic test slides from Hubei Maternal and Child Health Hospital) were scanned using two different scanners. A scanner called Scanner O was equipped with a 20x objective lens with a pixel size of 0.2930 um. The other called the Scanner T was equipped with a 40x objective lens with a pixel size of 0.1803 um. In this dataset, the images from Scanner O and Scanner T are carefully aligned. There are 2,257 image pairs in the training set and 966 image pairs in the test set as shown in Table 1. In this study, the images from Scanner O were used as the source images and the images from Scanner T were used as the target images.

3.1.2 Aligned Histopathology Dataset

The aligned histopathology dataset from StainNet [13] was used to verify the similarity between the normalized and target images. This dataset is originally from the publicly available part of the MITOS-ATYPIA ICPR’14 challenge [34]. This dataset was then aligned, registered, sampled and made publicly available by Kang et al. [13]. In this dataset, the same slides were scanned with two scanners (Aperio Scanscope XT called Scanner A and Hamamatsu Nanozoomer 2.0-HT called Scanner H). Among them, the training set contains 19,200 image pairs, and the test set contains 7,936 image pairs as shown in Table 1. In this study, the images from Scanner A were used as the source images and the images from Scanner H were used as the target images.

Table 1: The number of image pairs on the aligned cytopathology and histopathology dataset.
training set test set
The aligned cytopathology dataset 2,257 966
The aligned histopathology dataset 19,200 7,936
Table 2: The number of image patches on the cytopathology and histopathology classification dataset. The classification set is used to train and test the classifier and the stain normalization set is used to train the stain normalization methods.
The cytopathology classification dataset D1 D2 D3 D4 D5
The classification set 244,341 10,000 10,000 10,000 10,000
The stain normalization set 2,500 625 625 625 625
The histopathology classification dataset Uni16 C1 C2 C3 C4 C5
The classification set 40,000 13,613 12,355 14,076 9,235 9,696
The stain normalization set 6,000 1,200 1,200 1,200 1,200 1,200

3.1.3 Cytopathology Classification Dataset

The cytopathology classification dataset was used to verify the performance of the stain normalization algorithm on a multi-domain cytopathology classification diagnostic task. The slides in the cytopathology classification dataset are from Hubei Provincial Maternal and Child Health Hospital and Tongji Hospital. There are five different styles (D1-D5), among which D1-D4 come from Hubei Maternal and Child Health Hospital, and D5 comes from Tongji Hospital. In this dataset, the image patches that contain abnormal cells are labeled as abnormal patches, and the image patches that do not contain any abnormal cells are labeled as normal patches. There are 244,341 image patches in D1 and 40,000 image patches in D2-D5 (10,000 image patches in each center) as shown in Table 2. We randomly picked 2,500 image patches in D1 as the target images and 2,500 image patches in D2-D5 as the source images for training the stain normalization algorithm. For the classifier, we used D1 to train the classifier, and used D2-D5 to test the performance of the classifier with and without normalization.

3.1.4 Histopathology Classification Dataset

The histopathology classification datasets used the publicly available Camelyon16 [35] (399 whole slide images from two centers) and Camelyon17 [36] datasets (1,000 whole slide images from five centers). In our experiments, 100 WSIs from University Medical Center Utrecht (Uni16) in Camelyon16 training part were used to extract the training patches, and 500 WSIs from the five centers (C1-C5) in Camelyon17 training part were used to extract the test patches. In this dataset, the image patches that contain tumor cells are labeled as abnormal patches, and the image patches that do not contain any tumor cells are labeled as normal patches. There are 40,000 image patches from Uni16 in the training set of this dataset. And there are 58,975 image patches in the test set of this dataset as shown in Table 2. We randomly picked 6,000 image patches from the training set as the target images and 6,000 image patches from the test set as the source images for training stain normalization algorithms. For the classifier, we used the training set to train the classifier, and used the test set to test the performance of the classifier with and without normalization.

3.2 Evaluation Metrics

In order to evaluate different methods fairly and effectively, we considered in four aspects: similarity with target image, preservation of source image information, computational efficiency, and classifier accuracy.

In order to effectively measure the similarity to the target image and preserve the information of the source image, we used three similarity matrices, Quaternion Structural Similarity (QSSIM) [37], Structural Similarity index (SSIM) [38] and Peak Signal-to-Noise Ratio (PSNR). The similarity between the normalized image and the target image was evaluated using QSSIM, SSIM and PSNR, denoted as QSSIM Target, SSIM Target and PSNR Target. The preservation of source image information was evaluated using SSIM between the normalized image and the source image, denoted as SSIM Source. SSIM Target and PSNR Target were calculated using raw RGB values. Similar to [11,27], SSIM Source was calculated using grayscale images. The statistic results of SSIM Target, PSNR Target, and SSIM Source on the testing set in the aligned cytopathology dataset and the aligned histopathology dataset are shown in Tables 3 and 4, as mean ± standard deviation.

In order to evaluate the computational efficiency of different methods, the number of frames per second (FPS) was calculated on the system with 6-core Intel(R) Core (TM) i7-6850K CPU and NVidia GeForce GTX 1080Ti. The input and output (IO) time was not included. The FPS results of different methods are shown in Table 5 when the input image size is 512×512 pixels.

The accuracy was used to evaluate the classifier performance. The statistic results of accuracy on the cytopathology and histopathology classification datasets are shown in Tables 8 and 9, as mean ± standard deviation.

Refer to caption
Figure 3: Visual comparison among different methods on the aligned cytopathology dataset.

3.3 Implementation

For conventional methods (Reinhard [14], Macenko [16], Vahadane [5] and GCTI-SN [18]), one professionally selected image was used as the reference image. For the GAN based methods (SAASN [28], StainGAN [25], and StainNet [13]), we used their default training settings.

For ParamNet, four losses are included in the training framework. During training, we randomly transformed image resolutions (0.5\sim1.0) to stabilize learning. We used Adam [39] to optimize the network, and the learning rate is linearly increased from 0 to 2e-4 in the first 1,000 iterations, and then linearly decreased to 0 until the end of training. We trained ParamNet for 200k iterations on the aligned cytopathology dataset and aligned histopathology dataset, and 400k iterations on the cytopathology classification dataset and histopathology classification dataset.

For the classifier, we used SqueezeNet [40] pre-trained on ImageNet [41] as the classification network. The classifier was trained using cross entropy loss and Adam [39] optimizer. The initial learning rate was set to 0.0002, and it reduced by a factor of 0.1 at the 40th and 50th epoch. The training was stopped at the 60th epoch, which was chosen experimentally. The experiment was repeated 10 times in order to enhance reliability.

3.4 Results

In this section, we compared our method with state-of-art normalization methods, including Reinhard [14], Macenko [16], Vahadane [5], GCTI-SN [18], SAASN [28], StainGAN [25], and StainNet [13].

3.4.1 One-to-One Stain Normalization

Refer to caption
Figure 4: Visual comparison among different methods on the aligned histopathology dataset.
Table 3: Different evaluation metrics are reported for various stain normalization methods on the aligned cytopathology dataset.
Methods QSSIM Target SSIM Target PSNR Target SSIM Source
Source 0.825±0.027 0.765±0.028 16.9±1.89 1.000±0.000
Reinhard 0.814±0.051 0.739±0.046 19.8±3.3 0.885±0.042
Macenko 0.778±0.064 0.731±0.054 22.5±3.1 0.853±0.054
Vahadane 0.784±0.059 0.739±0.050 22.6±3.0 0.867±0.050
GCTI-SN 0.832±0.039 0.764±0.039 19.7±2.7 0.905±0.034
SAASN 0.855±0.025 0.778±0.026 21.5±2.0 0.990±0.001
StainGAN 0.832±0.029 0.764±0.030 29.7±1.6 0.905±0.021
StainNet 0.874±0.027 0.809±0.027 29.8±1.7 0.945±0.025
ParamNet 0.878±0.026 0.816±0.026 30.0±1.6 0.951±0.023
Table 4: Different evaluation metrics are reported for various stain normalization methods on the aligned histopathology dataset.
Methods QSSIM Target SSIM Target PSNR Target SSIM Source
Source 0.690±0.099 0.641±0.108 20.3±3.2 1.000±0.000
Reinhard 0.682±0.092 0.617±0.106 19.9±2.1 0.964±0.031
Macenko 0.703±0.101 0.656±0.115 20.7±2.7 0.966±0.049
Vahadane 0.708±0.102 0.664±0.116 21.1±2.8 0.967±0.046
GCTI-SN 0.685±0.093 0.638±0.100 19.7±2.2 0.947±0.043
SAASN 0.739±0.098 0.690±0.114 21.6±3.5 0.994±0.002
StainGAN 0.733±0.088 0.706±0.099 22.7±2.6 0.912±0.025
StainNet 0.724±0.097 0.691±0.107 22.5±3.3 0.957±0.007
ParamNet 0.720±0.098 0.687±0.106 22.7±3.0 0.966±0.001
Table 5: Quantitative comparison of computational efficiency among different methods. The input and output (IO) time is not included. The input image size is 512×512 pixels.
Methods Reinhard Macenko Vahadane GCTI-SN SAASN StainGAN StainNet ParamNet
FPS 54.8 4.0 0.5 6.4 5.2 19.6 881.8 1605.2

We compared our ParamNet with state-of-art normalization methods on the one-to-one stain normalization task. We reported the following results on the aligned cytopathology dataset and the aligned histopathology dataset: (1) Visual comparison among different methods. (2) Quantitative comparison of similarities among different methods. (3) Quantitative comparison of computational efficiency among different methods.

Visual comparison

Firstly, we compared our ParamNet with other stain normalization methods in the visual appearance. The visual comparison among different methods is shown in Fig. 3 and 4.

The results on the aligned cytopathology dataset are shown in Fig. 3. For conventional methods (Reinhard, Macenko, Vahadane and GCTI-SN), the normalized images are far from the target image in color and brightness. The conventional methods only extract information from a single reference image. Moreover, it is difficult to extract the precise color mapping relationship in this way, resulting in poor performance of the normalized results. For the image normalized by SAASN, its brightness is closer to the source image than to the target image. The images normalized by StainGAN, StainNet and ParamNet are very close to the target image. However, in terms of image content, the image normalized by StainGAN has a more complex background, while the image StainNet has a simpler background, only the image normalized by ParamNet is closest to the source image. StainGAN uses a complex convolutional network that tends to introduce additional noise. StainNet uses a single 1×1 convolutional network, which makes it not capable enough to handle complex color mapping relationships. ParamNet, on the other hand, can automatically determine the best parameters (weights and biases of convolutional layers) for each input image, thus achieving better performance.

The results on the aligned histopathology dataset are shown in Fig. 4. For conventional methods (Reinhard, Macenko, Vahadane and GCTI-SN), there are still some differences between the normalized image and the target image in color and brightness. However, the performance of conventional methods on aligned histopathology datasets is significantly better than that on aligned cytopathology datasets. This is because most of the conventional methods are designed for histopathology images and are not suitable for cytopathology images. For the image normalized by SAASN, its color is lighter to the target image. StainGAN introduces blur to the normalized image. The image normalized by StainNet is darker than the target image. Only the image normalized by ParamNet is most close to the target image.

Quantitative comparison of similarities

Secondly, we quantitatively compared our ParamNet with other stain normalization methods in the similarity to the target image and the similarity to the source image. The quantitative comparison among different methods is shown in Table 3 and 4.

The results on the aligned cytopathology dataset are shown in Table 3. The conventional methods (Reinhard, Macenko, Vahadane and GCTI-SN) have low QSSIM Target, SSIM Target and PSNR Target, which means that there is a large difference between the images normalized by the conventional methods and the target images. SAASN has the highest SSIM Source but low PSNR Target, which means that it preserves the information of the source images as much as possible but ignores the similarity with the target images. But the purpose of stain normalization is to be similar to the target images, so SAASN is inappropriate for stain normalization. StainGAN, StainNet and ParamNet have high PSNR Target, which means that the images normalized by them are similar enough to the target images. Among these three methods, StainGAN has the lowest SSIM Source, which means it may lose some information of the source images. For our ParamNet, it has the highest PSNR Target and also has a high SSIM Source, which means that it can achieve a high similarity to the target images while preserving the source images information as much as possible.

The results on the aligned histopathology dataset are shown in Table 4. The conventional methods (Reinhard, Macenko, Vahadane and GCTI-SN) have lower QSSIM Target, SSIM Target and PSNR Target than the GAN based methods (SAASN, StainGAN, StainNet and ParamNet) in Table 4, which means that the GAN based methods can achieve better performance. Among the GAN based methods, SAASN has the highest SSIM Source and the lowest PSNR Target, StainGAN and ParamNet have the highest PSNR Target, and ParamNet has the second highest SSIM Source. So, our ParamNet can achieve a high similarity to the target images while preserving the source images information as much as possible on the aligned histopathology dataset.

Quantitative comparison of computational efficiency

Finally, we quantitatively compared the computational efficiency of our method with other stain normalization methods. We tested the FPS of different methods with an input image size of 512×512 pixels and show the results in Table 5. As shown in Table 5, our ParamNet (1,605.2 FPS) is much faster than StainNet (881.8 FPS), which means that our ParamNet can normalize an 100,000×100,000 whole slide image (WSI) within 25s. The prediction sub-network of ParamNet computes at low resolution and the color mapping sub-network of ParamNet computes at high resolution. The sub-network has only two 1×1 convolutional layers with 8 channels, however, StainNet has three 1×1 convolutional layers with 32 channels. So our ParamNet is computationally more efficient than StainNet.

3.4.2 Multi-to-One Stain Normalization

Refer to caption
Figure 5: Visual comparison among different methods on the cytopathology classification dataset. The source images have multiple styles (D2-D5), and the target images are from D1. Red boxes mark obviously abnormal normalized results.
Table 6: SSIM Source for various stain normalization methods on the cytopathology classification dataset.
Methods D2 D3 D4 D5 Average
Reinhard 0.914±0.125 0.872±0.198 0.811±0.171 0.815±0.193 0.853±0.172
Macenko 0.866±0.169 0.857±0.230 0.766±0.203 0.727±0.192 0.804±0.198
Vahadane 0.848±0.189 0.850±0.247 0.751±0.205 0.715±0.199 0.791±0.210
GCTI-SN 0.905±0.175 0.901±0.199 0.893±0.150 0.878±0.181 0.894±0.176
SAASN 0.992±0.003 0.992±0.002 0.986±0.006 0.983±0.006 0.988±0.004
StainGAN 0.954±0.035 0.904±0.077 0.896±0.062 0.874±0.051 0.907±0.056
StainNet 0.941±0.032 0.968±0.012 0.881±0.040 0.898±0.029 0.922±0.028
ParamNet 0.993±0.010 0.994±0.001 0.989±0.006 0.987±0.008 0.991±0.006

Firstly, we compared our ParamNet with other stain normalization methods in the visual appearance as shown in Fig. 5 and 6.

The results on the cytopathology classification dataset are shown in Fig. 5. As can be seen, there are many wrong variations in color and brightness on the normalized images by the conventional methods (Reinhard, Macenko, Vahadane and GCTI-SN). For example, the cell in the source image of D2 is incorrectly normalized from dark red to bright red by GCTI-SN, and the brightness of the D5’s source image is over-enhanced by Reinhard and GCTI-SN. StainGAN and SAASN use complex convolutional networks for stain normalization, which not only reduces the computational efficiency of the algorithm but also brings artifacts to the normalized image. For example, SAASN and StainGAN add extra contents to the normalized images that do not belong to the source images in D4. And StainGAN removes some contents belonging to the source image in the normalized image of D5. As for StainNet, it has a single three-layer 1×1 convolutional network, which limits its color-mapping capability. So, there are many details missing, especially on the background of the normalized image by StainNet. Unlike StainNet, our ParamNet uses a dynamic parameter 1×1 convolutional network for color mapping, so our ParamNet performs well on the multi-to-one stain normalization task.

The results on the histopathology classification dataset are shown in Fig. 6. As can be seen, the conventional methods perform better on histopathology images, but there are still many wrong variations in color and brightness. For example, the light background in the source image of center3 is incorrectly normalized to the purple background by Reinhard and GCTI-SN, and the brightness of the center5’s source image is over-enhanced by Macenko and Vahadane. The GAN based methods all show good performance. However, the normalized images by StainGAN are a little blurry and the normalized images by StainNet are a bit purple. Our ParamNet can achieve the best performance among the GAN based methods.

Secondly, we compared our ParamNet with other stain normalization methods using SSIM Source as shown in Table 6 and 7.

Whether on the histopathology classification dataset or the cytopathology classification dataset, our ParamNet obtained the highest SSIM Source, which shows that our ParamNet retains the source image information best on the two datasets. StainNet uses a fixed parameter 1×1 convolutional network to perform normalization, whereas our ParamNet uses a dynamic parameter 1×1 convolutional network. The feature of dynamic parameters enable our network to use the best network parameters for each input image, so as to better retain the information of the source image. This is why the SSIM Source of our method can be better than StainNet.

Refer to caption
Figure 6: Visual comparison among different methods on the histopathology classification dataset. The source images are from five centers (C1-C5) of Camelyon17, and the target images are from Uni16 of Camelyon16. The source images of C3 and the target images are from same center but at different times. Red boxes mark obviously abnormal normalized results.
Table 7: SSIM Source for various stain normalization methods on the histopathology classification dataset.
Methods C1 C2 C3 C4 C5 Average
Reinhard 0.852±0.079 0.907±0.079 0.881±0.088 0.863±0.101 0.726±0.099 0.846±0.089
Macenko 0.902±0.084 0.722±0.152 0.939±0.064 0.918±0.085 0.803±0.092 0.857±0.096
Vahadane 0.930±0.078 0.843±0.159 0.941±0.065 0.931±0.081 0.819±0.087 0.893±0.094
GCTI-SN 0.835±0.148 0.887±0.103 0.899±0.076 0.867±0.127 0.726±0.095 0.843±0.110
SAASN 0.974±0.039 0.981±0.013 0.989±0.005 0.975±0.018 0.953±0.017 0.974±0.018
StainGAN 0.933±0.047 0.952±0.019 0.966±0.019 0.928±0.037 0.869±0.053 0.930±0.035
StainNet 0.945±0.039 0.947±0.016 0.886±0.031 0.947±0.018 0.898±0.017 0.925±0.024
ParamNet 0.974±0.038 0.988±0.011 0.983±0.019 0.976±0.023 0.958±0.066 0.976±0.032

3.4.3 Application results on the classification task

Table 8: Accuracy for various stain normalization methods on the cytopathology classification dataset.
Accuracy D2 D3 D4 D5 Average
Original 0.853±0.031 0.915±0.022 0.873±0.029 0.688±0.033 0.832±0.029
Reinhard 0.862±0.015 0.867±0.025 0.796±0.026 0.728±0.025 0.813±0.023
Macenko 0.929±0.014 0.937±0.023 0.915±0.010 0.818±0.034 0.900±0.020
Vahadane 0.905±0.022 0.940±0.031 0.907±0.019 0.745±0.037 0.875±0.026
GCTI-SN 0.932±0.016 0.944±0.022 0.924±0.017 0.850±0.039 0.912±0.024
SAASN 0.964±0.013 0.911±0.024 0.944±0.010 0.863±0.039 0.920±0.022
StainGAN 0.918±0.007 0.908±0.011 0.867±0.006 0.799±0.018 0.873±0.011
StainNet 0.928±0.034 0.963±0.014 0.924±0.032 0.820±0.059 0.909±0.035
ParamNet 0.968±0.013 0.965±0.017 0.943±0.006 0.867±0.039 0.936±0.019
Table 9: Accuracy for various stain normalization methods on the histopathology classification dataset.
Accuracy C1 C2 C3 C4 C5 Average
Original 0.500±0.016 0.561±0.030 0.907±0.003 0.528±0.014 0.765±0.041 0.652±0.021
Reinhard 0.871±0.014 0.855±0.008 0.815±0.006 0.823±0.027 0.808±0.015 0.834±0.014
Macenko 0.814±0.009 0.832±0.009 0.816±0.007 0.836±0.012 0.789±0.010 0.817±0.009
Vahadane 0.817±0.005 0.826±0.005 0.762±0.007 0.890±0.004 0.805±0.003 0.820±0.005
GCTI-SN 0.840±0.006 0.832±0.007 0.816±0.007 0.879±0.005 0.804±0.006 0.834±0.006
SAASN 0.883±0.004 0.887±0.002 0.908±0.003 0.888±0.004 0.819±0.010 0.877±0.005
StainGAN 0.898±0.004 0.871±0.001 0.921±0.003 0.903±0.004 0.881±0.005 0.895±0.003
StainNet 0.898±0.006 0.836±0.006 0.895±0.005 0.862±0.019 0.877±0.005 0.874±0.008
ParamNet 0.920±0.003 0.857±0.007 0.924±0.003 0.899±0.004 0.884±0.003 0.897±0.004

Moreover, we compared our ParamNet with other stain normalization methods on the classification task. The accuracy for various stain normalization methods is shown in Table 8,9.

The accuracy for various stain normalization methods on the cytopathology classification dataset is shown in Table 8. The classifier has poor performance on the original image of D2-D5, and the classifier has different performance improvements on the images normalized by different methods (except Reinhard). Among conventional methods, the classifier has the best accuracy (0.912 on average) on the images normalized by GCTI-SN. And among GAN based methods, the classifier has the best accuracy (0.936 on average) on the images normalized by our ParamNet. StainGAN and StainNet do not get competitive performance on the cytopathology classification dataset.

Refer to caption
Figure 7: GAN loss during training with different strategies. (a) The baseline framework’s stability of learning is poor. (b) The stability of learning is enhanced after the texture module is introduced in the framework. (c) The combination of texture module and random scale can effectively enhance the stability of learning.
Table 10: Accuracy of models with different tricks on the histopathology classification dataset.
Model C1 C2 C3 C4 C5 Average
Baseline 0.848±0.012 0.884±0.005 0.780±0.025 0.785±0.027 0.817±0.028 0.823±0.019
+Texture Module 0.921±0.004 0.860±0.006 0.811±0.019 0.900±0.004 0.848±0.010 0.868±0.009
+Random Scale 0.917±0.003 0.855±0.006 0.859±0.009 0.899±0.006 0.846±0.014 0.875±0.008
+Identity Loss 0.916±0.003 0.857±0.006 0.899±0.004 0.901±0.003 0.885±0.003 0.892±0.004
+Domain Consistency Loss 0.920±0.003 0.857±0.006 0.924±0.003 0.899±0.004 0.884±0.003 0.897±0.004

The accuracy for various stain normalization methods on the histopathology classification dataset is shown in Table 9. The classifier has poor performance on the original images of C1-C5, even the accuracy of the classifier is close to 0.5 in C1, C2 and C4. Overall, using the normalized images can improve the performance of the classifier compared to the original images. On the histopathology classification dataset, the classifier has better performance on images normalized by the GAN-based methods than the conventional methods. Among all the normalization methods in Table 9, our method has the highest accuracy (0.897 on average). The images in C3 and Uni16 are from the same medical center but at different times. For the classifier trained on Uni16, the accuracy is 0.907 on the original images of C3. And the accuracy can be increased to 0.924 on the normalized images by our ParamNet, which shows that our method can improve the accuracy of the classifier, even on images from the same center.

3.4.4 Ablation study

To verify the effectiveness of our proposed training framework, an ablation study was performed on the histopathology classification dataset. The accuracy of models trained with different strategies is shown in Table 10. In addition, we also visualize the changes in GAN loss during the training process as shown in Fig. 7. During the training, GAN loss is recorded every 100 iterations.

As can be seen from Fig. 7 (a), for the baseline model, the discriminator loss is almost 0, which shows that the discriminator can almost completely distinguish between real images and fake images. But for the least squares loss in our training framework, it is difficult for the discriminator to provide effective gradients to optimize the generator in this case. In order to solve this problem, a texture module is first introduced between the generator and the discriminator. As can be seen from Fig. 7 (b), although it helps to solve the problem of imbalance between the generator and the discriminator, there is still a slight imbalance. After introducing the texture module, we introduce random scales during the training process to make the discriminator focus more on the overall color difference. As can be seen from Fig. 7 (c), the combined use of random scale and texture modules can well solve the problem of imbalance between the generator and the discriminator.

The accuracy of models trained using different strategies on the histopathology classification dataset is given in Table 10, where the average accuracy of the baseline model is only 0.823. The texture modules and random scale are used to solve the problem of generator and discriminator imbalance. Combining the texture module with random scales increases the average accuracy to 0.875. Identity loss is used to enhance domain-invariant features, and the accuracy can be improved by 0.017. Domain consistency loss can enhance the consistency of the output of the generator and texture module, and can improve the accuracy from 0.892 to 0.897. Through the combination of the four strategies proposed in this article, the accuracy is finally increased from 0.823 to 0.897.

4 Discussion and Conclusion

In this paper, we proposed a novel stain normalization network, ParamNet, and an adversarial training framework. ParamNet contains two sub-networks, a modified ResNet18 [33] as the prediction sub-network and a fully 1×1 convolutional network as the color mapping sub-network, where the prediction sub-network can predict all parameters (weights and biases of convolutional layers) of the color mapping sub-network at a low resolution and the color mapping sub-network can directly normalize the source images at the original resolution. Moreover, we validated the effectiveness of our method on four datasets, the results show that ParamNet has better performance than other methods.

The feature of dynamic parameter ensures that our network has a sufficient capability for multi-domain normalization. And the color mapping sub-network has an extremely concise network structure, which allows our network to be computationally efficient and structurally consistent.

Acknowledgments

This work is supported by the NSFC project (grant 62375100, 62201221), China Postdoctoral Science Foundation (grant 2021M701320, 2022T150237) and Southern Medical University Research Start-up Foundation.

References

  • Tang et al. [2009] J. Tang, R. M. Rangayyan, J. Xu, I. E. Naqa, Y. Yang, Computer-aided detection and diagnosis of breast cancer with mammography: Recent advances, IEEE Transactions on Information Technology in Biomedicine 13 (2009) 236–251. doi:10.1109/TITB.2008.2009441.
  • Zhang et al. [2014] Z. Zhang, R. Srivastava, H. Liu, X. Chen, L. Duan, D. W. Kee Wong, C. K. Kwoh, T. Y. Wong, J. Liu, A survey on computer aided diagnosis for ocular diseases, BMC Medical Informatics and Decision Making 14 (2014) 1–29.
  • Cheng et al. [2021] S. Cheng, S. Liu, J. Yu, G. Rao, Y. Xiao, W. Han, W. Zhu, X. Lv, N. Li, J. Cai, et al., Robust whole slide image analysis for cervical cancer screening using deep learning, Nature Communications 12 (2021) 5639.
  • Jo et al. [2021] S. H. Jo, Y. Kim, Y. B. Lee, S. S. Oh, J.-r. Choi, A comparative study on machine learning-based classification to find photothrombotic lesion in histological rabbit brain images, Journal of Innovative Optical Health Sciences 14 (2021) 2150018.
  • Vahadane et al. [2016] A. Vahadane, T. Peng, A. Sethi, S. Albarqouni, L. Wang, M. Baust, K. Steiger, A. M. Schlitter, I. Esposito, N. Navab, Structure-preserving color normalization and sparse stain separation for histological images, IEEE Transactions on Medical Imaging 35 (2016) 1962–1971.
  • Bentaieb and Hamarneh [2018] A. Bentaieb, G. Hamarneh, Adversarial stain transfer for histopathology image analysis, IEEE Transactions on Medical Imaging 37 (2018) 792–802. doi:10.1109/TMI.2017.2781228.
  • Ismail et al. [1989] S. M. Ismail, A. B. Colclough, J. S. Dinnen, D. Eakins, D. M. Evans, E. Gradwell, J. P. O’Sullivan, J. M. Summerell, R. G. Newcombe, Observer variation in histopathological diagnosis and grading of cervical intraepithelial neoplasia, BMJ 298 (1989) 707–710. URL: https://www.bmj.com/content/298/6675/707. doi:10.1136/bmj.298.6675.707. arXiv:https://www.bmj.com/content/298/6675/707.full.pdf.
  • Ciompi et al. [2017] F. Ciompi, O. Geessink, B. E. Bejnordi, G. S. de Souza, A. Baidoshvili, G. Litjens, B. van Ginneken, I. Nagtegaal, J. van der Laak, The importance of stain normalization in colorectal tissue classification with convolutional networks, in: 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017), 2017, pp. 160–163. doi:10.1109/ISBI.2017.7950492.
  • Cai et al. [2019] S. Cai, Y. Xue, Q. Gao, M. Du, G. Chen, H. Zhang, T. Tong, Stain style transfer using transitive adversarial networks, in: F. Knoll, A. Maier, D. Rueckert, J. C. Ye (Eds.), Machine Learning for Medical Image Reconstruction, Springer International Publishing, Cham, 2019, pp. 163–172.
  • Anghel et al. [2019] A. Anghel, M. Stanisavljevic, S. Andani, N. Papandreou, J. H. Rüschoff, P. Wild, M. Gabrani, H. Pozidis, A high-performance system for robust stain normalization of whole-slide images in histopathology, Frontiers in Medicine 6 (2019) 193.
  • Khan et al. [2014] A. M. Khan, N. Rajpoot, D. Treanor, D. Magee, A nonlinear mapping approach to stain normalization in digital histopathology images using image-specific color deconvolution, IEEE transactions on Biomedical Engineering 61 (2014) 1729–1738.
  • Zheng et al. [2020] Y. Zheng, Z. Jiang, H. Zhang, F. Xie, D. Hu, S. Sun, J. Shi, C. Xue, Stain standardization capsule for application-driven histopathological image normalization, IEEE Journal of Biomedical and Health Informatics 25 (2020) 337–347.
  • Kang et al. [2021] H. Kang, D. Luo, W. Feng, S. Zeng, T. Quan, J. Hu, X. Liu, Stainnet: a fast and robust stain normalization network, Frontiers in Medicine 8 (2021) 746307.
  • Reinhard et al. [2001] E. Reinhard, M. Adhikhmin, B. Gooch, P. Shirley, Color transfer between images, IEEE Computer Graphics and Applications 21 (2001) 34–41.
  • Ruifrok et al. [2001] A. C. Ruifrok, D. A. Johnston, et al., Quantification of histochemical staining by color deconvolution, Analytical and Quantitative Cytology and Histology 23 (2001) 291–299.
  • Macenko et al. [2009] M. Macenko, M. Niethammer, J. S. Marron, D. Borland, J. T. Woosley, X. Guan, C. Schmitt, N. E. Thomas, A method for normalizing histology slides for quantitative analysis, in: 2009 IEEE International Symposium on Biomedical Imaging: from nano to macro, IEEE, 2009, pp. 1107–1110.
  • Salvi et al. [2020] M. Salvi, N. Michielli, F. Molinari, Stain color adaptive normalization (scan) algorithm: Separation and standardization of histological stains in digital pathology, Computer Methods and Programs in Biomedicine 193 (2020) 105506.
  • Gupta et al. [2020] A. Gupta, R. Duggal, S. Gehlot, R. Gupta, A. Mangal, L. Kumar, N. Thakkar, D. Satpathy, Gcti-sn: Geometry-inspired chemical and tissue invariant stain normalization of microscopic medical images, Medical Image Analysis 65 (2020) 101788.
  • Gill [2012] G. Gill, Cytopreparation: principles & practice, volume 12, Springer Science & Business Media, 2012.
  • Cho et al. [2017] H. Cho, S. Lim, G. Choi, H. Min, Neural stain-style transfer learning using gan for histopathological images, ArXiv Preprint arXiv:1710.08543 (2017).
  • Chen et al. [2021] X. Chen, J. Yu, S. Cheng, X. Geng, S. Liu, W. Han, J. Hu, L. Chen, X. Liu, S. Zeng, An unsupervised style normalization method for cytopathology images, Computational and Structural Biotechnology Journal 19 (2021) 3852–3863.
  • Salehi and Chalechale [2020] P. Salehi, A. Chalechale, Pix2pix-based stain-to-stain translation: A solution for robust stain normalization in histopathology images analysis, in: 2020 International Conference on Machine Vision and Image Processing (MVIP), IEEE, 2020, pp. 1–7.
  • Tellez et al. [2019] D. Tellez, G. Litjens, P. Bándi, W. Bulten, J.-M. Bokhorst, F. Ciompi, J. Van Der Laak, Quantifying the effects of data augmentation and stain color normalization in convolutional neural networks for computational pathology, Medical image analysis 58 (2019) 101544.
  • Zanjani et al. [2018] F. G. Zanjani, S. Zinger, B. E. Bejnordi, J. A. van der Laak, P. H. de With, Stain normalization of histopathology images using generative adversarial networks, in: 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), IEEE, 2018, pp. 573–577.
  • Shaban et al. [2019] M. T. Shaban, C. Baur, N. Navab, S. Albarqouni, Staingan: Stain style transfer for digital histological images, in: 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), IEEE, 2019, pp. 953–956.
  • Zhou et al. [2019] N. Zhou, D. Cai, X. Han, J. Yao, Enhanced cycle-consistent generative adversarial network for color normalization of h&e stained images, in: Medical Image Computing and Computer Assisted Intervention–MICCAI 2019, Springer, 2019, pp. 694–702.
  • Zhu et al. [2017] J.-Y. Zhu, T. Park, P. Isola, A. A. Efros, Unpaired image-to-image translation using cycle-consistent adversarial networks, in: Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 2223–2232.
  • Shrivastava et al. [2021] A. Shrivastava, W. Adorno, Y. Sharma, L. Ehsan, S. A. Ali, S. R. Moore, B. Amadi, P. Kelly, S. Syed, D. E. Brown, Self-attentive adversarial stain normalization, in: Pattern Recognition. ICPR International Workshops and Challenges, Springer, 2021, pp. 120–140.
  • Lei et al. [2020] G. Lei, Y. Xia, D.-H. Zhai, W. Zhang, D. Chen, D. Wang, Staincnns: An efficient stain feature learning method, Neurocomputing 406 (2020) 267–273.
  • Mahapatra et al. [2020] D. Mahapatra, B. Bozorgtabar, J.-P. Thiran, L. Shao, Structure preserving stain normalization of histopathology images using self supervised semantic guidance, in: Medical Image Computing and Computer Assisted Intervention–MICCAI 2020, Springer, 2020, pp. 309–319.
  • Lahiani et al. [2019] A. Lahiani, N. Navab, S. Albarqouni, E. Klaiman, Perceptual embedding consistency for seamless reconstruction of tilewise style transfer, in: Medical Image Computing and Computer Assisted Intervention–MICCAI 2019, Springer, 2019, pp. 568–576.
  • Anand et al. [2019] D. Anand, G. Ramakrishnan, A. Sethi, Fast gpu-enabled color normalization for digital pathology, in: 2019 International Conference on Systems, Signals and Image Processing (IWSSIP), IEEE, 2019, pp. 219–224.
  • He et al. [2016] K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, in: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, 2016, pp. 770–778.
  • Roux [2014] L. Roux, Detection of mitosis and evaluation of nuclear atypia score in breast cancer histological images, in: 22nd International Conference on Pattern Recognition (ICPR 2014), 2014.
  • Bejnordi et al. [2017] B. E. Bejnordi, M. Veta, P. J. Van Diest, B. Van Ginneken, N. Karssemeijer, G. Litjens, J. A. Van Der Laak, M. Hermsen, Q. F. Manson, M. Balkenhol, et al., Diagnostic assessment of deep learning algorithms for detection of lymph node metastases in women with breast cancer, JAMA 318 (2017) 2199–2210.
  • Bandi et al. [2018] P. Bandi, O. Geessink, Q. Manson, M. Van Dijk, M. Balkenhol, M. Hermsen, B. E. Bejnordi, B. Lee, K. Paeng, A. Zhong, et al., From detection of individual metastases to classification of lymph node status at the patient level: the camelyon17 challenge, IEEE Transactions on Medical Imaging 38 (2018) 550–560.
  • Kolaman and Yadid-Pecht [2012] A. Kolaman, O. Yadid-Pecht, Quaternion structural similarity: A new quality index for color images, IEEE Transactions on Image Processing 21 (2012) 1526–1536. doi:10.1109/TIP.2011.2181522.
  • Wang et al. [2004] Z. Wang, A. C. Bovik, H. R. Sheikh, E. P. Simoncelli, Image quality assessment: from error visibility to structural similarity, IEEE Transactions on Image Processing 13 (2004) 600–612.
  • Kingma and Ba [2014] D. P. Kingma, J. Ba, Adam: A method for stochastic optimization, ArXiv Preprint arXiv:1412.6980 (2014).
  • Iandola et al. [2016] F. N. Iandola, S. Han, M. W. Moskewicz, K. Ashraf, W. J. Dally, K. Keutzer, Squeezenet: Alexnet-level accuracy with 50x fewer parameters and¡ 0.5 mb model size, ArXiv Preprint arXiv:1602.07360 (2016).
  • Deng et al. [2009] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, L. Fei-Fei, Imagenet: A large-scale hierarchical image database, in: 2009 IEEE Conference on Computer Vision and Pattern Recognition, IEEE, 2009, pp. 248–255.