11email: {kaminyouho,vincentwu,uno}@aetherai.com
https://www.aetherai.com/
Ultra-high-resolution unpaired stain transformation via Kernelized Instance Normalization
Abstract
While hematoxylin and eosin (H&E) is a standard staining procedure, immunohistochemistry (IHC) staining further serves as a diagnostic and prognostic method. However, acquiring special staining results requires substantial costs. Hence, we proposed a strategy for ultra-high-resolution unpaired image-to-image translation: Kernelized Instance Normalization (KIN), which preserves local information and successfully achieves seamless stain transformation with constant GPU memory usage. Given a patch, corresponding position, and a kernel, KIN computes local statistics using convolution operation. In addition, KIN can be easily plugged into most currently developed frameworks without re-training. We demonstrate that KIN achieves state-of-the-art stain transformation by replacing instance normalization (IN) layers with KIN layers in three popular frameworks and testing on two histopathological datasets. Furthermore, we manifest the generalizability of KIN with high-resolution natural images. Finally, human evaluation and several objective metrics are used to compare the performance of different approaches. Overall, this is the first successful study for the ultra-high-resolution unpaired image-to-image translation with constant space complexity. Code is available at: https://github.com/Kaminyou/URUST.
Keywords:
Unpaired image-to-image translation, ultra-high-resolution, stain transformation, whole slide image1 Introduction

Histological staining, highlighting cellular components with dyes, is crucial in clinical diagnosis [1], which enables visualization of cells and extracellular matrix and abnormal identification. Since specific cellular components or biomarkers can be distinguished when particular dyes attach specific molecules in tissues, different staining methods are applied to diagnose various diseases and their subtypes [15, 11, 4]. The standard stain (or routine stain) is hematoxylin and eosin (H&E). While hematoxylin stains nuclei, eosin can stain cytoplasm. Immunohistochemistry (IHC) protocol is further developed to detect the presence of specific protein markers. For example, Ki-67 and ER staining can quantify the presence of Ki-67 and ER biomarkers, respectively. In clinical practice, high Ki-67 expression is considered a poor prognostic factor [23], while the presence of ER indicates the suitability of choosing specific target therapies that benefit related disease subtypes [26].
However, compared with H&E staining, the IHC staining process is much more expensive and requires extra biopsies, which are limited materials. With the development of deep learning-based image-to-image translation, virtually translating H&E into different IHC staining can be achieved. For example, de Haan et al. . proposed a supervised deep learning approach via CycleGAN [33] to transform stained images from H&E to Masson’s Trichrome (MT) staining [13]. While supervised pair-wise training is desirable, this approach requires perfectly paired staining images, necessitating de-staining and re-staining processes, and is not practically efficient in a clinical scenario. Most datasets are composed of unpaired H&E and IHC images from consecutive sections. Several methodologies have been proposed and successfully tackled the unpaired image-to-image translation problem [27]. Regardless of the astonishing performance, the existing methods are limited to low-resolution images and rarely explore the images with ultra-high-resolution. In histopathology, whole slide images (WSIs) are usually larger than pixels. The main challenge of transforming a WSI is the limitation of GPU memory capacity. Patch-wise inference with an assembly process can tackle ultra-high-resolution image-to-image translation, but tiling artifacts between adjacent patches would be a critical problem. Traditionally, overlapping windows have been leveraged to smooth the transitions but have limited effectiveness. Considering the mean and standard deviation calculated in instance normalization (IN) layers might influence hue and contrast, recently, Chen et al. developed Thumbnail Instance Normalization (TIN) for both ultra-high-resolution style transfer as well as image-to-image translation tasks [8]. Unfortunately, while their approach could overcome the resolution limitation, their erroneous assumption that all patches share global mean and standard deviation would lead to dramatically over/under-colorizing according to our comprehensive experiment, which is confirmed in Section 5.3.
To compensate for all the above limitations, we proposed a Kernelized Instance Normalization (KIN) layer that can replace the original IN layer during the inference process without re-training the models. With the help of KIN, images with arbitrary resolution can be translated with constant GPU memory space (as demonstrated in Fig. 1 and LABEL:fig:inference_usage). Moreover, utilizing the statistics of neighboring patches instead of global ones like TIN, our approach can further preserve the hue and contrast locally, which is especially paramount in stain transformation tasks. Besides the translation of H&E to four IHC staining, we additionally demonstrated the generalizability of KIN with natural images by translating summer to autumn style. Our novel contribution can be summarized as follows:
-
•
To the best of our knowledge, this is the first successful study for the ultra-high-resolution unpaired image-to-image translation with constant space complexity (GPU memory), which manifests state-of-the-art outcomes in stain transformation and can also be generalized to natural images.
- •
-
•
With the KIN module, local contrast and hue information in translated images can be well preserved. Besides, different kernels can be further applied to subtly adjust the translated images.
2 Related works
2.1 Unpaired image-to-image translation
Several frameworks have been proposed for unpaired image-to-image translation. CycleGAN [33], DiscoGAN [19], and DualGAN [31] were first presented to overcome the supervised pairing constraint via cycle consistency. However, subtle information was forced to be retained in the translated image to achieve better reconstruction, causing detrimental effects when two domains are substantially different such as dog-to-cat translation. Besides, a reverse mapping function might not always exist, which inevitably leads to artifacts in translated images. Recently, strategies beyond cyclic loss have been developed to reach one-sided unpaired image-to-image translation. While DistanceGAN [3] enforced distance consistency between different parts of the same sample in each domain, CUT [28] leveraged contrastive loss to maximize the patch-wise similarity between domains and achieved remarkable results. LSeSim [32] further utilized spatially correlation to maximize structural similarity and eliminate the domain-specific features.
2.2 Image-to-image translation for stain transformation
Transforming one stained tissue into another specific stain will dramatically save laboratory resources and money. Hence, growing research has leveraged unsupervised image-to-image translation to conduct stain transformation in several medical scenarios. Levy et al. . translated H&E to trichrome staining via CycleGAN for liver fibrosis staging [22]. Kapil et al. translated Cytokeratin to PD-L1 staining to bypass re-staining for segmentation [18]. de Haan et al. translated H&E to Masson’s Trichrome, periodic acid-Schiff, and Jones silver stain for improving preliminary diagnosis for kidney diseases via CycleGAN with perfectly paired images [13]. Lahiani et al. further broke the limitation of -pixel image patches by applying perceptual embedding consistency for H&E to FAP-CK transformation [21]. However, the lack of detailed description hampers the implementation of their methodology.
2.3 Ultra-high resolution image-to-image translation
Ultra-high-resolution images are ubiquitous in photography, artwork, posters, ultra-high (e.g., 8K) videos, and especially, WSIs in digital pathology (usually, larger than ). Due to the massive computational costs, transforming these images into different styles will be difficult. Traditionally, strategies have been proposed to address the problem of tiling artifacts created by patch-wise-based methods, including utilizing a larger overlapping window [20] or freezing IN layers during testing time [2]. By providing a patch-wise style transfer network with Thumbnail Instance Normalization (TIN), Chen et al. performed ultra-high-resolution style transformation with constant GPU memory usage [8]. Also, they applied their framework to an image-to-image translation task. However, according to our experiment, TIN may result in over/under-colorizing for their fallacious assumption that all patches can be normalized with the same global mean and standard deviation.

3 Proposed method
3.1 Overall Framework
Our framework targets one-sidedly translating an ultra-high-resolution image in domain (e.g., H&E stain domain) into image in domain (e.g., IHC domain), in which , and are the height and width of , via a mapping function, generator .
(1) |
Collections of unpaired X in and Y in would be first cropped into patches with the size of pixels to train a generator .
As our KIN module is only applied during the testing time and can be inserted into any framework with IN layers, we followed the original training process and hyperparameters proposed in the paper of CycleGAN [33], CUT [28], and LSeSim [32] to train the corresponding generators with their specific designed losses without any modification.
During the testing process, all the IN layers in are replaced with KIN layers. Given an image , non-overlapped patches are cropped with the size of . The coordinates of each patch corresponding to the original would be recorded simultaneously. For example, an image would be cropped into patches with coordinates of {} {}. Two caching tables of size , in which denotes the number of channels, would be initialized in each KIN for caching mean and standard deviation calculated.
As illustrated in Fig 2, we divide the testing process into two phases: caching and inference. During caching phase, each patch with its corresponding coordinates , are the input of the generator , and the calculated mean and standard deviation after passing the KIN will be cached. During the inference phase, , its corresponding coordinates , and a kernel are the input of the generator . The kernel , where and are the height and width of , is adjustable. When passing through the KIN layer, a region with the same size of kernel extended from will be extracted from the caching table and convolute with kernel k to compute mean and standard deviation which are used to normalize the feature maps. All the cropped patches will be passed to the in the aforementioned manner to yield translated patches . Eventually, all are assembled into an ultra-high-resolution translated image .
3.2 Kernelized Instance Normalization (KIN)
IN [29] has been widely used in GAN-based models for image generation and dramatically improved image quality [27]. Besides, multiple styles can be obtained by conditionally replacing the and in the IN layer [9]. IN can be formulated by:
(2) |
For each instance in a batch, and are calculated in a channel-wise manner, in which , denotes the batch size and and are trainable parameters.
We hypothesize that adjacent patches share similar statistics including the and computed in IN, and thus proposed KIN that could further alleviate the subtle incongruity that induces the tiling artifacts when adjacent patches are assembled. KIN is the extension of the original IN layer with extra two caching tables and to spatially store and values and additionally supports convolution operation on the caching tables with a given kernel . During the caching phase, KIN input a cropped patch with its spatial information, . and are computed as the original IN and cached.
(3) |
(4) |
During the inference phase, given a kernel with the size of , and are computed by convoluting on cache tables to generate translated images. To address the boundary cases, the cache tables would be padded initially with edge values.
(5) |
(6) |
(7) |
4 Datasets
4.1 Automatic Non-rigid Histological Image Registration (ANHIR)
Automatic Non-rigid Histological Image Registration (ANHIR) dataset [5, 6, 7, 10, 12, 24] consists of high-resolution WSIs from different tissue samples (lesions, lung lobes, breast tissue, kidney tissue, gastric tissue, colon adenocarcinoma, and mammary gland). The acquired images are organized in sets of consecutive tissue slices stained by various dyes, including H&E, Ki-67, ER/PR, CD4/CD8/CD68, etc., with sizes vary from to pixels. We randomly sampled three types of tissues to conduct our experiments. Each experiment comprises H&E stain and one target IHC stain: breast tissue (from H&E to PR), colon adenocarcinoma (COAD) (from H&E to CD4&CD68), and lung lesion (from H&E to Ki-67).
4.2 Glioma
The private glioma dataset was collected from H&E ( pixels) and epidermal growth factor receptor (EGFR) IHC ( pixels) stained tissue microarrays, and each comprised 105 tissue samples corresponding to 105 different patients. Totally 105 H&E stained tissue images with their consecutive EGFR counterparts were cropped from the microarrays and the image sizes vary from to pixels. We randomly selected 55 samples as the training set while the other 50 pairs as the testing set.
4.3 Kyoto summer2autumn
An extra natural image dataset was used to validate the generalizability of our methodology. We collected 17 and 20 high-resolution ( pixels) unpaired images taken in Tokyo during summer and autumn, respectively, as the training set and additional four summer images were used as a testing set. This Kyoto summer2autumn dataset111Kyoto summer2autumn dataset is available at: https://github.com/Kaminyou/Kyoto-summer2autumn was released to facilitate solving ultra-high-resolution-related problems that most computer vision studies might encounter.
5 Experiments
5.1 Experimental settings
Three popular unpaired image-to-image translation frameworks: CycleGAN [33], CUT [28], and L-LSeSim [32], were utilized to verify our approach. We followed the hyperparameter settings described in the original papers during the training process except for the model output size, which was changed from to . We trained the CycleGAN, CUT, and L-LSeSim for 50, 100, and 100 epochs. Models were trained and tested on three datasets: ANHIR, glioma, and Kyoto summer2autumn. Due to the insufficiency of WHIs in ANHIR dataset, we could only inference on the training set (note that training was in an unsupervised manner) while glioma and Kyoto summer2autumn datasets can be further split into training and testing sets. We replaced all IN layers with KIN layers in the generators during the inference process. One ultra-high-resolution image would be cropped into non-overlapped patches and pass through the KIN module. Translated patches were assembled to the final translated output. Constant and Gaussian kernels with sizes of 3, 7, and 11 were used to generate the best results. Translated images generated with KIN were compared with those from IN and TIN. Due to the GPU memory limitation, translated images generated with IN were also in a patch-wise (512512) manner, which is the same as the patch-wise IN in Chen et al. ’s work [8].
5.2 Metrics
In addition to the visualization of the translated images, we calculated Fréchet inception distance (FID) [14], histogram correlation, Sobel gradients [16] in YCbCr color domain, perception image quality evaluator (PIQE) [30], and natural image quality evaluator (NIQE) [25] to comprehensively evaluate the quality of translated ultra-high-resolution images.
However, due to the limitations of the available metrics that the tiling artifacts are difficult to be fairly graded and the unavailability of the perfectly matched counterpart, we conducted two human evaluation studies with five specialists: (a) quality challenge: given one source, one reference, and three translated images generated by patch-wise IN, TIN, and our KIN methods, respectively, specialists were asked to select the best among three translated images in 30 seconds. Since the images generated by CycleGAN and L-LSeSim were atrocious, we only chose images generated by CUT; (b) fidelity challenge: given one real image and one translated image, specialists were asked to select the one which is personally considered realistic in 10 seconds. We followed the protocol of the AMT perceptual studies from Isola et al. [17] but adjusted the time limitation as our images are extremely large. Since the data in ANHIR breast, COAD, and lung lesion subdatasets are insufficient, we combined these subdatasets as single ANHIR dataset and randomly selected pairs of real and translated WSIs from it.
5.3 Results




Breast | COAD | Lung lesion | ||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
FID | Corr. | Grad. | PIQE | NIQE | FID | Corr. | Grad. | PIQE | NIQE | FID | Corr. | Grad. | PIQE | NIQE | ||
CycleGAN | IN* | 98.60 | -0.07 | 13.62 | 4.95 | 9.39 | 103.25 | 75.25 | 15.08 | 5.26 | 9.08 | 76.15 | -4.49 | 9.92 | 62.69 | 13.14 |
TIN | 179.14 | -28.91 | 14.37 | 6.16 | 9.56 | 100.78 | 79.53 | 16.68 | 15.79 | 9.55 | 239.19 | 17.46 | 9.53 | 67.79 | 11.96 | |
KIN | 96.09 | 11.53 | 12.93 | 5.29 | 7.36 | 108.32 | 43.60 | 14.95 | 5.15 | 9.69 | 103.48 | -2.16 | 9.94 | 63.81 | 12.16 | |
CUT | IN* | 71.00 | 35.56 | 14.55 | 3.00 | 12.15 | 95.87 | 74.64 | 14.76 | 4.50 | 8.96 | 54.86 | -5.79 | 10.41 | 58.32 | 12.20 |
TIN | 125.18 | 39.50 | 17.04 | 3.42 | 10.98 | 91.81 | 33.29 | 15.49 | 11.96 | 9.02 | 251.48 | 80.14 | 11.80 | 32.08 | 12.20 | |
KIN | 72.59 | 36.32 | 14.05 | 3.27 | 10.66 | 93.68 | 76.45 | 14.60 | 4.49 | 8.94 | 56.38 | -9.63 | 10.58 | 60.13 | 12.08 | |
L-LSeSim | IN* | 65.82 | 31.57 | 15.04 | 3.03 | 13.36 | 100.42 | 48.15 | 13.64 | 4.23 | 8.45 | 56.30 | -3.82 | 9.71 | 46.40 | 13.88 |
TIN | 89.94 | 22.16 | 12.75 | 3.29 | 13.18 | 100.50 | 41.37 | 15.67 | 9.56 | 8.14 | 231.13 | 59.71 | 12.68 | 44.29 | 11.13 | |
KIN | 67.46 | 31.58 | 14.31 | 3.19 | 12.35 | 100.04 | 51.62 | 13.34 | 4.44 | 8.22 | 62.74 | -4.77 | 9.91 | 47.99 | 13.48 |
-
•
IN*: Patch-wise IN; Corr.: Histogram correlation; Grad.: Sobel gradients; : the lower the better; : the higher the better.
Glioma (training set) | Glioma (testing set) | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
FID | Corr. | Grad. | PIQE | NIQE | FID | Corr. | Grad. | PIQE | NIQE | ||
CycleGAN | IN* | 136.32 | 0.26 | 11.91 | 21.83 | 13.64 | 142.28 | 0.28 | 10.57 | 23.73 | 13.76 |
TIN | 220.00 | 0.28 | 5.42 | 39.05 | 12.01 | 207.03 | 0.38 | 4.65 | 41.16 | 11.99 | |
KIN | 157.26 | 0.14 | 7.22 | 27.89 | 11.27 | 150.93 | 0.19 | 6.31 | 29.87 | 11.54 | |
CUT | IN* | 105.22 | 0.85 | 14.81 | 23.76 | 13.99 | 105.66 | 0.85 | 13.48 | 24.02 | 14.05 |
TIN | 214.22 | 0.54 | 10.02 | 34.52 | 13.37 | 200.56 | 0.64 | 8.64 | 35.01 | 13.37 | |
KIN | 108.20 | 0.81 | 12.84 | 31.26 | 13.70 | 100.90 | 0.80 | 11.58 | 31.94 | 13.86 | |
L-LSeSim | IN* | 107.74 | 0.41 | 11.83 | 21.09 | 10.70 | 105.59 | 0.48 | 10.67 | 21.25 | 10.62 |
TIN | 203.70 | 0.10 | 8.22 | 24.87 | 10.94 | 191.44 | 0.19 | 7.58 | 23.34 | 10.75 | |
KIN | 113.92 | 0.41 | 8.64 | 26.00 | 10.63 | 106.90 | 0.46 | 7.69 | 26.85 | 10.40 |
-
•
IN*: Patch-wise IN; Corr.: Histogram correlation; Grad.: Sobel gradients; : the lower the better; : the higher the better.
Kyoto (training set) | Kyoto (testing set) | ||||||||
---|---|---|---|---|---|---|---|---|---|
FID | Grad. | PIQE | NIQE | FID | Grad. | PIQE | NIQE | ||
CycleGAN | IN* | 79.11 | 18.40 | 43.62 | 12.20 | 171.88 | 16.52 | 37.00 | 11.26 |
TIN | 87.10 | 12.39 | 54.29 | 12.24 | 180.50 | 10.49 | 52.56 | 11.92 | |
KIN | 93.60 | 17.08 | 44.65 | 12.11 | 192.25 | 15.12 | 39.53 | 11.10 | |
CUT | IN* | 77.59 | 17.87 | 43.81 | 13.40 | 157.04 | 18.29 | 37.74 | 11.41 |
TIN | 98.37 | 17.44 | 43.30 | 12.11 | 181.13 | 15.33 | 40.97 | 11.89 | |
KIN | 75.27 | 18.53 | 40.21 | 12.98 | 167.15 | 17.24 | 38.30 | 12.31 | |
L-LSeSim | IN* | 178.19 | 14.86 | 19.35 | 11.89 | 248.81 | 13.05 | 14.14 | 11.42 |
TIN | 178.98 | 11.27 | 19.13 | 12.07 | 253.14 | 9.74 | 12.21 | 11.18 | |
KIN | 192.42 | 16.41 | 19.19 | 12.01 | 265.07 | 16.12 | 13.00 | 10.54 |
-
•
IN*: Patch-wise IN; Grad.: Sobel gradients; : the lower the better


5.3.1 Stain transformation
Figs. 3, 5 and 4 and Fig. S5 and Fig. S6 show the translated images for three ANHIR subdatasets (breast tissue, COAD, and lung lesion) and glioma dataset, respectively. With only IN layers, CUT yields the images with best quality with some tiling artifacts, while CycleGAN led to checkerboard artifacts. L-LSeSim powerfully preserves spatial information but compromises color information. With TIN, all the translated images showed dramatically over/under-colorizing. With our KIN, translated images can have minor tiling artifacts and preserve their local features. However, if the original framework generated severe tiling artifacts, our KIN could alleviate but be hard to eliminate. Considering the similarity (FID and histogram correlation) and quality metrics (Sobel gradient, PIQE and NIQE), our KIN is superior to patch-wise IN and TIN in most cases (see Tabs. 1 and 2). Although KIN does not always obtain the best scores, a possible reason is that no appropriate metrics can reflect the performance of such unpaired WSIs stain transformation task. Thus, we established two human evaluation studies to pertinently evaluate the image quality and fidelity. As shown in Fig. 7, our KIN achieved the best performance in both.
5.3.2 Translation for natural images
Our KIN module also performed well on natural images (as shown in Fig. 6 and Fig. S7). As described above, KIN can alleviate the tiling artifacts generated by patch-wise IN while TIN would lead to over/under-colorizing. However, when it comes to natural images, over/under-colorizing would not be as obtrusive as in stain transformation cases, since people sometimes prefer over-stylized images. For example, High-Dynamic Range (HDR) or contrast adjustment techniques are popular to beautify photographs and render the photos more attractive. Tab. 3 and Fig. 7 provided the metrics evaluation results, and our KIN obtained the best performance among all methodologies in human evaluation. Considering the FID, CUT with our KIN is superior or competitive to other methods. Although Sobel gradients are higher in some cases, the high contrast level of one image might also contribute to higher gradients. On the other hand, there are only minor differences in PIQE and NIQE between methods. However, none of the metrics can effectively evaluate ultra-high-resolution images with tiling artifacts.
5.4 Ablation study
5.4.1 Kernel and kernel size
To elucidate the effect of different kernels and kernel size on the translated images, we applied constant and Gaussian kernels with the size of 1, 3, 7, 11, and in the KIN module (see Fig. S8 and S9). It is noteworthy that when kernel size is set to 1, the KIN module will operate in a manner of patch-wise IN, whereas it would be like TIN when kernel size is set to (bounded by the input image size). KIN is an eclectic approach that combines the advantages of patch-wise IN and TIN and avoids extremes of single and global features calculated in patch-wise IN and TIN. When kernel size increases from one to , the translated results gradually change from patch-wise IN to TIN. On the other hand, the constant kernel can help generate smoother results, while the Gaussian kernel will emphasize local features more.
6 Discussions
Our experiments showed that KIN performed well on multiple datasets, and unseen testing data can even be successfully inferred when sufficient training data are available. The over/under-colorizing problem caused by TIN is also revealed, which might be innocuous when natural images are used but would be detrimental when targeting stain transformation. Pathological features, which are essential for clinical judgment, would be compromised when global mean and standard deviation are applied in the TIN.
Although KIN can be inserted into any IN-based framework, the performance would be compromised if the original framework has amateurish performance, such as CycleGAN, which generates results diversely among adjacent patches. KIN can hardly eliminate all the tiling artifacts undertaking such cases. Interestingly, we found that CUT can yield more consistent results among adjacent patches, especially for the hue. On the other hand, LSeSim meticulously preserves all the structure but ignores the consistency of the hue, which is reasonable as CUT captures domain-specific features, but LSeSim focuses on spatial features according to their loss functions. Despite KIN achieving the best performance surpassing all previous approaches in human evaluation studies, its strength cannot be manifested due to the inadequacy of appropriate metrics for evaluating the quality and fidelity of unpaired ultra-high-resolution WSIs. Finally, ultra-high-resolution images are commonly used in daily life but there is no public dataset available for a fair comparison. To facilitate related researches, we released the Kyoto summer2autumn dataset.
7 Conclusion
This study presents Kernelized Instance Normalization (KIN) for ultra-high-resolution stain transformation with constant space complexity. KIN can be easily inserted into popular unpaired image-to-image translation frameworks without re-training the model. Comprehensive experiments with two WSI datasets were conducted and evaluated by human evaluation studies and appropriate metrics. An extra ultra-high-resolution natural image dataset was also utilized and demonstrated the generalizability of KIN. Overall, KIN surpassed all the previous approaches and generated state-of-the-art outcomes. Henceforth, ultra-high-resolution stain transformation or image-to-image translation, can be easily accomplished and applied in clinical practice.
Acknowledgements
We thank Chao-Yuan Yeh, the CEO of aetherAI, for providing computing resources, which enabled this study to be performed, and Cheng-Kun Yang for his revision suggestions.
References
- [1] Alturkistani, H.A., Tashkandi, F.M., Mohammedsaleh, Z.M.: Histological stains: a literature review and case study. Global journal of health science 8(3), 72 (2016)
- [2] de Bel, T., Hermsen, M., Kers, J., van der Laak, J., Litjens, G.: Stain-transforming cycle-consistent generative adversarial networks for improved segmentation of renal histopathology. In: International Conference on Medical Imaging with Deep Learning–Full Paper Track (2018)
- [3] Benaim, S., Wolf, L.: One-sided unsupervised domain mapping. In: NIPS (2017)
- [4] Birkman, E.M., Mansuri, N., Kurki, S., Ålgars, A., Lintunen, M., Ristamäki, R., Sundström, J., Carpén, O.: Gastric cancer: immunohistochemical classification of molecular subtypes and their association with clinicopathological characteristics. Virchows Archiv 472(3), 369–382 (2018)
- [5] Borovec, J., Kybic, J., Arganda-Carreras, I., Sorokin, D.V., Bueno, G., Khvostikov, A.V., Bakas, S., Eric, I., Chang, C., Heldmann, S., et al.: Anhir: automatic non-rigid histological image registration challenge. IEEE transactions on medical imaging 39(10), 3042–3052 (2020)
- [6] Borovec, J., Munoz-Barrutia, A., Kybic, J.: Benchmarking of image registration methods for differently stained histological slides. In: 2018 25th IEEE International Conference on Image Processing (ICIP). pp. 3368–3372. IEEE (2018)
- [7] Bueno, G., Deniz, O.: Aidpath: academia and industry collaboration for digital pathology (2019)
- [8] Chen, Z., Wang, W., Xie, E., Lu, T., Luo, P.: Towards ultra-resolution neural style transfer via thumbnail instance normalization. In: Proceedings of the AAAI Conference on Artificial Intelligence (2022)
- [9] Dumoulin, V., Shlens, J., Kudlur, M.: A learned representation for artistic style (2017)
- [10] Fernandez-Gonzalez, R., Jones, A., Garcia-Rodriguez, E., Chen, P.Y., Idica, A., Lockett, S.J., Barcellos-Hoff, M.H., Ortiz-De-Solorzano, C.: System for combined three-dimensional morphological and molecular analysis of thick tissue specimens. Microscopy research and technique 59(6), 522–530 (2002)
- [11] Fragomeni, S.M., Sciallis, A., Jeruss, J.S.: Molecular subtypes and local-regional control of breast cancer. Surgical Oncology Clinics 27(1), 95–120 (2018)
- [12] Gupta, L., Klinkhammer, B.M., Boor, P., Merhof, D., Gadermayr, M.: Stain independent segmentation of whole slide images: A case study in renal histology. In: 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018). pp. 1360–1364. IEEE (2018)
- [13] de Haan, K., Zhang, Y., Zuckerman, J.E., Liu, T., Sisk, A.E., Diaz, M.F., Jen, K.Y., Nobori, A., Liou, S., Zhang, S., et al.: Deep learning-based transformation of h&e stained tissues into special stains. Nature communications 12(1), 1–13 (2021)
- [14] Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems 30 (2017)
- [15] Inamura, K.: Update on immunohistochemistry for the diagnosis of lung cancer. Cancers 10(3), 72 (2018)
- [16] Irwin, F., et al.: An isotropic 3x3 image gradient operator. Presentation at Stanford AI Project 2014(02) (1968)
- [17] Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 1125–1134 (2017)
- [18] Kapil, A., Wiestler, T., Lanzmich, S., Silva, A., Steele, K., Rebelatto, M., Schmidt, G., Brieu, N.: Dasgan–joint domain adaptation and segmentation for the analysis of epithelial regions in histopathology pd-l1 images. arXiv preprint arXiv:1906.11118 (2019)
- [19] Kim, T., Cha, M., Kim, H., Lee, J.K., Kim, J.: Learning to discover cross-domain relations with generative adversarial networks. In: International conference on machine learning. pp. 1857–1865. PMLR (2017)
- [20] Lahiani, A., Gildenblat, J., Klaman, I., Albarqouni, S., Navab, N., Klaiman, E.: Virtualization of tissue staining in digital pathology using an unsupervised deep learning approach. In: European Congress on Digital Pathology. pp. 47–55. Springer (2019)
- [21] Lahiani, A., Klaman, I., Navab, N., Albarqouni, S., Klaiman, E.: Seamless virtual whole slide image synthesis and validation using perceptual embedding consistency. IEEE Journal of Biomedical and Health Informatics 25(2), 403–411 (2020)
- [22] Levy, J.J., Jackson, C.R., Sriharan, A., Christensen, B.C., Vaickus, L.J.: Preliminary evaluation of the utility of deep generative histopathology image translation at a mid-sized nci cancer center. bioRxiv (2020)
- [23] Luo, Z.W., Zhu, M.G., Zhang, Z.Q., Ye, F.J., Huang, W.H., Luo, X.Z.: Increased expression of ki-67 is a poor prognostic marker for colorectal cancer patients: a meta analysis. BMC cancer 19(1), 1–13 (2019)
- [24] Mikhailov, I., Danilova, N., Malkov, P.: The immune microenvironment of various histological types of ebv-associated gastric cancer. In: Virchows Archiv. vol. 473, pp. S168–S168. Springer 233 SPRING ST, NEW YORK, NY 10013 USA (2018)
- [25] Mittal, A., Soundararajan, R., Bovik, A.C.: Making a “completely blind” image quality analyzer. IEEE Signal processing letters 20(3), 209–212 (2012)
- [26] Oshi, M., Tokumaru, Y., Angarita, F.A., Yan, L., Matsuyama, R., Endo, I., Takabe, K.: Degree of early estrogen response predict survival after endocrine therapy in primary and metastatic er-positive breast cancer. Cancers 12(12), 3557 (2020)
- [27] Pang, Y., Lin, J., Qin, T., Chen, Z.: Image-to-image translation: Methods and applications. IEEE Transactions on Multimedia (2021)
- [28] Park, T., Efros, A.A., Zhang, R., Zhu, J.Y.: Contrastive learning for unpaired image-to-image translation. In: European Conference on Computer Vision. pp. 319–345. Springer (2020)
- [29] Ulyanov, D., Vedaldi, A., Lempitsky, V.: Instance normalization: The missing ingredient for fast stylization. arXiv preprint arXiv:1607.08022 (2016)
- [30] Venkatanath, N., Praneeth, D., Bh, M.C., Channappayya, S.S., Medasani, S.S.: Blind image quality evaluation using perception based features. In: 2015 Twenty First National Conference on Communications (NCC). pp. 1–6. IEEE (2015)
- [31] Yi, Z., Zhang, H., Tan, P., Gong, M.: Dualgan: Unsupervised dual learning for image-to-image translation. In: Proceedings of the IEEE international conference on computer vision. pp. 2849–2857 (2017)
- [32] Zheng, C., Cham, T.J., Cai, J.: The spatially-correlative loss for various image translation tasks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2021)
- [33] Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Computer Vision (ICCV), 2017 IEEE International Conference on (2017)