This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Reconstruct high-resolution multi-focal plane images from a single 2D wide field image

Jiabo Ma    \IEEEmembershipStudent, IEEE    Sibo Liu    \IEEEmembershipStudent, IEEE    Shenghua Cheng    Xiuli Liu    Li Chen    Shaoqun Zeng Corresponding author: Shenghua Cheng.J. Ma and S. Liu contributed equally to this work. J. Ma, S. Liu, S. Cheng, X. Liu and S. Zeng are with Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, Huazhong University of Science and Technology, Wuhan, Hubei 430074, China and MoE Key Laboratory for Biomedical Photonics, School of Engineering Sciences, Huazhong University of Science and Technology, Wuhan, Hubei 430074, China. (email: [email protected]; [email protected]; [email protected]; [email protected]; [email protected])L. Chen is with Department of Clinical Laboratory, Tongji Hospital, Huazhong University of Science and Technology, Wuhan, Hubei 430030, China. (email: [email protected])
Abstract

High-resolution 3D medical images are important for analysis and diagnosis, but axial scanning to acquire them is very time-consuming. In this paper, we propose a fast end-to-end multi-focal plane imaging network (MFPINet) to reconstruct high-resolution multi-focal plane images from a single 2D low-resolution wild filed image without relying on scanning. To acquire realistic MFP images fast, the proposed MFPINet adopts generative adversarial network framework and the strategies of post-sampling and refocusing all focal planes at one time. We conduct a series experiments on cytology microscopy images and demonstrate that MFPINet performs well on both axial refocusing and horizontal super resolution. Furthermore, MFPINet is approximately 24 times faster than current refocusing methods for reconstructing the same volume images. The proposed method has the potential to greatly increase the speed of high-resolution 3D imaging and expand the application of low-resolution wide-field images.

{IEEEkeywords}

2D-to-3D super resolution, multi-focal plane imaging, deep learning

1 Introduction

\IEEEPARstart

In the field of medical images, high-resolution (HR) 3D images are helpful for analysis and diagnosis [1]. Therefore, how to obtain HR 3D images with high throughput is important. The main methods to obtain HR 3D optical images are to scan point by point or plane by plane through two-photon[2], confocal [3], and optical sheet [4]. By using these techniques, the information outside focal planes can be effectively suppressed from entering the imaging system, thus greatly improving the axial resolution of 3D optical images. But these methods are generally slow in imaging due to axial scanning. At the same time, phototoxicity and photobleaching will affect observed samples due to long time scanning [5]. In order to alleviate the above problem, some methods optimize scanning strategy[6] and improve PSF[7] to speed up imaging. With the development of computer technologies, 3D fluorescence information can also be obtained by non-scanning methods. These methods can map the axial information to 2D images at one time, such as fluorescence field microscope [8] and Fresnel hologram [9], then use iterative algorithms to reconstruct 3D images. These non-scanning methods are usually time-consuming in image reconstruction, and generally need specific optical elements or instruments as assistance, which increases the complexity of methods.

In recent years, more and more methods apply deep learning to improve imaging resolution and increase imaging speed. Most super-resolution microscopy methods [10, 11, 12] implement on 2D images, which improves the resolution of images collected by low-end instruments and expands the application of low-resolution (LR) images in research and diagnosis. Further, some 3D super-resolution microscopy methods [13, 14, 15, 16] are also developed. 3D super-resolution technologies can greatly increase imaging speed while ensuring image resolution since the acquisition of 3D LR images requires fewer scans and produces less data. However, axial scanning is still required to acquire 3D images even if these super-resolution technologies are used. To overcome the problem, 2D-to-3D super-resolution methods are proposed, such as deep learning based virtual refocusing [17]. The method can refocus a single wide-field image to acquire multi-focal plane (MFP) images. With the help of deep learning’s powerful fitting ability and efficient forward inference, 3D images can be obtained fast without axial-scanning or special hardware.

However, the current method of obtaining MFP images from a single 2D wide-field image focuses on the axial refocusing and neglects the improvement of horizontal resolution. Therefore, the speed of acquiring HR MFP images still depends on horizontal fine scanning. To get rid of the dependence on HR images, in this paper, we propose a model called MFPINet (multi-focal plane imaging network) to reconstruct HR MFP wide-field images from a single LR 2D wide-field image based on deep learning. The designed MFPINet is end-to-end and does not need any iterative calculation. In order to generate realistic MFP images, the generative adversarial network (GAN) framework [18] is introduced in the proposed MFPINet. Specially, the generator of MFPINet is modified from U-Net [19] based on the characteristics of super resolution task. To reduce training difficulty and eliminate artifacts [20], we introduce residual connections [21] and remove common batch normalization (BN) [22] layers. In addition, to obtain HR MFP images more efficiently, we further optimize the U-Net in two ways. First, for horizontal super resolution, we adopt a post-sampling strategy instead of up-sampling the input image, which greatly reduces memory consumption and computational complexity. Second, for axial refocusing, the designed generator outputs NN focal plane images at one time instead of outputting specific focal plane image each time and then stitching them together. As for discriminator, it is a common classification network used to distinguish real and fake MFP images. But to make it easier for discriminator to be aware of the difference of all focal planes, the entire MFP images are used as the input of discriminator. Since the input MFP images are 3D, to save memory, the axial information in the 3D image (W×H×Z×CW\times H\times Z\times C) is converted to the color space to obtain a multi-channel 2D image (W×H×(Z×C)W\times H\times(Z\times C)). Then 2D convolutions are used to process input images. In summary, we propose a fast end-to-end MFPINet to transform a LR 2D image into HR MFP images. No additional optical hardware is needed in the whole process, so it can be applied to all kinds of wide field images without major adjustment. We conduct a series of experiments on cytology images to demonstrate MFPINet can generate high-quality HR MFP images.

The main contributions of this paper are as follows:

  • We first propose a generative model MFPINet for reconstructing high-resolution multi-focal plane images from a single 2D wide field image.

  • We verify the effectiveness of MFPINet for both axial refocusing and horizontal super resolution on cytology images.

  • Experimental results further show that MFPINet is approximately 24×\times faster than current refocusing methods.

2 Related Works

2.1 2D-to-2D Super Resolution

In recent years, the super resolution of 2D natural images [23, 24, 25, 26] has been widely studied. Dong et al. [23] introduced convolutional neural network into super resolution, making deep learning play a key role in super resolution. Kim et al. proposed VDSR [27], which greatly improved the effect of super resolution due to using deeper network and a strategy of learning global residuals. EDSR [24] removed unnecessary modules of conventional residual networks, and achieved better results by expanding model architecture and stabilizing training process. To obtain more realistic HR images, Ledig et al. [25] introduced GAN into super resolution and proposed SRGAN to obtain super-resolved images with richer details. But the introduction of GAN caused more artifacts. To alleviate artifacts, ESRGAN [20] introduced residual-in-residual dense blocks and removed batch normalization, which further improved the quality of reconstructed HR images.

At the same time, super-resolution methods designed for various medical images have also emerged. Wang et al. [11] used an U-Net-like GAN model to improve the resolution of low numerical aperture(NA) fluorescence microscopy images. Ma et al. [10] proposed a multi-stage and multi-supervised model PathSRGAN to expand the application of 4x/0.1-NA pathology images. Rivenson et al. [28] extended deep learning based super-resolution to mobile-phone-based microscopy images.

2.2 3D-to-3D super resolution

For 3D medical images, Pham et al. [16] proved that LR 3D MRI images can be reconstructed into high-quality HR 3D MRI images by using deep convolution neural networks, which greatly improves the speed of obtaining HR MRI images. Zhou et al. [29] improved resolution of 3D fluorescence microscopic images by using a dual super resolution model. You et al. [14] proposed a semi-supervised 3D CT super-resolution method by introducing Wassertein constraints to GAN and node constraints to loss functions. In summary, the above 3D super-resolution methods improves resolution and imaging speed of 3D medical images.

2.3 2D-to-3D super resolution

Although 3D super resolution is helpful in expanding the application of LR medical images, the acquisition of 3D LR images still requires time-consuming axial scanning. With the progress of computational imaging, deep learning based 2D-to-3D super-resolution methods are gradually emerging. Page et al. [30] proposed LFMNet to reconstruct confocal microscopy stacks from single light filed images. Wu et al. [17] proposed Deep-Z model based on GAN, which can refocus single 2D wide-field fluorescence images to obtain 3D images.

Although both MFPINet and Deep-Z can reconstruct 3D images from a single 2D wide field image, they are very different. First, MFPINet enhances the horizontal and axial resolution at the same time. So it does not require HR 2D images as input, which greatly improves the imaging speed and expands the applications of MFPINet. Second, MFPINet is designed to output 3D images at one time instead of generating single focal plane image one-by-one like Deep-Z, which improves reconstruction efficiency.

3 Method

In this part, we elaborate the network architecture, optimization functions and evaluation metrics of the proposed method.

3.1 Network Architecture

Refer to caption

Figure 1: The network architecture of MFPINet (example for 384×384×3384\times 384\times 3 sRGB image). The green arrow in the figure represents the flow of data, and the meaning of all other boxes and arrows is at the bottom of the figure. a. The architecture of the generator, which receives 2D wide-field images and outputs high-resolution MFP images. b. The architecture of the discriminator, which takes different MFP images as inputs and outputs corresponding categories of different images . c. The basic module of the generator.

The proposed MFPINet is a GAN framework model, which consists of two parts: generator (GG) and discriminator (DD). Fig. 1 shows the architecture of GG and DD.

The generator consists of an encoder and a decoder, which is inspired by U-Net [19] and ResNet [21]. The encoder is composed of a preprocessing module, three residual encoding modules (REM) and a stacked convolutional layers (SCL), which are expressed as the following equation:

xk=SCL(REM(REM(REM(P(Conv64(x3))))))x_{k}=SCL(REM(REM(REM(P(Conv_{64}(x_{3})))))) (1)

where x3x_{3} stands for sRGB input image, Convk()Conv_{k}(\cdot) stands for a convolutional layer with kk channels. P()P(\cdot) stands for parametric rectified linear unit. SCL()SCL(\cdot) stands for a stacked convolutional layers composed of three covolutional layers. REMREM represents the residual encoding module. SCLSCL and REMREM can be expressed as equation 2 and 3:

SCL=Conv64(Conv128(Conv128(xk)))SCL=Conv_{64}(Conv_{128}(Conv_{128}(x_{k}))) (2)
REM=AP(RB(RB(xk)))REM=AP(RB(RB(x_{k}))) (3)

where xkx_{k} stands for a feature map with k channels. AP()AP(\cdot) stands for 2D average pooling, which is used to halve the width and height of the feature map. RB()RB(\cdot) represents a residual block composed of four convolutional layers. Detailed topological connections can be found in Fig. 1.

The encoder outputs feature maps of four different scales, which are f1f_{1}, f2f_{2}, f3f_{3}, f4f_{4} from shallow to deep. Then decoder takes all output feature maps of the encoder as input for reconstruction. Specifically, the decoder consists of three residual decoding modules (RDM) and one super-resolution reconstruction (SRR) module, which are expressed as the following equation:

MFPI=SRR(RDM(RDM(RDM(f4,f3),f2),f1))\begin{split}MFPI=SRR(RDM(RDM(RDM(\\ f_{4},f_{3}),f_{2}),f_{1}))\end{split} (4)

where RDM()RDM(\cdot) represents the residual decoder module, which is used to integrate feature information of different scales. SRR()SRR(\cdot) represents the super-resolution reconstruction module, which is used to reconstruct HR MFP images from multi-scale feature maps fused by RDM()RDM(\cdot). SRRSRR and RDMRDM are expressed as equation 5 and 6:

RDM=RB(RB(CAT(NI(fk+1)),fk))RDM=RB(RB(CAT(NI(f_{k+1})),f_{k})) (5)
SSR=Tanh(Conv3×n(PS(Conv256(ffused))))SSR=Tanh(Conv_{3\times n}(PS(Conv_{256}(f_{fused})))) (6)

where fkf_{k} stands for the kthk^{th} feature map output by the encoder. NI()NI(\cdot) stands for nearest interpolation at width and height dimension of feature maps. CAT()CAT(\cdot) stands for the concatenation of tensors along channel dimension. ffusedf_{fused} represents feature maps fused by the three RDMRDM modules. PS()PS(\cdot) represents pixel shuffle [31] layer, which is used to double width and height dimension of feature maps. Tanh()Tanh(\cdot) represents tanh activation. nn represents the number of different focal plane 2D images in the output MFP images.

The above-mentioned encoder and decoder constitute the proposed generator. Although it borrows the U-shaped structure of U-Net, it is quite different from the original and common modified U-Net. First, the common BN module is discarded because BN is reported to cause artifacts [20]. Second, residual blocks are introduced to alleviate gradient problem caused by deep networks. Third, according to the characteristics of super-resolution tasks, only three down-sampling blocks are kept. Because super-resolution tasks are different from high-level tasks such as semantic segmentation, pixels in super-resolution tasks generally only have a greater relationship with their neighboring pixels. At the same time, because the model becomes smaller, the model can process larger images at a faster speed, thereby enabling the model to quickly generate MFP images. Fourth, the post-sampling strategy is adopted instead of interpolating the input image at the beginning, which greatly reduces the consumption of memory and computational complexity. At last, the proposed generator directly outputs 3D MFP images instead of outputting refocused images [17] one by one, which greatly improves the efficiency of acquiring MFP images.

The discriminator of the proposed MFPINet is a simple classification network, and its architecture is shown in Fig. 1. In order for the discriminator to perceive the difference of all focal plane images at the same time, the designed discriminator takes 3D MFP images as input. In addition, to reduce consumption of memory and computation, 3D MFP images are converted into 2D images, and then use 2D convolution to process them. Specifically, the original 3D MFP image represented by a four-dimensional array H×W×Z×CH\times W\times Z\times C is converted into a three-dimensional array H×W×(Z×C)H\times W\times(Z\times C).

3.2 Model Optimization

The proposed generator is designed to learn a transformation mapping from LR 2D images to corresponding 3D HR MFP images. In the training phase, generator GθgG_{\theta_{g}} with parameter θg\theta_{g} takes a 2D LR image XX as the input and output 3D HR MFP images Gθg(X)G_{\theta_{g}}(X). The discriminator DθdD_{\theta_{d}} with parameter θd\theta_{d} takes real HR MFP images and forged MFP images by GθdG_{\theta_{d}} respectively. Discriminator DθdD_{\theta_{d}} outputs a probability value for input MFP images, which represents the model’s confidence. The loss LGL_{G} of generator is defined as following equation:

LG=1Nn=1Nlog(Dθd(Gθg(X{n})))×α+|Gθg(X{n})Y{n}|\begin{split}L_{G}=\frac{1}{N}\sum\limits_{n=1}^{N}{-log(D_{\theta_{d}}(G_{\theta_{g}}(X^{\{n\}})))}\times\alpha\\ +|G_{\theta_{g}}(X^{\{n\}})-Y^{\{n\}}|\end{split} (7)

where NN represents the number of samples of mini-batch in each iteration. X{n}X^{\{n\}} is the nthn^{th} 2D input image. Gθg(X{n})G_{\theta_{g}}{(X^{\{n\}})} is the HR MFP images obtained by the generator, and Y{n}Y^{\{n\}} is the real HR MFP images corresponding to X{n}X^{\{n\}}. α\alpha is a hyper-parameter used to balance absolute error loss and adversarial loss, which is set to 0.01 in the experiments.

As for the discriminator loss LDL_{D}, it is defined as the following equation:

LD=1Nn=1Nlog(Dθd(Y{n}))+log(1Dθd(Gθg(X{n})))\begin{split}L_{D}=-\frac{1}{N}\sum\limits_{n=1}^{N}{log(D_{\theta_{d}}(Y^{\{n\}}))}\\ +log(1-D_{\theta_{d}}(G_{\theta_{g}}(X^{\{n\}})))\end{split} (8)

The meaning of the parameters in equation 8 is similar to that in equation 7. In order to improve the training speed and optimization stability, Adam [32] optimization algorithm is used to optimize both generator GθgG_{\theta_{g}} and discriminator DθdD_{\theta_{d}}. The learning rate is set to 0.0001 and other parameters are set to the default value. More details can be found in Experiments section.

3.3 Evaluation Metrics

In order to evaluate the quality of images generated by MFPINet, four common metrics are used to compare the differences between the generated MFP images IgI_{g} and the real mechanical scanning MFP images IgtI_{gt}. In order to explore the reconstruction effect of different focal planes, the MFP images are not evaluated as a whole, but as individual focal plane.Specifically, the following four metrics are used.

3.3.1 Mean Square Error

Mean square error (MSE) is widely used to evaluate the error of two sets. For evaluation of MFP images, first split it into NN 2D images along the axial axis, and then NN corresponding real mechanical scanning 2D images and the NN reconstructed 2D images are evaluated separately. Specifically, MSE is defined as following equation:

MSE(Ig,Igt)=1Nn=1N1H×W×CIg(n)Igt(n)22MSE(I_{g},I_{gt})=\frac{1}{N}\sum\limits_{n=1}^{N}{\frac{1}{H\times W\times C}{||I_{g}^{(n)}-I_{gt}^{(n)}||_{2}^{2}}} (9)

where HH and WW are the height and width of images. CC is the channel number of images (e.g. 3 for sRGB image). NN is the total number of images. Regarding MSE, the smaller the value, the better the generated image.

3.3.2 Peak Signal to Noise Ratio

Peak signal to noise ratio (PSNR) is the most widely used objective method to evaluate image quality. Similar to MSE, PSNR is defined as following equation:

PSNR(Ig,Igt)=10×log10((2n1)2MSE(Ig,Igt))PSNR(I_{g},I_{gt})=10\times log_{10}{(\frac{(2^{n}-1)^{2}}{MSE(I_{g},I_{gt})})} (10)

where nn is the bits number of an image (e.g. 8 for uint8). Compared with MSE, the larger the PSNR value, the better the generated image.

3.3.3 Mean Absolute Error

Mean absolute error (MAE) uses L1 distance, which is less sensitive to outliers than L2 distance of MSE. It is defined as follows:

MAE(Ig,Igt)=1Nn=1N1H×W×CIg(n)Igt(n)1MAE(I_{g},I_{gt})=\frac{1}{N}\sum\limits_{n=1}^{N}{\frac{1}{H\times W\times C}{||I_{g}^{(n)}-I_{gt}^{(n)}||_{1}}} (11)

The meaning of parameters in equation 11 is the same as that in equation 9. Similar to MSE, the smaller the MAE value, the better the generated image.

3.3.4 Structure Similarity Index

Structural similarity index (SSIM) is a metric to measure the similarity of two images. It considers more attributes of images, such as brightness, contrast and structure similarity. So it is more consistent with human vision than PSNR or MSE. SSIM is defined as following equation:

SSIM(Ig,Igt)=(2μgμgt+c1)(2σg,gt+c2)(μg2+μgt2+c1)(σg2+σgt2+c2)SSIM(I_{g},I_{gt})=\frac{(2\mu_{g}\mu_{gt}+c_{1})(2\sigma_{g,gt}+c_{2})}{(\mu_{g}^{2}+\mu_{gt}^{2}+c_{1})(\sigma_{g}^{2}+\sigma_{gt}^{2}+c_{2})} (12)

where μg\mu_{g} and μgt\mu_{gt} are average brightness of the generated image IgI_{g} and mechanical scanning image IgtI_{gt} respectively. σg\sigma_{g} and σgt\sigma_{gt} are standard deviation of IgI_{g} and IgtI_{gt} respectively. σg,gt\sigma_{g,gt} is covariance of two images. c1c_{1} and c2c_{2} are constants used to stabilize the division process.

4 Experiments

To verify the effectiveness of MFPINet, we evaluate the quality of images generated by MFPINet in cytology images. As there is no method that fully matches our work at present, we compare our method with Deep-Z [17] in reconstructing 3D images. At the same time, to evaluate the super-resolution effect of our model, the classical SRGAN [25] is used as a baseline.

4.1 Cytology Images

4.1.1 Data Collection

In our experiments, 5 liquid-based cytology slides are used, which come from the department of clinical laboratory, Tongji Hospital, Huazhong University of Science and Technology. For each glass slide, 11 digital whole image slides (WSI) with a step of 2.7 μm\mu m are scanned under 20×\times/0.75-NA objective lens using 3D Histech scanner. The resolution of WSIs is 0.243 μm/pixel\mu m/pixel with a size of about 100K × 100K pixels.

4.1.2 Data Preparation

The 5 slides are randomly divided into training set (3 slides) and test set (2 slides). For each slide, 5000 patches of 768×\times768 size with cell foreground are cropped out as ground truth (GT). Then bicubic interpolation are used to resize the patch size from 768×\times768 into 384×\times384. In order to further reduce the quality of input data, Gaussian blur with a window of 5×\times5 and a standard deviation of 3 is used to degrade the input LR images. Finally, the input LR images and HR GT images are normalized to -1\sim1. In experiments, the middle LR layer image (the 6th of ll layers, defined as 0 μ\mum) is taken as the input and all 11 HR layers are taken as the GT.

4.2 Training Details

All experiments are completed in PyTorch 1.4 [33] and Ubuntu 18.04 equipped with two NVIDIA RTX 2080ti GPUs. For SRGAN and Deep-Z, we retrained these methods using the parameters provided in original papers on our dataset.

4.2.1 MFPINet

When training MFPINet, Adam optimizer is used with an initial learning rate 0.0001. The model is trained 20 epochs on the training set and the learning rate is halved every two epochs.

4.2.2 SRGAN

When training SRGAN, the GT and input are the middle layer HR image (0 μm\mu m) and corresponding degraded LR image respectively. The degraded method is consistent with our model.

4.2.3 Deep-Z

As for Deep-Z, the input is a 384×384384\times 384 HR image and a digital propagation matrix (DPM, used to indicate the focal depth). GT is a HR image of specific focal plane corresponded to DPM.

4.3 Experiment Results

In this part, we show the overall effect of our model, as well as the individual effect of horizontal super resolution and axial refocusing.

4.3.1 Multi-focal Plane Imaging Using MFPINet

Refer to caption


Figure 2: The process and results of multi-focal plane imaging using MFPINet. Columns from left to right are horizontal super resolution of different positions, reconstructed HR MFP images by MFPINet, mechanical scanning HR MFP images GT, reconstruction error maps between MFPINet’s output and GT, reconstruction error maps between the interpolated input and GT. The red curve in the second and third columns represents the gray intensity distribution of pixels between two points.

Fig. 2 shows the reconstructed HR MFP images from a single 2D LR image. The focal depth range of the reconstructed MFP images is -13.5 μm\mu m to 13.5 μm\mu m with an interval of 2.7 μm\mu m. The Output and GT columns in Fig. 2 represent HR MFP images of MFPINet reconstruction and mechanical scanning separately. The right two columns in Fig. 2 are reconstruction error maps. |OutputGT||\textit{Output}-\textit{GT}| represents the MAE between MFPINet’s output and GT. |BicubicGT||\textit{Bicubic}-\textit{GT}| represents the MAE between the interpolated input and GT. SSIM, PSNR and MAE in the error maps are based on the current images. From the error maps and corresponding metrics, it can be seen that MFPINet can effectively reconstruct HR MFP images. In addition, the red curves in Output and GT columns represent the gray intensity distribution of pixels between two points. The results indicate that MFPINet can reconstruct well local details. On the left side of Fig. 2, some small patches were taken from different positions. The interpolated LR images Bicubic, Output and GT patches corresponding to the same position were spliced together. The results reflect the super-resolution effect of MFPINet. Notably, the error maps are obtained on corresponding gray images and the metrics were computed on sRGB images.

We provided the reconstructed images of all focal planes of Fig. 2 in Supplementary A. Besides the reconstruction error map, we also analyzed the reconstruction quality from frequency spectrum. In addition, we further analyzed the influence of input image resolution on axial refocusing reconstruction. The experimental results (see Supplementary B) show that the reconstruction effect of MFPINet with HR input is better than that of MFPINet with LR input. This indicates that LR input increases the difficulty of axial refocusing reconstruction.

4.3.2 Super Resolution

MFPINet combines the reconstruction of MFP images and horizontal super resolution. In order to verify the individual effect of super resolution, we compared SRGAN and our method on the test set. Table 1 shows the statistical values on SSIM, PSNR, MSE and MAE metrics. The results indicate that MFPINet performs slightly better than SRGAN on the four metrics. Fig. 3 shows the images generated by the two models. The left four columns are the original sRGB images, and the right three columns correspond to reconstruction error maps with SSIM, PSNR and MAE metrics. From the error maps, it can be seen that our method has smaller error in cytoplasm while SRGAN has smaller error in nucleus. Overall, our method is consistent with SRGAN. The results demonstrate that MFPINet is effective for horizontal super resolution while generating MFP images. Notably, SRGAN was trained and tested on the middle layer (0 μm\mu m) images.

Table 1: Comparison of SRGAN and MFPINet
Method SSIM PSNR MSE MAE
SRGAN 0.952 38.05 10.67 2.16
MFPINet 0.954 38.63 9.65 1.79

Refer to caption

Figure 3: HR images generated by bicubic interpolation, SRGAN and our method. The images of our method shown in the figure are taken from 0 μm\mu m layer of reconstructed MFP images by MFPINet. The meaning of other elements in the figure is similar to Fig. 2

4.3.3 Multi-focal Plane Reconstruction

In this section, we compared Deep-Z and our method on the test set to demonstrate the refocusing ability of MFPINet. Table 2 shows the statistical values of four metrics.

Table 2: Comparison of Deep-Z and MFPINet
Focal Depth Deep-Z MFPINet
PSNR SSIM MSE MAE PSNR SSIM MSE MAE
-13.5μ\mum 37.01 0.942 17.61 2.42 38.20 0.957 12.30 2.02
-10.8μ\mum 36.92 0.942 18.32 2.40 37.91 0.955 13.33 2.04
-8.1μ\mum 36.71 0.938 19.25 2.41 37.46 0.950 14.67 2.11
-5.4μ\mum 36.43 0.933 19.86 2.45 36.98 0.944 15.68 2.19
-2.7μ\mum 36.75 0.935 17.13 2.36 36.78 0.940 15.43 2.22
0.0μ\mum 46.50 0.993 1.64 0.73 38.64 0.954 9.65 1.79
2.7μ\mum 35.78 0.922 20.53 2.60 36.00 0.928 17.90 2.39
5.4μ\mum 36.67 0.932 17.13 2.42 37.63 0.944 12.80 2.15
8.1μ\mum 38.22 0.953 11.65 2.11 39.56 0.963 7.82 1.84
10.8μ\mum 38.68 0.957 10.36 2.03 40.24 0.967 6.56 1.73
13.5μ\mum 38.66 0.958 10.45 2.04 40.32 0.968 6.42 1.72

As can been seen from the table, MFPINet slightly exceeds Deep-Z on all focal planes except the layer 0 μm\mu m. This phenomenon is caused by different inputs. The input of Deep-Z is the layer 0 μm\mu m HR image, while the input of MFPINet is the layer 0 μm\mu m LR image. For the focal plane of 0 μm\mu m, Deep-Z degenerates into identity mapping, while our method degenerates into super resolution. So the effect of Deep-Z is much higher than our method in this focal plane. In order to further explore the difference between the two methods, we visualized their generated images in Fig. 4.

Refer to caption

Figure 4: MFP images generated by Deep-Z and our method. The meaning of other elements in the figure is similar to Fig. 2

The left three columns in Fig. 4 represent the original sRGB images and the right two columns represent reconstruction error maps. It can be seen that our method perform better than Deep-Z on all focal planes except the layer 0 μm\mu m. The error maps also show that the images reconstructed by the two methods have their own characteristics. Deep-Z’s reconstructed nuclei near the layer 0 μm\mu m are slightly better. Because the input image of Deep-Z has not been degraded and adjacent focal planes can learn more details from the HR input. However, as the focal plane becomes farther and farther, Deep-Z gradually loses the advantage of HR input, and the quality of generated images decreases. On the contrary, MFPINet is better than Deep-Z on non-middle focal planes. Notably, the images on focal planes more than 10 μm\mu m from the middle generally are overall blurred, thus showing higher metrics. In summary, although the input of our model is a LR image, the quality of reconstructed MFP images is consistent with that of Deep-Z. This results indicates that it is feasible to reconstruct high-quality MFP images using a single LR wide-field image.

4.3.4 Computational Efficiency

The purpose of the proposed MFPINet is to speed up 3D imaging. In order to verify the actual performance of MFPINet, we analyzed the complexity of MFPINet and tested the actual running time. Specifically, we tested the running time for obtaining a 768×\times768×\times11×\times3 size MFP image on a single RTX 2080ti GPU, and analyzed parameters and giga floating-point operations per second (GFLOPS) of MFPINet. Table 3 shows the results. Time column is the average time of one thousand calculations. It can be seen that MFPINet is approximately 24 times faster than Deep-Z when reconstructing 11-layer MFP images of the same size. At the same time, the parameters of MFPINet is only 1/8 of Deep-Z. In summary, the designed MFPINet is efficient and has a potential to be applied in actual scenarios.

Table 3: Computational efficiency of Deep-Z and MFPINet
Method GFLOPS Params (M) Time (ms)
Deep-Z (Single layer) 762.4 19.41 61.3
Deep-Z (MFP) 8386.4 19.41 673.8
MFPINet 264.2 2.50 27.8

5 Conclusion

The acquisition speed of HR 3D images is limited by axial scanning. To address the challenge, we propose a generative adversarial network named MFPINet to reconstruct HR MFP images from a single 2D wide filed microscopic image. The generator and discriminator are designed to refocus all focal planes at one time and post-sampling strategy is adopted to improve reconstruction efficiency. We conduct extensive experiments on cytology microscopy images and demonstrate that MFPINet performs well on both axial refocusing and horizontal super resolution. Experiments further show MFPINet is approximately 24 times faster than current refocusing methods for reconstructing the same volume images.

We demonstrate that MFPINet can reconstruct a 3D image with a depth of 29.7 μm\mu m from a single 2D image, but the experimental results show that the reconstruction effect far away from the center layer will be attenuated. Therefore, how to reconstruct deeper 3D images with high quality is still a challenge. For example, the density of objects in input images may influence the effect and depth of 3D reconstruction. We will further explore the problem in the future.

Acknowledgment

We thank the Optical Bioimaging Core Facility of Wuhan National Laboratory for Optoelectronics, Huazhong University of Science and Technology for the support in data acquisition.

References

  • [1] A. Bar-Zion, C. Tremblay-Darveau, O. Solomon, D. Adam, and Y. C. Eldar, “Fast vascular ultrasound imaging with enhanced spatial resolution and background rejection,” IEEE Transactions on Medical Imaging, vol. 36, no. 1, pp. 169–180, 2017.
  • [2] W. Denk, J. H. Strickler, and W. W. Webb, “Two-photon laser scanning fluorescence microscopy,” Science, vol. 248, no. 4951, pp. 73–76, 1990.
  • [3] J. Pawley, Handbook of biological confocal microscopy.   Springer Science & Business Media, 2006, vol. 236.
  • [4] R. Tomer, M. Lovett-Barron, I. Kauvar, A. Andalman, V. M. Burns, S. Sankaran, L. Grosenick, M. Broxton, S. Yang, and K. Deisseroth, “Sped light sheet microscopy: fast mapping of biological system structure and function,” Cell, vol. 163, no. 7, pp. 1796–1806, 2015.
  • [5] P. P. Laissue, R. A. Alghamdi, P. Tomancak, E. G. Reynaud, and H. Shroff, “Assessing phototoxicity in live fluorescence imaging,” Nature Methods, vol. 14, no. 7, pp. 657–661, 2017.
  • [6] J. P. Nguyen, F. B. Shipley, A. N. Linder, G. S. Plummer, M. Liu, S. U. Setru, J. W. Shaevitz, and A. M. Leifer, “Whole-brain calcium imaging with cellular resolution in freely behaving caenorhabditis elegans,” Proceedings of the National Academy of Sciences, vol. 113, no. 8, pp. E1074–E1081, 2016.
  • [7] S. R. P. Pavani, M. A. Thompson, J. S. Biteen, S. J. Lord, N. Liu, R. J. Twieg, R. Piestun, and W. Moerner, “Three-dimensional, single-molecule fluorescence imaging beyond the diffraction limit by using a double-helix point spread function,” Proceedings of the National Academy of Sciences, vol. 106, no. 9, pp. 2995–2999, 2009.
  • [8] M. Levoy, R. Ng, A. Adams, M. Footer, and M. Horowitz, “Light field microscopy,” in ACM SIGGRAPH 2006 Papers, 2006, pp. 924–934.
  • [9] J. Rosen and G. Brooker, “Non-scanning motionless fluorescence three-dimensional holographic microscopy,” Nature Photonics, vol. 2, no. 3, p. 190, 2008.
  • [10] J. Ma, J. Yu, S. Liu, L. Chen, X. Li, J. Feng, Z. Chen, S. Zeng, X. Liu, and S. Cheng, “Pathsrgan: Multi-supervised super-resolution for cytopathological images using generative adversarial network,” IEEE Transactions on Medical Imaging, 2020.
  • [11] H. Wang, Y. Rivenson, Y. Jin, Z. Wei, R. Gao, H. Günaydın, L. A. Bentolila, C. Kural, and A. Ozcan, “Deep learning enables cross-modality super-resolution in fluorescence microscopy,” Nature Methods, vol. 16, no. 1, pp. 103–110, 2019.
  • [12] H. Zhang, C. Fang, X. Xie, Y. Yang, W. Mei, D. Jin, and P. Fei, “High-throughput, high-resolution deep learning microscopy based on registration-free generative adversarial network,” Biomedical Optics Express, vol. 10, no. 3, pp. 1044–1063, 2019.
  • [13] J. Hatvani, A. Basarab, J. Tourneret, M. Gyöngy, and D. Kouamé, “A tensor factorization method for 3-d super resolution with application to dental ct,” IEEE Transactions on Medical Imaging, vol. 38, no. 6, pp. 1524–1531, 2019.
  • [14] C. You, G. Li, Y. Zhang, X. Zhang, H. Shan, M. Li, S. Ju, Z. Zhao, Z. Zhang, and W. Cong, “Ct super-resolution gan constrained by the identical, residual, and cycle learning ensemble (gan-circle),” IEEE Transactions on Medical Imaging, vol. 39, no. 1, pp. 188–203, 2019.
  • [15] Q. Lyu, H. Shan, C. Steber, C. Helis, C. T. Whitlow, M. Chan, and G. Wang, “Multi-contrast super-resolution mri through a progressive network,” IEEE Transactions on Medical Imaging, pp. 1–1, 2020.
  • [16] C.-H. Pham, A. Ducournau, R. Fablet, and F. Rousseau, “Brain mri super-resolution using deep 3d convolutional networks,” in 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017).   IEEE, Conference Proceedings, pp. 197–200.
  • [17] Y. Wu, Y. Rivenson, H. Wang, Y. Luo, E. Ben-David, L. A. Bentolila, C. Pritz, and A. Ozcan, “Three-dimensional virtual refocusing of fluorescence microscopy images using deep learning,” Nature Methods, vol. 16, no. 12, pp. 1323–1331, 2019.
  • [18] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” in Advances in Neural Information Processing Systems, Conference Proceedings, pp. 2672–2680.
  • [19] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical Image Computing and Computer-Assisted Intervention.   Springer, Conference Proceedings, pp. 234–241.
  • [20] X. Wang, K. Yu, S. Wu, J. Gu, Y. Liu, C. Dong, Y. Qiao, and C. Change Loy, “Esrgan: Enhanced super-resolution generative adversarial networks,” in Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 0–0.
  • [21] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Conference Proceedings, pp. 770–778.
  • [22] S. Ioffe and C. Szegedy, “Batch normalization: Accelerating deep network training by reducing internal covariate shift,” arXiv preprint arXiv:1502.03167, 2015.
  • [23] C. Dong, C. C. Loy, K. He, and X. Tang, “Image super-resolution using deep convolutional networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 38, no. 2, pp. 295–307, 2015.
  • [24] B. Lim, S. Son, H. Kim, S. Nah, and K. Mu Lee, “Enhanced deep residual networks for single image super-resolution,” in Proceedings of the IEEE conference on computer vision and pattern recognition workshops, 2017, pp. 136–144.
  • [25] C. Ledig, L. Theis, F. Huszár, J. Caballero, A. Cunningham, A. Acosta, A. Aitken, A. Tejani, J. Totz, Z. Wang et al., “Photo-realistic single image super-resolution using a generative adversarial network,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 4681–4690.
  • [26] Y. Guo, J. Chen, J. Wang, Q. Chen, J. Cao, Z. Deng, Y. Xu, and M. Tan, “Closed-loop matters: Dual regression networks for single image super-resolution,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 5407–5416.
  • [27] J. Kim, J. Kwon Lee, and K. Mu Lee, “Accurate image super-resolution using very deep convolutional networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 1646–1654.
  • [28] Y. Rivenson, H. Ceylan Koydemir, H. Wang, Z. Wei, Z. Ren, Y. Zhang, Z. Gorocs, K. Liang, D. Tseng et al., “Deep learning enhanced mobile-phone microscopy,” Acs Photonics, vol. 5, no. 6, pp. 2354–2364, 2018.
  • [29] H. Zhou, R. Cai, T. Quan, S. Liu, S. Li, Q. Huang, A. Ertürk, and S. Zeng, “3d high resolution generative deep-learning network for fluorescence microscopy imaging,” Optics Letters, vol. 45, no. 7, pp. 1695–1698, 2020.
  • [30] J. Page, F. Saltarin, Y. Belyaev, R. Lyck, and P. Favaro, “Learning to reconstruct confocal microscopy stacks from single light field images,” arXiv preprint arXiv:2003.11004, 2020.
  • [31] W. Shi, J. Caballero, F. Huszár, J. Totz, A. P. Aitken, R. Bishop, D. Rueckert, and Z. Wang, “Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Conference Proceedings, pp. 1874–1883.
  • [32] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.
  • [33] A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, and L. Antiga, “Pytorch: An imperative style, high-performance deep learning library,” in Advances in Neural Information Processing Systems, Conference Proceedings, pp. 8024–8035.