Physics-enhanced machine learning for virtual fluorescence microscopy
Abstract
This paper introduces a new method of data-driven microscope design for virtual fluorescence microscopy. Our results show that by including a model of illumination within the first layers of a deep convolutional neural network, it is possible to learn task-specific LED patterns that substantially improve the ability to infer fluorescence image information from unstained transmission microscopy images. We validated our method on two different experimental setups, with different magnifications and different sample types, to show a consistent improvement in performance as compared to conventional illumination methods. Additionally, to understand the importance of learned illumination on inference task, we varied the dynamic range of the fluorescent image targets (from one to seven bits), and showed that the margin of improvement for learned patterns increased with the information content of the target. This work demonstrates the power of programmable optical elements at enabling better machine learning algorithm performance and at providing physical insight into next generation of machine-controlled imaging systems.
I Introduction

The optical microscope remains a critical tool across a wide variety of disciplines. Examples include high-content screening in biology labs, quality control and defect detection in factories, and automated digitization of pathology slides in clinics. With the continued growth of automated software analysis tools, many microscope images are now rarely viewed directly in their raw format by humans, but are instead commonly processed by a computer first. Examples include the automatic classification of different cell types within large cell cultures [1], segmentation of cancerous areas from thin pathology tissue sections [2], and, as focused upon here, the automatic creation of fluorescent images from bright-field data [3].
Despite the continued rapid development of automated image analysis software, the microscope’s hardware has changed relatively little over the past several centuries. Most current microscopes still consist of standard illumination units and objective lenses that are optimized for direct human inspection. The physics of optical microscopes enforces several physical limitations, including a limited resolution, field-of-view, image contrast, and depth-of-field, for example, which restrict the amount of information that can be captured within each image. The standard design of a microscope biases this limited information towards human analysis, potentially impacting the accuracy of automated analysis.
Here, we attempt to optimize the hardware of a new microscope design to improve the accuracy of automated image processing by a deep neural network (DNN). Our aim in this work is to establish optimal hardware settings to improve the particular DNN task of image inference. To achieve this goal, we present a modified learning network that includes a physical model of our experimental microscope, which we jointly optimize during DNN training. In this work, we limit our physical model to include only the spectral and angular properties of the microscope’s illumination, provided by a programmable LED array, but leave open the possibility of also considering many other important parameters (focus setting, lens design, detector properties) in future work. Our proposed network models the microscope illumination pattern as a set of linear weights that are directly integrated into the DNN, allowing the calculation of gradients through back-propagation and end-to-end optimization during supervised training. After training, the optimized “physical” weights can be interpreted as the distribution of optimal LED brightnesses and colors to use in our experimental imaging setup, which transfers performance gains seen in training to a physical setup.
For the specific goal of DNN-optimized microscope illumination for image-to-image inference, we aim here to train a DNN to convert optimally illuminated microscope imagery into data obtained from a simultaneously captured fluorescence image. This goal of bright-field to fluorescence image inference, also termed “in-silico labeling”[3], has recently received interest as a promising means to avoid the need to fluorescently label specimens and to instead simply rely on post-processing standard bright-field image data. Prior work with this effort has typically required a relatively large data overhead (i.e., acquisition of 10 or more bright-field images per inference task), and has offered a limited amount of physical insight into performance variations as a function of collected data.
In this work, we use a DNN to jointly optimize the illumination to both reduce the number of required images for accurate fluorescence image inference, and to explore how performance scales with the amount of inferred information. To achieve this latter goal, we vary the precision of the desired fluorescent output to consider a range of inference tasks, from binary image segmentation (one bit) to the prediction of complete fluorescent images (seven bits). For each level of precision, we examine how the converged LED pattern changes, and how these patterns trend with precision. The optimized patterns not only yield higher accuracy results, but also provide a certain degree of physical intuition between scattered bright-field light and fluorescent emission that can be used to improve future data collection strategies.
II Related work


In recent years, convolutional neural networks (CNNs) have become commonplace for both medical and natural image processing [4]. Segmenting images to find specific cell types or sub-cellular features (e.g. cell nuclei), for example, is now a common biomedical image analysis task that CNNs excel at. The U-net structure [5], perhaps one of the most widely used CNN architectures, has been applied across a wide variety of segmentation tasks [6] and makes efficient use of annotated data during training.
As the use of neural networks continues to increase in popularity, many researchers are now also applying them to automatically analyze fluorescent imagery. Belthangady et al. [7] recently reviewed this increasing body of work and summarized it into two general categories: virtual staining and fluorescent image enhancement.
“In-silico” labelling is the process of using bright-field images to predict fluorescent images. While still a relatively new concept, several recent works demonstrate that DNNs can be quite effective at this task, which suggests that certain future experiments may forgo fluorescent imaging and staining entirely. Christiansen et al. [3] first demonstrated this concept by developing a CNN which predicted seven distinct fluorescent channels from a set of several dozen uniquely focused bright-field images. Ounkomol et al. [8] then used a modified U-net structure to predict 3D fluorescent images using 3D bright-field data. Chen et al. subsequently [9] applied this technology through the design of an Augmented Reality Microscope which overlays a neural network’s predictions on top of the underlying bright-field data. In a related effort, Rivenson et al. [10] used a neural network to infer histologically stained images from unstained tissue with high accuracy and visual quality. These works all demonstrate that it is possible to infer information revealed by a fluorescent or histological stain from unmodified bright-field data, albeit at varying levels of accuracy.
Fluorescent image enhancement focuses on improving the quality of existing fluorescent images. Weigert et al. [11] developed a content-aware image restoration method powered by a CNN. This work showed that fluorescent image restoration was possible by predicting high resolution fluorescent images from ones which were under-sampled. In an earlier work, Weigert et al. [12] used a CNN to perform isotropic reconstruction of 3D fluorescent data. Through these works, we can infer that data contained within fluorescent images is potentially redundant, and through the assumption of key underlying features, it is at times possible to enhance image quality via CNN post-processing.
However, in most studies, the focus is on processing data that has already been captured, rather than attempting to influence or improve the image acquisition process. While an early work used simple neural networks to effectively design components of optical systems [13], the first work (to the best of our knowledge) to examine hardware optimization in the context of CNNs was by Chakrabarti [14], who presented an optimal pixel level color-filter layout for color image reconstruction. A number of subsequent works have considered how to merge the optimization of various imaging hardware components into a differentiable optimization network [15, 16, 17, 18, 19, 20, 21, 22, 23]. However, few of these works have proposed the use of CNNs to optimize the image capture process for automated decision tasks (e.g., classification, object detection, image segmentation, etc.).
Such an approach was recently considered by Horstmeyer et al. [24], who suggested the use of a ”physical layer” in a CNN to identify optimized illumination patterns that can improve automated classification accuracy of the malaria infected blood cell detection. Other related work has used DNNs to design custom optical elements to improve an image’s depth-of-field [25], to inform new types of illumination for improved phase contrast imaging [26, 27], achieve superior resolution [28, 29], or infer color from a gray-scale image [19] . The goal of these works was not to improve the accuracy of automated image inference, as considered here.
In terms of using a microscope’s illumination to achieve new functionalities, prior work has clearly shown the benefits of applying programmable LED array lighting. This includes variable bright-field and dark-field imaging [30], measurement of a specimen’s surface gradient[26, 27], and quantitative phase imaging [31] to name a few. Mathematically rigorous methods have also been used to combine variably-illuminated images to increase image resolution. Two prominent examples are Fourier ptychographic microscopy (FPM) [32] and structured illumination microscopy (SIM) [33]. These works highlight the benefit of controlling illumination and provide ample evidence that it should be targeted for optimization.
III Methods
III-A Image Formation
In this work, we collected and processed data across two different experimental setups. Both setups used an off-the-shelf LED array (Adafruit product ID 607), which provided the capability to turn on specific LED array patterns at different colors (e.g., red, green and blue) for multi-angle and multi-spectral specimen illumination (Figure 1A). In each setup, the LED array was placed sufficiently far from the sample plane such that the illumination from each LED could be modelled as a plane wave at the sample, propagating at a unique angle corresponding to its relative position. This type of illumination has been previously used for FPM [32, 34], as well as phase contrast [26, 27] and super-resolved 3D imaging [35, 36, 37], among other enhancement techniques. As noted above, illumination from a wide variety of angles and colors provides diverse information about biological specimens, which are primarily transparent, that is not available under normal illumination (see Figure 2). However, it is not directly obvious which type of illumination is best-suited for mapping non-fluorescent image data into fluorescence stained image data. The optimal patterns will likely depend upon which sample features are fluorescently labeled, as well as properties of the specimen’s complex index of refraction. In this work, we consider two separate fluorescent labeled sample categories - one in which the cell nucleus is labeled, and a second in which the cell membrane is labeled - to verify this hypothesis. In addition, the specific inference task may also impact the optimal illumination distributions (i.e., segmentation versus image formation) in a non-obvious way. To solve for such task specific illumination patterns, we propose to use a modified DNN, which includes a ”physical layer” [24] for joint hardware optimization.
In this type of approach, the supervised machine learning network itself jointly determines the optimal distribution of LEDs’ brightness and colors to use for sample illumination, and to perform the subsequent image-to-image inference task. As we describe below, the process of LED illumination can be modeled by a single physical layer, which we prepend to the front of a DNN U-Net architecture for improved image-to-image inference. After the network training is complete, the optimized weights within the physical layer (in this case, representing the angular and spectral distribution of the sample illumination) inform us of a better optical design for each specific inference task at hand. Figure 3 shows how this process is realized for both training and inference.
To mathematically model image formation, we express the th specimen of interest as the complex function . Given LEDs within the array, we can denote the amplitude of the plane wave emitted by each colored LED emitting at center wavelength as a weight for . Each LED in the array illuminates the sample with a coherent plane wave from a unique angle . Since, the LEDs are mutually incoherent with respect to each other, the image formed by an illumination composed of multiple LEDs is equivalent to the incoherent sum of images obtained by illuminating the member LEDs individually. Assuming a thin specimen, for example, we may write the th detected quasi-monochromatic image as the incoherent sum,
(1) |
where, describes the plane wave generated by the th LED with intensity at wavelength across the sample plane with coordinate , denotes the plane wave transverse wavevector with respect to the optical axis, and denotes the microscope’s coherent point-spread function and describes the imaging system blur.
As the LED brightness is a scalar quantity, we can factor it out of the summation in Eq. 1. Furthermore, if we denote the image of the th sample formed when it is illuminated by the th LED at a fixed brightness and wavelength as , then we arrive at a simple linear model for image formation for the th sample:
(2) |
The detected image under illumination from a particular LED pattern, where each LED has a brightness , is equal to the weighted sum of images captured by turning on each LED individually. The detected image is then entered into the neural network for processing. Equation 2 represents our physical layer for microscope illumination, and the fully differentiable optimization pipeline is shown in Figure 3.
To find the set of weights which best parameterize the transform in Eq. 2 for the subsequent inference task, we must have access to all uniquely illuminated images during network training. However, once training is complete, the set of optimized weights are then mapped onto the physical LED matrix to allow for acquisition of a single optimally illuminated image, which is then processed by the remainder of the DNN for the fluorescent image inference task.
III-B Network Design
We used a consistent U-Net architecture that we prepended with a physical layer to model Equation. 2. The exact architecture and layer configurations, used in all the reported experiments, is detailed in Appendix B. The weights within the physical layer were unconstrained, meaning they could take on both positive and negative values. To experimentally realize an unconstrained set of illumination weights , we captured two images (instead of one) - a first with the positive set of weights, and a second with the negative set, before subtracting the second image from the first for the final result.
Although the experimentally captured measurements inherently contain Poisson noise, the digital simulation of multi-LED images used during training will have an artifactually inflated SNR (e.g., averaging images produces a improvement in SNR). To compensate for this we introduced a Noise Layer which adds dynamically generated Gaussian random noise to the data after the first physical layer:
(3) |
Additive noise is modelled on a per-pixel level with Equation 3, using a hyperparameter to control the scale of the noise in proportion to the pixel intensity. Note that the variance of the random noise is proportional to the image pixel intensity itself, and thus is consistent with a Poisson noise model.
Finally, an penalty within the network cost function (absolute deviation from zero) was applied to the physical layer weights . This penalty term is given a small weight proportional to the magnitude of the gradients to drive weights to zero, if and only if they are not significantly contributing to the task performance (see details in Appendix B). The addition of this type of penalty reduces variance across random seeds and aids in interpretation of the resulting LED patterns.
IV Experiments
We examined two types of fluorescently labeled biological specimen in the following experiments. For each experiment, a square grid of multi-color LEDs was used for illumination, where each multi-color LED included 3 spectral channels (with center wavelengths 480, 540, and 632 nm), representing different illumination sources. For each imaging experiment, an LED image stack was acquired by turning on each LED individually and capturing a unique image. To provide target labels, each sample was also illuminated with a UV fluorescent excitation source and captured via a separate path containing a fluorescent emission filter, but using the same optical imaging setup that collects the bright-field image data. By using the same optical setup to capture the same FOV, we ensured that sample positioning and lens distortion factors remained constant across bright-field and fluorescent channels. Multiple non-overlapping fields-of-view were captured for each specimen of interest.
After data capture, each full field-of-view dataset was split into pixel sections, with each bright-field section having channels (one channel for each LED configuration). These formed the datacubes, which were fed into our modified DNN for network training. The matching fluorescent channel was used as a target label for training, and is not present during inference. After training, the network produced a 675 element optimizable LED weight vector . To compare our optimized illumination results with other image collection approaches, we ran the same training process using a fixed (i.e., without optimizing the physical layer), where the values correspond to common illumination configurations often seen in microscopy [38]:
-
1.
Center is spatially coherent illumination from only the center LED and 3 colors
-
2.
All is spatially incoherent illumination from all 225 LEDs and 3 colors
-
3.
DC stands for the differential contrast illumination method
-
4.
Off-axis is from an LED located 4 mm off the optical axis, illuminating at approximately
-
5.
Random is a randomly selected set of LED brightness values
For each configuration and examined dataset, the neural network architecture remained unchanged and was trained using the same hyperparameters.

IV-A Data Label Generation
To understand the impact of the optimized illumination pattern , we varied the task difficulty by modifying the fidelity of the target label. The 7 bit ground truth fluorescence image was converted to 7 different images with different bit depths, ranging from 1 bit to 7 bits of dynamic range. Here, we hypothesize that reconstruction of a 1-bit fluorescence image, which is effectively a segmentation mask, is easier than accurately recovering a 7 bit fluorescence image. It should be noted that the input bright-field images (the training data) remains same for all tested cases. This allowed the same data to be used to train a neural network to perform both binary image segmentation task (1 bit precision), and to separately train another network to perform a full fluorescent image inference task (7 bit precision), and to also consider all tasks that span these two common efforts. Figure 4 shows how by varying the precision of the desired network output, it is possible to vary the difficulty of the subsequent image inference task, from producing a prediction that contains bits to bits per pixel.
To achieve a balance of pixel values across each discretized histogram within the label images, we employed a global k-means quantization strategy, where . To find the mean values a naive k-means algorithm was applied on a flattened version of the entire dataset, treating each pixel value as an independent value. The initial mean values were uniformly distributed across the space between zero and one: . Mean values were iteratively adjusted until either all values had converged (with a threshold of: ) or iterations passed.
The inference results were only rounded to meet the appropriate precision of the target bit level, with no further post processing (k-means or otherwise).
IV-B Cell Nuclei
The data used in the first experiment was originally acquired by a FPM microscope [39] and contains 90% confluent HeLa cells, stained with a fluorescent nuclear label (DAPI). The employed microscope included a two-lens arrangement with an f = 200 mm tube lens (ITL200, Thorlabs) and an f = 50 mm Nikon lens (f/1.8D AF Nikkor). This two-lens setup has a collection numerical aperture (NA) of 0.085 with 3.87x magnification. The sample was placed at the front focal plane of the f = 50 mm lens and a CCD detector (pixel size 5.5 m, Prosilica GX6600) captured images of the sample under LED matrix illumination (one LED at a time, as noted above). The LED array was placed 80 mm behind the sample with 32x32 individually addressable elements (pitch size 4 mm), of which only the inner 15x15 were utilized. Full-FOV images from all illumination angles and colors were divided into 442 unique datacubes. The set of datacubes was randomly split into training, testing, and validation sets containing 356, 48, and 48 images respectively.




IV-C Cell Membrane
For the second set of experiments, we captured images of Pan 16 pancreatic cancer cells stained using CellMask Green plasma membrane stain (C37608). Similarly to the dataset described in Section IV-B, an identical LED array was placed 60mm beneath the biological sample and of which the inner 15x15 grid was used. An Olympus PlanN 0.25NA 10x objective lens was used in conjunction with a Basler Ace (acA4024-29um) CMOS sensor. To capture the matching fluorescence data a set of Thorlabs filters (MDF-GFP - GFP Excitation, Emission, and Dichroic Filters) were inserted into the optical setup. Figure 1 shows a photo of the setup used. The data was split into test/train/validation sets such that data from each sample only existed in one of the sets, providing a robust means of gauging generalization and overfitting. When split into images there were 820/108/108 samples in the train, validation and test sets respectively.
V Results and Discussion
V-A Performance
Across both tasks and all bit depth configurations, we found that the optimized illumination patterns determined by our physical layer outperformed all other illumination arrangements (coherent and incoherent bright-field, dark-field, random and phase contrast). Figure 5 plots the mean squared error (MSE) of each inference task as a function of inference bit depth (i.e., inference difficulty), cell type, and physical layer parameterization. This main result shows that jointly optimized microscope hardware can lead to superior image inference performance across a wide range of tasks - ranging from image segmentation to virtual fluorescence imaging.
Although MSE provides an accurate measure of the task performance, it is less suited to understand the perceptual quality of the results. We thus also computed the structural similarity index (SSIM) [40] across the test set for both tasks (results shown in Figure 5). Here we can see that the reduction in mean squared error brought by the optimized illumination carries onto a higher perceptual quality.

V-B Performance versus inference task
By varying the bit depth of the output label, we were able to test our approach with seven different configurations. We first observe that the gap in performance (both MSE and SSIM – Figure 5) between the inference results created by the optimized LED pattern (”learned”), and the alternative standard LED illumination patterns, is quite consistent. Given that the pixel values within all of the output labels are normalized to a range, this result is somewhat consistent with our expectations. The increase in difficulty of the task is partially offset through the decrease in the ”cost” of being wrong.
To get a better understanding of the interaction of task difficulty and performance we examined the relative performance of the ”learned” method compared to the best of the ”standard” methods. To do this we constructed a relative performance metric, , which is the quotient of the minimum MSE among the standard patterns’ MSE and the learned configurations MSE, for each bit depth . See equations 4 and 5.
(4) | ||||
(5) |
Within Figure 6 we show that the relative performance of the learned LED configuration improves with difficulty when compared to the best standard configuration. This illustrates how the joint optimization process provides a consistent performance improvement that becomes more important as the difficulty of the task increases.



V-C LED Parameterization
In addition to improved performance, the DNN-optimized LED illumination patterns also offer several interesting physical insights. To aid in interpretation of the patterns, each LED was normalized by it’s exposure (as shown in Figure 7), as using different exposures per LED enabled us to fully utilize the dynamic range of our image sensor. Figure 8 displays the learned patterns as a function of inference task (1-bit to 7-bit) on a per-color-channel basis for both tested specimens.
These illumination patterns highlight the importance of two primary features for transmitting visible light information that is most relevant to a particular fluorescent specimen and stain. First, producing phase contrast by illuminating differently within the bright-field (the inner 3x3 LEDs for the HeLa task, and the inner 5x5 LEDs for the PAN task) and dark-field appears important across all tasks. Second, providing color contrast (e.g., green for the positive image and red for the negative image) also appears important, most notably for the HeLa nuclei prediction task. The pattern differences on a per-task (HeLa vs. PAN) can be attributed both to the differences between setups (notably that the PAN task setup included more bright-field LEDs) as well as the differences between samples. It should be noted that for the PAN task the LED patterns appear shifted, this is due to an indexing error which occurred during the data-preprocessing steps and does not hold any physical significance.
We are also able to compare the distribution of optimized LED patterns as a function of task difficulty (i.e., output label bit depth). Although the importance of establishing a learned pattern to improve network performance is clear (section V-B), the trend among the spatial patterns themselves is less so. We found that, in general, the lower bit-depth patterns generated images which put more emphasis on higher frequency information, and vice versa. Figure 9 shows how the average resolved spatial frequency power decreases with increased bit depth, although this trend was noisy and several bit-depths were exceptions. The values were determined by finding the first moment of each illuminated images’ spatial frequency power representation along it’s radius, more details are provided in Appendix A. We hope to examine more possible trends in future work.


Although there is only a small amount of variation present in the overall system performance, the LED patterns themselves do vary across random seeds. Figure 12 shows examples and statistics of this variance across both tasks. We hypothesize that there are two factors driving this variance. First, images acquired from high-angle illumination (particularly dark-field) had lower intensity values on average and therefore contributed less to the synthesized image. This translates to a higher degree of allowed variance in the optimized result, since as the brightness value of an LED illuminating from a high angle doesn’t impact the image as much as the more centrally located LEDs, and hence the gradients driving its value will be fairly weak. The use of the L1 norm penalty reduced this variance across both tasks.
Second, there is clearly a coupling between the solution of the neural network and the parameterization of the LED array. The joint optimization procedure followed in this work uses this coupling to get higher performance results, and shows that higher performance is consistently obtained. But, as an adverse effect, it also decreases our ability to draw conclusions from the end parameterization of the LED array. This is because when a deep neural network is being trained, it can arrive at one of many local minima (finding the global minima is highly unlikely). Due to the network arriving and terminating at a local minima, the LED parameterization is such that the lighting is optimized for that specific local minima, instead of what might be globally optimal. Within section V-D, we comment on one method of alleviating the impact of arriving at local optima through the use of targeted regularization.
V-D Impact of Regularization on LED Parameterization
Many of the trials presented here were run both with and without the addition of regularization on the weights representing the illumination pattern. Upon comparison, we found that the addition of this regularization did not impact our target performance statistic (MSE on the test set), however it did have a large effect on the LED patterns themselves. Most notably, when the LED weights are left unconstrained, the network tends towards an illumination pattern configuration using many LEDs at high dark-field angles. When constrained to use fewer LEDs, the optimized pattern instead utilizes bright-field and dark-field LEDs at lower angles. Although information contributed from higher angles may still be useful overall, the L1 norm prioritizes illumination angles that consistently contribute large amounts of information. In this respect a standard L1 norm may not be ideal for this kind of optimization, investigation into a more suitable regularization strategy will be investigated in future work.
In Figure 10, we show side-by-side comparisons for both cases across a representative set of bit depths. This comparison clearly shows that not only do the regularized patterns contain less within-trial spatial variance, but also they are potentially more spatially interpretable. We postulate that this is because the added regularization is having the desired effect of forcing the optimization to find a parameterization that is robust to perturbations (caused by the addition of noise) while simultaneously having the minimum overall energy (caused by the L1 penalty). While performance in the simulated environment remains the same (with and without regularization), we hypothesize that these spatially smoother LED patterns are more robust to changes in the eventual physical setup. However, we leave testing this hypothesis for future work.


VI Conclusion
In summary, we have presented a novel method for developing image capture systems that can be optimized for deep learning based image-to-image inferencing tasks. By placing the physical parameters of the microscope in the gradient pathway, we jointly optimized the way an image was captured with the way it was processed. This allows imaging systems to sample data, not based on what a human experimenter prefers, but governed through optimization.
The two experiments we performed show that this technique is robust in its improvement, with MSE under each configuration being minimized by the ”learned” or jointly optimized approach. Furthermore, our experiments show that this technique provides performance improvement for even the simplest version of inference and this improvement increases with problem difficulty and illumination sensitivity.
Our experimental results show that the physical parameters of a microscope play an important role in deep learning image-to-image inference systems. By allowing the joint optimization of illumination and image processing we achieve consistently better performance than all tested alternatives. We hope our results continue to motivate the imaging and machine learning community to re-examine how they capture data and continue to develop understanding of the connection between data capture and data processing.
VII Acknowledgements
We would like to thank Jaebum Chung for collecting and providing the data used for the HeLa task, as well as Syliva Ceballos and William Eldridge for preparing the Pan 16 samples. Research reported in this publication was supported by the National Institute Of Neurological Disorders And Stroke of the National Institutes of Health under award number RF1NS113287, as well as the Duke-Coulter Translational Partnership.
References
- [1] D. A. Van Valen, T. Kudo, K. M. Lane, D. N. Macklin, N. T. Quach, M. M. DeFelice, I. Maayan, Y. Tanouchi, E. A. Ashley, and M. W. Covert, “Deep learning automates the quantitative analysis of individual cells in live-cell imaging experiments,” PLoS computational biology, vol. 12, no. 11, 2016.
- [2] J. Xu, C. Zhou, B. Lang, and Q. Liu, “Deep learning for histopathological image analysis: Towards computerized diagnosis on cancers,” in Deep Learning and Convolutional Neural Networks for Medical Image Computing, pp. 73–95, Springer, 2017.
- [3] E. M. Christiansen, S. J. Yang, D. M. Ando, A. Javaherian, G. Skibinski, S. Lipnick, E. Mount, A. O’Neil, K. Shah, A. K. Lee, et al., “In silico labeling: predicting fluorescent labels in unlabeled images,” Cell, vol. 173, no. 3, pp. 792–803, 2018.
- [4] G. Litjens, T. Kooi, B. E. Bejnordi, A. A. A. Setio, F. Ciompi, M. Ghafoorian, J. A. Van Der Laak, B. Van Ginneken, and C. I. Sánchez, “A survey on deep learning in medical image analysis,” Medical image analysis, vol. 42, pp. 60–88, 2017.
- [5] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical image computing and computer-assisted intervention, pp. 234–241, Springer, 2015.
- [6] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015 (N. Navab, J. Hornegger, W. M. Wells, and A. F. Frangi, eds.), (Cham), pp. 234–241, Springer International Publishing, 2015.
- [7] C. Belthangady and L. A. Royer, “Applications, promises, and pitfalls of deep learning for fluorescence image reconstruction,” Nature methods, pp. 1–11, 2019.
- [8] C. Ounkomol, S. Seshamani, M. M. Maleckar, F. Collman, and G. R. Johnson, “Label-free prediction of three-dimensional fluorescence images from transmitted-light microscopy,” Nature methods, vol. 15, no. 11, p. 917, 2018.
- [9] P.-H. C. Chen, K. Gadepalli, R. MacDonald, Y. Liu, S. Kadowaki, K. Nagpal, T. Kohlberger, J. Dean, G. S. Corrado, J. D. Hipp, et al., “An augmented reality microscope with real-time artificial intelligence integration for cancer diagnosis,” Nature medicine, vol. 25, no. 9, pp. 1453–1457, 2019.
- [10] Y. Rivenson, H. Wang, Z. Wei, K. de Haan, Y. Zhang, Y. Wu, H. Günaydın, J. E. Zuckerman, T. Chong, A. E. Sisk, et al., “Virtual histological staining of unlabelled tissue-autofluorescence images via deep learning,” Nature biomedical engineering, vol. 3, no. 6, p. 466, 2019.
- [11] M. Weigert, U. Schmidt, T. Boothe, A. Müller, A. Dibrov, A. Jain, B. Wilhelm, D. Schmidt, C. Broaddus, S. Culley, et al., “Content-aware image restoration: pushing the limits of fluorescence microscopy,” Nature methods, vol. 15, no. 12, pp. 1090–1097, 2018.
- [12] M. Weigert, L. Royer, F. Jug, and G. Myers, “Isotropic reconstruction of 3d fluorescence microscopy images using convolutional neural networks,” in International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 126–134, Springer, 2017.
- [13] J. Macdonald, A. J. Breese, and N. L. Hanbury, “Optimization of a lens design using a neural network,” in Applications of Artificial Neural Networks IV, vol. 1965, pp. 431–442, International Society for Optics and Photonics, 1993.
- [14] A. Chakrabarti, “Learning sensor multiplexing design through back-propagation,” in Advances in Neural Information Processing Systems, pp. 3081–3089, 2016.
- [15] S. Diamond, V. Sitzmann, S. Boyd, G. Wetzstein, and F. Heide, “Dirty pixels: Optimizing image classification architectures for raw sensor data,” arXiv preprint arXiv:1701.06487, 2017.
- [16] K. Kulkarni, S. Lohit, P. Turaga, R. Kerviche, and A. Ashok, “Reconnet: Non-iterative reconstruction of images from compressively sensed measurements,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 449–458, 2016.
- [17] V. Sitzmann, S. Diamond, Y. Peng, X. Dun, S. Boyd, W. Heidrich, F. Heide, and G. Wetzstein, “End-to-end optimization of optics and image processing for achromatic extended depth of field and super-resolution imaging,” ACM Transactions on Graphics (TOG), vol. 37, no. 4, pp. 1–13, 2018.
- [18] J. Chang, V. Sitzmann, X. Dun, W. Heidrich, and G. Wetzstein, “Hybrid optical-electronic convolutional neural networks with optimized diffractive optics for image classification,” Scientific reports, vol. 8, no. 1, pp. 1–10, 2018.
- [19] E. Hershko, L. E. Weiss, T. Michaeli, and Y. Shechtman, “Multicolor localization microscopy and point-spread-function engineering by deep learning,” Optics express, vol. 27, no. 5, pp. 6158–6183, 2019.
- [20] Y. Xue, S. Cheng, Y. Li, and L. Tian, “Reliable deep-learning-based phase imaging with uncertainty quantification,” Optica, vol. 6, no. 5, pp. 618–629, 2019.
- [21] P. del Hougne, M. F. Imani, A. V. Diebold, R. Horstmeyer, and D. R. Smith, “Learned integrated sensing pipeline: Reconfigurable metasurface transceivers as trainable physical layer in an artificial neural network,” Advanced Science, vol. 7, no. 3, p. 1901913, 2020.
- [22] H. Sun, A. V. Dalca, and K. L. Bouman, “Learning a probabilistic strategy for computational imaging sensor selection,” arXiv preprint arXiv:2003.10424, 2020.
- [23] G. Barbastathis, A. Ozcan, and G. Situ, “On the use of deep learning for computational imaging,” Optica, vol. 6, no. 8, pp. 921–943, 2019.
- [24] A. Muthumbi, A. Chaware, K. Kim, K. C. Zhou, P. C. Konda, R. Chen, B. Judkewitz, A. Erdmann, B. Kappes, and R. Horstmeyer, “Learned sensing: jointly optimized microscope hardware for accurate image classification,” Biomedical Optics Express, vol. 10, no. 12, pp. 6351–6369, 2019.
- [25] V. Sitzmann, S. Diamond, Y. Peng, X. Dun, S. Boyd, W. Heidrich, F. Heide, and G. Wetzstein, “End-to-end optimization of optics and image processing for achromatic extended depth of field and super-resolution imaging,” ACM Transactions on Graphics (TOG), vol. 37, no. 4, p. 114, 2018.
- [26] M. Kellman, E. Bostan, N. Repina, and L. Waller, “Physics-based learned design: Optimized coded-illumination for quantitative phase imaging,” IEEE Transactions on Computational Imaging, 2019.
- [27] B. Diederich, R. Wartmann, H. Schadwinkel, and R. Heintzmann, “Using machine-learning to optimize phase contrast in a low-cost cellphone microscope,” PloS one, vol. 13, no. 3, p. e0192937, 2018.
- [28] Y. F. Cheng, M. Strachan, Z. Weiss, M. Deb, D. Carone, and V. Ganapati, “Illumination pattern design with deep learning for single-shot fourier ptychographic microscopy,” Optics express, vol. 27, no. 2, pp. 644–656, 2019.
- [29] S. Jiang, K. Guo, J. Liao, and G. Zheng, “Solving fourier ptychographic imaging problems via neural network modeling and tensorflow,” Biomedical optics express, vol. 9, no. 7, pp. 3306–3319, 2018.
- [30] G. Zheng, C. Kolner, and C. Yang, “Microscopy refocusing and dark-field imaging by using a simple led array,” Optics letters, vol. 36, no. 20, pp. 3987–3989, 2011.
- [31] X. Ou, R. Horstmeyer, C. Yang, and G. Zheng, “Quantitative phase imaging via fourier ptychographic microscopy,” Optics letters, vol. 38, no. 22, pp. 4845–4848, 2013.
- [32] G. Zheng, R. Horstmeyer, and C. Yang, “Wide-field, high-resolution fourier ptychographic microscopy,” Nature photonics, vol. 7, no. 9, p. 739, 2013.
- [33] M. G. Gustafsson, “Surpassing the lateral resolution limit by a factor of two using structured illumination microscopy,” Journal of microscopy, vol. 198, no. 2, pp. 82–87, 2000.
- [34] P. C. Konda, L. Loetgering, K. C. Zhou, S. Xu, A. R. Harvey, and R. Horstmeyer, “Fourier ptychography: current applications and future promises,” Optics Express, vol. 28, no. 7, pp. 9603–9630, 2020.
- [35] R. Horstmeyer, J. Chung, X. Ou, G. Zheng, and C. Yang, “Diffraction tomography with fourier ptychography,” Optica, vol. 3, no. 8, pp. 827–835, 2016.
- [36] L. Tian and L. Waller, “3d intensity and phase imaging from light field measurements in an led array microscope,” optica, vol. 2, no. 2, pp. 104–111, 2015.
- [37] K. C. Zhou and R. Horstmeyer, “Diffraction tomography with a deep image prior,” Opt. Express, vol. 28, pp. 12872–12896, Apr 2020.
- [38] L. Tian and L. Waller, “Quantitative differential phase contrast imaging in an led array microscope,” Optics express, vol. 23, no. 9, pp. 11394–11403, 2015.
- [39] J. Chung, J. Kim, X. Ou, R. Horstmeyer, and C. Yang, “Wide field-of-view fluorescence image deconvolution with aberration-estimation from fourier ptychography,” Biomedical optics express, vol. 7, no. 2, pp. 352–368, 2016.
- [40] Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image quality assessment: from error visibility to structural similarity,” IEEE transactions on image processing, vol. 13, no. 4, pp. 600–612, 2004.
Appendix A Illuminated Image Frequency Analysis
The illumination patterns shown in this work offer a limited amount of insight into how the DNN affects image formation. To develop a better understanding of this process, we examined the illuminated images themselves, and examined how they change as we manipulated the difficulty of each inference task. To do this we define a metric called Average Spatial Frequency Power which summarizes the spatial frequency of an image. Figure 11 shows a diagram illustrating the process of computing this metric.
First, the images from a representative sample of the dataset (in this case, the test set) are illuminated with the optimized LED pattern, creating a set of illuminated images. Each of these images is then put through a 2D Fourier transform (2D FFT) to establish a spatial frequency domain representation, which we squared (i.e., element-wise square) to form a per-image power spectrum. We then computed the average power spectrum across all representative images per sample. Finally the average spatial frequency power is calculated by taking the first moment of the frequency power representation along the radius of the spatial frequency distribution, resulting in the 1D plot shown in the main text.

Appendix B Neural Network Configuration Parameters
Throughout all experiments the same neural network architecture was used. We detail the key parameters of our U-Net architecture in Table I. Within Table II we report the hyperparameters used during training. With the exception of the noise level these hyperparameters were the same across tasks.
Parameter | Value |
---|---|
Initial Filters | |
Filter Expansion Ratio | |
Convolutional Layers per Block | |
Convolutional Down-sampling Blocks | |
Convolutional Up-sampling Blocks | |
Convolutional Kernel Size | |
Activation Function | ReLU |
Final Activation Function | Sigmoid |
BatchNorm Frequency | After every convolution |
Convolutional Padding Amount | (1,1) zeros |
Hyper Parameters | HeLa Task | PAN Task |
---|---|---|
Optimizer | Adam | Adam |
Initial Learning Rate | ||
LR Reduction Factor | ||
LR Reduction Patience | 5 | 5 |
Noise Level (k) | 0.1 | 0.3 |
L1 Penalty | 0.0004 | 0.0004 |
Batch Size | 4 | 4 |
Appendix C Additional LED Patterns and Illumination Examples
To supplement the LED patterns shown in the main text Figure 12 shows every LED pattern across three random seeds on a per-color basis. The variance for each individual LED is also plotted to illustrate how random seeds effect converged LED patterns. In general we observe that the average variance across the patterns decreases with bit-depth.






To highlight the differences in the final illuminated image, as well as the impact on the inference results Figure 13 shows examples for both tasks across a sub-set of bit depths. The learned (DNN-optimized) patterns consistently out-perform the alternatives, yielding inference results which are closer to the discretized fluorescent label.



