onlynew
Differences between human and machine perception in medical diagnosis
Abstract
Deep neural networks (DNNs) show promise in image-based medical diagnosis, but cannot be fully trusted since their performance can be severely degraded by dataset shifts to which human perception remains invariant. If we can better understand the differences between human and machine perception, we can potentially characterize and mitigate this effect. We therefore propose a framework for comparing human and machine perception in medical diagnosis. The two are compared with respect to their sensitivity to the removal of clinically meaningful information, and to the regions of an image deemed most suspicious. Drawing inspiration from the natural image domain, we frame both comparisons in terms of perturbation robustness. The novelty of our framework is that separate analyses are performed for subgroups with clinically meaningful differences. We argue that this is necessary in order to avert Simpson’s paradox and draw correct conclusions. We demonstrate our framework with a case study in breast cancer screening, and reveal significant differences between radiologists and DNNs. We compare the two with respect to their robustness to Gaussian low-pass filtering, performing a subgroup analysis for microcalcifications and soft tissue lesions. For microcalcifications, DNNs use a separate set of high frequency components than radiologists, some of which lie outside the image regions considered most suspicious by radiologists. These features run the risk of being spurious, but if not, could represent potential new biomarkers. For soft tissue lesions, the divergence between radiologists and DNNs is even starker, with DNNs relying heavily on spurious high frequency components ignored by radiologists. Importantly, this deviation in soft tissue lesions was only observable through subgroup analysis, which highlights the importance of incorporating medical domain knowledge into our comparison framework.
Following their success in the natural image domain [1, 2, 3, 4, 5, 6, 7], deep neural networks (DNNs) have achieved human-level performance in various tasks of image-based medical diagnosis [8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20]. DNNs have a number of additional benefits: they can reach diagnoses quickly, do not suffer from fatigue, and can be deployed anywhere in the world. However, they currently possess an Achilles’ heel which severely limits their clinical applicability. They cannot be fully trusted, given their extreme vulnerability to dataset shifts to which human perception is robustly invariant. For example, a dermatologist-level skin cancer classifier, approved for use as a medical device in Europe, learned to spuriously associate surgical skin markings with malignant melanoma [21]. As a result, the classifier’s false positive rate increased by 40% during an external validation. Also, in three out of five cases, a pneumonia classifier performed substantially worse in cross-institutional settings than in the institution in which it was trained [22].
A promising path to establishing trust in DNNs for medical diagnosis is to understand whether and how their perception is different to that of humans. With this understanding in hand, we can recognize the limits of DNNs and better anticipate their failures. An effective empirical means of comparing human and machine perception is to study their robustness to input perturbation. By removing certain information from the input and analyzing the resulting change in prediction, we can infer the degree to which that information was utilized. This approach has been successfully applied in the natural image domain, and has exposed fundamental differences between human and machine perception [23, 24, 25, 26, 27, 28]. We extend this line of work, taking into account a critically important consideration for medical diagnosis.


We argue that subgroup analysis is necessary in order to draw precise and correct conclusions regarding perception in medical diagnosis. From a performance standpoint, subgroups are important in medical diagnosis because different types of errors can vary greatly in clinical significance [29]. When models are trained within the dominant paradigm of empirical risk minimization, all errors are treated equally, which can lead to a large disparity in performance across subgroups. Robust optimization addresses this issue by considering the performance of various subgroups, and optimizing the worst-case subgroup performance [30, 31]. In addition to performance, subgroups are also important for understanding perception. A failure to incorporate subgroups can lead to incorrect conclusions due to Simpson’s paradox [32], which states that subgroup-specific relationships can disappear or even reverse when the subgroups are aggregated. We therefore propose a framework which uses subgroup-specific perturbation robustness to compare human and machine perception in medical diagnosis. While others have separately studied perturbation robustness and subgroup performance, we combine the two ideas to draw precise conclusions regarding perception. We demonstrate our framework with a case study in breast cancer screening, and show that failure to account for subgroups would indeed result in incorrect conclusions. It is important to note that while we analyze perturbation robustness, our purpose here is not specifically to improve the robustness of machine-derived diagnosis. Rather, we aim to use perturbation robustness as a lens to understand human and machine perception.
In our framework, human and machine perception are compared along two axes, represented by two distinct questions. First, if we subject humans and machines to a perturbation which removes clinically meaningful information, do they share the same robustness characteristics? Second, to what degree do humans and machines agree on the most suspicious regions of an image?
The first step in this framework is to identify subgroups that are diagnosed in a significantly different manner, and to find an input perturbation that removes clinically relevant information from each of these subgroups. See Figure 1 for an illustration of this step applied to breast cancer screening. With the subgroups and perturbation identified, humans and machines can then be compared with respect to their perturbation robustness. Predictions are collected from humans and machines based on medical images to which the perturbation has been applied with varying severity. Probabilistic modeling is applied to these predictions, capturing the isolated effect of the perturbation, while factoring out confounding effects such as individual idiosyncrasies. The probabilistic model is used to compare the perturbation robustness of humans and machines in terms of two criteria that are important for diagnosis: predictive confidence and class separability. Predictive confidence measures the strength of predictions, and is independent of correctness. Class separability represents correctness, and is quantified as the distance between the distributions of predictions for positive and negative cases. See Figure 2 (a–e) for a visualization of this procedure for comparing human and machine predictions in the setting of breast cancer screening.
Humans and machines are then compared with respect to the regions of an image they find most suspicious. This is also framed in terms of perturbation robustness. First, humans annotate the most suspicious regions of interest (ROIs) for each image. The perturbation is then applied in separate steps to the interior of these ROIs, to the exterior regions only, and to the entire image. By analyzing the robustness of machines to these perturbations, we can infer whether humans and machines find the same regions informative. Figure 2 (g–j) illustrates this process for mammographic images.
Our framework enables a nuanced comparison of human and machine perception for medical diagnosis through the analysis of subgroup-specific perturbation robustness. In our case study, we examined the robustness of radiologists and DNNs to Gaussian low-pass filtering, separately analyzing two subgroups called microcalcifications and soft tissue lesions. We discovered significant differences between radiologists and DNNs in both subgroups. For microcalcifications, DNNs use a different set of high frequency components than radiologists, and some of these components lie outside image regions identified as most suspicious by radiologists. It is unclear from the current analysis whether these divergent frequency components represent spurious features or potential new biomarkers of disease. For soft tissue lesions, the difference in perception is even more significant, with DNNs using clearly spurious high frequency components ignored by radiologists. Importantly, we show that without subgroup analysis, we would have failed to observe this divergent behavior in soft tissue lesions, thus artificially inflating the similarity of radiologists and DNNs.
Results
Experimental setup.
We experimented with the NYU Breast Cancer Screening Dataset [33] developed by our research team and used in a number of prior studies [12, 13, 14, 34, 35], and we applied the same training, validation, and test set data split as has been reported previously. This dataset consists of 229,426 screening mammography exams from 141,473 patients. Each exam contains at least four images, with one or more images for each of the four standard views of screening mammography: left craniocaudal (L-CC), left mediolateral oblique (L-MLO), right craniocaudal (R-CC), and right mediolateral oblique (R-MLO). Each exam is paired with labels indicating whether there is a malignant or benign finding in each breast. See Extended Data Figure 7 for an example of a screening mammogram. We used a subset of the test set for our reader study, which is also the same subset used in the reader study of [12]. For our DNN experiments, we used two architectures in order to draw general conclusions: the deep multi-view classifier (DMV) [12], and the globally-aware multiple instance classifier (GMIC) [13, 14]. We primarily report results for GMIC, since it is the more recent and better-performing model. The corresponding results for DMV, which support the generality of our findings, are provided in the Extended Data section.
Perturbation reader study.
In order to compare the perception of radiologists and DNNs, we applied Gaussian low-pass filtering to mammograms, and analyzed the resulting effect on their predictions. We selected nine filter severities ranging from unperturbed to severe, where severity was represented as a wavelength in units of millimeters on the physical breast. Details regarding the calculation of the filter severity are provided in the Methods section. Figure 1 demonstrates how low-pass filtering affects the appearance of malignant breast lesions.
We conducted a reader study in order to collect predictions for low-pass filtered images from radiologists. This reader study was designed to be identical to that of [12], except that the mammograms were randomly low-pass filtered in our case. We assigned the same set of 720 exams to ten radiologists with varying levels of experience. The images were presented to the radiologists in a conventional format, and an example is shown in Extended Data Figure 7. Each radiologist read each exam once, and for each exam, we uniformly sampled one severity level out of our set of nine, and applied it to all images in the exam. The radiologists made binary predictions indicating the presence of a malignant lesion in each breast. We describe the details of the reader study in the Methods section.
We then trained five DNNs from random weight initializations, and made predictions on the same set of 720 exams. We repeated this nine times, where the set of exams was low-pass filtered with each of the nine filter severities. We note that for each DNN, we made a prediction for every pair of exam and filter severity. In contrast, for each radiologist, we only had predictions for a subset of the possible combinations of exam and filter severity. This means that if we arrange the predictions in a matrix where each row represents a filter severity and each column an exam, the matrix of predictions is sparse for each radiologist, and dense for each DNN. This fact is visualized in Figure 2 (b–c). The sparsity of the radiologist predictions is by design; we were careful to ensure that each radiologist only read each exam once, since if they were to have seen the same exam perturbed with multiple filter severities, their predictions would have been unlikely to be independent. However, the sparsity prevents us from comparing radiologists and DNNs using evaluation metrics that use predictions for the complete set of exams. We therefore utilized probabilistic modeling to use the available radiologist predictions to infer values for the missing predictions.
A probabilistic model of predictions.

We applied probabilistic modeling to achieve two purposes. The first is to study the effect of low-pass filtering on specific subgroups of lesions in isolation, after factoring out various confounding effects such as the idiosyncracies of individual radiologists and DNNs. The second is to infer the radiologists’ predictions for each pair of exam and filter severity, since some pairs were missing by design. We modeled the radiologists’ and DNNs’ predictions as i.i.d. Bernoulli random variables. Let us denote radiologist (DNN) ’s prediction on case filtered with severity as . We parameterized our model as
(1) |
where is the logistic function. There are four latent variables with the following interpretation: represents the bias of exams in subgroup , is the bias of exam , is the effect that low-pass filtering with severity has on exams in subgroup , and is the idiosyncrasy of radiologist (DNN) on exams in subgroup . See Figure 3 for a graphical representation of our model. We considered several parameterizations of varying complexity, and selected the one with the maximum marginal likelihood. See the Methods section for details regarding our probabilistic model.
Comparing humans and machines with respect to their sensitivity to the removal of clinically meaningful information.

Using the probabilistic model, we compared how low-pass filtering affects the predictions of radiologists and DNNs, separately analyzing microcalcifications and soft tissue lesions. We performed each comparison with respect to two metrics: predictive confidence and class separability. Since the latent variable represents the effect of low-pass filtering on each prediction, we examined its posterior distribution in order to measure the effect on predictive confidence. We sampled values of from the posterior predictive distribution in order to quantify how low-pass filtering affects class separability. We computed the Kolmogorov-Smirnov (KS) statistic between the sampled predictions for the positive and negative class. This represents the distance between the two distributions of predictions, or how separated the two classes are. Sampling from the posterior predictive distribution was necessary for radiologists, since we did not have a complete set of predictions from them. Although such sampling was not strictly necessary for DNNs given the full set of available predictions, we performed the same posterior sampling for DNNs in order to ensure a fair comparison.
Figure 4a presents the results for microcalcifications. We only consider DNNs that are trained with unperturbed data in this section. The results for training DNNs on low-pass filtered data are discussed in the next section. The left subfigure represents predictive confidence, as measured by the posterior expectation of . Since low-pass filtering removes the visual cues of malignant lesions, we hypothesized that it should decrease predictive confidence. In other words, we expected to see . Above the left subfigure, we report in order to quantify how much the posterior distributions align with this hypothesis. Small values indicate a significant negative effect on predictive confidence. We note that these values are not intended to be interpreted as the -values of a statistical test. Instead, they quantify the degree to which each is negative. We observe that for microcalcifications, low-pass filtering decreases the predictive confidence of both radiologists and DNNs. There is, however, an interesting difference in that for the range of most severe filters, the effect is constant for radiologists, while DNNs continue to become less confident.
The right subfigure of Figure 4a depicts the effect of low-pass filtering on class separability. This is quantified by the KS statistic between the predictions for the positive and negative class, where the positive class is restricted to malignant microcalcifications. Similar to our hypothesis that low-pass filtering decreases predictive confidence, we hypothesized that it should also reduce class separability. That is, we expected the KS statistics for severity to be smaller than those for . This is because removing the visual cues for malignant lesions should make it more difficult to distinguish between malignant and nonmalignant cases. We tested this hypothesis using the one-tailed KS test between the KS statistics for and . The -values for this test are reported above the right subfigures, where small values mean that filtering significantly decreases class separability. We found that low-pass filtering decreases class separability for both radiologists and DNNs, but in substantially different ways. The radiologists’ class separability steadily declines for the range of less severe filters, while it is constant for DNNs. Meanwhile, similar to what we observed for predictive confidence, the radiologists’ class separability is constant for the range of most severe filters, while it continues to decline for DNNs. In summary, for microcalcifications, low-pass filtering decreases the predictive confidence and class separability for both radiologists and DNNs, but differences suggest that they use a different set of high frequency components.
Next, we compared how low-pass filtering affects the predictions of radiologists and DNNs on soft tissue lesions (Figure 4b). The results show that low-pass filtering degrades the predictive confidence and class separability of DNNs, while having almost no effect on radiologists. The invariance of radiologists’ impressions to filtering implies that the high frequency components used by DNNs in this context are spurious, since there exist alternative features used by radiologists that are robust to low-pass filtering. The DNNs’ reliance on such spurious features is a clear vulnerability, and must be addressed before they can be trusted to make clinical diagnoses. In the Discussion section, we describe a potential explanation for this phenomenon, as well as a plausible scenario in which this behavior would cause DNNs’ performance to degrade.
Training DNNs with low-pass filtered data.
We observed that low-pass filtering decreases the predictive confidence and class separability of DNNs for all lesion subgroups. However, since the DNNs only encountered low-pass filtering during testing, it is possible that this effect is solely due to the dataset shift between training and testing. We therefore repeated the previous experiments for DNNs, where the same filtering was applied during both training and testing. We then examined whether the effects of low-pass filtering on the DNNs’ perception could be attributed to information loss rather than solely to dataset shift.
For microcalcifications (Figure 4a), training on filtered data slightly reduced the effect of low-pass filtering on predictive confidence and class separability, but the effect was still present, particularly for the most severe filters. This implies that the effect of filtering on microcalcifications can be attributed to information loss, and not solely to dataset shift. In other words, high frequency components in microcalcifications contain information that is important to the perception of DNNs.
Meanwhile, for soft tissue lesions (Figure 4b), training on low-pass-filtered data significantly reduces the effect on predictive confidence and class separability, even for severe filters. This suggests that the effect of low-pass filtering on soft tissue lesions can primarily be attributed to dataset shift rather than information loss. In fact, DNNs trained with low-pass-filtered data maintain a similar level of class separability compared to networks trained on the original data. This confirms what we observed for radiologists, which is that high frequency components in soft tissue lesions are largely dispensable, and that more robust features exist.
Annotation reader study.
Our results thus far show that radiologists and DNNs use a different set of high frequency components in both microcalcifications and soft tissue lesions. Since this analysis has purely been in the frequency domain, we extend our comparison to the spatial domain by examining the degree to which radiologists and DNNs agree on the most suspicious regions of an image. Towards this end, we conducted a reader study in which seven radiologists annotated up to three regions of interest (ROIs) containing the most suspicious features of each image. 120 exams were used in this study, which is a subset of the 720 exams in the perturbation reader study. See the Methods section for details regarding this reader study. We then applied low-pass filtering to the interior and exterior of the ROIs, as well as to the entire image. Examples of the annotation and the low-pass filtering schemes are shown in Figure 2 (g–j). We made predictions using DNNs trained with the original data in order to understand the relationship between the high frequency components utilized by DNNs, and the regions of mammograms that are most suspicious to the radiologists.
Comparing humans and machines with respect to the regions of an image deemed most suspicious.

We began by comparing the effect of the three low-pass filtering schemes on the DNNs’ predictions for microcalcifications (Figure 5a). We observed that filtering the ROI interior has a similar effect to filtering the entire image for mild filter severities. This suggests that for the frequencies in question, DNNs primarily rely on the same regions that radiologists consider suspicious. Meanwhile, class separability for the two ROI-based filtering schemes diverge significantly for high severities: filtering the ROI interior ceases to further decrease class separability at some threshold filter severity, whereas exterior filtering continues to degrade class separability beyond this threshold. The implication is that a range of high frequency components utilized by DNNs exist in the exterior of the ROIs deemed most important by human radioloigsts.
For soft tissue lesions (Figure 5b), filtering the ROI interior decreases class separability, but to a lesser degree compared to filtering the entire image. This means that DNNs do utilize high frequency components in regions that radiologists find suspicious, but only to a limited degree. Meanwhile, filtering the ROI exterior has a similar effect on class separability as filtering the entire image. These observations suggest that the high frequency components that DNNs use for soft tissue lesions may be scattered across the image, rather than being localized in the areas that radiologists consider suspicious.
Discussion
Arguably, the most significant difference we observed between human and machine perception in our case study was that for soft tissue lesions, DNNs use spurious high frequency components that are ignored by radiologists. Interestingly, we showed that by training DNNs on low-pass-filtered data, the difference narrows significantly in terms of predictive confidence, and nearly disappears for class separability. Here, we present an explanation for this phenomenon, as well as a realistic scenario in which it is harmful. We hypothesize that DNNs rely heavily on the borders of soft tissue lesions, while radiologists make more use of the radiodense or bright areas within the borders. Low-pass filtering removes the borders without affecting the interior brightness, which can explain why DNNs are severely affected while radiologists are invariant. If our hypothesis is true, this is potentially harmful since the borders of soft tissue lesions are often obscured in patients with high breast fibroglandular tissue density [36]. In other words, DNNs that rely on borders can be compromised by a dataset shift towards patients with breasts that are more dense than those in the training set. As for the effect of training on low-pass-filtered data, our interpretation is that DNNs are biased towards learning high frequency components, and this causes them to associate spurious high frequency patterns with malignant soft tissue lesions. When these spurious patterns are absent from the data, DNNs are able to learn alternative features with little effect on their class separability. Based on our hypothesis regarding the borders of soft tissue lesions and their bright interiors, we suspect that by training on low-pass-filtered data, DNNs are able to become more like radiologists by using the radiodense interiors of lesions rather than relying excessively on the borders. Training on low-pass-filtered data can therefore be interpreted as inducing a prior to favor low-frequency features.
Our framework draws inspiration from perturbation robustness studies in the adjacent domain of natural images, where there is also an ongoing crisis regarding the trustworthiness of machine perception. One key innovation in our work is to incorporate subgroup analysis to draw precise conclusions regarding perception. This is critically important for establishing trust in DNNs in the medical domain, where incorrect conclusions can directly harm patients’ physical and mental well-being. Not accounting for subgroups can be very dangerous, as it can lead to drawing erroneous conclusions due to Simpson’s paradox. In our case study, our conclusions would change significantly if we treated microcalcifications and soft tissue lesions as a single subgroup. As shown in Figure 6, we would incorrectly conclude that radiologists and DNNs have comparable perturbation robustness in terms of both predictive confidence and class separability, thus artificially inflating the similarity between human and machine perception.

The identification of subgroups with clinically meaningful differences is a crucially important component of our framework, as it strongly influences the conclusions. It requires domain knowledge, and is not as simple as enumerating all possible subgroups. The reason is that, due to the rarity of some subgroups, there is a balance to strike between the number of specified subgroups and the amount of available data. In our case study, we combined architectural distortions, asymmetries, and masses into the soft tissue lesion subgroup because there are only 30 cases of malignant soft tissue lesions in our reader study dataset. By doing this, we addressed data scarcity while accounting for the fact that soft tissue lesions as a whole are much less localized than microcalcifications, and thus require a significantly different diagnostic approach.
The choice of input perturbation is another key consideration in our framework. For the purpose of understanding perception, it is more important for the removed information to be clearly characterizable than for the perturbation to be clinically realistic. For example, analyzing robustness in cross-institutional settings is clinically realistic, but it does not allow us to draw precise conclusions regarding perception, since it is unclear what information may have changed between institutions. Having said that, if the perturbation removes clinically relevant information and is additionally clinically realistic, this is beneficial because it allows us to reason about robustness in a plausible scenario. Our choice of Gaussian low-pass filtering is clinically relevant, as another type of low-pass filtering called motion blurring does occur in practice. Mammograms can be blurred by motion caused by patients or imaging devices [37], and [38] demonstrated that it can degrade the ability of radiologists to detect malignant lesions. While there exist differences between Gaussian low-pass filtering and motion blurring, we expect that robustness to the former will translate to the latter. This is because DNNs have been shown to exhibit similar robustness characteristics to various types of blurring [27]. We noticed that when comparing the class separability between radiologists and DNNs trained with low-pass-filtered data (Figure 4), DNNs are more robust to low-pass filtering for microcalcifications, while both are largely invariant for soft tissue lesions. This may be a significant advantage of DNNs in clinical practice.
In summary, we proposed a framework for establishing trust in machine perception for medical diagnosis, which we expect to be applicable to a variety of clinical tasks and imaging technologies. We demonstrated the efficacy of this framework with a case study in breast cancer screening. The framework uses subgroup-specific perturbation robustness to compare human and machine perception along two axes: their sensitivity to the removal of clinically meaningful information, and the regions of an image deemed most suspicious. In our case study, we revealed significant differences between radiologists and DNNs along both axes of comparison. We found that for microcalcifications, DNNs use a different set of high frequency components than radiologists, and some of these components lie outside the image regions that radiologists find most suspicious. Meanwhile, for soft tissue lesions, DNNs utilize spurious high frequency components that are ignored by radiologists. We also showed that we would have missed this stark divergence between radiologists and DNNs in soft tissue lesions if we failed to perform subgroup analysis. This is evidence that future studies comparing human and machine perception in other medical domains should separately analyze subgroups with clinically meaningful differences. By utilizing appropriate subgroup analysis driven by clinical domain knowledge, we can draw precise conclusions regarding machine perception, and potentially accelerate the widespread adoption of DNNs in clinical practice.
References
- [1] Alex Krizhevsky, Ilya Sutskever and Geoffrey E. Hinton “ImageNet classification with deep convolutional neural networks” In NIPS, 2012, pp. 1106–1114
- [2] Karen Simonyan and Andrew Zisserman “Very deep convolutional networks for large-scale image recognition” In ICLR, 2015
- [3] Shaoqing Ren, Kaiming He, Ross B. Girshick and Jian Sun “Faster R-CNN: towards real-time object detection with region proposal networks” In NIPS, 2015, pp. 91–99
- [4] Joseph Redmon, Santosh Kumar Divvala, Ross B. Girshick and Ali Farhadi “You only look once: unified, real-time object detection” In CVPR, 2016, pp. 779–788
- [5] Kaiming He, Xiangyu Zhang, Shaoqing Ren and Jian Sun “Deep residual learning for image recognition” In CVPR, 2016, pp. 770–778
- [6] Gao Huang, Zhuang Liu, Laurens Maaten and Kilian Q. Weinberger “Densely connected convolutional networks” In CVPR, 2017, pp. 2261–2269
- [7] Kaiming He, Georgia Gkioxari, Piotr Dollár and Ross B. Girshick “Mask R-CNN” In ICCV, 2017, pp. 2980–2988
- [8] Andre Esteva et al. “Dermatologist-level classification of skin cancer with deep neural networks” In Nature 542.7639, 2017, pp. 115–118
- [9] Robert V. Lindsey et al. “Deep neural network improves fracture detection by clinicians” In Proc. Natl. Acad. Sci. USA 115.45, 2018, pp. 11591–11596
- [10] Nicolas Coudray et al. “Classification and mutation prediction from non–small cell lung cancer histopathology images using deep learning” In Nat. Med. 24.10 Nature Publishing Group, 2018, pp. 1559–1567
- [11] Xiaoxuan Liu et al. “A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: a systematic review and meta-analysis” In The Lancet Digital Health 1.6 Elsevier, 2019, pp. e271–e297
- [12] Nan Wu et al. “Deep neural networks improve radiologists’ performance in breast cancer screening” In IEEE Trans. Med. Imaging. 39.4 IEEE, 2019, pp. 1184–1194
- [13] Yiqiu Shen et al. “Globally-aware multiple instance classifier for breast cancer screening” In International Workshop on Machine Learning in Medical Imaging, 2019, pp. 18–26 Springer
- [14] Yiqiu Shen et al. “An interpretable classifier for high-resolution breast cancer screening images utilizing weakly supervised localization” In arXiv 2002.07613, 2020
- [15] Alejandro Rodriguez-Ruiz et al. “Stand-alone artificial intelligence for breast cancer detection in mammography: comparison with 101 radiologists” In JNCI 111.9 Narnia, 2019, pp. 916–922
- [16] Diego Ardila et al. “End-to-end lung cancer screening with three-dimensional deep learning on low-dose chest computed tomography” In Nat. Med. 25.6 Nature Publishing Group, 2019, pp. 954–961
- [17] Scott Mayer McKinney et al. “International evaluation of an AI system for breast cancer screening” In Nature 577.7788 Nature Publishing Group, 2020, pp. 89–94
- [18] Hyo-Eun Kim et al. “Changes in cancer detection and false-positive recall in mammography using artificial intelligence: a retrospective, multireader study” In The Lancet Digital Health 2.3 Elsevier, 2020, pp. e138–e148
- [19] Thomas Schaffter et al. “Evaluation of combined artificial intelligence and radiologist assessment to interpret screening mammograms” In JAMA Netw. Open 3.3 American Medical Association, 2020, pp. e200265–e200265
- [20] Yuan Liu et al. “A deep learning system for differential diagnosis of skin diseases” In Nat. Med. Nature Publishing Group, 2020, pp. 1–9
- [21] Julia K Winkler et al. “Association between surgical skin markings in dermoscopic images and diagnostic performance of a deep learning convolutional neural network for melanoma recognition” In JAMA Dermatol. 155.10 American Medical Association, 2019, pp. 1135–1141
- [22] John R Zech et al. “Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: a cross-sectional study” In PLoS Med. 15.11 Public Library of Science San Francisco, CA USA, 2018, pp. e1002683
- [23] Christian Szegedy et al. “Intriguing properties of neural networks” In ICLR, 2014
- [24] Jason Jo and Yoshua Bengio “Measuring the tendency of CNNs to learn surface statistical regularities” In arXiv 1711.11561, 2017
- [25] Samuel Dodge and Lina Karam “A study and comparison of human and deep learning recognition performance under visual distortions” In ICCCN, 2017, pp. 1–7 IEEE
- [26] Robert Geirhos et al. “Generalisation in humans and deep neural networks” In NeurIPS, 2018, pp. 7549–7561
- [27] Dan Hendrycks and Thomas G. Dietterich “Benchmarking neural network robustness to common corruptions and perturbations” In ICLR, 2019
- [28] Dong Yin, Raphael Gontijo Lopes, Jon Shlens, Ekin Dogus Cubuk and Justin Gilmer “A fourier perspective on model robustness in computer vision” In NeurIPS, 2019, pp. 13255–13265
- [29] Luke Oakden-Rayner, Jared Dunnmon, Gustavo Carneiro and Christopher Ré “Hidden stratification causes clinically meaningful failures in machine learning for medical imaging” In CHIL ACM, 2020, pp. 151–159
- [30] Shiori Sagawa, Pang Wei Koh, Tatsunori B Hashimoto and Percy Liang “Distributionally robust neural networks for group shifts: On the importance of regularization for worst-case generalization” In ICLR, 2020
- [31] Karan Goel, Albert Gu, Yixuan Li and Christopher Ré “Model patching: closing the subgroup performance gap with data augmentation” In arXiv 2008.06775, 2020
- [32] Judea Pearl “Comment: understanding Simpson’s paradox” In Am. Stat. 68.1 Taylor & Francis, 2014, pp. 8–13
- [33] Nan Wu et al. “The NYU breast cancer screening dataset v1”, 2019
- [34] Thibault Févry et al. “Improving localization-based approaches for breast cancer screening exam classification” In arXiv 1908.00615, 2019
- [35] Nan Wu et al. “Improving the ability of deep neural networks to use information from multiple views in breast cancer screening” In Medical Imaging with Deep Learning, 2020, pp. 827–842 PMLR
- [36] Min Sun Bae et al. “Breast cancer detected with screening US: reasons for nondetection at mammography” In Radiology 270.2 Radiological Society of North America, 2014, pp. 369–377
- [37] Jae Jeong Choi et al. “Mammographic artifacts on full-field digital mammography” In JDI 27.2 Springer, 2014, pp. 231–236
- [38] Ahmed K Abdullah et al. “The impact of simulated motion blur on lesion detection performance in full-field digital mammography” In Brit. J. Radiol. 90.1075 The British Institute of Radiology., 2017, pp. 20160871
- [39] Krzysztof J Geras et al. “High-resolution breast cancer screening with multi-view deep convolutional neural networks” In arXiv 1703.07047, 2017
- [40] Meelis Kull et al. “Beyond temperature scaling: obtaining well-calibrated multi-class probabilities with Dirichlet calibration” In NeurIPS, 2019, pp. 12295–12305
- [41] Alp Kucukelbir, Rajesh Ranganath, Andrew Gelman and David Blei “Automatic variational inference in Stan” In NIPS, 2015, pp. 568–576
- [42] Adam Paszke et al. “PyTorch: an imperative style, high-performance deep learning library” In NeurIPS, 2019, pp. 8024–8035
- [43] Bob Carpenter et al. “Stan: a probabilistic programming language” In J. Stat. Softw. 76.1 Columbia Univ., New York, NY (United States); Harvard Univ., Cambridge, MA …, 2017
Methods
DNN training methodology.
We conducted our DNN experiments using two architectures: the Deep Multi-View Classifer [12], and the Globally-Aware Multiple Instance Classifier [13, 14]. With both architectures, we trained an ensemble of five models. A subset of each model’s weights was initialized using weights pretrained on the BI-RADS label optimization task [39], while the remaining weights were randomly initialized. For each architecture, we adopted the same training methodology used by the corresponding authors.
Probability calibration.
We applied Dirichlet calibration [40] to the predictions of DNNs used in our probabilistic modeling. This amounts to using logistic regression to fit the log predictions to the targets. We trained the logistic regression model using the validation set, and applied it to the log predictions on the test set to obtain the predictions used in our analysis. We used L2 regularization when fitting the logistic regression model, and tuned the regularization hyperparameter via an internal 10-fold cross-validation where we further split the validation set into “training” and “validation” sets. In the cross-validation, we minimized the classwise expected calibration error [40].
Gaussian low-pass filtering.
Low-pass filtering is a method for removing information from images that allows us to interpolate between the original image and, in the most extreme case, an image where every pixel has the value of the mean pixel value of the original image. We experimented with nine filter severities selected to span a large range of the frequency spectrum. We implemented the Gaussian low-pass filter by first applying the shifted two-dimensional discrete Fourier transform to transform images to the frequency domain. The images were multiplied element-wise by a mask with values in . The values of this mask are given by the Gaussian function
(2) |
where and are horizontal and vertical coordinates, is the Euclidian distance from the origin, and is the cutoff frequency. represents the severity of the filter, where frequencies are reduced to of their original values when . Since the mammograms in our dataset vary in terms of spatial resolution as well as the physical length represented by each pixel, we expressed the filter severity in terms of a normalized unit of cycles per millimeter on the breast. Let where and are the height and width of the image, and let denote the physical length in millimeters represented by each pixel. Then we can convert cycles per millimeter to cycles per frame length of the image using
(3) |
Perturbation reader study.
In order to compare humans and machines with respect to their sensitivity to the removal of clinically meaningful information, we conducted a reader study in which ten radiologists read 720 exams selected from the test set. While all radiologists read the same set of exams, each exam was low-pass filtered with a different severity for each radiologist. Except for the low-pass filtering, the setup of this reader study is identical to that of [12]. Each exam consists of at least four images, with one or more images for each of the four views of mammography: L-CC, R-CC, L-MLO, and R-MLO. All images in the exam were concatenated into a single image such that the right breast faces left and is presented on the left, and the left breast faces right and is displayed on the right. Additionally, the craniocaudal (CC) views are on the top row, while the mediolateral oblique (MLO) views are on the bottom row. An example of this is shown in Extended Data Figure 7. Among the 1,440 breasts, 62 are malignant, 356 are benign, and the remaining 1,022 are nonbiopsied. Among the malignant breasts, there are 26 microcalcifications, 21 masses, 12 asymmetries, and 4 architectural distortions, while in the benign breasts, the corresponding counts are: 102, 87, 36, and 6. For each exam, radiologists make a binary prediction for each breast, indicating their diagnosis of malignancy.
Probabilistic modeling.
We modeled the radiologists’ and DNNs’ binary malignancy predictions with the Bernoulli distribution
(4) |
where indexes the breast, the reader, and the low-pass filter severity. Each distribution’s parameter is a function of four latent variables
where indexes the subgroup of the lesion. We included the following subgroups: unambiguous microcalcifications, unambiguous soft tissue lesions, ambiguous microcalcifications and soft tissue lesions, mammographically occult, and nonbiopsied. We considered these five subgroups in order to make use of all of our data, but only used the first two in our analysis. We assigned the generic weakly informative prior to each latent variable. We chose the Bernoulli distribution because it has a single parameter, and thus makes the latent variables interpretable. Additionally, radiologists are accustomed to making discrete predictions in clinical practice. The posterior distribution of the latent variables is given by
The exact computation of the posterior is intractable, since the marginal likelihood involves a four-dimensional integral. We therefore applied automatic differentiation variational inference (ADVI) [41] in order to approximate the posterior. ADVI, and variational inference in general, optimizes over a class of tractable distributions in order to find the closest match to the posterior. For our choice of tractable distributions, we used the mean-field approximation, meaning that we optimized over multivariate Gaussians with diagonal covariance matrices.
In practice, while radiologists made binary predictions, DNNs made continuous predictions in that we then calibrated. Despite the DNN predictions not being binary, we used equivalent procedures to specify the probabilistic model for radiologists and DNNs. To see how, let denote a DNN’s unobserved binary prediction for case , and let denote its observed real-valued prediction for the same case. In order to obtain , we could treat it as a random variable and obtain values for it through sampling. We instead used directly, specifying the log joint density as
Annotation reader study.
In order to compare humans and machines with respect to the regions of an image deemed most suspicious, we conducted a reader study in which seven radiologists read the same set of 120 unperturbed exams. The exams in this study were a subset of the 720 exams from the perturbation reader study, and also included all malignant exams from the test set. This study had two stages. In the first stage, the radiologists were presented with all views of the mammogram, and they made a malignancy diagnosis for each breast. This stage was identical to the reader study in [12]. In the second stage, for breasts that were diagnosed as malignant, the radiologists annotated up to three ROIs around the regions they found most suspicious. The radiologists annotated each view individually, and the limit of three ROIs applied separately to each view. For exams that contained multiple images per view, the radiologists annotated the image where the malignancy was most visible. The radiologists annotated the images using Paintbrush on MacOS, or Microsoft Paint on Windows. In order to constrain the maximum area that is annotated for each image, we included a pixel blue ROI template in the bottom corner of each image to serve as a reference. The radiologists then drew up to three red ROIs such that each box approximately matched the dimensions of the reference blue ROI template.
Data availability
The radiologist and DNN predictions used in our analysis are available at https://github.com/nyukat/perception_comparison under the GNU AGPLv3 license.
Code availability
The code used in this research is available at https://github.com/nyukat/perception_comparison under the GNU AGPLv3 license. This, combined with the predictions which we also open-sourced, makes our results fully reproducible. We used several open-source libraries to conduct our experiments. The DNN experiments were performed using PyTorch [42], and the probabilistic modeling was done with PyStan (https://github.com/stan-dev/pystan), the Python interface to Stan [43]. The code for the DNNs used in our experiments is also open-source, where GMIC is available at https://github.com/nyukat/GMIC, and DMV at https://github.com/nyukat/breast_cancer_classifier.
Acknowledgements
The authors would like to thank Mario Videna, Abdul Khaja and Michael Costantino for supporting our computing environment and Eric K. Oermann for helpful comments on the draft of this paper. We also gratefully acknowledge the support of NVIDIA Corporation, which donated some of the GPUs used in this research. This work was supported in part by grants from the National Institutes of Health (P41EB017183 and R21CA225175) and the National Science Foundation (HDR-1922658).
Author contributions
SJ conceived the initial idea and designed the first set of experiments. TM, SJ, WO, KC and KJG designed the final version of the experiments. TM and WO conducted the experiments with neural networks. TM and WO preprocessed the screening mammograms. TM, SJ, LM and LH conducted the reader study. CeC, NS, RE, DQ, ChC, LD, EK, AK, JL, JP, KP, BR and HT collected the data. TM, SJ and WO conducted literature search. TM, KC and KJG designed the probabilistic model. TM performed the probabilistic inference. LM, LH and BR analyzed the results from a clinical perspective. DS contributed to framing the paper. SJ, KC and KJG supervised the execution of all elements of the project. DS, LM, LH, KC and KJG acquired funding for this research. All authors provided critical feedback and helped shape the manuscript.
Competing interests
The authors declare no competing interests.
Extended data


