Deep Evidential Learning for Radiotherapy Dose Prediction
Abstract
Background: As we navigate towards integrating deep learning methods in the real clinic, a safety concern lies in whether and how the model can express its own uncertainty when making predictions.
In this work, we present a novel application of an uncertainty-quantification framework called Deep Evidential Learning in the domain of radiotherapy dose prediction.
Method:
Using medical images of the Open Knowledge-Based Planning Challenge dataset, we found that this model can be effectively harnessed to yield uncertainty estimates
that inherited correlations with prediction errors upon completion of network training. This was achieved only after reformulating the original loss function for a stable implementation.
Results:
We found that (i)epistemic uncertainty was highly correlated with prediction errors, with various association indices comparable or stronger than those for Monte-Carlo Dropout and Deep Ensemble methods, (ii)the median error varied with uncertainty threshold much more linearly for epistemic uncertainty in Deep Evidential Learning relative to these other two conventional frameworks, indicative of a more uniformly calibrated sensitivity to model errors,
(iii)relative to epistemic uncertainty, aleatoric uncertainty demonstrated a more significant shift in its distribution in response to Gaussian noise added to CT intensity, compatible with its interpretation as reflecting data noise.
Conclusion:
Collectively, our results suggest that Deep Evidential Learning is a promising approach that can endow deep-learning models in radiotherapy dose prediction with statistical robustness. We have also demonstrated how this framework leads to uncertainty heatmaps that correlate strongly with model errors, and how it can be used to equip the predicted Dose-Volume-Histograms with confidence intervals.
keywords:
uncertainty quantification , deep evidential learning , radiotherapy dose prediction[inst1]Gryphon Center for Artificial Intelligence and Theoretical Sciences, country=Singapore
[inst2]Georgia Institute of Technology, Atlanta, GA, country=USA
[inst3]University of Pennsylvania, Perelman School of Medicine, Department of Radiation Oncology, Philadelphia, country=USA
1 Introduction
Radiotherapy treatment planning is a highly intricate process requiring collaborative interactions among medical physicists, dosimetrists and radiation oncologists in formulating the optimal treatment plan (see e.g. [1, 2]) for each individual patient. It is a procedure that is characterized by a potentially large degree of user-variability arising from different inter-institutional guidelines, the fundamental origin of which lies in the subjective nature of tradeoffs like those between tumor control and sparing of normal tissues, etc. [1, 3, 4] Automation by machine learning methods can help to homogenize these variations among the medical professionals, while attaining improvements in the consistency of overall plan quality (see e.g. [5]). The adoption of artificial intelligence (A.I.) related techniques led to what is known as knowledge-based planning (KBP) [6], which leverages knowledge inherent in past clinical treatment plans to generate new ones with minimal intervention from human experts.
A complete KBP method can normally be regarded as a two-stage pipeline [7] : (i)prediction of the dose distribution that should be delivered to patient (ii)conversion of the prediction into a deliverable treatment plan via optimization. Recent dose prediction models cover quite a range of sites and modalities [5, 8], including prostate intensity-modulated-radiation-therapy (IMRT) [9, 10], prostate volumetric modulated arc therapy (VMAT) [11], lung IMRT [12] and head and neck VMAT [13]. These models are designed to predict volumetric dose distributions from which dose-volume histogram (DVH) and other dose constraint-related statistics can be deduced. Within the literature, it has been noted [6] that the vast majority of published works was performed using large private datasets which made comparison of model quality challenging. Towards alleviating this issue, the Open Knowledge-Based Planning (OpenKBP) Grand Challenge was organized [7, 14] to enable an international effort for the comparison of dose prediction models on a single open dataset involving 340 patients of head and neck cancer treated by IMRT. To date, there has been many excellent dose prediction models trained on the OpenKBP dataset as reviewed in [5]. Notable examples include DeepDoseNet of [15] which combined features of ResNet and DenseNet, the MtAA-NET of [16] which is based on a generative adversarial network, and TrDosePred of [17] and Swin UNETR++ of [18] which are Transformer-based frameworks.iiiiiiFor a more complete list, see Table 5 of [5].. Like the above-mentioned projects, our work here leverages upon the OpenKBP dataset and attempts to furnish dose prediction. However, unlike the majority of these papers, the major focus of our work here lies in developing a dose prediction model that also encapsulates an uncertainty quantification framework.
As we navigate towards integrating deep learning methods in the real clinic, a safety-related concern lies in whether and how the model can express its own uncertainty when making predictions based on real-life data distributions beyond those that they were trained on. Uncertainty quantification has been a major theme in the broader context of machine learning [19], and has recently gathered increasing attention in the domain of medical image analysis [20]. However, to our knowledge, there is a scarcity of works devoted specifically to uncertainty quantification in dose prediction models. As already noted in papers dedicated to medical image analysis [20, 21, 22, 23, 24], uncertainty estimates equip deep learning models with statistical robustness and a measure of reliability that can be useful in discovering regions of the model’s ignorance, thereby alerting the human expert to potentially erroneous predictions. In the domain of dose prediction, uncertainty estimates can be used to characterize the reliability of a model. In the KBP pipeline, dose prediction models are themselves inputs to some dose mimicking model [7] which generates deliverable treatments via optimization of a set of objective functions [25, 26]. Apart from using reliability as a selection criterion for various dose prediction models, it was pointed out in [7, 27] that probabilistic dose distributions can serve as robust inputs to the objective functions. At the time of writing, we are only aware of [8] and [28] being the only publications that examined the role of uncertainty quantification frameworks for dose prediction. The work in [8] studied how Monte-Carlo Dropout (MC Dropout) and a Deep Ensemble-based bagging method can furnish useful uncertainty estimates for a U-Net-based dose prediction model, whereas [28] proposed a Gaussian mixture model in which the standard deviation of each Gaussian mixture is part of the network’s outputs and can be used to reflect the tradeoff implicit in data from two different treatment protocols.
In this work, we examined the applicability of a relatively new uncertainty quantification framework known as Deep Evidential Learning [29, 30] for dose prediction. Its fundamental principle lies in asserting a higher-order Bayesian prior over the probabilistic neural network output, embedding it within the loss function and subsequently generating uncertainty estimates together with the completion of model training. There are two major variants of it as proposed in the seminal works of Sensoy et al. in [30] [‘Deep Evidential Classification’] and Amini et al. in [29] [‘Deep Evidential Regression’]. We apply the formalism of [29] to construct a dose prediction model with 3D U-Net [31] as the backbone model architecture. In the process, we found that this approach required additional crucial refinements for an effective and stable implementation. These refinements are related to the form of the final model layer which connects inputs and weights to the parameters of the higher-order Bayesian prior distribution. These parameters are formulated as the model’s outputs, and are thus naturally obtained when training is completed. They are then used to furnish estimates of model uncertainty. This is in contrast to two major categories of uncertainty quantification frameworks in the current literatureiiiiiiiiiA general survey can be found in [19] whereas comprehensive recent reviews in the field of medical imaging can be found in [20] and [32]. : (i) MC Dropout (ii) Deep Ensemble method. For MC Dropout [33], one typically inserts stochastic dropout variables at various points of the neural networks, and after training, a number of feedforward passes are performed to evaluate the mean and variance of the output. The latter is then regarded as the model uncertainty. For Deep Ensemble method [34, 35], one typically collects a family of models sharing the main network structure but differing in hyperparameters or simply the initial random weight distributions. From the ensemble set of outputs, the model prediction is taken as the average with uncertainty being the variance.
A prominent feature of Deep Evidential Learning is that the higher-order distribution parameters can be used to compute two different types of uncertainties: the aleatoric and epistemic uncertainties. The former probes the level of noise in data while the latter characterizes the uncertainty intrinsic to the model (see e.g. [36] for a detailed explanation). Distinguishing between these two classes of uncertainties is useful since aleatoric uncertainty is largely indicative of noise level in data whereas epistemic uncertainty points to model insufficiency to generalize beyond the training dataset, be it a problem of data insufficiency itself or an unsuitable order of model complexity for the task, etc. In our work, we demonstrated that these two uncertainty types can indeed be differentiated by a noise-sensitivity test and the degree to which each was associated with prediction errors.
Like in [8, 15, 16, 17], our study is based on the OpenKBP dataset, and involved constructing a dose prediction model as part of the KBP pipeline. As emphasized in [7], although the mean absolute error (MAE) is an indicator of the model’s feasibility as a dose prediction system, one should bear in mind that dose prediction models enact an intermediate role in the complete KBP pipeline. Ultimately, it is the deliverable treatment plan most compatible with various clinical criteria that we wish to generate. For the OpenKBP Challenge, while prediction models with better MAE dose score generally led to treatment plans with higher criteria satisfaction, the best KBP treatment plan in [7] turned out to be associated with the one that ranked 16 on the dose score chart with a MAE of Gy. In comparison, our Deep Evidential model’s MAE score was Gy. We also found that for a stable and effective implementation of the Deep Evidential model, much fine-tuning of model hyperparameters was required. Thus, we picked a vanilla 3D U-Net as our backbone architecture of which learning curve approximately flattened within a couple of hours’ training. This enabled us to perform extensive ablation experiments for completing a reformulation of the original theory in [29], so that it can be efficiently adapted for dose prediction with excellent uncertainty-error correlation. The primary focus of our work here is to elucidate the extent to which uncertainty estimates are correlated with model prediction errors, as they can be further harnessed to characterize reliability of a model.
Our paper is organized as follows. In Section 2, we furnish a brief exposition of the theoretical basis underpinning Deep Evidential Learning [29, 30], our reformulation of the loss function and some brief comments on related works in [8, 28]. In Section 3, we outline details of our model structure and implementation. This is followed by a presentation of various results in Section 4. We end with a summary of key findings, main limitations and significance of our work in Sections 5 and 6.
2 Preliminaries
2.1 Model uncertainties from a Bayesian prior probability distribution
In the standard supervised learning formalism, typically the model is trained via minimizing a loss function such as the mean squared error so that the neural networks’ weights converge to at least a local minimum point, yielding a good accuracy defined in terms of the error term implied by the loss function. This standard approach does not directly give any estimates of model or data noise uncertainties. In a more refined approach, one can attempt to provide an estimate of uncertainty by the following assertion:
-
•
The target outputs can be modeled as being drawn from a probability distribution. For example, in our context of dose prediction, we let the predicted dose in voxel be described by a Gaussian probability distribution function (PDF) with mean and variance parameters .
-
•
We then attempt to train the model to infer the set of (for every voxel ) by the principle of maximum likelihood estimation. Typically, this is performed by minimizing the negative log-likelihood function defined as
(1) For an ensemble of models being trained upon the same dataset to yield estimates of the parameters , one would obtain a spectrum of differing values of , the distribution being stochastic in nature due to distinct initial weight distributions, and their sensitivity to the random noise present in the training dataset.
The Deep Evidential Learning framework of [30] and [29] posits that there is another PDF which describes the distribution of the Gaussian mean and variance parameters. This can be regarded as a ‘higher-order’ PDF that in turn describes the distribution of of the ‘first-order’ PDF in eqn. (1). In Bayesian theory, there is a natural mathematical entity that enacts this role – the prior probability distribution that treats as random rather than deterministic variables. In [29], various model training and simulations were performed by taking the prior distribution to be a product of a Gaussian distribution (for ) and an inverse-gamma distribution (for ). Each of them is separately a two-parameter distribution. Suppressing the voxel index , we have
(2) | |||
(3) |
where is the inverse gamma distribution. The overall prior distribution is a four-parameter normal-inverse-gamma distribution defined as the product of the two PDFs above.
(4) |
Given this prior distribution, we can take expectation values and other statistical moments with respect to it to compute mean and variances. In the simplest form, Deep Evidential Learning implies attaching a final feedforward layer (e.g. of the convolutional type) to an existing model architecture such that we have as the eventual model outputs.
The mean prediction, epistemic and aleatoric uncertainties are then defined with respect to as follows.
(5) |
where is the ‘first-order’ PDF of eqn. (1). It is in this sense that is the ‘higher-order’ PDF providing the description of of as random variables. We note that the aleatoric and epistemic uncertainties are obtained after integrating over using eqn. (4). The factorized form of enables an analytic expressions for each of these uncertainties. In Deep Evidential Learning, the model output parameters are . From these parameters, we can then obtain the dose prediction (given by ) as well as the uncertainty estimates for each voxel.
2.2 Deep Evidential Regression for a dose prediction model: a refined loss function
We now apply such an uncertainty quantification framework in the context of a dose prediction model which estimates the dose delivered to each CT voxel based on an input set of CT images and various binary-valued masks representing the radiotherapy’s target regions and the organs-at-risk. The OpenKBP dataset of [7, 14] was used as our source, and a vanilla 3D U-Net was employed as the backbone architecture.
In [29], the first-order PDF pertaining to the model output was assumed to be the normal distribution , with a 4-parameter normal-inverse-gamma distribution ( eqn. (4) ) adopted to be the higher-order evidential distribution. For our case, a crucial caveat is that the radiation dose value is definitively constrained within a finite interval, and not unbounded. To incorporate this physical nature of the radiation dose, we first define a dimensionless dose parameter where is the physical dose in units of Gy. We then pass to its logit representation which we take as the form of output for the neural network (rather than the physical dose ). Ablation experiments using the validation dataset were used to determine the coefficients in the linear map such that the logit representation yielded the optimal model performance. Our choice of the linear map also avoided the singular points of the logit function by construction.
Formally, our choice of the output data representation implies that the first-order PDF for each voxel is related to the logit-normal distribution more precisely instead of a Gaussian. Its full form reads
(6) |
This is a close cousin of the normal distribution, and has a bounded domain as the support. The parameters are now dimensionless parameters defined as the Gaussian mean and standard deviation of the logit of the (dimensionless) dose . For their higher-order evidential distributions, we can still follow the procedure of [29], i.e. we adopt a normal distribution for and an inverse-gamma distribution for the variance to yield a normal-inverse-gamma distribution for (6). Also, similar to the approach in [29], we can derive the aleatoric and epistemic uncertainties. In their original forms as defined in eqn. (5), they are dimensionless variances associated with the logit representation. To convert them to the corresponding dimensionful ones in units of Gy2, we first invoke the standard approximation technique of relating the uncertainty in the (logit of ) to that in itself, before using the linear map to deduce the physical aleatoric and epistemic uncertainties in units of Gy2. For clarity, we refer to them separately as which are related to in eqn. (5) via .
Following [29], the primary loss function can be taken to be the maximum likelihood loss defined as the marginal likelihood obtained when we integrate over the first-order mean , and variance . After some algebra, we find that we obtain the likelihood function
(7) | |||||
(8) |
where the expression lying on the right of the term can be identified as the Student’s t distribution . Unfortunately, we found that in this form, the loss function did not lead to stable learning curves that could generate uncertainty heatmaps that correlated with prediction errors. After some experimentation, we found that making the following refinements to eqn. (8) significantly enhanced the effectiveness of the model:
-
•
removal of the factor from eqn. (8),
-
•
passing the negative log-likelihood function through a regularizing positive-definite activation function (our choice: standard logistic function),
-
•
adding a mean-squared-error loss function term.
These refinements led to the eventual form of our loss function being as follows.
(9) | |||||
(10) |
where are crucial hyperparameters of which optimal values for our purpose will be discussed in Sec. 3.2. The first regularization term preceded by was proposed in [29] to penalize statistical evidence for incorrectly labeled terms. We note in passing that in [29], it was argued that the prior (the function in eqn. (4) ) has a singular limit when we take parameter values corresponding to ‘zero evidence’, and that softening this limit by introducing finite yet small values for , did not appear to be effective. We found this to be the case as well. Another possibility that we explored is using Jeffrey’s prior (which in this case involves an improper uniform prior for the mean of the model output). But this turned out to be similarly ineffective relative to the choice adopted in [29].
2.3 Related Work
Among the many papers devoted to the subject of dose prediction models (see e.g. [5] for an extensive review and compilation), to our knowledge, there has been only two which directly delved into the notion of uncertainty quantification frameworks: [8] and [28]. These papers contained highly interesting results which we would like to briefly discuss in relation to our work.
In [8], the authors studied the frameworks of MC Dropout and bootstrap aggregation which is essentially a Deep Ensemble method where each model is trained on only a portion of the entire training dataset. Like in this work, they used the OpenKBP Challenge dataset of [14] for validating the performance of each uncertainty quantification framework, concluding that Deep Ensemble yielded a lower mean absolute error (MAE) while showing better correlation between uncertainty and prediction error. In particular, the authors proposed that with the aid of an additional scaling factor, the models generated uncertainty heatmaps and DVH confidence intervals which appeared reasonable. This scaling factor was defined differently for each region of interest (ROI) and was argued to be necessary to bring the uncertainty values into a more ‘interpretable scale’ as the raw ones were only providing ‘relative measures’. For each set of points belonging to a ROI in the validation dataset, this ROI-dependent factor was defined simply as the standard deviation of the ratio between the prediction error and the uncertainty value (see eqn. (3),(4) of [8]). In our work, we did not incorporate such an empirical scaling factor in our definitions of various uncertainties, including our final uncertainty heatmaps and the DVH confidence intervals. For Deep Evidential Learning where the predictive variance can be expressed precisely in terms of learned output variables of the model, in principle, it is not clear to us how an additional empirical scaling factor should be justified. By construction, this rescaling factor naturally enhances the correlation between uncertainty and error since the uncertainty values themselves are part of the factor’s definition, assuming that it generalizes well for unknown (e.g. testing) data. Conversely, we took the resulting average value of such a ratio (across all voxels) to assess whether the uncertainty distribution generated by the model was reasonable in terms of the overall scale of magnitude. Through the numerical values (see Table 1) and the DVH certainty intervals in Fig. 8, we found that this was indeed the case.
The work in [28] proposes the use of Gaussian mixture models of which parameters are the neural network’s outputs, and with U-Net as the backbone architecture. Such a mixture density network is very similar in form to our Deep Evidential Learning model since the output variables of our model are also parameters of a PDF. In our case, the PDF is a higher-order normal-inverse-gamma distribution whereas in [28], the parameters are the means and variances of the component normal distributions (see eqn. (1)) and the coefficients defining their linear combination. Each component of the Gaussian mixture pertained to a treatment protocol representing specific priorities (e.g. OAR sparing over target coverage, etc.), and the relative values of the variances quantified the trade-offs between these protocols. The cross-entropy loss function associated with the cumulative distribution function (CDF) of the predicted dose was then taken as part of the objective function for generating a dose mimicking model. In [28], deliverable treatment plans were the endpoints of their dose prediction algorithm, with the optimization problem solved by RayStation’s native sequential quadratic programming solver and final dose computed via a collapsed cone algorithm [28]. Although we noted that there was no scrutiny of the correlation between uncertainty and error, the authors of [28] (and its follow-up work [27]) notably pointed out that probabilistic dose prediction models such as [8] and ours could play a crucial role in their pipeline since our model would yield dose CDFs that can be used to define the cross-entropy loss function as part of the dose-mimicking model’s objective function. It would thus be interesting to extend results of our work here by integrating our probabilistic dose predictions into a pipeline like [28] where deliverable treatment plans with specified machine parameter settings can be anticipated as final outcomes.
3 Methodology
3.1 On the OpenKBP Challenge Dataset
The OpenKBP Challenge dataset of [7, 14] is devoted towards establishing an open framework for the development of plan optimization models for knowledge-based planning (KBP) in radiotherapy. It is an augmented variation of real clinical data constructed as follows: medical images were taken from several institutions available on The Cancer Imaging Archive (TCIA) [37] which hosts open-source data that has undergone de-identification compliant with DICOM standard.ivivivIn particular, all metadata and Protected Health Information (PHI) has been removed from the datasets provided in OpenKBP Challenge [7, 14]. After removing highly incomplete imaging datasets to obtain a final competition dataset comprising of 340 patients, synthetic radiation plans were generated for each of them using a published automated planning method [38]. This final dataset is then split randomly by the organizers of OpenKBP Challenge to be as follows: 200 (training), 40 (validation), 100 (testing). In our work, we followed this decomposition of data. As in a conventional machine learning pipeline, the validation dataset was used for tuning of hyperparameters whereas the testing dataset was used for reporting various results in our work. We refer the reader to [39, 40] for more pedagogical expositions of data-splitting practices and algorithms.
The radiotherapy plans were delivered using nine equidistant coplanar beams at various angles with a 6 MV step-and-shoot intensity-modulated-radiation-therapy (IMRT) in 35 fractions. The organizational team of the Challenge used the IMRTP library from A Computational Environment for Radiotherapy Research implemented in MATLAB [41] to generate the dose deposited at each voxel. These dose distributions were from fluence-based treatment plans with similar degrees of complexity [14, 42].
We used a -dimensional input representation consisting of (i)CT in Hounsfield units(clipped to be within [0, 4095] before being normalized to [0,1]), (ii)structure masks of three planning target volumes (PTV) and seven organs-at-risk (OAR) regions : each is a Boolean tensor labeling any voxel contained within the respective structure. The OARs are brainstem, spinal cord, right and left parotids, larynx, mandible and esophagus. The three PTV regions are: PTV56, PTV63, PTV70 which are targets that should receive 56 Gy, 63 Gy and 70 Gy of radiation dose respectively.
3.2 On model architecture and hyperparameters
Since our primary goal is to examine the uncertainty estimation aspects of the model, in this work, we adopt a simple vanilla 3D U-Net as the backbone architecture upon which we insert two additional layers so that they are compatible with the framework of Deep Evidential Learning. U-Nets and their variants have featured heavily in many deep-learning-guided tasks related to radiotherapy dose prediction [8, 14]. They are essentially convolutional neural networks equipped with an encoder-decoder architecture that allows effective learning of both image features at various levels of resolution (see e.g. [43, 44] for reviews on U-Net and [45, 46] for pedagogical introductions to convolutional neural networks).
Our U-Net has the following structure (see Fig. 1):
-
•
it takes in a 11-channel input of size voxels,
-
•
each downsampling level consists of two convolutional layers each with a kernel, ReLU activation and equipped with a dropout unit, followed by maximum pooling with kernel size ,
-
•
the downsampling operation is performed 4 times, with the dropout rates increasing consecutively as , and the number of convolutional filters for each layer being respectively,
-
•
the bottleneck layer has 256 filters and dropout rate of 0.30,
-
•
at each of the four upsampling levels, features from the contraction path are concatenated with the corresponding upsampled features, and the dropout rates decrease similarly in the reverse order.
-
•
a 8-channel pointwise convolution is then applied followed by a final 4-channel pointwise convolution that yields the four parameters of the Bayesian prior distribution of eqn. (4).
This model carries about free weight parameters. Various hyperparameters such as the dropout rate at each level, etc. were eventually adopted after running ablation experiments to determine their optimal values.
Imposing some choice of numerical bounds to in the final layer is a crucial aspect of the hyperparameter tuning process unique to this framework. The parameter sets the overall (dimensionless) scale to both uncertainties of which ratio is set by which has to be positive. For this dataset, we found good uncertainty-error correlation after imposing . In [29], the authors imposed the constraint which we also followed here, as for this range of values, the aleatoric and epistemic uncertainties can be conveniently described by simple analytic expressions. More broadly speaking, is the larger admissible range for as a parameter of the inverse-gamma distribution. We also found it useful to map the zero dose to a small positive value in the (dimensionless) normalized dose defined to be where is the physical dose in Gy. In the loss function, we work in the logit-representation of , with the predicted dose . Since this logistic function is monotonic, is bounded from below as with our choice of .

For the loss function originally proposed in [29], we found that unfortunately, it led to training curves with frequent oscillations and poor eventual outcomes. As explained in detail in Sec. 2.2, we reformulated the loss function for a logistic representation of the dose variable such that one could obtain a stable implementation. Our refined loss function in eqn. (10) is characterized by the hyperparameters which describe couplings of the loss function to a mean-squared-error (MSE) term and a KL-divergence-like regularization term respectively. Results from ablation experiments with validation dataset indicated that optimal values of these hyperparameters lie within the intervals , . For the results reported here for the rest of our paper, we took . In Fig. 2, we plot the learning curves pertaining to the original and our refined loss functions for comparison.

Towards comparing the Deep Evidential model against more common approaches of uncertainty quantification in literature, we implemented a Deep Ensemble model and Monte-Carlo Dropout model for the same dataset. Each of the 5 neural networks used in the ensemble model shared the same 3D U-Net architecture with all hyperparameters preserved. Random weight initialization was all performed via the He-uniform initializer [47]. Similarly, for the Monte-Carlo Dropout model, the same base model was used for passing 30 forward passes with the dropout layers activated for model prediction. All models were trained for 200 epochs with a learning rate of using Adam optimizer. Our focus in this work lies in studying aspects of uncertainty estimation, and thus our choice of a relatively simple backbone architecture equipped with only weight parameters. Compared to other more complicated setups, each of these models converged quickly in about a couple of hours while attaining a mean-absolute-error of dose prediction that laid within the top 20 of the score chart of the OpenKBP Challenge [14].
4 Results
We found that our Deep Evidential Learning model yielded uncertainty estimates which demonstrated strong correlation with prediction errors, while achieving a similar level of accuracy relative to the methods of Monte-Carlo Dropout and Deep Ensemble. The metrics that we used as measures of any putative uncertainty-error associations were (i)the Spearman’s rank correlation coefficient between the uncertainty and error values across the testing dataset and their patient-averaged distributions, (ii)the mutual information between them.
Model | MAE (Gy) | (Gy2) | |||
---|---|---|---|---|---|
Deep Evidential | 3.09 | (2.80, 0.20) | (0.83, 0.69) | (0.69, 0.63) | (0.67, 0.61) |
Monte-Carlo Dropout | 3.28 | 2.12 | 0.73 | 0.62 | 0.60 |
Deep Ensemble | 3.10 | 3.42 | 0.86 | 0.66 | 0.56 |
As summarized in Table 1, one observes that across the various correlation indices, the epistemic uncertainty’s degree of association with error was higher than Monte-Carlo Dropout and Deep Ensemble, with the only exception being its slightly lower than that of Deep Ensemble. For all three models, the correlation between patient-averaged uncertainties and errors was always higher, suggesting that there is a certain degree of stochasticity characterizing this relationship within each patient’s dataset. Relative to epistemic uncertainty , aleatoric uncertainty in the Deep Evidential Learning model consistently demonstrated lower correlations with error, and its value was less affected by averaging within individual patient (compared to other models). We also note that the order-of-magnitudes of the various averaged uncertainty values were such that , the exception being which was generally an order-of-magnitude below and uncertainties of the other two models.
4.1 Variation of prediction error with uncertainty threshold
The results in Table 1 indicated that Deep Evidential Learning model yielded spectra of uncertainty values which were more correlated with prediction error, relative to MC Dropout and Deep Ensemble methods. To further scrutinize the difference among the various frameworks, we examined how error distributions varied with changing threshold values of uncertainty. Since the error distributions were generically found to be quite skewed, we used the median of the error distribution as its characterizing parameter. Fig. 3 and 4 below reveal how median error changed with increasing levels of uncertainty thresholds for the different models. Assuming that uncertainty values correlate positively with prediction error implies that one expects a monotonically increasing curve in such a plot, a trend that was indeed manifest for these models as shown in Fig. 3 and 4.
In particular, from Fig. 4, the visibly more evident linearity of the curve for Deep Evidential Learning indicated a sensitivity of epistemic uncertainty to prediction errors that was relatively more uniformly calibrated compared to Monte-Carlo Dropout and Deep Ensemble methods.


4.2 Probing noise-sensitivity of uncertainty measures
A prominent feature of Deep Evidential Learning is a clear principled approach towards distinguishing between aleatoric and epistemic uncertainties. They can be defined precisely using the higher-order PDF as expressed in eqn. (5). Traditionally, aleatoric uncertainty is commonly interpreted as arising from inherent random noise in the data, in contrast to epistemic uncertainty which presumably captures uncertainty originating from model suitability and data sufficiency (see e.g. [32, 36]).
We would like to probe the level of sensitivity of various uncertainty measures to noise perturbations of the input data. In particular, we would like to identify if there would be any difference in noise-sensitivity between and . Our choice of perturbation is the addition of Gaussian noise [48] of mean zero and standard deviation 0.5 to each CT voxel (of which intensity was preprocessed to lie within ). To enable a more holistic description of response to noise, we plot the empirical cumulative distribution function (eCDF) of various uncertainty measures preceding and following the noise addition.
In Fig. 5, the plot of various eCDFs collectively illustrated the differences among the uncertainty measures in their responses towards the Gaussian perturbation. Since different measures yielded distinct overall scale of uncertainty magnitudes, we normalized each with respect to the maximum value for each uncertainty type. In the limit of the Gaussian noise completely dominating over any pre-existing noisiness, and assuming that the uncertainty value of each voxel is correlated only with the noise PDF (identical for every voxel), we would expect the uncertainty eCDF to approach a symmetric distribution with mean (normalized) uncertainty at 0.5 in this limiting scenario. As a simple reference distribution, in Fig. 5, we included the line corresponding to the uniform distribution. From just visual inspection, aleatoric uncertainty distribution and that of Deep Ensemble changed more significantly relative to others. The noise-induced fractional changes in the relative entropy (measuring the Kullback-Leibler divergence between each eCDF and the uniform distribution) turned out to be for aleatoric uncertainty and Deep Ensemble, about ten times larger than those for epistemic uncertainty and Monte-Carlo dropout.

The various curves appeared to suggest the following.
-
•
The relative responsiveness of aleatoric (stronger) and epistemic (weaker) uncertainties are indeed compatible with their conventional interpretations as reflecting inherent data noise and model-related uncertainties respectively.
-
•
Monte-Carlo Dropout yielded an uncertainty distribution that, like epistemic uncertainty, was not sensitive to the addition of the Gaussian noise. This favored its interpretation as an epistemic uncertainty in nature.
-
•
Deep Ensemble method yielded an uncertainty distribution that was responsive to the addition of noise. Like the aleatoric uncertainty in Deep Evidential Learning, the eCDF shifted to being less long-tailed and leaning towards being more uniform.
For the Deep Ensemble method, it was explained in [34, 49] that one can often use a Gaussian mixture model as an effective description of the ensemble prediction. By Eve’s law of total variance [49], the predictive variance admits a natural decomposition into aleatoric and epistemic components. This appears to be consistent with the noise sensitivity displayed by Deep Ensemble method in Fig. 5.
4.3 Applications
4.3.1 Uncertainty heatmaps
Heatmaps of the dose uncertainty distributions can be used through direct visual inspection to discover regions of potentially acute errors made by the dose prediction model. This capability hinges on the strength of the correlation between uncertainty and error distributions. In Table 1, epistemic uncertainty in the Deep Evidential model appeared to perform comparatively well, if not better, than the two more conventional uncertainty quantification frameworks as measured by mutual information and Spearman’s correlations. These statistical indices were defined at the level of the entire (testing) dataset. For each of the seven OAR and three PTV regions, one could in principle compute the uncertainty distribution localized within its interior. Thus, in Table 2 we collect for various ROIs their mean uncertainty () values together with the errors in predicted dose . The Spearman’s coefficients between each uncertainty type and the corresponding dose prediction error are also indicated (all with p-values ). Among all the ROIs, the L,R-parotid regions exhibited the highest correlations between and epistemic uncertainty, while the larynx region showed the lowest measures of correlation for both types of uncertainties. The mandible region carried the highest aleatoric uncertainty and was associated with the highest . Across the spectrum of ROIs in Table 2, we found each mean epistemic uncertainty to be strongly correlated with , with Spearman’s correlation coefficient of , whereas that for the aleatoric uncertainty was much less significant at .
ROI | (Gy2) | (Gy2) | (Gy) | ||
---|---|---|---|---|---|
PTV70 | 0.21 | 2.63 | 0.56 | 0.68 | 2.23 |
PTV63 | 0.21 | 2.65 | 0.65 | 0.74 | 2.16 |
PTV56 | 0.20 | 2.36 | 0.68 | 0.78 | 1.87 |
Brainstem | 0.10 | 0.94 | 0.69 | 0.69 | 0.65 |
Spinal Cord | 0.27 | 2.49 | 0.79 | 0.82 | 1.59 |
R-Parotid | 0.13 | 2.43 | 0.80 | 0.84 | 1.38 |
L-Parotid | 0.10 | 2.78 | 0.80 | 0.84 | 1.45 |
Esophagus | 0.06 | 1.18 | 0.71 | 0.71 | 0.98 |
Larynx | 0.17 | 1.95 | 0.43 | 0.48 | 2.46 |
Mandible | 0.21 | 3.77 | 0.83 | 0.79 | 4.02 |
At the level of each ROI, apart from the larynx, (1 s.f.) for all other OAR and PTV regions. Coupled with the error-uncertainty threshold study in Sec. 4.1 and the noise-sensitivity test in Sec. 4.2, this tapestry of results equips us with the basis for interpreting elevated heatmap regions of epistemic uncertainty as indicators of potential regions of prediction errors.
In Fig. 6, we display an illustrative set of axial CT images portrayed alongside their corresponding uncertainty heatmaps. In principle, the aleatoric uncertainty probes the level of inherent data noise, while the epistemic uncertainty is more strongly correlated with the elevated regions in the (leftmost) dose prediction error heatmap. Indeed, in Fig. 6, we observe that the relative intensity map of epistemic uncertainty tends to parallel the corresponding error heatmap more sensitively.

4.3.2 DVH with confidence bands
For each patient, we can combine the aleatoric and epistemic uncertainties learned by the neural network to construct confidence intervals for the Dose-Volume-Histogram for each of the region-of-interest. Denoting the predictive variance of the dose in each voxel by , and the predicted dose variable by , one can invoke the general Eve’s law of total variance to obtain
(11) |
where are the mean and variance defined in (3). To see this, we recall that in general, Eve’s law of total variance (see e.g. [50]) states the following relation for random variables and :
(12) |
where conditional expectations such as are random variables themselves. To apply (12) appropriately to our context, we can identify , and to collectively represent all model-dependent and data-dependent variables. The relation (11) then follows directly from (12) by virtue of the definitions of . In the following, we furnish a contextual proof of this result taking into account our framework explicitly.vvvOur style of proof is similar in spirit to that presented in [49] for a class of Deep Ensemble models that admit interpretations as Gaussian mixture models, and of which outputs are engineered to be . We begin with the definition
(13) |
Recall that in the Deep Evidential Learning framework, a higher-order prior distribution describes the randomness of . Also, taking into account the definition of , we can then write
(14) |
Substituting (14) into (13), we then have
(15) | |||||
(16) |
hence recovering eqn. (11) expected from more general arguments. Thus, we see that the predictive variance is the sum of the aleatoric and epistemic uncertainties. Our derivation is consistent with a similar result expressed in [36], and provides the theoretical basis for an operational definition of from knowledge of aleatoric and epistemic uncertainties.
In Fig. 7, we plot examples of individual patient’s DVH for various PTV and OAR regions, each equipped with a confidence interval. In each diagram of Fig. 7, refers to the standard error in dose in units of Gy obtained by averaging over all voxels of the ROI in the patient whose ID shown is defined in the dataset of [14]. A visible feature of these bands is that they tend to contain the more extreme deviations of each groundtruth DVH from its corresponding predicted one along their peripheral edges. This is consistent with the order-of-magnitude estimate of as drawn from numerical values of the mean error and uncertainties in Table 1. Thickness and various features of these DVH confidence bands for the OARs and PTV regions can be used to characterize the extent of reliability of various dose prediction models, differentiating among them in this aspect.




Fig. 8 presents an illustrative DVH containing a number of organs-at-risk and target regions for an individual patient in the dataset of [14]. Such a confidence band-enhanced DVH can be employed in the radiotherapy clinic to enable a more reliable interpretation of the model-predicted DVH.

5 Discussion
We have refined and applied Deep Evidential Learning framework to radiotherapy dose prediction, our most prominent finding being that the uncertainty estimates obtained from model training inherited strong correlations with prediction errors. In the following, we conclude with a summary of key findings accompanied by a discussion of various limitations of our work.
5.1 Summary of key findings
The original loss function proposed in [29] did not work in this particular context of dose prediction. We found that regularizing the final layer with a sigmoid function and adding a mean-squared-error term led to a stable and effective implementation of Deep Evidential Learning. Upon completion of model training, we found that the model inherited uncertainty estimates at a more granular level compared to conventional MC Dropout and Deep Ensemble methods. Epistemic uncertainty () in Deep Evidential Learning was highly correlated with prediction errors. Correlation indices (mutual information, Spearman’s coefficients) were comparable or higher than those for MC Dropout and Deep Ensemble methods (see Table 1). Aleatoric uncertainty demonstrated a more significant shift in its empirical CDF upon addition of Gaussian noise to CT intensity distribution compared to (see Fig. 5). It also displayed visibly weaker correlation with prediction errors relative to . These traits of appeared to be supportive of their conventional interpretations as reflecting data noise and model-related uncertainties respectively.
Another aspect of error-uncertainty relationship that appeared to distinguish Deep Evidential Learning was that the median error varied with uncertainty threshold much more linearly for as compared to MC Dropout and Deep Ensemble methods, indicative of a more uniformly calibrated sensitivity to model errors (see Fig. 3, 4). This is compatible with the fact that the uncertainty heatmaps generated by the model were found to be highly effective in identifying potential regions of model’s inaccuracies.
5.2 Towards incorporating our models in the radiotherapy clinic
An essential plan evaluation tool for the radiation oncologist is the DVH that characterizes a dose distribution. We demonstrated how Eve’s law of total variance enables one to express the predictive variance in terms of , and used this result to construct confidence intervals for DVH (see Fig. 8). This technique can be harnessed by the radiotherapy treatment planning team to enable a more statistically informed interpretation of DVH in deep learning-guided treatment planning. In the Knowledge-Based-Planning pipeline, dose prediction models can be used as inputs to some dose mimicking model [7] that produces deliverable treatments via optimization of a set of objective functions [25, 26]. The probabilistic dose distributions generated by our models can serve as robust inputs to the objective functions of [25, 26] to generate details of deliverable radiotherapy treatments. In this regard, it would be interesting to explore how the uncertainty estimates generated by our models translate to those of related quantities in the final treatment plan, such as the positions and motion dynamics of multileaf collimators and other machine parameters. This would pave the way for completing an uncertainty-aware AI-guided pipeline for radiotherapy treatment planning.
5.3 Limitations and Future Directions
Finally, we outline several limitations of our work together with corresponding suggestions for future directions. Although the simple 3D U-Net backbone architecture led to efficient and fast convergence, it would be interesting to study how more complicated network structures like the transformers-based models of [17, 18] perform after modifying their final layers so that there are four output heads corresponding to the parameters of in Deep Evidential Learning.
As a simple explicit example, a U-Net-like transformer model (Swin-UNet) was proposed in [51] and [52] for segmentation tasks. To enable Deep Evidential Learning on such backbone models, one can append a final 4-channel pointwise convolution layer to these transformer block-based architectures, which will yield the four parameters of the prior normal-inverse-gamma distribution in eqn. (4) as model outputs. Model training should employ the maximum likelihood loss function of [29] or the refined version that we have proposed here in eqn. (10). More generally, this simple minimal prescription of appending a final layer of 4-channel pointwise convolution layer applies for other transformer-based vision models adapted for segmentation. If higher model complexity is needed, this final layer can be preceded by more multi-channel pointwise convolution layers, with other essential traits of transformer models left intact, e.g. multi-head self-attention modules, shifted window-based attention mechanism, etc. Details of these inner layers are unaffected and do not require additional modifications when encapsulating the transformer-based vision model within the framework of Deep Evidential Learning.
Enabling Deep Evidential Learning on other backbone model structures will provide a more extensive study of the degree to which a low MAE can be attained while not compromising the uncertainty-error correlation. Indeed it will be interesting to explore how our refined Deep Evidential Learning model works for other regression tasks beyond the context of dose prediction, such as traffic forecasting in telecommunication networks in [53, 54]. It would be interesting to pursue whether the variety of model architectures examined in [53, 54] can be extended to uncertainty-aware ones using our techniques.
The MC Dropout and Deep Ensemble methods we studied for comparison purpose admit more complex variants which, in principle, may also yield aleatoric and epistemic uncertainties [36, 49, 55]. It would be interesting to perform a similar analysis for them, although they would bring with them additional challenges (e.g. the use of Laplacian priors for MC Dropout setting as described in [36]; computational cost in picking the optimal ensemble parameters as explained in [55] where a novel automated approach for Deep Ensemble was proposed).
We had used the same regularization term in the loss function following [29], which is independent of the parameter of the prior distribution in eqn. (4). It would be interesting to explore if there are other alternatives which express full dependence on all four parameters of . In an analogous framework for classification, Sensoy et al. in [30] showed that the regularization term measuring the Kullback-Leibler divergence from an uninformative prior (e.g. uniform distribution) worked well for classification problems, and notably, the authors of [29] had attempted to find the corresponding version for regression yet without success.
Fundamentally, our modeling of dose prediction was based on the simplifying assumption that the patient’s CT images and the oncologist’s specification of the target regions are sufficient inputs for predicting the pareto-optimal dose distribution. In reality, effectiveness of radiation is often sensitive to a more complex web of biological factors. Nonetheless, we hope that our work can be a good starting point towards precision radiation oncology. Towards this goal, a pertinent future direction would be to expand the set of model’s input features to include genetic, molecular and other unique clinical characteristics of each patient, beyond just using medical images.
6 Conclusion
In this work, we have presented a novel application of Deep Evidential Learning in the domain of radiotherapy dose prediction. Using medical images of the OpenKBP Challenge dataset, we found that this model can be effectively harnessed to yield uncertainty estimates upon completion of network training. This was achieved only after reformulating the original loss function of [29] for a stable implementation. Since epistemic uncertainty was found to be highly correlated with prediction errors, its distribution could be used to discover and highlight areas of potential inaccuracies of the neural network, apart from being a diagnostic indicator of model reliability. Towards enhancing its clinical relevance, we demonstrated how to construct the predicted Dose-Volume-Histograms’ confidence intervals. We hope that this work has furnished the crucial preliminary steps towards realizing Deep Evidential Learning for dose prediction models, paving another path towards quantifying the reliability of treatment plans in the context of knowledge-based-planning.
Declaration of competing interest
All authors have no conflict of interest to declare.
Acknowledgments
H.S.Tan wishes to acknowledge that the primary development of this work was conducted while he was affiliated with University of Pennsylvania, Perelman School of Medicine, Department of Radiation Oncology, to whom he is grateful for support.
References
- [1] Woody NM, Gregory M M Videtic MDCMF, Vassil AD. Handbook of Treatment Planning in Radiation Oncology. Springer Publishing Company; 2014.
- [2] Fraass B, Doppke K, Hunt M, Kutcher G, Starkschall G, Stern R, et al. American Association of Physicists in Medicine Radiation Therapy Committee Task Group 53: Quality assurance for clinical radiotherapy treatment planning. Medical Physics. 1998;25(10):1773-829.
- [3] Nelms BE, Robinson G, Markham J, Velasco K, Boyd S, Narayan S, et al. Variation in external beam treatment plan quality: An inter-institutional study of planners and planning systems. Pract Radiat Oncol. 2012 Oct-Dec;2(4):296-305.
- [4] Craft DL, Hong TS, Shih HA, Bortfeld TR. Improved planning time and plan quality through multicriteria optimization for intensity-modulated radiotherapy. Int J Radiat Oncol Biol Phys. 2012 Jan;82(1):e83-90.
- [5] Kui X, Liu F, Yang M, Wang H, Liu C, Huang D, et al. A review of dose prediction methods for tumor radiation therapy. Meta-Radiology. 2024;2(1):100057.
- [6] Wu B, Ricchetti F, Sanguineti G, Kazhdan M, Simari P, Chuang M, et al. Patient geometry-driven information retrieval for IMRT treatment plan quality control. Medical Physics. 2009;36(12):5497-505.
- [7] Babier A, Mahmood R, Zhang B, Alves VGL, Barragán-Montero AM, Beaudry J, et al. OpenKBP-Opt: an international and reproducible evaluation of 76 knowledge-based planning pipelines. Phys Med Biol. 2022 Sep;67(18).
- [8] Nguyen D, Barkousaraie AS, Bohara G, Balagopal A, McBeth R, Lin MH, et al. A comparison of Monte Carlo dropout and bootstrap aggregation on the performance and uncertainty estimation in radiation therapy dose prediction with deep learning neural networks. Physics in Medicine and Biology. 2021;66(5):054002.
- [9] Nguyen D, Long T, Jia X, Lu W, Gu X, Iqbal Z, et al. A feasibility study for predicting optimal radiation therapy dose distributions of prostate cancer patients from patient anatomy using deep learning. Sci Rep. 2019 Jan;9(1):1076.
- [10] Kearney V, Chan JW, Haaf S, Descovich M, Solberg TD. DoseNet: a volumetric dose prediction algorithm using 3D fully-convolutional neural networks. Physics in Medicine and Biology. 2018 dec;63(23):235022. Available from: https://dx.doi.org/10.1088/1361-6560/aaef74.
- [11] Shiraishi S, Moore KL. Knowledge-based prediction of three-dimensional dose distributions for external beam radiotherapy. Medical Physics. 2016;43(1):378-87.
- [12] Barragán-Montero AM, Nguyen D, Lu W, Lin MH, Norouzi-Kandalan R, Geets X, et al. Three-dimensional dose prediction for lung IMRT patients with deep neural networks: robust learning from heterogeneous beam configurations. Medical Physics. 2019;46(8):3679-91.
- [13] Nguyen D, Jia X, Sher D, Lin MH, Iqbal Z, Liu H, et al. 3D radiotherapy dose prediction on head and neck cancer patients with a hierarchically densely connected U-net deep learning architecture. Physics in Medicine and Biology. 2019 mar;64(6):065020.
- [14] Babier A, Zhang B, Mahmood R, Moore KL, Purdie TG, McNiven AL, et al. OpenKBP: The open‐access knowledge‐based planning grand challenge and dataset. Medical Physics. 2021 Jun;48(9):5549–5561.
- [15] Soomro MH, Alves VGL, Nourzadeh H, Siebers JV. DeepDoseNet: a deep learning model for 3D dose prediction in radiation therapy. arXiv preprint arXiv:211100077. 2021.
- [16] Li H, Peng X, Zeng J, Xiao J, Nie D, Zu C, et al. Explainable attention guided adversarial deep network for 3D radiotherapy dose distribution prediction. Knowledge-Based Systems. 2022;241:108324.
- [17] Hu C, Wang H, Zhang W, Xie Y, Jiao L, Cui S. TrDosePred: A deep learning dose prediction algorithm based on transformers for head and neck cancer radiotherapy. J Appl Clin Med Phys. 2023 Jul;24(7):e13942.
- [18] Wang K, Tan HS, Mcbeth R. Swin UNETR++: Advancing Transformer-Based Dense Dose Prediction Towards Fully Automated Radiation Oncology Treatments; 2024.
- [19] Abdar M, Pourpanah F, Hussain S, Rezazadegan D, Liu L, Ghavamzadeh M, et al. A review of uncertainty quantification in deep learning: Techniques, applications and challenges. Information Fusion. 2021 dec;76:243-97.
- [20] Lambert B, Forbes F, Doyle S, Dehaene H, Dojat M. Trustworthy clinical AI solutions: A unified review of uncertainty quantification in Deep Learning models for medical image analysis. Artificial Intelligence in Medicine. 2024;150:102830.
- [21] Ghesu FC, Georgescu B, Mansoor A, Yoo Y, Gibson E, Vishwanath RS, et al. Quantifying and leveraging predictive uncertainty for medical image assessment. Medical Image Analysis. 2021;68:101855.
- [22] Irvin J, Rajpurkar P, Ko M, Yu Y, Ciurea-Ilcus S, Chute C, et al. CheXpert: a large chest radiograph dataset with uncertainty labels and expert comparison. In: Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence and Thirty-First Innovative Applications of Artificial Intelligence Conference and Ninth AAAI Symposium on Educational Advances in Artificial Intelligence. AAAI’19/IAAI’19/EAAI’19. AAAI Press; 2019. Available from: https://doi.org/10.1609/aaai.v33i01.3301590.
- [23] Zou K, Yuan X, Shen X, Wang M, Fu H. TBraTS: Trusted Brain Tumor Segmentation. In: Medical Image Computing and Computer Assisted Intervention – MICCAI 2022: 25th International Conference, Singapore, September 18–22, 2022, Proceedings, Part VIII. Berlin, Heidelberg: Springer-Verlag; 2022. p. 503–513. Available from: https://doi.org/10.1007/978-3-031-16452-1_48.
- [24] Jones CK, Wang G, Yedavalli V, Sair H. Direct quantification of epistemic and aleatoric uncertainty in 3D U-net segmentation. J Med Imaging (Bellingham). 2022 May;9(3):034002.
- [25] Benson HP. Existence of efficient solutions for vector maximization problems. Journal of Optimization Theory and Applications. 1978;26(4):569-80.
- [26] Chan TCY, Craig T, Lee T, Sharpe MB. Generalized Inverse Multiobjective Optimization with Application to Cancer Therapy. Operations Research. 2014;62(3):680-95.
- [27] Zhang T, Bokrantz R, Olsson J. Probabilistic feature extraction, dose statistic prediction and dose mimicking for automated radiation therapy treatment planning. Med Phys. 2021 Sep;48(9):4730-42.
- [28] Nilsson V, Gruselius H, Zhang T, De Kerf G, Claessens M. Probabilistic dose prediction using mixture density networks for automated radiation therapy treatment planning. Phys Med Biol. 2021 Feb;66(5):055003.
- [29] Amini A, Schwarting W, Soleimany A, Rus D. Deep evidential regression. Advances in neural information processing systems. 2020;33:14927-37.
- [30] Sensoy M, Kaplan L, Kandemir M. Evidential deep learning to quantify classification uncertainty. Advances in neural information processing systems. 2018;31.
- [31] Çiçek Ö, Abdulkadir A, Lienkamp SS, Brox T, Ronneberger O. 3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation. In: Ourselin S, Joskowicz L, Sabuncu MR, Unal G, Wells W, editors. Medical Image Computing and Computer-Assisted Intervention – MICCAI 2016. Cham: Springer International Publishing; 2016. p. 424-32.
- [32] Zou K, Chen Z, Yuan X, Shen X, Wang M, Fu H. A review of uncertainty estimation and its application in medical imaging. Meta-Radiology. 2023;1(1):100003.
- [33] Gal Y, Ghahramani Z. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning. In: Balcan MF, Weinberger KQ, editors. Proceedings of The 33rd International Conference on Machine Learning. vol. 48 of Proceedings of Machine Learning Research. New York, New York, USA: PMLR; 2016. p. 1050-9. Available from: https://proceedings.mlr.press/v48/gal16.html.
- [34] Lakshminarayanan B, Pritzel A, Blundell C. Simple and scalable predictive uncertainty estimation using deep ensembles. In: Proceedings of the 31st International Conference on Neural Information Processing Systems. NIPS’17. Red Hook, NY, USA: Curran Associates Inc.; 2017. p. 6405–6416.
- [35] Ganaie MA, Hu M, Malik AK, Tanveer M, Suganthan PN. Ensemble deep learning: A review. Engineering Applications of Artificial Intelligence. 2022 Oct;115:105151. Available from: http://dx.doi.org/10.1016/j.engappai.2022.105151.
- [36] Kendall A, Gal Y. What uncertainties do we need in Bayesian deep learning for computer vision? In: Proceedings of the 31st International Conference on Neural Information Processing Systems. NIPS’17. Red Hook, NY, USA: Curran Associates Inc.; 2017. p. 5580–5590.
- [37] Clark K, Vendt B, Smith K, Freymann J, Kirby J, Koppel P, et al. The Cancer Imaging Archive (TCIA): maintaining and operating a public information repository. J Digit Imaging. 2013 Dec;26(6):1045-57.
- [38] Babier A, Mahmood R, McNiven AL, Diamant A, Chan TCY. Knowledge-based automated planning with three-dimensional generative adversarial networks. Med Phys. 2020 Feb;47(2):297-306.
- [39] Walston SL, Seki H, Takita H, Mitsuyama Y, Sato S, Hagiwara A, et al. Data set terminology of deep learning in medicine: a historical review and recommendation. Japanese Journal of Radiology. 2024 Jun. Available from: http://dx.doi.org/10.1007/s11604-024-01608-1.
- [40] Bradshaw TJ, Huemann Z, Hu J, Rahmim A. A Guide to Cross-Validation for Artificial Intelligence in Medical Imaging. Radiol Artif Intell. 2023 Jul;5(4):e220232.
- [41] Deasy JO, Blanco AI, Clark VH. CERR: a computational environment for radiotherapy research. Med Phys. 2003 May;30(5):979-85.
- [42] Craft D, Süss P, Bortfeld T. The tradeoff between treatment plan quality and required number of monitor units in intensity-modulated radiotherapy. Int J Radiat Oncol Biol Phys. 2007 Apr;67(5):1596-605.
- [43] Azad R, Aghdam EK, Rauland A, Jia Y, Avval AH, Bozorgpour A, et al.. Medical Image Segmentation Review: The success of U-Net; 2022. Available from: https://arxiv.org/abs/2211.14830.
- [44] Yin XX, Sun L, Fu Y, Lu R, Zhang Y. U-Net-Based Medical Image Segmentation. J Healthc Eng. 2022;2022:4189781.
- [45] O’Shea K, Nash R. An Introduction to Convolutional Neural Networks; 2015. Available from: https://arxiv.org/abs/1511.08458.
- [46] Sarvamangala DR, Kulkarni RV. Convolutional neural networks in medical image understanding: a survey. Evol Intell. 2022;15(1):1-22.
- [47] He K, Zhang X, Ren S, Sun J. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification. In: 2015 IEEE International Conference on Computer Vision (ICCV); 2015. p. 1026-34.
- [48] Gravel P, Beaudoin G, De Guise JA. A method for modeling noise in medical images. IEEE Trans Med Imaging. 2004 Oct;23(10):1221-32.
- [49] Valdenegro-Toro M, Saromo D. A Deeper Look into Aleatoric and Epistemic Uncertainty Disentanglement. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). 2022:1508-16.
- [50] Weiss NA, Holmes PT, Hardy M. A Course in Probability. Pearson Addison Wesley; 2006. Available from: https://books.google.com/books?id=Be9fJwAACAAJ.
- [51] Cao H, Wang Y, Chen J, Jiang D, Zhang X, Tian Q, et al.. Swin-Unet: Unet-like Pure Transformer for Medical Image Segmentation; 2021. Available from: https://arxiv.org/abs/2105.05537.
- [52] Cai Y, Long Y, Han Z, Liu M, Zheng Y, Yang W, et al. Swin Unet3D: a three-dimensional medical image segmentation network combining vision transformer and convolution. BMC Medical Informatics and Decision Making. 2023;23(1):33.
- [53] Alizadeh M, Beheshti MTH, Ramezani A, Bolouki S. An optimized hybrid methodology for short-term traffic forecasting in telecommunication networks. Transactions on Emerging Telecommunications Technologies. 2023;34(12):e4860. Available from: https://onlinelibrary.wiley.com/doi/abs/10.1002/ett.4860.
- [54] Alizadeh M, Beheshti MTH, Ramezani A, Saadatinezhad H. Network Traffic Forecasting Based on Fixed Telecommunication Data Using Deep Learning. In: 2020 6th Iranian Conference on Signal Processing and Intelligent Systems (ICSPIS); 2020. p. 1-7.
- [55] Egele R, Maulik R, Raghavan K, Lusch B, Guyon I, Balaprakash P. AutoDEUQ: Automated Deep Ensemble with Uncertainty Quantification. In: 2022 26th International Conference on Pattern Recognition (ICPR). Los Alamitos, CA, USA: IEEE Computer Society; 2022. p. 1908-14. Available from: https://doi.ieeecomputersociety.org/10.1109/ICPR56361.2022.9956231.