This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Deep Learning Analysis of Deeply Virtual Exclusive Photoproduction

Jake Grigsby [email protected] Department of Physics, University of Virginia, Charlottesville, VA 22904, USA.    Brandon Kriesten [email protected] Department of Physics, University of Virginia, Charlottesville, VA 22904, USA.    Peter Alonzi [email protected] School of Data Science, University of Virginia, Charlottesville, VA 22904, USA    Matthias Burkardt [email protected] New Mexico State University - Department of Physics, Box 30001 MSC 3D, Las Cruces NM, 88003 - USA    Joshua Hoskins [email protected] Department of Physics, University of Virginia, Charlottesville, VA 22904, USA.    Simonetta Liuti [email protected] Department of Physics, University of Virginia, Charlottesville, VA 22904, USA.
Abstract

We present a Machine Learning based approach to the cross section and asymmetries for deeply virtual Compton scattering from an unpolarized proton target using both an unpolarized and polarized electron beam. Machine learning methods are needed to study and eventually interpret the outcome of deeply virtual exclusive experiments since these reactions are characterized by a complex final state with a larger number of kinematic variables and observables, exponentially increasing the difficulty of quantitative analyses. Our deep neural network (FemtoNet) uncovers emergent features in the data and learns an accurate approximation of the cross section that outperforms standard baselines. FemtoNet reveals that the predictions in the unpolarized case systematically show a smaller relative median error than the polarized that can be ascribed to the presence of the Bethe Heitler process. It also suggests that the tt dependence can be more easily extrapolated than for the other variables, namely the skewness, ξ\xi and four-momentum transfer, Q2Q^{2}. Our approach is fully scalable and will be capable of handling larger data sets as they are released from future experiments.

I Introduction

Artificial Intelligence (AI) methods including machine learning (ML), deep learning (DL) and neural networks (NN) have been increasingly applied to attack several of the outstanding questions on strongly interacting systems in Quantum Chromodynamics (QCD), ranging from both the sign problem (see Ohnishi et al. (2019); Mori et al. (2017); Alexandru et al. (2018) and references therein), and the inverse problem in lattice QCD, to extracting the behavior of parton distributions functions (PDFs) from data. However, the systematic application of AI to problems in hadronic physics can still be considered to be in its initial stages, which involves exploring a variety of methodologies Bedaque et al. (2020). One important aspect that is still being developed concerns the treatment of uncertainty, in particular, studying the accuracy of AI based predictions within a well defined statistical approach and with precise benchmarking.

In this paper we address the recently developed science of nuclear femtography, which aims at reconstructing the spatial structure of hadrons, or nucleon tomography, using information from deeply virtual electron proton exclusive photoproduction, epepγep\rightarrow e^{\prime}p^{\prime}\gamma, and various other related reactions. The open questions that we address concern, in particular, the choice of a ML method that is suitable to attack the issues specific to nuclear femtography, including the impact of noise and experimental uncertainty.

Deeply virtual exclusive processes have been identified as the cleanest experimental probe of the 3D structure of the proton as well as its mechanical properties including the quark and gluon distributions of angular momentum, pressure, and shear forces (see reviews in Goeke et al. (2001); Diehl (2001); Belitsky and Radyushkin (2005); Kumericki et al. (2016)). Information on these distributions is encoded in the Generalized Parton Distributions (GPDs) which are functions of the longitudinal momentum fraction, xx, and of the longitudinal momentum transfer, ξ\xi (skewness parameter) and four momentum transfer squared, t=Δ2t=\Delta^{2}, at a given scale determined by the photon virtuality, Q2Q^{2}. GPDs can be defined for all the different polarization configurations of the struck quark/gluon and proton. Their Fourier transform with respect to the transverse component of Δ\Delta produces faithful images of the proton constituents structure on the transverse plane.

In exclusive photoproduction we access GPDs from the matrix elements for deeply virtual Compton scattering (DVCS), where the photon is produced at the proton vertex. A competing background process given by the Bethe-Heitler (BH) reaction is also present, where the photon is emitted from the electron and the matrix elements measure the proton elastic form factors. Note that for this prototypical process, the cross section is a function of four independent kinematic variables: the angle between the lepton and hadron planes, ϕ\phi, the scale, Q2Q^{2}, the skewness, ξ\xi, which is related to Bjorken xx, and tt; GPDs for all quark flavors appear in the cross section embedded in convolutions over xx known as Compton form factors (CFFs), with complex kernels depending on (x,ξ,Q2x,\xi,Q^{2}). Several CFF combinations corresponding to various quark-proton polarization configurations appear simultaneously in the cross section terms for each beam-target polarization configuration Kriesten et al. (2020). The complexity associated with the large number of kinematic variables, and the complex structure of the cross section makes the problem of extracting tomographic images of quark and gluon structure of the proton virtually impossible to solve with traditional methods; for high precision femtography, which is required to obtain proton images, we need to develop more sophisticated analyses. The success of ML methodologies in modeling complex natural phenomenon make this a prime choice for GPD extraction.

The ultimate goal is development of a tool set for the extraction of GPD from experimental data however, as a first step, we develop a framework for the extraction of the scattering cross-section. The results of our analysis show that the network can effectively generalize in tt, even in the regions with no data, but is less effective in ξ\xi, or Q2Q^{2}. These results are not only beneficial to the physics community but provide interesting overlays with the data science community. One important example is discussion of handling experimental error bars found in Section III.2.1, which is ubiquitous to physics analyses but less commonly considered in building machine learning models.

We use industry standard ML techniques to fit a cross section model based on currently available DVCS experimental data, allowing us to make efficient and accurate predictions across a wide kinematic range represented by the data. By estimating model uncertainty, we can make informed decisions about predictions well outside of the region defined by our data. Several of the practical challenges of fitting a sparse Neural Network, with significant experimental uncertainty, are addressed in this paper. All of our software tools are open-source and can be found Group (2020).

This paper is organized as follows: in Section II we present the theoretical framework for DVCS and give background on Deep Learning and Neural Networks. In Section III we detail the specifics of our cross section network and training process. Then, in Section IV we use our networks to make predictions across a wide range of kinematic bins, comparing them against both experimental measurements and previous theory-driven models. Finally, our conclusions and outlook are presented in Section V.

II Background

II.1 DVCS Theoretical Framework

The cross section for epepγep\rightarrow e^{\prime}p^{\prime}\gamma reads,

d5σdxBjdQ2d|t|dφdφS=Γ|T|2.\frac{d^{5}\sigma}{dx_{Bj}dQ^{2}d|t|d\varphi d\varphi_{S}}=\Gamma\,\big{|}T\big{|}^{2}\;. (1)

where TT is a superposition of the DVCS amplitude and it’s background process (Bethe-Heitler) BH,

|T|2=|TBH+TDVCS|2=|TBH|2+|TDVCS|2+|T|^{2}=|T_{\rm BH}+T_{\rm DVCS}|^{2}=|T_{\rm BH}|^{2}+|T_{\rm DVCS}|^{2}+\mathcal{I}\; (2)
\displaystyle\mathcal{I} =\displaystyle= TBHTDVCS+TDVCSTBH\displaystyle T_{BH}^{*}T_{DVCS}+T_{DVCS}^{*}T_{BH} (3)

The cross section is a differential form of five independent variables, the four momentum transfer squared between the initial (ee) and final (ee^{\prime}) electrons, Q2Q^{2}; xBj=Q2/(2Mν)x_{Bj}=Q^{2}/(2M\nu) where ν\nu is the energy transfer between the initial and final electrons and MM is the proton mass; tt, the (negative) four momentum transfer squared between the initial and final proton; the angle φ\varphi between the lepton and hadron planes, and φS\varphi_{S}, the azimuthal angle of the transverse proton spin vector. We will consider in our analysis available fixed target data for an unpolarized target where the measured set of observables includes

Ee,Q2,xBj,t,φ,E_{e},Q^{2},x_{Bj},t,\varphi,

EeE_{e} being the initial electron energy in the lab (the angle φS\varphi_{S} is integrated over since we are not considering scattering from a polarized target). The ultimate physics goal is to extract from experiment the QCD CFFs which appear in the DVCS matrix element. At leading order one can define eight CFFs, which correspond to the real and imaginary parts of the two vector and two axial vector current components, respectively. Before considering the full extraction of all eight CFFs, an important first step is to understand what is required to learn the form of the cross section. Our ML analysis yields model independent results for the cross section independent from the separation into the BH, DVCS and {\cal I} components given in Eq.(2). Because of the nature of the scattering experiments there are many more data on the ϕ\phi dependence than on the other variables (see Table I in Section III.1). In this paper we focus on the simplest case of unpolarized target scattering for which we also have the largest DVCS dataset.

II.2 Deep Neural Networks

A Deep Neural Network (DNN) is an effective function approximation method consisting of several layers of computation, in which each layer computes a new representation of the previous layer’s outputs. In the simplest case, each layer consists of an affine transformation of the previous layer’s representation of the data, followed by an element-wise nonlinearity. For example, the output of the first layer is:

𝐱1=z(𝐖0𝐱+𝐛0)\displaystyle\mathbf{x}_{1}=z(\mathbf{W}_{0}\mathbf{x}+\mathbf{b}_{0}) (4)

Where zz is a nonlinear activation function applied to each element of its input vector. 𝐖\mathbf{W} is an m×nm\times n matrix where mm is the dimension of the previous layer’s representation and nn is the dimension of the current layer’s output. 𝐛\mathbf{b} is a vector of size nn. The entire model can be written as a composition of functions

fθ(𝐱)=\displaystyle f_{\theta}(\mathbf{x})= 𝐖k(z(𝐖k1((z(𝐖0𝐱\displaystyle\mathbf{W}_{k}(z(\mathbf{W}_{k-1}(\dots(z(\mathbf{W}_{0}\mathbf{x} +𝐛0)))+𝐛k1))+𝐛k,\displaystyle+\mathbf{b}_{0})))+\mathbf{b}_{k-1}))+\mathbf{b}_{k},

where kk is the number of hidden layers in the model. Model ‘architectures’ are descriptions of the number (kk) and size (nn) of a network’s layers, as well as the choice of nonlinearity (zz). θ\theta is the set of all the entries in the transformation weights 𝐖\mathbf{W} and biases 𝐛\mathbf{b}. These parameters are initialized randomly, and then adjusted by an optimization algorithm. The goal is to minimize some metric of prediction error - usually called the loss function. A standard choice for regression problems is the Mean Squared Error (MSE):

θMSE(x,y)=12(yfθ(x))2\displaystyle\mathcal{L}_{\theta}^{MSE}(\textbf{x},y)=\frac{1}{2}(y-f_{\theta}(\textbf{x}))^{2} (6)

Where yy is the true value of our predicted variable, often called the ‘label’ or ‘target’ value. In our case, backpropagation is used to compute the partial derivatives of the loss function on a batch of BB samples w.r.t each of the parameters in θ\theta, and the resulting gradient vector is used to update the parameters via Stochastic Gradient Descent (SGD):

θt+1=θtλθ1BΣi=0Bθt(x(i),σi)\displaystyle\theta_{t+1}=\theta_{t}-\lambda\nabla_{\theta}\frac{1}{B}\displaystyle\Sigma_{i=0}^{B}{\mathcal{L}_{\theta_{t}}(\textbf{x}^{(i)},\sigma^{i})} (7)

Where λ\lambda is a learning rate parameter. Modern optimizers (Kingma and Ba, 2014) (Hinton et al., ) often dynamically adjust λ\lambda based on heuristics like momentum, where consecutive gradient steps in a similar direction increase the learning rate and rapid changes in direction push it towards zero.

III Training the Deep Network

III.1 Experimental Data

We compiled a dataset from several DVCS experiments Georges (2018); Defurne et al. (2015, 2017); Jo (2012), covering the range of kinematic bins shown in Figure 1. We discarded data sets that listed no explicit φ\varphi dependence. We used a common format for collecting and recording our data sets such that they list the kinematic dependence of the cross section points, the phi dependence, the polarization of the observable, and the statistical and systematic errors.

Experiment Q2Q^{2} (GeV2) tmintmaxt_{\text{min}}-t_{\text{max}} (GeV2) xBjminxBjmaxx_{\text{Bj}}^{\text{min}}-x_{\text{Bj}}^{\text{max}} EbeamE_{\text{beam}} (GeV) Polarization # points Reference
Hall B 1.11 / 3.77 -0.11 / -0.45 0.13 / 0.48 5.75 UU / LU 3862 Jo (2012)
Hall A 1.45 / 2.38 -0.17 / -0.37 0.34 / 0.40 5.75 UU / LU 936 Defurne et al. (2015)
Hall A 1.51 / 2.00 -0.18 / -0.36 0.36 5.55 UU / LU 788 Defurne et al. (2017)
Hall A 2.71 / 8.51 -0.21 / -1.28 0.34 / 0.61 4.49 / 10.99 UU / LU 2160 Georges (2018)
Table 1: Experimental data used in our neural network analysis. No data from HERMES or COMPAS was used because it is integrated in φ\varphi and therefore inconsistent dimensionally with a majority of the JLAB DVCS data.
Refer to caption
Refer to caption
Figure 1: (Top) Kinematic region in xBjx_{Bj} and tt and (Bottom) xBjx_{Bj} and Q2Q^{2} where experimental measurements used in this analysis have been taken at 6 GeV and 12 GeV.

Each entry consists of the input kinematic variables (xBjx_{Bj}, tt, Q2Q^{2}, EbE_{b}, φ\varphi) as well as the value and experimental error of the cross section. In total, we are working with 3,862 unpolarized and 3,884 polarized data points. Because these measurements are typically taken by sweeping across φ\varphi values in a fixed kinematic bin, most points are unique only in φ\varphi, meaning there are relatively few values of xBj,t,Q2,Ebx_{Bj},t,Q^{2},E_{b}. This presents an interesting challenge for ML approaches, which become much more capable of predicting across φ\varphi values than they are of extrapolating these patterns to distant kinematic space.

To help measure this effect, we use a 70%/20%/10%70\%/20\%/10\% split to separate the kinematic bins into three groups, which we’ll call the training set, validation set and test set, respectively. Note that this differs from the standard ML practice of splitting the datapoints themselves into random groups. The training set is used to directly adjust the parameters of the model. The validation set is used as an indicator of performance during training - either for early stopping (see Sec III.3), or for hyperparameter selection111A hyperparameter is a fixed constant that controls some aspect of an algorithm’s behavior. They can have a significant impact on performance, but cannot be learned like many of the model’s core parameters. As an example, SGD optimizes the parameters of a model, but the initial learning rate and batch size of the optimizer are predetermined hyperparameters.. Final performance will be reported on the test set, which is made up of kinematic bins with which the model has never been trained or evaluated. A model that performs well on the test set is capable of predicting entire φ\varphi curves in unseen regions of kinematic space, under the assumption that we are testing on inputs drawn from the same distribution as our training data. This assumption often breaks down in practice, as we’ll see in Section III.4. Generalizing to unseen kinematic bins is a more difficult challenge than generalizing to held-out datapoints in bins that were part of the training set, because the overall shape of the φ\varphi curve can be quite predictable as long as nearby points can provide a good frame of reference.

Based on the analysis of (Kriesten et al., 2020), we experimented with transforming the dataset of (xBjx_{Bj}, tt, Q2Q^{2}, EbE_{b}, φ\varphi) by concatenating (sin(φ\varphi), cos(φ\varphi), sin(2φ2\varphi), cos(2φ2\varphi), sin(3φ3\varphi), cos(3φ)3\varphi)). These harmonics could serve as additional features that assist the model in learning the cross section’s φ\varphi dependence. We’ll refer to the use of this additional input information as “Harmonic Features.”

We also scale the cross section values by 1000 to mitigate floating-point underflow and stabilize training. Unless otherwise noted, the plots below have been re-scaled to their original units.

III.2 Baselines

First, we examine the possibility of applying basic regression techniques to cross section fits to get an understanding of the difficulty of the problem and establish a performance baseline. We use the implementations provided in Pedregosa et al. (2011):

  • \bullet

    Linear is a simple linear regression model. A vector of weights 𝐖\mathbf{W} and a scalar bias bb map the input kinematics, xx, to the cross section prediction with σ^=𝐖x+b\hat{\sigma}=\mathbf{W}x+b. These parameters are the solution to a closed-form system of equations that minimizes the mean squared error (MSE) of the model’s predictions Kenney and Keeping (1962).

  • \bullet

    SVR is a Support Vector Machine Regression model Vapnik (1999). The model parameters, 𝐖\mathbf{W} and bb are the solution to the optimization problem Chang and Lin (2011):

    min𝐰,b,γ,γ\displaystyle\mathop{min}_{\mathbf{w},b,\gamma,\gamma^{*}} 12𝐰T𝐰+Ci=1lγi+Ci=1lγi\displaystyle\frac{1}{2}\mathbf{w}^{T}\mathbf{w}+C\sum_{i=1}^{l}\gamma_{i}+C\sum_{i=1}^{l}\gamma^{*}_{i}
    subject to 𝐰Tρ(𝐱i)+bσiϵ+γi,\displaystyle\mathbf{w}^{T}\rho(\mathbf{x}_{i})+b-\sigma_{i}\leq\epsilon+\gamma_{i},
    σi𝐰Tρ(𝐱i)bϵ+γ,\displaystyle\sigma_{i}-\mathbf{w}^{T}\rho(\mathbf{x}_{i})-b\leq\epsilon+\gamma^{*},
    γi,γi0,i=1,,l.\displaystyle\gamma_{i},\gamma^{*}_{i}\geq 0,i=1,\dots,l.

    Where CC and ϵ\epsilon are hyperparameters that control error tolerance and ρ\rho is a kernel function that is used to map input vectors to a space where they may be more linearly separable. We experiment with both polynomial and radial basis function kernels.

III.2.1 Evaluating Performance

Because of the large range of cross section targets present in our dataset, common regression metrics like mean squared error are biased towards a small number of kinematic bins. Therefore, we focus on statistics that normalize for the value of the true cross section. The mean absolute percentage error is a better comparison, but suffers from outliers when the model is inaccurate on very small cross sections. Instead, we compute the median absolute percentage error (MAPE). While this is an improvement, the data set has enough noise that relatively accurate fits can have high error. With this in mind, we also compute an accuracy metric, which we define to be the percentage of predictions that are close enough to the target value to be within the experiment’s error bars defined according to Fig. 2.

Accuracy=1NiΔσi\displaystyle\text{Accuracy}=\frac{1}{N}\sum_{i}\Delta_{\sigma_{i}}
Δσi={0,if{σiσ^i>0and|σiσ^i|>δσ^i+δσi,σiσ^i<0and|σiσ^i|>δσ^i+δσi,1,else{σiσ^i>0and|σiσ^i|<δσ^i+δσi,σiσ^i<0and|σiσ^i|<δσ^i+δσi,\displaystyle\Delta_{\sigma_{i}}=\begin{cases}0,\,\,\,\text{if}&\begin{cases}\sigma_{i}-\hat{\sigma}_{i}>0\,\,\,\text{and}\,\,\,|\sigma_{i}-\hat{\sigma}_{i}|>\delta_{\hat{\sigma}_{i}}+\delta_{\sigma_{i,\uparrow}}\\ \sigma_{i}-\hat{\sigma}_{i}<0\,\,\,\text{and}\,\,\,|\sigma_{i}-\hat{\sigma}_{i}|>\delta_{\hat{\sigma}_{i}}+\delta_{\sigma_{i,\downarrow}}\\ \end{cases}\\ \\ 1,\,\,\,\text{else}&\begin{cases}\sigma_{i}-\hat{\sigma}_{i}>0\,\,\,\text{and}\,\,\,|\sigma_{i}-\hat{\sigma}_{i}|<\delta_{\hat{\sigma}_{i}}+\delta_{\sigma_{i,\uparrow}}\\ \sigma_{i}-\hat{\sigma}_{i}<0\,\,\,\text{and}\,\,\,|\sigma_{i}-\hat{\sigma}_{i}|<\delta_{\hat{\sigma}_{i}}+\delta_{\sigma_{i,\downarrow}}\\ \end{cases}\\ \end{cases}
Figure 2: We use an accuracy metric which is defined by the number of points that lie within experimental error bars. Δσi\Delta_{\sigma_{i}} is assigned a value of +1 or 0 based on the difference between the cross section prediction, σ^i\hat{\sigma}_{i}, and the data, σi\sigma_{i}. The assignment is determined by a flowchart of criteria. First we determine whether the difference is positive or negative, indicating whether we are looking at the positive side of the error bar or the negative. Then we compare the absolute value of the difference |σiσ^i||\sigma_{i}-\hat{\sigma}_{i}| to the size of the experimental error, δσ^i\delta_{\hat{\sigma}_{i}}, plus asymmetric systematic errors, δσi,,δσi,\delta_{\sigma_{i,\uparrow}},\delta_{\sigma_{i,\downarrow}}. If the absolute difference is less than the size of the error bar then we assign a +1 value to Δσi\Delta_{\sigma_{i}}, else we assign a 0. The accuracy is then defined as the sum of all Δσi\Delta_{\sigma_{i}} divided by total number of data points.

After a hyperparameter search, the best model is selected according to its MAPE on the validation set; it is then evaluated for the first time on the test set. The results are shown in Table 2.

UU LU
Standard Harmonic Features Standard Harmonic Features
Method Median % Error Accuracy (%) Median % Error Accuracy (%) Median % Error Accuracy (%) Median % Error Accuracy (%)
Linear 238.87 2.10 343.68 1.0 293.03 19.68 333.23 22.34
SVR 45.73 19.14 37.57 27.13 68.09 57.71 72.97 57.98
Theory 13.39 49.50 N/A N/A 62.46 62.67 N/A N/A
FemtoNet 9.99 61.97 N/A N/A 56.06 63.03 N/A N/A
Table 2: Results of baseline methods on the Test set. Each training run uses an identical Train/Val/Test split. We focus on median absolute percentage error because the large range of cross section sizes biases metrics like mean squared error towards a small number of kinematic bins. Accuracy is the % of predictions that lie within experimental error bars. Best results in bold for clarity.

Overall, the baselines perform poorly. While the linear models’ high error is hardly a surprise given the cross section’s trigonometric shape across φ\varphi, the SVR model with access to high degree polynomial features still struggles to find an accurate fit. Results on the use of additional harmonic features are mixed, and we decide against using them in the experiments that follow. In general, high error metrics like these highlight the complexity of the cross section function across the whole range of kinematic bins. Simple approaches are not sufficient for accurate predictions.

III.3 Training a Deep Network on DVCS Data

Refer to caption
Refer to caption
Figure 3: Training curves for FemtoNet LU (top) and UU (bottom). The train loss is computed across batches of the training set while the validation loss is measured at the end of each epoch. When the validation loss has not hit a new minimum in the last 20 epochs, training ends to avoid overfitting.

We train two separate NNs, one to predict σUU\sigma_{UU} and the other to predict σLU\sigma_{LU}. There are several major challenges in applying Deep Learning to this problem. The first is the range of the cross section data values; the target values for our networks’ predictions span seven orders of magnitude in both the polarized and unpolarized cross sections. While the last layer of our architecture uses a linear activation function that can output any scalar, this wide range in targets destabilizes training. When kinematic points with large cross sections are sampled in a batch of training data, they can create outlier loss values that increase the size of our update gradients and make large adjustments to model parameters that decrease performance in regions with a smaller cross section. This effect is magnified by the MSE function, which is quadratic in the absolute error of our models’ predictions. We address this by changing our loss function to the Huber loss:

θH(x,σ)={12(σfθ(x))2,if|σfθ(x)|δ;12|σfθ(x)|,if|σfθ(x)|>δ;\displaystyle\mathcal{L}_{\theta}^{H}(\textbf{x},\sigma)=\begin{cases}\frac{1}{2}(\sigma-f_{\theta}(\textbf{x}))^{2},&\text{if}\left|\sigma-f_{\theta}(\textbf{x})\right|\leq\delta;\\ \\ \frac{1}{2}|\sigma-f_{\theta}(\textbf{x})|,&\text{if}\left|\sigma-f_{\theta}(\textbf{x})\right|>\delta;\\ \end{cases} (8)

The Huber loss is quadratic in the model’s error when that error is less than a hyperparemter δ\delta, and linear otherwise. This limits the effect of large cross section values relative to MSE. The median value of the cross section data is relatively small, so the squared error term is stable for most datapoints and therefore the specific choice of δ\delta is unimportant; we fix δ=1\delta=1. Many data points have error bars that are listed well above 100%100\% of the cross section value. Fitting to the center of such a wide range can create irregularities that generalize poorly and encourage overfitting. Ideally, our model would pay less attention to highly uncertain measurements. Towards this goal, we consider a variation on the standard Huber Loss that incorporates the relative size of the experimental error bars for each datapoint. Let Δσ\Delta\sigma represent the size of the symmetric error bar, while Δσ\Delta\sigma_{\downarrow} and Δσ\Delta\sigma_{\uparrow} represent the negative and positive asymmetric error, respectively. The Error Weighted Huber Loss (EW Huber) is then defined:

θEW(x,σ)=θH(x,σ)1+((2Δσ+Δσ+Δσ)/σ)\displaystyle\mathcal{L}_{\theta}^{EW}(\textbf{x},\sigma)=\frac{\mathcal{L}_{\theta}^{H}(\textbf{x},\sigma)}{1+((2\Delta\sigma+\Delta\sigma_{\downarrow}+\Delta\sigma_{\uparrow})/\sigma)} (9)

The EW Huber discounts the loss value of datapoints where the total size of the error bar is large relative to the scale of the cross section. Points with no experimental error stay unchanged and the effect of the weighting increases with the experimental uncertainty.

In many regions of kinematic space, the unpolarized cross section is dominated by the Bethe-Heitler component, which can be computed as a known function of the network’s inputs. We take advantage of this by implementing the Bethe-Heitler calculation as a custom neural network layer with fixed parameters that cannot be adjusted by gradient descent. This layer computes the Bethe-Heitler component of the input point, which is then multiplied by the output of the standard NN architecture to produce the predicted cross section. This effectively changes the function the NN must learn to approximate from σ\sigma to the simpler σ/BH\sigma/BH. We test these approaches on increasingly large NN sizes to see how they affect performance, and the results are shown in Figure 4. The Huber loss turns out to be essential for successful training. Performance improves with model size - perhaps the biggest advantage of Deep Learning techniques. The BH Ratio simplifies the NN function enough to help small models, but as models grow larger their representational capacity improves and this simplification becomes unnecessary.

Refer to caption
Figure 4: Comparing several of our design choices across varying model sizes. ’MSE’ uses a standard Mean Squared Error loss functions. ‘Huber‘ switches to a clipped variant of MSE that protects against outliers. ‘BH Ratio‘ generates cross sections by multiplying the output of the network by the value of the Bethe Heitler component, which is calculated based on known theory (Kriesten et al., 2020) inside of a custom layer. ‘EW Huber‘ reweights terms in the huber loss based on the size of the experimental error bars - reducing the importance of imprecise datapoints. Gradient clipping is used across all experiments.

III.3.1 Overfitting and Early Stopping

Our models’ goal is to make accurate predictions in previously unseen kinematic bins. By improving their performance on the training set through gradient descent, we tend to improve their performance on the test set. However, over-optimizing on those training points can eventually come at the expense of test set performance. This effect is known as ‘overfitting’ and is an important concern when training deep networks. A simple and effective heuristic to stop training before overfitting occurs is to periodically evaluate the model on the validation set. When the validation loss stops decreasing, we are in danger of overfitting, and end the training run. This approach, known as Early Stopping, uses the validation set as a stand-in for the test set. SGD assumes the training, testing, and validation sets are sampled from the same underlying data distribution. Therefore, we can use the performance on the validation set as a sample of our performance on the testing set, as long as we aren’t making gradient updates on the validation data. We use Early Stopping in every training run, ending training after the validation loss has not decreased in more than 20 full iterations through the training set (epochs).

The size of the dataset is another obstacle. While many of the most successful applications of DL use training sets of millions (Deng et al., 2009) or even billions (Brown et al., 2020) of samples, the DVCS experiments at our disposal leave us with a little under 4,0004,000 datapoints. Small datasets increase the risk of overfitting and often force us to regularize the model, or to restrict it’s ability to move freely in parameter space and arrive at a solution that is hyper-specific to the training set. We accomplish this with Dropout regularization (Srivastava et al., 2014). With Dropout, connections between neurons in the network are severed before each forward pass with some probability pp, which becomes an important hyperparameter. This is typically implemented by sampling binary weight matrices that serve as masks of the network weights - setting them to zero when the connection should be removed and one otherwise. This is shown pictorially in Fig. 6. Intuitively, Dropout reduces overfitting by preventing the network from being overly reliant on the outputs of a small number of neurons. There are many other approaches for regularization. We focus on Dropout because of it’s use in predicting model uncertainty, which will be discussed further in Sec IV.

III.3.2 FemtoNet

The final architectures of our cross section prediction networks are determined with by a large search, guided by the Hyperband algorithm (Li et al., 2018). Hyperband trains many possible variations of the network architecture and uses the results to decide which kinds of networks it should dedicate more time/resources to investigating and which kinds to disregard. The final UU and LU models both use the Adam optimizer (Kingma and Ba, 2014) with learning rate 7.5x1047.5\mathrm{x}10^{-4}. The dropout probability p=.25p=.25. FemtoNet UU uses two hidden layers, both of size 20482048, while FemotoNet LU uses two hidden layers with sizes of 20482048 and 10241024. Both networks use the EW Huber Loss. They also normalize inputs before the first layer, with mean and std statistics determined from the training set.

Refer to caption
Figure 5: A schematic of the FemtoNet model architecture. Two large hidden layers with ReLU activations feed into a linear output layer that determines the cross section prediction. Dropout is applied after each Fully Connected (FC) layer.
Refer to caption
Figure 6: Example of the application of a binary dropout mask to a fully connected neural network. The 1’s and 0’s are assigned based on a probability distribution over each weight. When we re-sample the binary mask over each forward pass of the network, the variance of the output can be used as an approximation of statistical uncertainty of the neural network predictions.

We evaluate FemtoNet on the same test set as the baseline approaches, with the results also listed in Table 2. FemtoNet comfortably outperforms the ML baselines, though the improvement is much more significant in the unpolarized cross section than the polarized data. We discuss this finding further in Sec IV.3. In addition to comparing against other ML techniques, we compare against the theoretical cross section model developed in (Kriesten et al., 2020). The NN is even more accurate over the test set - adding DVCS cross section prediction to a long list of problems in ML in which the performance of hand crafted or manually specified heuristics or formulae can be matched by learned features that are automatically discovered. However, the theory model has the important advantage of consistency across kinematics far outside the region of the training data, while the NN is less constrained.

Refer to caption
Refer to caption
Figure 7: Agreement between the theory-based cross section model and the NN approximation before (Top) and after (Bottom) the addition of generated pseudo-data. This is calculated fixing Q2=2Q^{2}=2 GeV2 plotted as a function of the variables momentum transfer variables ξ\xi and tt. The shaded rectangle represents where current DVCS exclusive data sit at a Q22Q^{2}\approx 2 GeV2.

In the upper plot of Figure 7, we compare the cross section predictions of FemtoNet to the theoretical predictions across a slice of tt and ξ\xi. The region of existing experimental data is indicated by the shaded rectangle. Unsurprisingly, the two agree quite strongly in this area, where both have already been proven to be accurate based on the test set evaluation. However, as we move away from existing data we see a dramatic spike in the difference in the NN’s predictions relative to the theory-based model. Note that this experiment says nothing about the accuracy of the theory’s predictions over these kinematics - there is no experimental data here to serve as the ground-truth. All we can conclude is that the theory and ML significantly disagree. The ability of NNs to generalize outside their training distributions is an active area of research in ML. Deep Learning is often concerned with high dimensional data, such as images or audio files, where any concept of distance between datapoints is essentially meaningless. In contrast, the xBjx_{Bj}, tt, and Q2Q^{2} variables of kinematic space make it easy to think about “how far” our model can generalize, and may make this problem of interest to the broader ML community. Fig 7 suggests that FemtoNet learns a generalizable representation of the tt dependence, but struggles with other kinematic variables.

III.4 Scaling to Larger Datasets

While the cross section measurments used in our analysis consist of around 4,000 datapoints, that total will likely see significant growth from future experiments. We are interested in developing a method that can scale as new measurements are collected. This can be tested by expanding the kinematic region to include a larger range of xBj,tx_{Bj},t, and Q2Q^{2}. In particular, we generate pseudo-data in regions outside the range where data currently exists to measure the network’s compatibility with theoretical calculations of the cross section.

III.4.1 Pseudo data

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 8: Process for creating pseudo-data at the kinematics Q2=1.820GeV2,xBj=0.343,t=0.172GeV2,Eb=5.75GeVQ^{2}=1.820\,\,\text{GeV}^{2},\,\,x_{Bj}=0.343,\,\,t=-0.172\,\,\text{GeV}^{2},\,\,\text{E}_{b}=5.75\,\,\text{GeV}. ( 1st{}^{\text{st}} a) construction of GPD envelopes, (2nd{}^{\text{nd}} b) generation of CFFs from GPD envelopes, (3rd{}^{\text{rd}} c) creation of an envelope of cross section values, (4th{}^{\text{th}} d) final pseudo data with simulated errors.

In Figure 8 we demonstrate the process for generating pseudo-data using several diquark model calculations to parametrize the GPDs. The panels are described in descending order in detail in the text from subsections (a - d).

a) GPD envelopes are generated using two parametric forms following Goldstein et al. (2012) as HVAH_{VA}, EVAE_{VA}, H~VA\widetilde{H}_{VA}, E~VA\widetilde{E}_{VA} and Vanderhaeghen et al. (1999) as HVGGH_{VGG}, EVGGE_{VGG}, H~VGG\widetilde{H}_{VGG}, E~VGG\widetilde{E}_{VGG}. To ensure an unbiased and suitably “random” choice of the parametric form for the GPD we vary the parameters PioP_{i}^{o} of each model by a random fluctuation, Δi\Delta_{i}, of up to ±\pm 10 % of PioP_{i}^{o} dictated by a random number generator. We create the parameters of the model in the following way.

Pi=Pio+ΔiP_{i}=P_{i}^{o}+\Delta_{i} (10)

This is done for both parametric forms of the GPDs. The final envelope is generated by combining the two models in a random combination,

F\displaystyle F =\displaystyle= αFVA+βFVGG\displaystyle\alpha\,F_{VA}+\beta\,F_{VGG} (11)

where F=[H,E,H~,E~]F=[H,E,\widetilde{H},\widetilde{E}], α\alpha is a randomly generated number between 0 and 1, and β=1α\beta=1-\alpha. Using this method, for each kinematic bin we generate 25 GPD curves as shown in the top panel of Fig.8.

b) We evolve the GPDs using pQCD evolution equations and convolute the GPDs with the corresponding Wilson coefficient function to calculate an envelope of 25 real and imaginary parts of the Compton form factors. The imaginary part of the Compton form factor is directly related to the value of the GPD at the cross over point between ERBL and DGLAP regions multiplied by π\pi, therefore it is more susceptible to the random fluctuations of the GPDs. This is in contrast to the real part of the Compton form factor which is constrained by the enveloping Wilson coefficient function and integral over the longitudinal momentum variable xx.

c) The CFF envelope enters into the cross section following the theory discussed in section II above and Kriesten et al. (2020); Kriesten and Liuti (2020). We generate pseudo-data by evaluating the cross section at azimuthal angle, φ\varphi, values starting at 7.5 and going in increments of 15 degrees until we reach 360. For each of the 24 φ\varphi values we generate 25 cross section points using each of the 25 generated CFFs. This creates an envelope of cross section calculations for each kinematic bin. The spread of the cross section values can be seen to be suitable enough to be random, but it is approximately centered around the actual data values.

d) For each φ\varphi value of the data, we randomly pick one data point from the 25 in the cross section envelope. The random choosing of the data also simulates some of the statistical fluctuations naturally present in the experimental data. We generate the pseudo-data errors by comparing to the data set in the regions where there is data and controlling the error to match the relative size of the cross section errors. Using pseudo data, we have control over the size of the error bars giving us a unique opportunity to study the effects of experimental error when using a neural network analysis. We then use this metric to extend into regions where there is no data, simulating approximate errors in regions unexplored by experiment. The coverage of our pseudo-dataset relative to existing experimental data is shown in Figure 9.

Refer to caption
Refer to caption
Figure 9: Kinematic regions in (Top) xBjx_{Bj} versus Q2Q^{2} and (Bottom) xBjx_{Bj} versus tt where we generate pseudo-data compared to where current DVCS data exists.

III.4.2 Evaluating Performance on the Pseudo-dataset

We evaluate our Deep Learning approach on an expanded pseudo-dataset of more than 30,00030,000 unpolarized cross section measurements. One of the most important advantages of NNs is their scalability; it’s not uncommon for hard problems in ML to be solved by simply scaling up a model’s parameter count. This is certainly an option when fitting cross sections. However, for the sake of comparison, we use the same training process and model architecture developed for the experimental datasets. The results are shown in Table 3. Our NN pipeline adapts well, highlighting its ability to learn an accurate approximation of the cross section across a large region of kinematic space where it’s more difficult to rely on overfitting or ‘memorization.’

In the bottom of Fig. 7, we investigate the pseudo-data’s impact on the shape of the NN’s approximation relative to the theory-based model described in (Kriesten et al., 2020). As discussed in Sec III.3, the NN can diverge significantly from the theoretical model when predicting on kinematic bins that are far from existing measurements. When given access to cross sections in more distant regions of kinematic space (as part of the pseudo-dataset), this problem vanishes, underscoring the importance of collecting DVCS measurements across a wide range of kinematics in future experiments.

Median % Error Accuracy (%)
Linear 3090.56 1.50
SVR 42.46 47.75
FemtoNet 12.50 73.90
Table 3: Pseudo-dataset results. NN approaches scale well to a dataset of more than 30,00030,000 samples.

IV Predicting the Cross Section

IV.1 Estimating Model Uncertainty During Inference

When making predictions that have an effect on decisions or later calculations, we’d like to know the confidence level of our model - how sure are we that our predictions are accurate? This confidence is quantified as model uncertainty. Recent literature (see Refs.Kumericki et al. (2011); Moutarde et al. (2019) for DVCS, Ball et al. (2017, 2012) and references therein for inclusive scattering) has adopted the “replica method” as a means of estimating uncertainty. The replica method is similar to the Bootstrap aggregating (Breiman, 1996) approach that is common in ML, in which a set of models is trained by sampling from the dataset with replacement and using the disagreement of their predictions to estimate uncertainty. The replica method consists of running the learning algorithm many times on versions of the dataset where points are randomly perturbed relative to their experimental error bars. This is meant to propagate the experimental error into the results of the model. After many cycles of resampling and retraining, we have a set of models trained on slightly different datasets. The mean output of this set of models is used as the prediction, and the standard deviation serves as an approximation of uncertainty. While this is a valid way to incorporate experimental/epistemic uncertainty into the modeling process, the repeated training can be prohibitively expensive when working with large models or datasets. Using multiple copies of the model parameters can be memory-intensive, difficult to vectorize, and less computationally efficient when making predictions. Ideally, we could recover model uncertainty from a single neural network.

One possibility is a Bayesian Neural Network, where we optimize a distribution over weights and sample from that distribution between each prediction Blundell et al. (2015). While effective, these networks have significantly more parameters and can be more difficult to train. A simple alternative is to leave dropout turned on during inference. By resampling our binary mask over the weights between each forward pass, we create a prediction distribution that can be used to approximate model uncertainty. This technique can be shown Gal and Ghahramani (2015) to be an approximation of a Gaussian Process. Our model generates a batch of kk predictions over the same datapoint, each with different dropout masks. Unless otherwise noted, we set k=1000k=1000. This smooths the uncertainty bands for plotting, but this value can be decreased if inference speed is a priority.

IV.2 Learning the φ\varphi dependence

First, we examine the model’s ability to fit the range of φ\varphi values in existing kinematic bins (xBjx_{Bj}, tt, Q2Q^{2}, EbE_{b}). FemtoNet’s predictions in unseen kinematic bins from the test set are shown in Figure 10. The models learn to predict the cross section within experimental error bars, and are robust to noise and outliers, particularly in the Hall B datasets.

Refer to caption Refer to caption
Refer to caption Refer to caption
Figure 10: The φ\varphi dependence of σLU\sigma_{LU} (top row), and σUU\sigma_{UU} (bottom row) in previously unseen kinematic bins. Shading indicates two standard deviations of 1000 predictions.

IV.3 Unpolarized vs. Polarized Results

The total unpolarized cross section at leading twist contains 3 pieces as shown in Eq.(2): a piece that is bilinear in Compton form factors (DVCS), an interference piece that is linear in Compton form factors (\mathcal{I}), and a piece that has no dependence on Compton form factors and is an exactly calculated background parametrized by the elastic form factors (BH). The Bethe-Heitler cross section term, in most kinematic regimes, dominates the total cross section; often an order of magnitude or larger than the interference and DVCS terms combined. We can calculate the background to a well known degree of accuracy using covariant four vector products. The error and the noise of the cross section is thus due entirely to the terms which are parametrized by the Compton form factors. Since the background dominates the cross section, it then also dominates the noise inherent in scattering cross section data, effectively reducing the errors and simplifying the functional form. It is demonstrated in Table 2 that the neural network can predict the unpolarized cross section with a relatively small median percentage discrepancy.

This is in contrast to the polarized cross section which has a drastically different result. The functional form of the polarized cross section we find is more difficult for the neural network to learn. We can explain this by the presence of a large contribution of noise in the polarized data. By parity constraints, the longitudinally polarized beam on an unpolarized target (LU) cross section has no Bethe-Heitler background nor a term bilinear in Compton form factors at leading twist. It is linear in Compton form factors; as such, there is no error / noise suppression as in the unpolarized case.

We can compare the LU polarized cross section to an equivalent reduced unpolarized cross section where we have subtracted off the Bethe-Heitler background to investigate the impact of the dominant term on the errors as shown in Figure 11. Notice that the unpolarized cross section tends to converge to a relative error similar to that of the polarized cross section at similar kinematics. Where the relative error is defined as

δ¯σ=1Niδσiσi\displaystyle\overline{\delta}_{\sigma}=\frac{1}{N}\sum_{i}\frac{\delta_{\sigma_{i}}}{\sigma_{i}} (12)
Refer to caption
Refer to caption
Figure 11: (Top) The total polarized cross section which at leading twist is linearly dependent on Compton form factors. (Bottom) The reduced unpolarized cross section which we define as the total cross section subtracting the Bethe-Heitler cross section. The reduced cross section has 2 terms both dependent on Compton form factors. Notice that the relative errors of these two converge without the presence of a background process in the unpolarized case.

The background in the unpolarized cross section tends to simplify the functional form of the cross section while simultaneously smoothing out the noise and decreasing relative error. This, however, presents a problem for extracting Compton form factors where the cross section terms they parametrize tend to have large noise components.

IV.4 Predictions in New Kinematic Regions

We aren’t restricted to kinematic bins that have already been measured - the model learns to extrapolate between existing bins. We can sweep through regions where data does not exist, and use the model’s uncertainty estimate to advise our use of its predictions. In Figure 12 we illustrate this point for two particular kinematic values.

This result is consistent with what reported in the upper panel of Fig.7 where the model’s uncertainty in the experimental data domain is very low. Notice that the neural network seems to be able to predict the tt dependence much more accurately than the xixi dependence. This result could mean, on one side, that the tt dependence has a simpler and consistent functional form throughout the phase space than the ξ\xi dependence. The latter will, in fact, be affected by perturbative QCD evolution. On the other hand, the kinematic spread of the data for the tt variable covers a wider section than for ξ\xi: this feature might also influence the outcome. Clearly, the availability of more data in a wider range in ξ\xi for fixed tt will help interpreting this result.

Refer to caption
Refer to caption
Figure 12: Predictions across a range of ξ\xi values. ξ\xi is converted to xBjx_{Bj} before being sent to the model. The blue dots in the figure represent the experimental data. The uncertainty in the curves was discussed in Section IV.1. Notice that for ξ<0.1\xi<0.1 the cross section is evaluated in a kinematically unphysical region where the neural network can still make predictions.

While the dataset contains thousands of datapoints, it contains significantly fewer unique kinematic bins; most of the individual points come from kinematic sweeps in the azimuthal angle φ\varphi. The impact of this can be seen empirically in the model’s uncertainty estimates. The uncertainty grows quickly as we move between different kinematic regions, while it remained relatively constant between different φ\varphi values in a fixed region.

V Conclusions and outlook

In summary, a deep learning neural network analysis is essential to extract information from deeply virtual exclusive scattering data in a model independent way. These reactions are characterized by final states with a large number of kinematic variables and observables, and a consequent growth in complexity of the mapping between the initial variables into the output states.

Our analysis is a necessary first step in a future global extraction of Compton form factors from data. The extraction of information from the data requires an understanding of the features that directly impact such an analysis. In our case, the significant contribution of noise and the relatively small kinematics spread of the data made predicting the cross section a difficult problem even using a deep neural network. Nevertheless, as shown in Table 2, an analysis using a DNN was proven the most efficient and accurate method, as compared to other baselines, for extracting information. Several additional challenges in our analysis stem from the current lack of an extensive centralized experimental database for exclusive scattering data. Another difficulty with this data set is the mismatch in the quantity of data between the φ\varphi dependence and the rest of the kinematic variables (tt, Q2Q^{2}, and xBjx_{Bj}). This results in the neural network was not generalizing well across kinematic space. Such challenges can be curtailed through the establishment of a common data set with bench-marking requirements for the publishing of data. This would optimize the usage of data from different experiments in global analyses.

Our results from FemtoNet suggest many interesting characteristics of DVCS data. In particular, we have shown that there is a substantial and consistent signal in the unpolarized cross section. This is at variance with the polarized cross section where the median absolute percentage error is much larger (see Table 2). Furthermore, our results extrapolate to kinematics outside of the range covered in experiment: the DNN seems to be able to generalize in the kinematic variable tt more precisely and for a wider range than for the other variables ξ\xi and Q2Q^{2}. This suggests, one side, interesting implications for Fourier transforms for 3D visualization from DVCS data. On the other hand, the model’s apparent inability to extrapolate in ξ\xi could imply an inconsistent function for ξ\xi across the kinematic phase space.

We also studied the question of how to use experimental errors in neural networks by incorporating them in the error weighted Huber loss (EW Huber, Eq.(9)), and also in our accuracy metric (Fig.2). For this purpose we used pseudo data. With pseudo data we have control over the error bars and can establish relationships between experimental errors and the uncertainty on the extracted Compton form factors.

In future studies we will, on one side, extend our approach to analyze other exclusive experiments at Jlab@12 GeV, as well as at the Electron Ion Collider (EIC), while also addressing remaining open questions. We plan to further study both the accuracy and the uncertainty obtained from the DNN analysis of DVCS data. In particular, we will analyze the contribution of the background Bethe-Heitler process in the unpolarized cross section that seems to dominate the accuracy of the predictions with respect to the polarized electron scattering. The framework provided in this paper to treat the experimental error in ML affords us the ability to next study the effect of the data errors on the uncertainty of the extracted Compton form factors. This will, in turn, allow us to optimize both the precision and the kinematic domain for experimental measurements at future colliders needed to efficiently disentangle the observables from the data.

Finally, our method is flexible to be extended to incorporate lattice QCD results in the data analysis. The latter will play an even more important role in future exclusive experiments analyses than in deep inelastic scattering, towards the ultimate extraction of GPDs from data.

Acknowledgements.
This work was funded by DOE grant DE-SC0016286 and in part by the DOE Topical Collaboration on TMDs (B.K. and S.L.), DOE grant DE-FG03-95ER40965 (M.B.) and SURA grant C2019-FEMT-002-04 and C2020-FEMT-006-05. We gratefully acknowledge the University of Virginia School of Data Science for logistical support; technical assistance by Bryan Wright and Rick Marshall (UVA, Physics), and many insightful physics discussions with Gordon Cates.

References

  • Ohnishi et al. (2019) A. Ohnishi, Y. Mori,  and K. Kashiwa, JPS Conf. Proc. 26, 024011 (2019).
  • Mori et al. (2017) Y. Mori, K. Kashiwa,  and A. Ohnishi, Phys. Rev. D 96, 111501 (2017)arXiv:1705.05605 [hep-lat] .
  • Alexandru et al. (2018) A. Alexandru, P. F. Bedaque, H. Lamm,  and S. Lawrence, Phys. Rev. D 97, 094510 (2018)arXiv:1804.00697 [hep-lat] .
  • Bedaque et al. (2020) P. Bedaque et al.,   (2020), arXiv:2006.05422 [physics.comp-ph] .
  • Goeke et al. (2001) K. Goeke, M. V. Polyakov,  and M. Vanderhaeghen, Prog. Part. Nucl. Phys. 47, 401 (2001)arXiv:hep-ph/0106012 [hep-ph] .
  • Diehl (2001) M. Diehl, Eur.Phys.J. C19, 485 (2001)arXiv:hep-ph/0101335 [hep-ph] .
  • Belitsky and Radyushkin (2005) A. Belitsky and A. Radyushkin, Phys.Rept. 418, 1 (2005)arXiv:hep-ph/0504030 [hep-ph] .
  • Kumericki et al. (2016) K. Kumericki, S. Liuti,  and H. Moutarde,   (2016), arXiv:1602.02763 [hep-ph] .
  • Kriesten et al. (2020) B. Kriesten, S. Liuti, L. Calero-Diaz, D. Keller, A. Meyer, G. R. Goldstein,  and J. Osvaldo Gonzalez-Hernandez, Phys. Rev. D 101, 054021 (2020)arXiv:1903.05742 [hep-ph] .
  • Group (2020) F. Group, “Nuclear Femtography Group,” https://pages.shanti.virginia.edu/Femtography/ (2020).
  • Kingma and Ba (2014) D. P. Kingma and J. Ba, arXiv preprint arXiv:1412.6980  (2014).
  • (12) G. Hinton, N. Srivastava,  and K. Swersky,  .
  • Georges (2018) F. Georges, Deeply virtual Compton scattering at Jefferson Lab, Ph.D. thesis, Institut de Physique Nucléaire d’Orsay, France (2018).
  • Defurne et al. (2015) M. Defurne et al. (Jefferson Lab Hall A), Phys. Rev. C92, 055202 (2015)arXiv:1504.05453 [nucl-ex] .
  • Defurne et al. (2017) M. Defurne et al.Nature Commun. 8, 1408 (2017)arXiv:1703.09442 [hep-ex] .
  • Jo (2012) H.-S. Jo, PoS QNP2012, 052 (2012)arXiv:1207.3709 [nucl-ex] .
  • Pedregosa et al. (2011) F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot,  and E. Duchesnay, Journal of Machine Learning Research 12, 2825 (2011).
  • Kenney and Keeping (1962) J. F. Kenney and E. Keeping, Mathematics of statistics 1, 252 (1962).
  • Vapnik (1999) V. N. Vapnik, IEEE transactions on neural networks 10, 988 (1999).
  • Chang and Lin (2011) C.-C. Chang and C.-J. Lin, ACM transactions on intelligent systems and technology (TIST) 2, 1 (2011).
  • Deng et al. (2009) J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li,  and L. Fei-Fei, in CVPR09 (2009).
  • Brown et al. (2020) T. B. Brown, B. Mann, N. Ryder, M. Subbiah, J. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam, G. Sastry, A. Askell, S. Agarwal, A. Herbert-Voss, G. Krueger, T. Henighan, R. Child, A. Ramesh, D. M. Ziegler, J. Wu, C. Winter, C. Hesse, M. Chen, E. Sigler, M. Litwin, S. Gray, B. Chess, J. Clark, C. Berner, S. McCandlish, A. Radford, I. Sutskever,  and D. Amodei, “Language models are few-shot learners,”  (2020), arXiv:2005.14165 [cs.CL] .
  • Srivastava et al. (2014) N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever,  and R. Salakhutdinov, The journal of machine learning research 15, 1929 (2014).
  • Li et al. (2018) L. Li, K. Jamieson, G. DeSalvo, A. Rostamizadeh,  and A. Talwalkar, “Hyperband: A novel bandit-based approach to hyperparameter optimization,”  (2018), arXiv:1603.06560 [cs.LG] .
  • Goldstein et al. (2012) G. R. Goldstein, J. O. G. Hernandez,  and S. Liuti, J.Phys. G39, 115001 (2012)arXiv:1201.6088 [hep-ph] .
  • Vanderhaeghen et al. (1999) M. Vanderhaeghen, P. A. M. Guichon,  and M. Guidal, Phys. Rev. D 60, 094017 (1999).
  • Kriesten and Liuti (2020) B. Kriesten and S. Liuti,   (2020), arXiv:2004.08890 [hep-ph] .
  • Kumericki et al. (2011) K. Kumericki, D. Mueller,  and A. Schafer, “Studying deeply virtual compton scattering with neural networks,”  (2011), arXiv:1110.3798 [hep-ph] .
  • Moutarde et al. (2019) H. Moutarde, P. Sznajder,  and J. Wagner, The European Physical Journal C 79, 614 (2019).
  • Ball et al. (2017) R. D. Ball et al. (NNPDF), Eur. Phys. J. C 77, 663 (2017)arXiv:1706.00428 [hep-ph] .
  • Ball et al. (2012) R. D. Ball, V. Bertone, F. Cerutti, L. Del Debbio, S. Forte, A. Guffanti, J. I. Latorre, J. Rojo,  and M. Ubiali (NNPDF), Nucl. Phys. B 855, 153 (2012)arXiv:1107.2652 [hep-ph] .
  • Breiman (1996) L. Breiman, Machine learning 24, 123 (1996).
  • Blundell et al. (2015) C. Blundell, J. Cornebise, K. Kavukcuoglu,  and D. Wierstra (PMLR, Lille, France, 2015) pp. 1613–1622.
  • Gal and Ghahramani (2015) Y. Gal and Z. Ghahramani, “Dropout as a bayesian approximation: Representing model uncertainty in deep learning,”  (2015), arXiv:1506.02142 [stat.ML] .