Understanding and visualizing the statistical analysis of SN1987A neutrino data
Abstract
The SN1987A detection through neutrinos was an event of great importance in neutrino physics, being the first detection of neutrinos created outside our solar system, and then inaugurating the era of experimental neutrino astronomy. The data have been largely studied in many different analysis, and has presented several challenges in different aspects, since both supernova explosion dynamics and neutrino flavour conversion in such extreme environment still have many unknowns. In addition, the low statistics also invoke the need of unbinned statistical methods to compare any model proposal with data. In this paper we focus on a discussion about the most used statistical analysis interpretation, presenting a pedagogical way to understand and visualize this comparison.
Keywords: supernova; neutrino.
1 Introduction
Particle physics provides a fertile ground to a vast number of methods to statistically compare theory and data, giving a quantitative filling in order to guide the prospects of a theory, or even reveal important accomplishments or tensions in experimental efforts.
In general, the theory can visually (and intuitively) be compared to data for most of the methods, but one interesting phenomena, the neutrino burst detected from the supernova SN1987A, that is particularly affected by low statistics, evidence difficulties to such parallel.
In this paper we propose a visual understanding to handle this difficulty through an animation of a parameterized model and SN1987A antineutrino data.
2 Statistical Analysis
The quantitative comparison between theoretical predictions and experimental results is a major part of any scientific endeavour. As in many other areas, in particle physics an important quantity around such comparison is made is the detection rate of a specific event. As an example, theoretical models of solar neutrino production provide us with a steady theoretical neutrino flux. Solar neutrino experiments provides us with a detected event rate. The comparison between these two quantities can be done by translating the theoretical flux in an expected event rate, or inversely, translating the detected event rate in a compatible expected flux.
Besides, a lot of information can be extracted from the dependence of neutrino flux with its energy or time of detection. The most straightforward way to include this information on statistical analysis would be to split the total data into bins of specific energy or time intervals. Maybe the most spread statistical tool to implement these kind of analysis would be to calculate the following :
(1) |
where the indices and track the binning of the data, is the theoretical prediction and is the experimental data. This analysis has one great advantage: it allows to visually grasp how good is the accordance between theory and experiment in a figure where experimental data points, uncertainties and theoretical predictions can be plotted together. If theoretical predictions are contained inside the region around the experimental data points delimited by the uncertainties, than we expect a good fit.
Several examples in neutrino physics can illustrate such procedure. Again taking solar neutrinos as an example, the data presented by the neutrino detector Super-Kamiokande is divided both in energy and time bins. In Figure 1 the data of 1496 days of running experiment are presented, and the binning on energy and time can be seen, although the binning in time is indirect, through a binning in the position of the Sun when the data was taken. The continuous line represent the prediction for the best fit point of the statistical analysis when flavour conversion is considered in two scenarios, the LMA and the LOW solutions.

As discussed, it is possible to visually get a feeling about how good the accordance between experimental data and theoretical predictions by how the theoretical curves cross the region around the experimental data within the error bars. For this particular example, two solutions to the solar neutrino problem are presented, with large mixing angle and large (LMA) and lower (LOW). We can expect by visually inspecting the figure that both solutions would fit the data quite well, fact that is confirmed by a more careful statistical analysis.
The problem of this visualization is when there is no efficient way to collect the data to form bins, for instance, due to the low event rate. In particular, when the event rate is very low it is necessary to take the experimental data event by event.
In this context, what we propose here is to recover a way to visually access how good the accordance between experimental data and theoretical prediction in a particular scenario when the statistical analysis is done event by event: the neutrino data from Supernova 1987A.
3 Supernova 1987A
The Core Collapse Supernova is a remarkable end of life of a star and one of the most peculiar astrophysical phenomena. Despite being a prominent optical phenomena, its most outstanding property is the powerful release of of gravitational binding energy, generally in the order of erg, from a progenitor star in (anti)neutrinos of all flavors in MeV scale.
Nevertheless the high neutrino luminosity, a limitation in the neutrino observation on Earth is related to the high distance from the source, with decreasing flux proportional to , restricting the possible region for neutrino burst detection to a galactic or nearby the Milky Way, that possesses a low supernova rate of per century [2].
Even though, in 1987, three detectors, Kamiokande II [3], IMB [4, 5] and Baksan [6], were capable to observe a neutrino signal associated to a SN in the Large Magellanic Cloud ( kpc). These data are presented in Figure 2. In contrast with Super-Kamiokande data in Figure 1, these are individual events, and there is no obvious way to overlap a theoretical curve to them. Since the theoretical models provide a flux density and any kind of binning to convert this density into a event probability would be quite arbitrary, the approach through an unbinned maximum likelihood estimation is a robust alternative to confront the theoretical hypothesis with these individual events. In next section we describe such procedure.

4 Modelling SN1987A event-by-event likelihood
Frequently the likelihood treatment in particle physics involves the usage of Poisson distribution , that fits well to phenomena that has small probability to occur, but a large number of tries. Given a measured variable set , the Poisson likelihood is given by:
(2) |
where can be a particular number of events that occurs in a interval of our variable, in a number of intervals, or bins, and is the expected value in the same interval. It is convenient to write as a distribution function on the variable , or , with a given events rate in an equally spaced bin of variable and number of counts . Including it in equation (2):
(3) | |||||
However, binning the data to use a single expected value of a set of points requires to assume a given statistical distribution of such a bin, that generally is assumed to be Gaussian for higher number of entries. The low statistics scenario does not allow this assumption, then it is possible to model the likelihood (3) to account for each event apart. This can be made by taking the bin to an infinitesimal width and number of counts , so we consider only infinitesimal bins with one event and drop the others, then (3) becomes
(4) |
that also has the change from total number of bins to total number of observed events . The idea behind maximum likelihood is to maximize the quantity in (4), or given the correspondence , minimize the to respect to a free set of parameters . If we have a single event at , this expressions reduces to . For different models with a normalized expected event rate , the likelihood is maximized for the model with the highest value of . And letting the normalization runs freely, it is maximized for . It is straightforward to note that if we consider more than one single event this maximum occurs on the total number of events.
In a supernova detection, such as SN1987A, the variables are the neutrino energy, the detection time and events scattering angle, i.e. . Then eq. (4) becomes:
(5) |
where is a triply differential equation, and is the expected number of events at the detector. For simplicity we did not include the scattering angle dependence on the animations presented in the following, although they were used in the likelihood calculation. A complete analysis, including other details such as background, energy resolution and efficiency can be seen in [7], [8] and [9].

5 Single event distribution
The main ingredient to construct the likelihood is the theoretical triply differential expected rate. However, since there is no way to convert the theoretical predictions into some quantity to be compared with individual events, we can instead modify the events to match the theoretical probability distribution. For instance, all SN1987A events are published with an uncertainty in energy, so the true information we can take from each event is a probability distribution around some most probable result. Assuming such distribution to be Gaussian, a specific event with energy and uncertainty on time with uncertainty is related to the following probability distribution:
This can be compared with the theoretical probability of inducing an event on the detector:
(6) |
where is a proportionality factor that take into account the detector size and efficiency.
To proper visualize the data points being collected, we can create an animation with the detected event probability integrated on time. Since the uncertainty on time is very small, the distribution converges to a -function, and such animation would advance in steps while the data gets collected:
(7) |
Such animation is presented in Figure 3 (red curve). Since what is presented is the cumulative result after integrating on time, the final moment of this animation, when integrated also on energy, provides all the 29 events detected by the three experiments. The comparison with theoretical predictions can be made visually if we produce a similar animation for the expected number of events, integrating Eq. (6) on time, also presented in Figure 3 (dashed curve).
The parameterization of the antineutrino111Once the detectors in 1987 were capable to measure a single channel, the inverse beta decay (), only electron antineutrinos could be detected. flux in Eq. (6) consists in a two-component emission (accretion + cooling) and nine free parameters, that come from the proposed flux , with , where () is the initial antineutrino (positron) temperature from the cooling (accretion) phase, is the radius of the neutrinosphere, () is the characteristic time from the cooling (accretion) phase and is the initial accreting mass [8]. The remaining three free parameters are a time offset to be adjusted independently for each detector.
These parameters are estimated from an event-by-event maximum likelihood, and the best fit values of our analysis, used in Figure 3, are:
(8) | |||
(9) |
As described before, the maximization on the likelihood depends on two terms. The term in the exponential factor factor is related to the number of events, and drives the theoretical parameters to those who provides the right expected number of events, i.e., the area under the curves in the end of the animation in Figure 3. It is quite easy to grasp if our theoretical model fits well the data by this aspect.
The second term access how close the theoretical curve is to the experimental one at the data central points, both in energy and in time. Since the uncertainty in time is negligible, we can visually compare the curves at the moments a new data is collected, providing us with a visual tool to this second ingredient of the statistical analysis. By performing these two analysis on Figure 3, we can expect that, although not perfect, the theoretical prediction would provide a reasonably good fit to the data.

It is useful now to analyse a set of theoretical parameters that do not fit well the data. This is done in Fig 4, where we chosed two set of parameters that are excluded at 90% C.L. according to our analysis. These parameters were chosen in a way to not change the total number of predicted events, so we can focus on the energy spectrum information. It is clear, again using a visual comparison, that these new set of parameters produce a worse fit to the data, fact that is confirmed by a full statistical analysis.
As it was pointed out earlier, the two main neutrino observables that we are taking into consideration are the neutrino energy and the time of detection. After discussing the first on the above analysis, we will now focus on the second, and the best way to do this is the limits in neutrino mass that can be achieved using this statistical method.
6 Neutrino mass limits
An important remark is that the neutrino detection spread in time is an important source of physical information, allowing us to probe both Supernova explosion mechanisms and neutrino properties. The most important neutrino property that can be probed by such time spread is its mass.
The first difficulty in these kind of analysis is that the data itself does not allow us to correlate the time of arrival of the neutrino burst at the detectors with the unknown time at which the neutrinos left the Supernova. The solution is to use the data itself to establish, through statistical analysis, the match between the neutrino flux theoretical prediction and the data, taking the time of arrival of the first neutrino event in each detector as a marker. The time of the following events, are taken as relative ones to the time of arrival of the first event, :
and is left to vary freely to best match the theoretical prediction in a previously established time scale.
This simple picture arises when we assume massless neutrinos. In this case the relative time between events is identical to the relative time between the emission of these detected neutrinos on the production site, since the time delay due to the travel between the supernova and the detectors does not depend on the neutrino properties. In Fig. 3 it is assumed a vanishing mass neutrino, and the time showed on the animation correspond to the time since the supernova offset.

However, since the neutrinos have mass, neutrinos with different energies have different velocities, which changes the described scenario. More energetic neutrinos travels faster then less energetic neutrinos, meaning that the relative times between events does not correspond to relative times of the neutrinos emission. The correction is done by a simple kinematic analysis:
where is the distance to the supernova, and and are the neutrino mass and event energy. The sub-index () refers to the time at production (detection). The emission time of each event is then calculated from the relative times , and the kinematic corrections:
(10) |
Instead of making the correction on the time of the production, presented here to give proper credit to the authors that proposed and performed this analysis, we prefer to correct the theoretical predictions by continuous spread in time on the neutrino flux spectrum at the detector. So, instead of converting the time of the detected events to the supernova emission, we adjust the theoretical prediction to the detector site. Clearly both choices are equivalent, but with this second procedure we can use the same data animation presented in Fig. 3, and adjust the theoretical curve by making the replacement:
in Eq. (6).
An animation evidencing this model dependent limit is shown in Fig. 5, for a eV neutrino mass, and the same astrophysical parameters used to produce Fig. 3, and then with the same neutrino flux at the source. But due to the different time lag of neutrinos traveling to Earth with different energies, the time history of the expected number of events changes significantly, allowing us to place a limit on neutrino analysis using a proper statistical analysis.
7 Conclusion
This paper intended to present a pedagogical view of how to understand the likelihood analysis when an event-by-event treatment is necessary. The detection of SN1987A is a perfect example for that, once a lot of physics can be extracted by the few events that were collected through neutrino detection. It also has the interesting feature that different information can be extracted from the total expected number of events, its spectral distortion or its time structure. We present some animations as a visual tool to understand the statistical procedure, and produce a first impression on how different models fit the data.
Acknowledgements
This study was financed in part by the Coordenação de Aperfeiçoamento de Pessoal de Nível Superior - Brasil (CAPES) - Finance Code 001. The authors are also thankful for the support of FAPESP funding Grant 2014/19164-6. The authors are thankful to Pedro Dedin for usefull discussions during the production of this article.
References
- [1] M. B. Smy. Solar neutrino precision measurements using all 1496 days of Super-Kamiokande I data. Nucl. Phys. B Proc. Suppl., 118:25–32, 2003.
- [2] Karolina Rozwadowska, Francesco Vissani, and Enrico Cappellaro. On the rate of core collapse supernovae in the milky way. New Astronomy, 83:101498, 2021.
- [3] K. Hirata et al. Observation of a Neutrino Burst from the Supernova SN 1987a. Phys. Rev. Lett., 58:1490–1493, 1987.
- [4] R. M. Bionta et al. Observation of a Neutrino Burst in Coincidence with Supernova SN 1987a in the Large Magellanic Cloud. Phys. Rev. Lett., 58:1494, 1987.
- [5] Todd Haines et al. Neutrinos From SN1987A in the Imb Detector. Nucl. Instrum. Meth. A, 264:28–31, 1988.
- [6] E. N. Alekseev, L. N. Alekseeva, I. V. Krivosheina, and V. I. Volchenko. Detection of the Neutrino Signal From SN1987A in the LMC Using the Inr Baksan Underground Scintillation Telescope. Phys. Lett. B, 205:209–214, 1988.
- [7] A Ianni, G Pagliaroli, Alessandro Strumia, FR Torres, FL Villante, and F Vissani. Likelihood for supernova neutrino analyses. Physical Review D, 80(4):043007, 2009.
- [8] Giulia Pagliaroli, F Vissani, ML Costantini, and A Ianni. Improved analysis of sn1987a antineutrino events. Astroparticle Physics, 31(3):163–176, 2009.
- [9] Beat Jegerlehner, Frank Neubig, and Georg Raffelt. Neutrino oscillations and the supernova 1987a signal. Physical Review D, 54(1):1194, 1996.
- [10] G. Pagliaroli, F. Rossi-Torres, and F. Vissani. Neutrino mass bound in the standard scenario for supernova electronic antineutrino emission. Astropart. Phys., 33:287–291, 2010.