This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

From masses and radii of neutron stars to EOS of nuclear matter through neural network

Zehan Wu School of Physics and Optoelectronics, South China University of Technology, Guangzhou 510641, P.R. China    Dehua Wen111Corresponding author. [email protected] School of Physics and Optoelectronics, South China University of Technology, Guangzhou 510641, P.R. China
Abstract

The equation of state (EOS) of dense nuclear matter is a key factor to determine the internal structure and properties of neutron stars. However, the EOS of high-density nuclear matter has great uncertainty mainly because the terrestrial nuclear experiments cannot reproduce matter as dense as that in the inner core of a neutron star. Fortunately, continuous improvements in astronomical observations of neutron stars provide the opportunity to inversely constrain the EOS of high-density nuclear matter. A number of methods have been proposed to implement this inverse constraint, such as the Bayesian analysis algorithm, the Lindblom’s approach, and so on. Neural network algorithm is an effective new method developed in recent years. By employing a set of isospin-dependent parametric EOSs as the training sample of neural network algorithm, we set up an effective way to reconstruct the EOS with relative accuracy through a few mass-radius data. Based on the obtained neural network algorithms and according to the NICER observations on masses and radii of neutron stars with assumed precision, we get the inversely constrained EOS and further calculate the corresponding macroscopic properties of the neutron star. The results are basically consistent with the constraint on EOS from the Huth etal.et~{}al. based on Bayesian analysis. Moreover, the results show that even though the neural network algorithm was obtained by using the finite parameterized EOS as the training set, it is valid for any rational parameter combination of the parameterized EOS model.

pacs:
97.60.Jd; 04.40.Dg; 04.30.-w; 95.30.Sf

I Introduction

Understanding the nature of neutron stars and determining the equation of state (EOS) of high-density asymmetric nuclear matter are both long-standing problems in nuclear physics and astrophysics Chen1 ; Li25 ; Li26 ; Li2019EPJA . The EOS of asymmetric nuclear matter is a key input for probing the structure and properties of neutron stars. However, the theoretical predictions on the EOS at supra-saturation densities diverge broadly. Except possible phase transitions, density dependence of the symmetry energy is the most uncertain part of the EOS, especially at supra-saturation densities Dieperink03PRC ; Li26 ; Li2019EPJA . Although significant progresses have been made in probing the symmetry energy in terrestrial nuclear laboratories Russotto2 ; LeFevre3 ; Adhikari4 , the inability to reproduce the authentic high-density neutron-rich nuclear matter restricts the full and accurate understanding of symmetry energy and EOS at supra-saturation densities.

Fortunately, many groundbreaking observational discoveries from scientific facilities in recent years have led to a new upsurge in understanding the neutron stars and EOS of high-density asymmetric nuclear matter. For example, the mass-radius measurement on neutron stars from the Neutron Star Interior Composition Explorer (NICER) and the gravitational radiation detection of binary neutron star mergers from LIGO and Virgo have taken the understanding of neutron stars and EOS of the asymmetric nuclear matter to a new level Miller5 ; Riley6 ; Pang7 ; Abbott8 . In the near future, the Large Observatory for X-ray Timing (LOFT), the Advanced Telescope for High Energy Astrophysics (ATHENA), the Square Kilometre Array (SKA) and Einstein Telescope (ET) will also bring more observations of neutron stars when they become operational Feroci40 ; Nandra41 ; Punturo42 ; Dewdney43 ; Rezzolla44 . It is believed that they will provide powerful insights in understanding the neutron stars and the EOS.

If we regard obtaining the properties of neutron stars from the EOS of nuclear matter as forward method, which normally can be numerically calculated according to the EOS and TOV equations for static spherically symmetrical stars. In combination with the inverse derivation from the mass-radius to EOS given in Ref. Lindblom9 using enthalpy, there is a one-to-one mapping between EOS and the mass-radius relation of neutron stars. That is, if a set of accurate mass-radius observations of neutron stars is obtained, the EOS of nuclear matter can be reversely achieved. But how can we reversely obtain the EOS from the properties of neutron stars? At present, there is no reliable and widely accepted method. The purpose of this work is to try to find an effective way reversely mapping the EOS of nuclear matter through the mass-radius relations of neutron stars.

Lindblom had pioneered the inverse TOV mapping to obtain the pressure-density sequence P(ρ\rho) of the EOS from the accurate mass-radius sequence of a neutron star Lindblom9 . The reality is that the accurate observation data of the mass-radius of the neutron star is still very rare. As more advanced algorithms are introduced, a small amount of observation data can provide effective constraint on the EOS. For example, Ref. XuJPRC21 implemented constraints on EOS-related parameters within the framework of Bayesian analysis. Bayesian analysis is an algorithm for updating information by correcting existing information based on observed events. The use of this algorithm therefore requires ensuring both the correctness of the prior probabilities and the reasonableness of the observed events. This year, Huth etal.et~{}al. also used this algorithm to achieve constraints on the EOS by combining data from heavy-ion collisions (HIC) with results from multi-messenger astronomical observations Huth11 . Furthermore, a substantial amount of Bayesian analysis work has been contributed to neutron star research YX12 ; Silva13 ; Xie14 .

Neural networks (NN) is a branch of machine learning (ML) and a computing system inspired by biological neural networks Amari17 . It has the advantages of Bayesian analysis algorithm and polynomial fitting algorithm. Only a few less precise observations are needed to obtain the effective constraint through NN, which can also be used to calculate other properties of neutron stars. In the 1990s, NN was pioneered by Clark and collaborators for research related to nuclear physics Gazula92 ; Gazula93 ; Gazula95 ; Gazula99 . This algorithm is now widely used in theoretical prediction work to improve the accuracy of nuclear mass and nuclear charge radii measurements ZhaoNPA22 ; Niu18 ; Utama16PRC ; Utama16JPG . In the field of nuclear astrophysics, this algorithms have also been introduced into neutron star research. For example, the inverse TOV mapping of NN was constructed by using the EOS with sound velocity segmentation in Ref. Fujimoto16 , the reconstructed EOS was implemented with NN at 1 to 7 times the saturated nuclear density ρ00.16fm3\rho_{0}\approx 0.16\ \mathrm{fm}^{-3} in Ref. Soma22AX , Krastev constructed NNs from mass-radius and mass-tidal deformations to symmetry energy Krastev15 , and NN from nuclear parameters to neutron star properties was constructed in Ref. Thete22AX . Additionally, the mapping from X-ray spectra to EOS parameters was achieved by using NN in Ref. Farrell22AX .

Although much of the NN work has given exciting results, many questions remain unsolved. For example, the plausibility of the symmetry energy parameters of the EOS after noise addition, the implementation of the NN prediction function and its general applicability, and the validation of the effectiveness. In this work, we will construct a new NN based on the parametric EOS, which can be efficiently implemented for any combination of parameters satisfying terrestrial experiments and multi-messenger astronomical observations with a small number of mass-radius relations to inverse TOV mapping.

This paper is organized as follows. In Sec. II we briefly review the isospin-dependent parametric EOS and the basic formula for calculating the properties of neutron stars. The NN implementation of the inverse TOV mapping is presented in Sec. III. NN combined with NICER observational constraints are presented in Sec. IV. The summary and outlook are in Sec. V.

Unless otherwise stated, in the formulae we use the gravitational units (G=c=1G=c=1).

II Isospin-dependent parametric EOS and the fundamental formula of neutron stars

II.1 Isospin-dependent parametric EOS

Implementing a NN for inverse TOV mapping requires a large number of the EOS as training samples. Moreover, these EOS also need to meet the range of nuclear parameters as much as possible, such as LL, KsymK_{\mathrm{sym}}, etc. The isospin-dependent parametric EOS is able to satisfy both the need to generate a large number of EOS and the rationalization of nuclear parameters.

For asymmetric nuclear matter, the energy per nucleon can be expanded by isospin asymmetry Li25 ; Li26

E(ρ,δ)E0(ρ)+Esym(ρ)δ2,E\left(\rho,\delta\right)\approx E_{0}\left(\rho\right)+E_{\mathrm{sym}}\left(\rho\right)\delta^{2}, (1)

where δ=(ρnρp)/(ρn+ρp)\delta=\left(\rho_{\mathrm{n}}-\rho_{\mathrm{p}}\right)/\left(\rho_{\mathrm{n}}+\rho_{\mathrm{p}}\right) is the isospin asymmetry. The first term on the right side E0(ρ)E_{0}\left(\rho\right) is usually referred to as the energy of symmetric nuclear matter, and the second term Esym(ρ)E_{\mathrm{sym}}\left(\rho\right) is referred to as the symmetry energy, which has the physical meaning of the energy difference per nucleon between asymmetric nuclear matter and pure neutron matter.

Usually, the two terms on the right side in Eq. (1) can be expanded at the saturation density ρ0\rho_{0} Zhang18APJ ; Xie14

E0(ρ)=E0(ρ0)+K02(ρρ03ρ0)2+J06(ρρ03ρ0)3,\displaystyle E_{0}\left(\rho\right)=E_{0}\left(\rho_{0}\right)+\frac{K_{0}}{2}\left(\frac{\rho-\rho_{0}}{3\rho_{0}}\right)^{2}+\frac{J_{0}}{6}\left(\frac{\rho-\rho_{0}}{3\rho_{0}}\right)^{3}, (2.2)
Esym(ρ)=Esym(ρ0)+L(ρρ03ρ0)+Ksym2(ρρ03ρ0)2+Jsym6(ρρ03ρ0)3.\displaystyle E_{\mathrm{sym}}\left(\rho\right)=E_{\mathrm{sym}}\left(\rho_{0}\right)+L\left(\frac{\rho-\rho_{0}}{3\rho_{0}}\right)+\frac{K_{\mathrm{sym}}}{2}\left(\frac{\rho-\rho_{0}}{3\rho_{0}}\right)^{2}+\frac{J_{\mathrm{sym}}}{6}\left(\frac{\rho-\rho_{0}}{3\rho_{0}}\right)^{3}. (2.3)

Constrained by terrestrial nuclear experiments, the most probable values of the parameters in the above equations are as follows: L=58.7±28.1MeVL=58.7\pm 28.1\ \mathrm{MeV}, 400Ksym100MeV-400\leqslant K_{\mathrm{sym}}\leqslant 100\ \mathrm{MeV}, 200Jsym800MeV-200\leqslant J_{\mathrm{sym}}\leqslant 800\ \mathrm{MeV}, 300J0400MeV-300\leqslant J_{0}\leqslant 400\ \mathrm{MeV}, K0=240±20MeVK_{0}=240\pm 20\ \mathrm{MeV}, Esym(ρ0)=31.7±3.2MeVE_{\mathrm{sym}}\left(\rho_{0}\right)=31.7\pm 3.2\ \mathrm{MeV} Shlomo33 ; Piekarewicz34 ; Zhang35 ; Oertel36 ; Li37 . It is worth pointing out that in recent years, PREX-II has given higher LL values (L=106±37MeVL=106\pm 37\ \mathrm{MeV} Adhikari4 ). This result is also considered later in this work.

The pressure of the system can be calculated numerically from

P(ρ,δ)=ρ2d(E/ρ)dρ.P\left(\rho,\delta\right)=\rho^{2}\frac{d\left(E/\rho\right)}{d\rho}. (2.4)

To construct the EOS in the whole density range, the NV EOS model Negele38 and BPS EOS model Baym39 are used for the inner crust and the outer crust, respectively.

II.2 Fundamental formula of statically spherically symmetric neutron stars

Neutron stars in this work are considered to be isolated, non-rotating, and statically spherically symmetric. The Tolman-Oppenheimer-Volkoff (TOV) equations consist of the equation of hydrostatic equilibrium Oppenheimer19 ; Tolman20

dPdr=[m(r)+4πr3P(r)][ρ(r)+P(r)]r[r2m(r)],\frac{dP}{dr}=-\frac{\left[m\left(r\right)+4\pi r^{3}P\left(r\right)\right]\left[\rho\left(r\right)+P\left(r\right)\right]}{r\left[r-2m\left(r\right)\right]}, (2.5)

and

dm(r)dr=4πr2ρ(r).\frac{dm\left(r\right)}{dr}=4\pi r^{2}\rho\left(r\right). (2.6)

The external boundary condition of the star is as follows: ρ=p=0\rho=p=0. Given the center density of the neutron star, it can be solved layer by layer from the interior to the exterior of the star using the method of solving high-precision initial value problems.

Tidal deformation of neutron star is an EOS-dependent property which can be measured through the gravitational wave events such as GW170817 Abbott18PRL . To linear order, the tidal deformation Λ\varLambda is defined as Flanagan22

Qij=Λεij,Q_{ij}=-\varLambda\varepsilon_{ij}, (2.7)

where QijQ_{\mathrm{ij}} and εij\varepsilon_{\mathrm{ij}} represent the induced quadrupole moment and the static tidal field, respectively. The second tidal Love number k2k_{2} is defined as

k2=32ΛR5,k_{2}=\frac{3}{2}\varLambda R^{-5}, (2.8)

which can be solved together with the TOV equation Chamel23 ; Chamel24 .

III Neural network implementation of inverse TOV mapping

To clearly demonstrate the process of implementing an NN for inverse TOV mapping, we draw the flow chart as shown in Fig. 1. The first step requires the provision of a relatively complete training set. After determining the range of symmetry energy parameters, the massive EOS (output of NN) is generated by the method in Section II.1, while the corresponding mass-radius points (input of NN) are calculated by the TOV equation. The training set of NN is filtered by astronomical observations. In the second step, the training set is substituted into the initialized NN training to examine the loss. Adjust the key parameters of NN until the loss converges. The final step is to find the minimum number of neurons at the input, which means comparing the relative errors under different mass-radius point conditions. The different mass-radius points come from the non-training sample, which is the EOS in the non-training set and the corresponding mass-radius (consistent with the parameter range and filtering in the first step).

\pgfmathresultptStartSelect the range of parameters (such as LKsymK0J0L-K_{sym}-K_{0}-J_{0})Calculation of EOS with parametric EOSCalculation of M-R using TOV equationFiltering by astronomical observationsInitialized NNLoss convergence?Adjusting the parameters of NNReduction of neurons at the inputComparative Relative ErrorEndNon-training set samplesNY
Figure 1: Flowchart of inverse TOV mapping neural network implementation

The most significant prerequisite for the implementation of this algorithm is the training sample processing. Four groups of variable parameters are used to generate training samples to cover as large EOS range as possible. The variable parameter space is as follows: 30L143MeV30\leqslant L\leqslant 143\ \mathrm{MeV}, 400Ksym100MeV-400\leqslant K_{\mathrm{sym}}\leqslant 100\ \mathrm{MeV}, 200Jsym800MeV-200\leqslant J_{\mathrm{sym}}\leqslant 800\ \mathrm{MeV}, 300J0400MeV-300\leqslant J_{0}\leqslant 400\ \mathrm{MeV}, K0=240MeVK_{0}=240\ \mathrm{MeV}, Esym(ρ0)=31.7MeVE_{\mathrm{sym}}\left(\rho_{0}\right)=31.7\ \mathrm{MeV} Shlomo33 ; Piekarewicz34 ; Zhang35 ; Oertel36 ; Li37 . It is especially worth noting that the slope LL range was developed by combining terrestrial Adhikari4 and astrophysical data. To obtain preliminary training samples, we generate EOS (Section II.1) by taking points at equal intervals in the above parameter interval, then solve the corresponding M-R with the TOV equation. Moreover, in order to make the output more reasonable, we used astronomical observation results during the sample generation stage: Mmax2.14MM_{\max}\geqslant 2.14\ M_{\odot} Cromartie20NatAs , Λ1.4=190120+390\varLambda_{1.4}=190_{-120}^{+390} Abbott18PRL , 1.44±0.15M1.44\pm 0.15\ M_{\odot} Miller5 corresponding to [11.2, 13.3]km\left[11.2,\ 13.3\right]\ \mathrm{km}, 2.08±0.07M2.08\pm 0.07\ M_{\odot} Riley6 corresponding to [12.2, 16.3]km\left[12.2,\ 16.3\right]\ \mathrm{km}. Similarly to Ref. Fujimoto_JHEP2021 considering the causality condition (cs<cc_{s}<c) and the stability condition (dp/dρ>0dp/d\rho>0), if the maximum mass filter is then added, the sample size is comparable to that after the screening of astronomical observations as described above, with a limited improvement in the generalization ability. In the sample processing stage, due to the randomness of observations, we randomly sampled the mass-radius sequences over the whole mass range. Note that because it is more difficult to form low-mass neutron stars, as in Ref. Fujimoto_JHEP2021 the sampling point is set to be greater than 1 MM_{\odot}. But, considering the theoretical lower mass limit of neutron stars that can reach 0.1 MM_{\odot} Haensel_AA2002 , as well as recent observations of the HESS J1731-347 Doroshenko_NA2022 , a wider mass sampling interval was adopted. The final sample size involved in the NN is approximately 880,000, where M-R points are used as input and P-ρ\rho points are used as output.

Table 1: Architecture of neural network
Layer          Number of neurons          Activation function
0(Input)          80          N/A
1          100          ReLU
2          300          ReLU
3          400          ReLU
4          400          ReLU
5          400          ReLU
6          300          ReLU
7(Output)          182          tanh

The NN was initialized with reference to the framework of Ref. Fujimoto16 and Krastev15 and according to the background of our training set. By adjusting the relevant parameters, such as the number of layers, the optimization function, the epoch size, etc., the loss convergence of the NN is achieved. We finalized the NN architecture as shown in Table 1. The 80 neurons in the input layer means that the NN can provide accurate predictions under the condition of 40 pairs of (M,R) points. The other key parameters for this NN are shown in Appendix A and the predicted results for the different input points are shown in Appendix B.

To verify the accuracy, we used several mass-radius sequences of the EOS to substitute into this NN for prediction. In particular, we use APR3 and APR4 to test the effectiveness of this NN for non-parameter EOS.

Fig. 2 shows the NN predictions for three adopted EOS. It is shown that our NN method achieves a relatively accurate prediction function with approximately 40 mass-radius points as input. The solid black line for the (a) panel equation of state is randomly generated and substituted into Eq.(2.5) to give the solid black line for the mass-radius sequence. Additionally, We also test some EOS other than the parameterized EOS and find that some of them can also be predicted well, as shown in (b) and (c) in Fig. 2 for APR3 and APR4 EOS. This result means that our NN method will no longer be limited to parametric EOS even though we use only parametric EOS as training samples.

Refer to caption
Figure 2: The EOS predictions of NN from 40 sampled points for the LJ0JsymKsymL-J_{0}-J_{\mathrm{sym}}-K_{\mathrm{sym}} panel, where the solid black line represents the original EOS, and the blue dots denote the NN predictions. The three sample EOSs are, (a) parametric EOS with randomly generated parameters as L=40.5MeVL=40.5\ \mathrm{MeV}, Ksym=200.5MeVK_{\mathrm{sym}}=-200.5\ \mathrm{MeV}, Jsym=320MeVJ_{\mathrm{sym}}=320\ \mathrm{MeV}, J0=110MeVJ_{0}=110\ \mathrm{MeV}, K0=240MeVK_{0}=240\ \mathrm{MeV}, Esym(ρ0)=31.7MeVE_{\mathrm{sym}}\left(\rho_{0}\right)=31.7\ \mathrm{MeV}, (b) APR3 EOS and (c) APR4 EOS. The inset shows the NN sampling from the mass-radius sequence. The lower panels are the relative error of the fitting.

For comparison with the results in Ref. Krastev15 , we also probe the output of NN by setting the parameters of EOS in the LKsymL-K_{\mathrm{sym}} panel, where J0=Jsym=0J_{0}=J_{\mathrm{sym}}=0, and LL and KsymK_{\mathrm{sym}} take a range of [30, 90]MeV\left[30,\ 90\right]\ \mathrm{MeV} and [400, 100]MeV\left[-400,\ 100\right]\ \mathrm{MeV}, respectively. Using the same sample generation process as for the LJ0JsymKsymL-J_{0}-J_{\mathrm{sym}}-K_{\mathrm{sym}} panel, the sample size at this point is approximately 8,000. Fig. 3 shows the NN predictions for the randomly selected test EOS with the parameters limited in the previously mentioned range. We try to use as few input samples as possible to achieve the prediction function. Here the prediction function for the full density segment is implemented in the NN with 4 rows of mass-radius sequences as input (the data of the input is shown in Table 2). Since the low-dimensional parameter space is simpler compared to the high-dimensional, e.g., the same (M, R) point corresponds to fewer EOS, only four output points and some parameters of the NN in the simplified Table 1 are needed to implement the function. The above results show that our NN design is more suitable for inverse TOV mapping under isospin-dependent parametric EOS conditions.

Refer to caption
Figure 3: The EOS predictions of NN from 4 sampled points for the LKsymL-K_{\mathrm{sym}} panel (The EOS parameters are as follows: L=40.5MeVL=40.5\ \mathrm{MeV}, Ksym=80.5MeVK_{\mathrm{sym}}=80.5\ \mathrm{MeV}, Jsym=0MeVJ_{\mathrm{sym}}=0\ \mathrm{MeV}, J0=0MeVJ_{0}=0\ \mathrm{MeV}, K0=240MeVK_{0}=240\ \mathrm{MeV}, Esym(ρ0)=31.7MeVE_{\mathrm{sym}}\left(\rho_{0}\right)=31.7\ \mathrm{MeV}). The solid black line is the original EOS from which the sample points shown as the red dots were generated. The blue dots are the EOS data output by NN. The inset shows the NN sampling from the mass-radius sequence (Greater than 1 MM_{\odot}).
Table 2: Mass-radius sampling points for Fig. 3
Mass (MM_{\odot})          Radius (km)
1.650030          11.8838
1.916370          11.6289
2.046519          11.3741
2.201511          10.3368

Compared to Fig. 2, Fig. 3 demonstrates a greater accuracy of the neural network predictions after reducing the two parameter dimensions. The fact that other different kinds of the EOS can be valid suggests that this NN prediction has certain universality.

IV Combining NICER observations to verify the validity of NN

Refer to caption
Figure 4: NN predictions combining the most probable points of the two NICER observations. (a) The red curve is NN’s predicted EOS result, and the blue area presents the constraint of EOS from Ref. Huth11 in 95% (light) and 68% (dark) credible intervals. (b) The red curve represents the dimensionless tidal deformation and stellar mass calculated using the EOS from (a). The constraint on the Λ\Lambda of canonical neutron star from GW170817 Abbott8 is also presented.

To validate the predictions of the inverse TOV mapping-type NNs, prediction EOS from NICER observations are shown and discussed in this section within the parametric EOS framework as well as adopting the results of Refs. Abbott8 and Huth11 . Furthermore, due to the insufficient number of observations at the neutron star mass-radius, the following approach was adopted to satisfy the NN input points: 1. The EOS is generated in batch with parametric EOS within the range of nuclear parameters (LL, KsK_{s}, JsJ_{s}, J0J_{0}) mentioned in Sec. III; 2. Take one EOS that fits near the most probable point of the two NICER observations (R1.44M=13.0kmR_{1.44M\odot}=13.0\ km Miller21APJ , R2.08M=13.7kmR_{2.08M\odot}=13.7\ km Riley19APJ ); 3. Forty points were randomly taken on the M-R curve corresponding to this EOS. However, for the real situation, 40 precise (M,R) observations from different neutron stars are necessary for NN input. It is expected that such accurate and sufficient observations will be available in the near future.

The NN-based predicted EOS together with further calculated dimensionless tidal deformations as a function of stellar mass are displayed in Fig. 4(a). To demonstrate the effect of NN, we also plot the constraint on EOS from Huth etal.et~{}al. Huth11 and Λ\Lambda of canonical neutron star (M=1.4MM=1.4M_{\odot}) from GW170817 Abbott8 . It is shown that the predictions obtained under the assumption of observational accuracy prefer relatively stiff EOS at high density, which is basically in agreement with the results of Ref. Huth11 . Further calculation of the predicted EOS shows that the canonical neutron star has a dimensionless tidal deformation of Λ1.4507\varLambda_{1.4}\approx 507, indicating a high tidal deformability. From Fig. 4(b), we can see that when the stellar mass is less than 1.6 MM_{\odot}, the tidal deformability decreases relatively quickly as the mass increases, when the mass is higher than 1.6 MM_{\odot}, the tidal deformability decreases relatively slowly with the increase of mass. For massive neutron stars with M=2.0MM=2.0M_{\odot}, its tidal deformation is Λ2.051\varLambda_{2.0}\approx 51, which is far smaller than that of canonical neutron stars.

V Summary and Outlook

The NN algorithms to implement the inverse mapping TOV equation is constructed by employing the parametric EOS as the training sample. Through the practical application of the obtained neural network algorithm, we get the following results.

  • (i)

    We implemented the inverse constraint EOS of NN in 4-dimensional parameter space (LJ0JsymKsymL-J_{0}-J_{\mathrm{sym}}-K_{\mathrm{sym}}) and 2-dimensional parameter space (LKsymL-K_{\mathrm{sym}}), respectively. About 40 mass-radius points are required to output higher precision EOS under the condition that LJ0JsymKsymL-J_{0}-J_{\mathrm{sym}}-K_{\mathrm{sym}} is a variable. Four mass-radius points are required to output a higher precision EOS with LKsymL-K_{\mathrm{sym}} as the variable.

  • (ii)

    The EOS based on the obtained NN algorithm and the two sets of NICER observational values with assumed precision is predicted. Similar to the results from other method, the EOS constrained by the NICER observation and NN is relatively stiff, and the corresponding tidal deformability of canonical neutron star is relatively high. Our NN predictions are basically consistent with the results of Huth etal.et~{}al. Huth11 .

Despite the generalisation ability of the algorithm we provide, the output of the neural network is difficult to ensure credible results when the EOS is other than the isospin-dependent parametric EOS. To fill this gap, more types of EOSs need to be added to the training samples in the future. We believe that by combining precise observations from multiple sources, NN will be a promising tool for achieving more precise constraints on the EOS of nuclear matter.

Acknowledgements.
This work is supported by NSFC (Grants No. 12375144 and 11975101), Guangdong Natural Science Foundation (Grants No. 2022A1515011552 and 2020A151501820).

Appendix A Neural network implementation

NN, one of the common algorithms in ML, is now used in almost every aspects of scientific research and engineering Wright28 . Based on the universal approximation theorem, NN can implement complex nonlinear mappings Hornik29 ; Leshno30 . Its construction is mainly divided into an input layer, a hidden layer and an output layer. Each neuron can be treated as a container of numbers. Similar to the transmission of electrical signals between synapses, neurons are using numbers as signals between them. Starting from the input layer, each layer of neurons goes through the following process between each other

zl=w1a2l1++wnanl1+b.z^{\mathrm{l}}=w_{1}a_{2}^{\mathrm{l-1}}+\cdots+w_{\mathrm{n}}a_{\mathrm{n}}^{\mathrm{l-1}}+b. (2)

where ww represents the weight, ll represents the serial number of layers, and bb represents the bias value Goodfellow . The calculated zz is then substituted into the activation function to obtain the value of the output of the neuron to the next layer:

ajl=f(zl),a_{\mathrm{j}}^{\mathrm{l}}=f\left(z^{\mathrm{l}}\right), (3)

where f(x)f(x) is called the activation function and is usually used as Sigmoid or ReLU, the latter being adopted in this work. When summarizing the operations of the whole layer of neurons, we can obtain the following expression

f([w0,0w0,1w0,nw1,0w1,1w1,nwk,0wk,1wk,n][a00a10an0]+[b0b1bn])=output.f\left(\left[\begin{matrix}w_{0,0}&w_{0,1}&\cdots&w_{0,\mathrm{n}}\\ w_{1,0}&w_{1,1}&\cdots&w_{1,\mathrm{n}}\\ \vdots&\vdots&\ddots&\vdots\\ w_{\mathrm{k},0}&w_{\mathrm{k},1}&\cdots&w_{\mathrm{k},\mathrm{n}}\\ \end{matrix}\right]\left[\begin{array}[]{c}a_{0}^{0}\\ a_{1}^{0}\\ \vdots\\ a_{\mathrm{n}}^{0}\\ \end{array}\right]+\left[\begin{array}[]{c}b_{0}\\ b_{1}\\ \vdots\\ b_{\mathrm{n}}\\ \end{array}\right]\right)=output. (4)

At this point, the NN has completed one forward propagation. We also need a metric to evaluate how well the output compares to the true value. This type of evaluation metric is called the loss function, in which we usually use the mean squared error (MSE=1ni=1n(YiYi^)2MSE=\frac{1}{\mathrm{n}}\sum_{\mathrm{i}=1}^{\mathrm{n}}{\left(Y_{\mathrm{i}}-\widehat{Y_{\mathrm{i}}}\right)^{2}}) method . The MSE of each batch is equivalent to the construction drawing, guiding the optimizer to continuously adjust a huge number of weights (ww) and biases (bb) in the NN. The available optimizers are Stochastic Gradient Descent (SGD) and Adam Adam , the latter being adopted in this paper.

The platform for our implementation of NN is Keras, with Tensorflow as its backend Keras ; Tensorflow . It integrates the excellent features of Compute Unified Device Architecture (CUDA) for parallel computing on GPU with Tensorflow as the backend thus providing a rich interface. The training of the NN was performed on NVIDIA GeForce GTX 1650. It can significantly save the time cost of computing while implementing complex NN. To prevent overfitting, Dropout is set to 0.3 between layers. Furthermore, we chose an initial learning rate of 0.0003, a batch size of 500, and an epoch of 800. The proportion of the training dataset used for the validation dataset was 0.1. The hidden layer is used as a fully connected layer. Based on the above conditions, we have designed six layers of NN to achieve inverse TOV mapping. All the data in the NN need to be normalized, which in this paper is used as (M//3, R//35) and ((log10p\log_{10}p)//40, (log10ρ)/20(\log_{10}\rho)/{20}). The EOS is taken logarithmic in order to avoid too large a gap in pressure magnitude between different density intervals affecting the prediction accuracy of the NN. For ease of calculation both P and ρ\rho are taken as logarithmic results in MeVfm3MeV\cdot fm^{-3} and kgm3kg\cdot m^{-3} respectively.

Appendix B Results for different input points

We gradually reduce the input points after achieving loss convergence, while improving the relevant parameters of NN. Within the conventional NN framework, most of the improvements to achieve convergence of the NN can only affect the convergence rate of the loss cannot affect the minimum number of input points, such as increasing the number of hidden layers. As the number of neurons in the input layer continues to decrease, a significant error occurs in the LJ0JsymKsymL-J_{0}-J_{\mathrm{sym}}-K_{\mathrm{sym}} panel for 35 pairs of (M,R) (see Fig. 5), while a significant error occurs in the LKsymL-K_{\mathrm{sym}} panel for 3 pairs of (M,R) (see Fig. 6). It is important to note that over-concentration of (M,R) sampling points affects the NN predictions. Such sampling is a small probability event in astronomical observations.

Refer to caption
Figure 5: Similar to Fig 2 (a), but for 35 sampled points.
Refer to caption
Figure 6: Similar to Fig. 3, but for 3 sampled points.

Data Availability Statement: The main code used in this work can be found in the Github repository https://github.com/zhscut/NN-for-Neutron-Star.

References

  • (1) L. W. Chen, C. M. Ko and B. A. Li, Phys. Rev. Lett. 94, 032701. (2005).
  • (2) B. A. Li, X. Han, Phys. Lett. B 727, 276-281 (2013).
  • (3) B. A. Li, L. W. Chen and C. M. Ko, Phys. Rep. 464, 113-281 (2008).
  • (4) B. A. Li, P. G. Krastev, D. H. Wen, and N. B. Zhang, (2019), arXiv:1905.13175.
  • (5) A.E.L. Dieperink, Y. Dewulf, D. Van Neck, M. Waroquier, V. Rodin, Phys. Rev. C 68 064307. (2003).
  • (6) P. Russotto, S. Gannon, and S. Kupny, etet alal., Phys. Rev. C 94, 034608 (2016).
  • (7) A. Le Fevre, Y. Leifels, etet alal., Nucl. Phys. A 945, 112-133 (2016).
  • (8) D. Adhikari, H. Albataineh, etet alal., Phys. Rev. Lett. 126, 172502. (2021).
  • (9) M. C. Miller etet alal., Astrophys J. Lett. 918, L28 (2021).
  • (10) T. E. Riley etet alal., Astrophysical J. Lett. 918, L27 (2021).
  • (11) P. T. Pang, I. Tews, and M. W. Coughlin etet alal., Astrophysical. J. 922, 14 (2021).
  • (12) B. P. Abbott etet alal., Phys. Rev. lett. 121, 161101 (2018).
  • (13) M. Feroci, L. Stella, etet alal., Experimental Astronomy 34, 415-444 (2012).
  • (14) http://www. the-athena-x-ray-observatory. eu.
  • (15) M. Punturo, M. Abernathy, etet alal., Classical and Quantum Gravity 27, 194002 (2010).
  • (16) P. E. Dewdney, P. J. Hall, etet alal., Proceedings of the IEEE 97, 1482-1496 (2009).
  • (17) L. Rezzolla, P. Pizzochero, etet alal., Springer International Publishing.(2018).
  • (18) L. Lindblom, Astrophysical. J. 398, 569-573 (1992).
  • (19) Y. Fujimoto, K. Fukushima, and K. Murase, Phys. Rev. D 98, 023019 (2018).
  • (20) J. Xu, Z. Zhang, B. A. Li, Phys. Rev. C 104(5), 054324 (2021).
  • (21) S .Huth, P. T. Pang, etet alal., Nature 606, 276-280 (2022).
  • (22) Y. Li, H. Chen, D. Wen, etet alal., Eur. Phys. J. A 57 1-10 (2021).
  • (23) H. O. Silva, A. M. Holgado, A. Cardenas-Avendano and N. Yunes, Phys. Rev. Lett. 126, 181101 (2021).
  • (24) W. J. Xie, B. A. Li, Astrophysical J. 883, 174 (2019).
  • (25) S. I. Amari, M. A. Arbib, Systems neuroscience, 119-165 (1977).
  • (26) S. Gazula, J. Clark, and H. Bohr, Nucl. Phys. A 540, 1 (1992).
  • (27) K. Gernoth, J. Clark, J. Prater, and H. Bohr, Phys. Lett. B 300, 1 (1993).
  • (28) K. Gernoth and J. Clark, Neural Networks 8, 291 (1995).
  • (29) J. W. Clark, T. Lindenau, and M. Ristig, Springer Lecture Notes in Physics, Vol. 522 (Springer-Verlag, Berlin, 1999).
  • (30) T. Zhao, H. Zhang, Nucl. Phys. A 1021, 122420 (2022).
  • (31) Z.M. Niu, H.Z. Liang, Phys. Lett. B 778 48. (2018).
  • (32) R. Utama, J. Piekarewicz, H.B. Prosper, Phys. Rev. C 93 014311 (2016).
  • (33) R. Utama, W.C. Chen, J. Piekarewicz, J. Phys. G, Nucl. Part. Phys. 43 114002 (2016).
  • (34) S. Soma, L. Wang, S. Shi, etet alal., arXiv preprint arXiv:2201.01756. (2022).
  • (35) P. G. Krastev, Galaxies 10, 16 (2022).
  • (36) A. Thete, K. Banerjee, T. Malik, arXiv:2208.13163. (2022).
  • (37) D. Farrell, P. Baldi, J. Ott, A. Ghosh, etet alal., arXiv:2209.02817. (2022).
  • (38) N. B. Zhang, B. A. Li, J. Xu, Astrophysical. J. 859(2), 90. (2018).
  • (39) N. B. Zhang, B. J. Cai, B. A. Li, W. G. Newton, and J. Xu, Nucl. Sci. Technol. 28, 181 (2017).
  • (40) S. Shlomo, V. M. Kolomietz, G. Colo, Eur. Phys. J. A, 30(1), 23-30 (2006).
  • (41) J. Piekarewicz. Journal of Physics G Nuclear Physics. 37, 064038 (2010).
  • (42) M. Oertel, M. Hempel, T. Klähn, and S. Typel, Rev. Mod. Phys. 89, 015007 (2017).
  • (43) B. A. Li, Nuclear Physics News, 27, 7 (2017).
  • (44) J. W. Negele, D. Vautherin, Nucl. Phys. A 207, 298-320 (1973).
  • (45) G. Baym, C. Pethick, P. Sutherland, Astrophysical. J. 170, 299. (1971).
  • (46) J. R. Oppenheimer, G. M. Volkoff, Phys. Rev. 55, 374 (1939).
  • (47) R. C. Tolman, Phys. Rev. 55, 364 (1939).
  • (48) B. P. Abbott, R. Abbott, T. D. Abbott, etet alal., Phys. Rev. Lett. 121, 161101. (2018).
  • (49) E. E. Flanagan, T. Hinderer, Phys. Rev. D 77, 021502(R) (2008).
  • (50) T. Hinderer, Astrophysical. J. 677, 1216.(2008).
  • (51) T. Hinderer, B. D. Lackey, R. N. Lang, J. S. Read, Phys. Rev. D 81, 123016 (2010).
  • (52) H. T. Cromartie, E. Fonseca, S.M Ransom, etet alal., Nat. Astron. 4: 72-76. (2020).
  • (53) Y. Fujimoto, M. Nitta, Journal of High Energy Physics, 2021(9), 1-36. (2021).
  • (54) P. Haensel, J. L. Zdunik, F. Douchin, Astron. Astrophys. 385, 301-307. (2002).
  • (55) V. Doroshenko, etet alal., Nat. Astron., 6(12), 1444-1451. (2022).
  • (56) M. C. Miller, F. K. Lamb, A. J. Dittmann, etet alal., Astrophys. J. Lett. 887, L24 (2019).
  • (57) T. E. Riley, A. L. Watts, S. Bogdanov, etet alal., Astrophys. J. Lett. 887, L21 (2019).
  • (58) L. G. Wright, T. Onodera, and M. M. Stein, etet alal., Nature, 601, 549-555 (2022).
  • (59) K. Hornik, M. Stinchcombe, and H. White, Neural networks, 3, 551-560. 171 (1990).
  • (60) M. Leshno, V. Y. Lin, etet alal., Neural Networks, 6, 861-867.(1993).
  • (61) I. Goodfellow, Y. Bengio, A. Courville Deep learning. (MIT press), (2016).
  • (62) D. P. Kingma,J. Ba, Adam: A method for stochastic optimization. arXiv:1412.6980, (2014).
  • (63) F. Chollet, "Keras: Deep learning library for theano and tensorflow," https://github.com/fchollet/keras (2015).
  • (64) M. Abadi, A. Agarwal, P. Barham, etet alal., Tensorflow: Large-scale machine learning on heterogeneous distributed systems. arXiv:1603.04467, (2016).