Free Access
Issue
A&A
Volume 642, October 2020
Article Number A78
Number of page(s) 8
Section Stellar structure and evolution
DOI https://doi.org/10.1051/0004-6361/202038130
Published online 07 October 2020

© ESO 2020

1. Introduction

Neutron stars (NS) are currently the best astrophysical sites for studying the details of dense matter physics in conditions that are inaccessible for terrestrial experiments (see e.g. Haensel et al. 2007 for a textbook introduction) Specifically, this refers to the equation of state (EOS) of dense, cold, neutron-rich matter at densities many times higher than the nuclear saturation density ρs ≃ 2.7 × 1014 g cm−3, corresponding to the nuclear saturation baryon density ns ≃  0.16 fm−3.

Because the complete theory of many-body nuclear interactions is not known in full, recent efforts have been focussed on inferring the EOS from astrophysical observations of NSs. Recent observations include the NICER simultaneous measurements of the mass and radius of PSR J0030+0451 (Riley et al. 2019; Miller et al. 2019), the 170817 binary NS inspiral detection and parameter estimation done by the LIGO and Virgo Collaborations (Abbott et al. 2017a, 2019, 2018), including the measurement of the masses and tidal deformability of the system components, accompanied by the observations of high-energy photons by the Fermi and INTEGRAL satellites (Abbott et al. 2017b), as well as several observations of massive ≃2 M NSs (Demorest et al. 2010; Fonseca et al. 2016; Antoniadis et al. 2013; Cromartie et al. 2020). These observations provide indirect but nevertheless informative answers to the question of how compact objects are structured and, hence, the nature of their internal composition. The procedure is based on solving the equations of stellar structure, typically the Tolman–Oppenheimer–Volkoff (TOV) equations (Tolman 1939; Oppenheimer & Volkoff 1939) for an assumed EOS (or class of EOSs) to subsequently compare the global observable NS parameters, such as gravitational mass M, radius R, and, recently, tidal deformability Λ as well (Flanagan & Hinderer 2008; Van Oeveren & Friedman 2017) to observed values; in the simplest case of the TOV equation, there is a strong relation between the sequence of global NS parameters (EOS functionals), such as M(R) or M(Λ), and the pressure-density p(ρ) relation defining the EOS. Therefore, given a set of astrophysical measurements, it is possible, in principle, to recover the EOS by inverting the TOV equations. In reality, however, astrophysical observations are affected by measurement errors and they are not distributed optimally in the parameter space, meaning that an observer doesn’t have any freedom in selecting the intrinsic parameters, such as the mass, M, of the observed object, to optimally cover the range of pressure and density so that the part of the EOS relation that is of interest may be revealed.

The most common strategy in the estimation of EOS utilises Bayesian inference, which is based on the inversion of the TOV equations and a limited number of observations. Examples of this approach were recently presented in the following works: Steiner et al. (2010, 2013), Raithel et al. (2016), Holt & Lim (2019), Fasano et al. (2019), Hernandez Vivanco et al. (2019) and Traversi et al. (2020). Here, instead of directly inverting the TOV equations, we study an alternative approach based on a machine-learning (ML) artificial neural network (ANN), inspired by the autoencoder (AE) architecture (Hinton & Zemel 1993; Goodfellow et al. 2016). Similar machine learning techniques applied to results of numerical simulations and measurements currently make up a field of active research; for example, Haegel & Husa (2019) show that the final mass and spin of a Kerr black hole can be predicted from the initial values of parameters of black hole components. Specifically, this has been explored in the field of the NS EOS, Fujimoto et al. (2018, 2020) presented an application of a feed-forward neural network to infer the EOS from NS mass-radius measurements, whereas Ferreira & Providência (2019) compare machine-learning neural networks and support vector machine regression methods in unveiling the nuclear EOS parameters from NS observations.

Here, we study the application of ML to infer the dense-matter EOS pressure-density p(ρ) relations from a simulated set of NS observations, using a neural network trained on a purposefully simple data set, based on piecewise relativistic polytrope EOS. We performed the analysis using simulated data containing electromagnetic as well as gravitational-waves observables: radii, masses, and tidal deformabilities, applying the current knowledge of the NS mass distribution function, and varying the number of simulated observations and measurement uncertainties. While trained and tested on the piecewise relativistic polytropic EOS data set, our ML model was also validated on realistic EOS examples: it successfully recovers the SLy4 EOS (Douchin & Haensel 2001) as well as the APR EOS (Akmal et al. 1998) and the BSK20 EOS (Goriely et al. 2010). Additionally, we show the ANN network is flexible enough to generalise the mapping of the mass-radius M(R) relation from the mass–tidal-deformability M(Λ) relation, effectively allowing for the possibility of inference of the NS radius from several GW-only measurements.

The outline of the article is as follows. In Sect. 2, we discuss the choice of the machine learning algorithms used. Section 3 is devoted to the description of the input and output data generation procedures, with a particular emphasis on the EOS and the NS structure. Section 4 contains results of the neural-networks estimation of the dense-matter EOS from NS observables: M(R) and M(Λ). We discuss the results in Sect. 5. We conclude in Sect. 6 with a summary of our study.

2. Machine learning

The machine learning field of computer science is based on the premise that algorithms can learn from examples in order to solve problems and make predictions without needing to be explicitly programmed (Samuel 1959). Among the many ML algorithms, the ANN currently belong to the most popular. Complex ANN consisting of many neurons combined with various training algorithms (such as the backpropagation and stochastic gradient descent – for textbook review, see e.g. Goodfellow et al. 2016 and references therein) are able to capture complicated non-linear relationships in the data by composing hierarchical internal representations. The complex (in other words, deeper) the algorithm is, the more abstract features it can, in principle, learn from the data.

The main motivation for employing ANN in our project is associated with non-linear potential of the this ML model. Since each neuron in the network is basically a non-linear function, it is possible to create a complex mapping between the input and the output of the model. This characteristic is necessary for the estimation of EOS based on observables, even when excluding uncertainties since the analytical relations between the input and output parameters are non-linear.

The input to our model are astrophysical measurements of NS parameters (presented to the ANN as two arrays of M, R or M, Λ concatenated into one), whereas at the output we obtain an array of similar shape (concatenated p, ρ values). Further in text, we present additional project in which we reconstruct radius based on gravitational observables (M and Λ concatenated into one vector), the ANN output consists only of radius values. By definition, the size of the output is half the size of the input.

3. Data preparation

In this section, we describe the design of parametric EOSs and methods used to obtain the stellar parameters.

3.1. Equations of state and stellar structure

In order to cover a sufficiently-broad and representative space of solutions corresponding to M(R) and M(Λ) sequences, we employ the following simplified, parametric approach to the EOS. We assume the measurement of the low-density part of the EOS and adopt the SLy4 EOS description of Haensel & Pichon (1994) and Douchin & Haensel (2001) up to some baryon density n0, comparable to and typically larger than the nuclear saturation density (ns ≡ 0.16 fm−3). At the n0 a relativistic polytrope (Tooper 1965),

p ( n ) = κ n γ , ρ c 2 = P ( n ) γ 1 + n m b c 2 , $$ \begin{aligned} p(n)=\kappa n^{\gamma },\quad \rho c^2 = \frac{P(n)}{\gamma - 1} + nm_{b}c^2, \end{aligned} $$(1)

replaces the SLy4 EOS. For each polytrope, the pressure p and the mass-energy density ρc2 are defined using the pressure coefficient κ, the polytropic index γ responsible for the stiffness of the matter, and the mass of the baryon mb. We select the γ index as a parameter of choice; consequently, κ and mb are determined by demanding the chemical and mechanical equilibrium at n0. The first polytrope with γ1 ends at some density, n1 >  n0, where a second relativistic polytrope with γ2 is attached, and continues until n2, where a polytrope with γ3 starts. The bottom-left panel of Fig. 1 shows a schematic plot of the EOS. The parameter ranges are collected in Table 1.

thumbnail Fig. 1.

Top panels: left – sample M(R) relations; right – corresponding M() relations. Red curve and corresponding observations: example of a training datum, consisting on 20 points from the M(R) and M(), selected assuming random normal distribution with the mean value equal to true values and standard deviations σM = 0.1 M, σR = 1.0 km and σ = 1.0 km. The configurations with the same M and R have, in general, different . Bottom panels: left – schematic of a model EOS, composed in the low-density part from the realistic crust of the SLy4 EOS (Haensel & Pichon 1994; Douchin & Haensel 2001) and piecewise relativistic polytropes (Tooper 1965), right – mass-radius M(R) relations generated using the piecewise relativistic polytrope model (thin solid grey curves) and astrophysical models of NS sequences, based on the following EOSs: the SLy4 EOS (solid blue curve), the APR EOS (dashed red curve) and the BSK20 EOS (dash-dotted green curve).

Table 1.

Ranges of piecewise polytrope EOS parameters used in the study.

For a given EOS, we solve the equations of hydrostatic equilibrium for a spherically symmetric distribution of mass. The space-time metric is:

d s 2 = e ν ( r ) c 2 d t 2 d r 2 1 2 G M ( r ) / r c 2 r 2 ( d θ 2 + sin 2 θ d ϕ 2 ) , $$ \begin{aligned} \mathrm{d}s^2=e^{\nu (r)} c^2 \mathrm{d}t^2 - \frac{\mathrm{d}r^2}{1-2GM(r)/rc^2} - r^2(\mathrm{d}\theta ^2 + \sin ^2\theta \mathrm{d}\phi ^2), \end{aligned} $$(2)

with the gravitational mass M(r) inside the radius r

d M ( r ) d r = 4 π ρ ( r ) r 2 . $$ \begin{aligned} \frac{\mathrm{d}M(r)}{\mathrm{d}r}=4 \pi \rho (r) r^2. \end{aligned} $$(3)

Then the resulting Tolman-Oppenheimer-Volkoff equations (Tolman 1939; Oppenheimer & Volkoff 1939),

d P ( r ) d r = G r 2 ( ρ ( r ) + P ( r ) c 2 ) ( M ( r ) + 4 π r 3 P ( r ) c 2 ) × ( 1 2 G M ( r ) c 2 r ) 1 , $$ \begin{aligned} \frac{\mathrm{d}P(r)}{\mathrm{d}r}&= -\frac{G}{r^2}\left(\rho (r)+\frac{P(r)}{c^2}\right) \left(M(r)+\frac{4\pi r^3 P(r)}{c^2}\right) \nonumber \\&\times&\left(1-\frac{2GM(r)}{c^2r}\right)^{-1}, \end{aligned} $$(4)

supplied with the equation for the metric function ν(r),

d ν ( r ) d r = ( 2 P ( r ) + ρ ( r ) c 2 ) d P ( r ) d r , $$ \begin{aligned} \frac{\mathrm{d}\nu (r)}{\mathrm{d}r}= -\left(\frac{2}{P(r)+\rho (r)c^2} \right) \frac{\mathrm{d}P(r)}{\mathrm{d}r}, \end{aligned} $$(5)

are integrated from the center towards the surface (where the pressure, P, vanishes, which defines the radius of the star, R) using a Runge-Kutta 4th order numerical scheme with a variable integration step (Press et al. 1992) for a range of central parameters of the EOS (e.g. the central pressures Pc) resulting in the M(R) sequence.

In addition to gravitational mass, M, and radius, R, we also calculate the static lowest-order tidal deformability of the star, defined as

λ = 2 3 R 5 k 2 . $$ \begin{aligned} \lambda = \frac{2}{3}R^5k_2. \end{aligned} $$(6)

The parameter λ represents the star’s reaction on the external tidal field (e.g. exerted by a companion in a tight binary system, as observed in Abbott et al. 2017a). It is obtained in the lowest-order approximation, by calculating the second (quadrupole) tidal Love number k2 (Love 1911), a function of stellar parameters and hence the EOS:

k 2 = 8 5 x 5 ( 1 2 x ) 2 ( 2 y + 2 x ( y 1 ) ) ( 2 x ( 6 3 y + 3 x ( 5 y 8 ) ) + 4 x 3 ( 13 11 y + x ( 3 y 2 ) + 2 x 2 ( 1 + y ) ) + 3 ( 1 2 x ) 2 ( 2 y + 2 x ( y 1 ) ) ln ( 1 2 x ) ) 1 , $$ \begin{aligned} k_2&= \frac{8}{5}x^5 (1-2x)^2 \bigl ( 2-{ y} + 2x({ y}-1)\bigr ) \Bigl ( 2x \bigl ( 6 -3{ y} +3x(5{ y}-8)\bigr ) \Bigr . \nonumber \\&\quad +\ 4x^3 \left( 13 -11{ y} + x(3{ y} -2) + 2x^2(1+{ y}) \right) \nonumber \\&\quad \left.+\ 3(1-2x)^2 \bigl ( 2 - { y} + 2x({ y}-1)\bigr ) \ln (1-2x) \right)^{-1}, \end{aligned} $$(7)

where x = GM/Rc2 denotes the star’s compactness, and y the solution of

d y d r = y 2 r 1 + 4 π G r 2 / c 2 ( P / c 2 ρ ) ( r 2 G M ( r ) / c 2 ) y + ( 2 G / c 2 ( M ( r ) + 4 π r 3 P / c 2 ) r ( r 2 G M ( r ) / c 2 ) ) 2 + 6 r 2 G M ( r ) / c 2 4 π G r 2 / c 2 r 2 G M ( r ) / c 2 ( 5 ρ + 9 P / c 2 + ( ρ + P / c 2 ) 2 c 2 ρ d P / d ρ ) , $$ \begin{aligned} \frac{\mathrm{d}y}{\mathrm{d}r}&= -\frac{{ y}^2}{r}-\frac{1+4\pi G r^2/c^2(P/c^2-\rho )}{(r-2GM(r)/c^2)}{ y} \nonumber \\&\quad + \left(\frac{2G/c^2(M(r) + 4\pi r^3 P/c^2)}{\sqrt{r}(r-2GM(r)/c^2)}\right)^2 + \frac{6}{r-2GM(r)/c^2} \nonumber \\&\quad - \frac{4\pi G r^2/c^2}{r-2GM(r)/c^2}\left(5\rho + 9P/ c^2 +\frac{\left( \rho + P/c^2\right)^2 c^2}{ \rho \mathrm{d}P/\mathrm{d}\rho }\right), \end{aligned} $$(8)

evaluated at the stellar surface (Flanagan & Hinderer 2008; Van Oeveren & Friedman 2017). In the following we use the mass-normalised value of the λ parameter,

Λ = λ ( G M / c 2 ) 5 . $$ \begin{aligned} \Lambda = \lambda \left( GM/c^2\right)^{-5}. \end{aligned} $$(9)

In order to relate the Λ parameter with the stellar radius R, we produce a following radius-like parameter (M, Λ), which we call the tidal radius (proposed in Wade et al. 2014):

R ̂ = 2 M Λ 1 / 5 . $$ \begin{aligned} \hat{R} = 2M\Lambda ^{1/5}. \end{aligned} $$(10)

This function of Λ and M is henceforth used in the study. Sample M(R) and M() relations are presented in the top panels of Fig. 1, along with simulated measurement points (the procedure of obtaining them is presented in Sect. 3.2). Moreover, in the bottom right panel of Fig. 1, we present a bundle of M(R) relations used in the training, generated for piecewise relativistic polytropes to compare the training set with the astrophysical models based on the SLy4, APR and BSK20 EOSs. The training data cover the space of parameters similar to astrophysical models; therefore, we expect that the algorithm will generalise the EOS reconstruction toward previously unseen types of curves (types of curves on which it wasn’t trained on).

3.2. Neutron-star mass function and simulated measurement errors

In order to investigate the influence of the amount and precision of data – the number of observations N and their measurement errors – we restrict the sample of M(R) and M() for masses from the astrophysically-realistic range above 1 M, which corresponds to observed NS masses in Galactic binary NS systems (Alsing et al. 2018), and in the GW170817 and GW190425 events (Abbott et al. 2017a, 2020). We exclude from further analysis all piecewise polytropic solutions that are not compatible with current state of observations, giving the NS maximum masses above 1.9 M (conservative choice motivated by the observations of massive NSs, see Demorest et al. 2010; Fonseca et al. 2016; Antoniadis et al. 2013; Cromartie et al. 2020).

To realistically recreate astrophysical observations, we select the measurement points from a realistic NS mass function (mass probability distribution) out of which the mass values are to be randomly selected. Consistently with current observations of NSs in the Galaxy, the mass function is represented by a double-Gaussian distribution (Alsing et al. 2018) with the main peak around the Chandrasekhar mass and the second, smaller peak corresponding to the NS masses close to 2 M, namely 𝒩(μ1, σ1)+𝒩(μ2, σ2), where μ1 = 1.34, σ1 = 0.07, μ2 = 1.80, σ2 = 0.21 (see Fig. 1 in Alsing et al. 2018 for details). This NS mass function is consistent with a recent GW observation of a heavy NS binary system (Abbott et al. 2020).

The training data set is prepared by assuming that the measurements are witness to a measurement error. After randomly choosing N values of the gravitational mass M from the above-mentioned mass distribution, we construct the training samples corresponding to a given M(R) or M() point by drawing values from normal distributions 𝒩(M(R),σi) or N ( M ( R ̂ ) , σ i ) $ \mathcal{N}(M(\hat{R}), \sigma_i) $, with i = M, R, respectively. For the σi parameters, we chose the values in the range of 0.01−0.1 M for σM and 0.01−1 km for σR. Uncertainties for tidal deformabilities are defined in terms of and were in range σ = 0.01−1 km, which corresponds to σΛ = 102−103. An example of the training sequence, obtained assuming the double-Gaussian mass distribution and σM = 0.1 M, σR = σ = 1.0 km is shown in Fig. 1 (marked red in top panels). Gray curves (and the red curve) correspond to M(R) and M() relations computed with the TOV equations for some examples of the piecewise polytropic EOS described in Sect. 3.1. The scattered points correspond to the actual input data fed to our model; they are based on the red curve values according to the procedure describe above.

In total the training dataset contains 13982 piecewise polytrope EOSs (see Table 1 for the details), out of which the M(R) and M() sequences were produced by solving the TOV equations. For each of these sequences, we then randomly selected N values of M (N equal to 10, 15, 20, 30, 40, or 50 observations) using the above-mentioned NS mass distribution, and recover the corresponding values of R and . For each input EOS, this procedure is repeated a fixed number of Ns = 30 times. As a result, each input EOS is represented in the training stage by Ns different realisations of N observations of M(R) or M(), subject to “observational errors” by drawing the values from normal distributions parametrised by σi. This step allows us to effectively estimate the errors that ANN makes in the prediction of output sequences, that is, the error of reconstructing pressures and densities. To compute these errors, we then calculate the differences between the estimated output and the ideal expected result (the “ground truth” values). The errors are averaged for each measurement in a given collection of realisations. This step is repeated for all the EOSs in the training dataset, returning the set of error distributions: in the case of 20 measurements, we recover 20 distributions. The error bars presented for the output values in Sect. 4 are the mean values of these distributions.

We contrast the reconstruction errors with the ANN loss function as they represent different features. Loss function is a metric defining overall performance of the ANN in terms of how well the predicted values are to original ground truth values in general. Reconstruction errors give detailed information about differences between predicted pressures and densities and their corresponding ground truth values. Furthermore, the reconstruction error changes with respect to the values of pressure and density.

At the last step of data preparation, pressures and densities were converted to the decimal logarithm values and scaled together with masses, radii, and tidal radii to the range (0, 1). Rescaling is required by the ANN non-linear functions since their domain is in the range (0, 1).

The data sets were then split into two separate subsets: a training set (70% of all instances from the total dataset) and the testing set (30% of the total dataset). In cases when the ANN was tested against the measurements corresponding to realistic tabulated EOS, the simulated measurement data was generated in the same way as for the piecewise polytropic EOS.

3.3. ANN

In the design of the ANN, we used parts of the AE architecture. The AE (Kramer 1991) is a specific type of network capable of learning how to efficiently compress and encode the data into the so-called latent space representation and, later, to decompress and reconstruct the initial data as closely as possible. The core functionality of the AE is data dimensionality reduction. During training, AE learns how to ignore the noise and extract only crucial features of the data. Dimensionality reduction is in particular useful in the application of AEs aiming for data clustering. Specifically, the features of latent representation of an AE may be used to characterise the data, for example, by employing the conditional training of the variational AE using the training data with parameter labels to subsequently study the distribution of parameters in the latent space of variables. In the present exploratory work, we employ the simplest encoder-decoder structure of AE and we do not use the properties of the latent space, leaving that aspect to a future work.

The final architecture of our ANN was chosen based on empirical tests based on the data. As an output criterion for the loss function we use the mean squared error (MSE). We tested architectures ranging from one to eight hidden layers. The optimal network, reaching the minimum value for MSE, was the one containing four hidden layers with the following number of neurons: 512, 256, 256, 512.

The final set of hyper-parameters used for the training was the following (parameters defined as in e.g. Goodfellow et al. 2016):

  • ReLU as the activation function for hidden layers,

  • sigmoid activation function for the output layer,

  • ADAM optimiser (Kingma & Ba 2014),

  • batch size of 128,

  • 0.001 learning rate.

The ANN architecture was implemented using the Python Keras library (Chollet 2015) on top of the TensorFlow library (Abadi et al. 2015), with support for the GPU. We developed the model on the NVidia Quadro P60001 and performed the production runs on the Cyfronet Prometheus cluster2 equipped with Tesla K40 GPUs, running CUDA 10.0 (Nickolls et al. 2008) and the cuDNN 7.3.0 (Chetlur et al. 2014).

4. Results

The results presented below are split into subsections. The first present the results of EOS reconstruction from M(R) and M() simulated measurements with errors using ANN trained on piecewise polytropic EOS results. The second subsection shows the application of ANN on the realistic EOS resulting from microscopic calculations (SLy4 EOS, Douchin & Haensel 2001), that is, a reconstruction of the EOS which is not a piecewise polytropic model. We also study an application of the ANN to a direct reconstruction of the NS radius with the GW-only observations of the tidal deformability.

4.1. Translating the NS observations, M(R) or M (Λ), to EOS

Here, we present the results of the ANN application to the reconstruction of the EOS based on the gravitational mass, M, and radius, R, observations, which may be a result of electromagnetic observations of, for example, the NICER mission, as well as the EOS reconstruction based on the gravitational-wave observations of mass, M, and tidal deformability, Λ (which we reparametrise as ; see Eq. (10)). The ANN described in Sect. 3.3 is trained on data sets with varying number of observations and measurement uncertainties. The resulting figures of merit – the ANN loss function MSE – are shown in Fig. 2 with the left plot corresponding to the EOS reconstruction using M(R) data and the right using M() data.

thumbnail Fig. 2.

Dependence of the MSE (ANN loss function described in Sect. 3.2) as a function of number of observations N and measurement uncertainties in the case of EOS estimation based on M(R) (left figure) and M() (right figure) observations.

The accuracy of EOS estimation is mostly influenced by the assumed measurement uncertainties in both presented cases. The value of MSE is proportional to the measurement errors; it reaches the highest value for the largest of considered uncertainties: σM = 0.1 M for mass, M, σR = 1 km for the radius, R, and σ = 1 km for the tidal radius, . Furthermore, the number of observations N had little effect on the MSE; the increase in N slightly decreased the MSE in all studied cases.

The top panels of Fig. 3 present two examples of the EOS reconstruction for the small and large measurement uncertainties in the case of N = 20 M(R) observations. Both EOSs are recovered correctly within reconstruction errors computed as specified in the Sect. 3.2 with respect to the ground-truth values of related input EOSs (marked with dashed lines on the right panel). The error ranges in case of EOS estimation using M(R) data for different measurement uncertainties are presented in the upper part of Table 2. The resulting σp and σρ spans increase proportionally, with increasing σM and σR. Furthermore, in all presented cases, the ranges for pressure errors were wider than density errors, indicating that ANN was more uncertain in the reconstruction of pressure values. The increase in the reconstruction errors is expected because the overall performance of the ANN was worse during the training (see the blue and violet curves in Fig. 2 for comparison). Another effect related to the worse performance of the EOS reconstruction is the significant increase of errors and decrease in the accuracy of reconstruction for higher p(ρ) values. Several effects may be responsible for this result, for example, the impact of adopted NS mass distribution. Naturally, if the dataset contains a smaller number of high (close to 2 M) M samples, the high p(ρ) values of the EOS are less efficiently probed. As a result, the EOS reconstruction is less certain overall in this range. We discuss alternative explanations in Sect. 5.

thumbnail Fig. 3.

Top panels: example of the input data (M(R) measurements with errors, left plot), and corresponding output data from ANN (p(ρ) relation, right plot) for the estimation of EOS from M(R). Both input samples consist of 20 observations with masses randomly selected from a mass distribution (Sect. 3.2) and measurement uncertainties equal to σM = 0.1 M, σR = 1 km (blue sample), σM = 0.01 M, σR = 0.01 km (red sample). For the description of the uncertainties on the output, see the text. Bottom panels: example of the input data (M() measurements with errors, left plot), and corresponding output data from ANN (p(ρ) relation, right plot) for the estimation of EOS from M(). Both input samples consist of 20 observations with masses randomly selected from a mass distribution (Sect. 3.2) and measurement uncertainties equal to σM = 0.1 M, σR = 1 km (blue sample), σM = 0.01 M, σ = 0.01 km (red sample). Dashed curves correspond to original (ground-truth, error-free) sequences of input and output of the TOV equations. Presented examples correspond to different EOSs.

Table 2.

Reconstruction error ranges for σp and σρ of the ANN for studied measurement uncertainties in case of EOS reconstruction for the M(R) data (upper table) and M() data (lower table).

The examples shown in the bottom panels of Fig. 3 corresponded to the EOS reconstruction using M() data for two cases of small and large measurement uncertainties and N = 20 observations. Both EOSs are estimated correctly within reconstruction errors with respect to the ground-truth values of corresponding EOSs (marked with dashed lines) and the errors are proportional to the values of the density and pressure, similarly to the M(R) case.

4.2. Application on realistic EOS

We test the ANN trained on piecewise polytropic EOS (and the TOV solutions obtained with them) on a realistic microscopic EOSs: the SLy4 EOS (Douchin & Haensel 2001), the APR EOS (Akmal et al. 1998) and the BSK20 EOS (Goriely et al. 2010). To generate data for this test, we followed the approach detailed in Sect. 3 as in the case of the polytropic EOSs. Figure 4 contain the results of EOS reconstruction using the M(R) data (top panels) and M() (bottom panels) for N = 20, σM = 0.1 M and σR = 1 km. Among the realistic microphysical EOS we have considered, the EOS relation reconstructed for the APR EOS and BSK20 EOS agree with original (ground truth) input values almost perfectly, whereas the SLy4 EOS model is reconstructed less precisely; however, the reconstructed EOS relation agrees with the ground truth values (dashed line) within reconstruction errors from Table 2.

thumbnail Fig. 4.

Top panels: ANN-reconstructed EOS from the M(R) data for the SLy4 EOS model (left plot), the APR EOS model (middle plot) and the BSK20 EOS model (right plot). Results are computed for the input M(R) data consisting of 20 observations with measurement uncertainties equal to merr = 0.1 M, err = 1 km. Bottom panels: ANN-reconstructed EOS from the M() data for the same EOS as in the top panels. Results for the input M() data consisting of 20 observations with measurement uncertainties equal to merr = 0.1 M, err = 1 km. Dashed lines correspond to the exact EOS relations.

These results show that the ANN trained on a relatively simple dataset of relativistic piecewise polytropes is able to generalise the task of EOS reconstruction towards an unknown during its training of realistic EOS.

4.3. Radius reconstruction using Λ measurements

We also present the results of an additional analysis which aims to directly reconstruct the NS radius R from GW-only observations of masses and tidal deformabilities. As Eq. (9) shows, the tidal deformability is related to M and R and to the second Love number k2, all of which are functionals on the EOS. In the general case, the Λ − R relation cannot be simply obtained (see e.g. Zhang et al. 2020; De et al. 2018 and references therein). From the point of view of the M(R) diagram, the relation between Λ and R depends on the slope of M(R), which is indirectly a function of the NS susceptibility to deformations (see Sieniawska et al. 2019 for examples of configurations with the same M and R, but different Λ values; their Sect. 3.2, Figs. 9 and 10).

In order to study the ability of reconstructing the R based on M and Λ observations, we modified the ANN described in the Sect. 3.3 since, for this case, the size of the output was twice smaller (M and concatenated at the input and R and at the output). We considered the same measurement uncertainties σM and σ as in the EOS reconstruction. The results of the ANN training are shown in Fig. 5 in terms of MSE.

thumbnail Fig. 5.

Evolution of MSE (ANN loss) in the function of number of observations and measurement uncertainties in case of R computation based on M().

Similarly as in the estimation of EOS, the strongest influence on the radius computation had the measurement uncertainties. The MSE changed in range between 5 × 10−5 and 10−2 for the data varying in uncertainties from σM = 0.01 M and σ = 0.01 km to σM = 0.1 M and σ = 1 km. Moreover, the impact of observations number was insignificant.

The examples of the radius estimation are presented in Fig. 6. Top panels show the radius computed by ANN using piecewise polytropic data for two cases of measurement uncertainties: σM = 0.01 M and σ = 0.01 km (red sample) and σM = 0.1 M and σ = 1.0 km (blue sample). Bottom panels present the estimated radius for the data corresponding to the realistic cases: the SLy4 EOS, APR EOS, and BSK20 EOS for σM = 0.1 M and σ = 1.0 km. Within the reconstruction errors, σR, all cases were correctly reconstructed, in comparison to the dash line representing the exact values of radii computed from the TOV equations. However, the σR increase proportionally to σM and σ. Furthermore, the errors varied randomly with respect to the value of radius. In contrast to pressure and density errors, no trend in radius errors was present.

thumbnail Fig. 6.

Top panels: example of the input data (left plot) and corresponding output data from ANN (right plot) in the case of R computation based on M() for the piecewise relativistic polytropes. Bottom panels: radius reconstructed from the M() data for the SLy4 EOS model (left plot), the APR EOS model (middle plot), and the BSK20 EOS model (right plot). All results were computed for the input M(R) data consisting of 20 observations with measurement uncertainties equal to σM = 0.1 M, σ = 1 km. Dashed lines correspond to exact values obtained by solving the TOV equations.

In general, the radius reconstruction from tidal deformability using ANN is possible, which demonstrates an additional ability on the part of ANN to build a non-linear mapping between astrophysical parameters of interest.

5. Discussion

The above results point us to a conclusion that the application of ANN in EOS reconstruction from astrophysical observations works for the majority of our data, with decreasing reliability for data with the largest measurement errors.

In comparison with similar approaches to the same problem (Fujimoto et al. 2018, 2020), our work extends the study of applications of the NN to NS multi-messenger astrophysics in several ways: we directly output the p(ρ) EOS table, not limiting the output to selected EOS parameters, meaning that our implementation is, in principle, not bound to specific prescription of the EOS. In addition, we study the application of the AE architecture to the problem of EOS reconstruction, investigate as input the tidal deformability parameters as a function of mass, not only M(R), that is, we try to simulate a situation in which the data comes exclusively from GW measurements, and we also investigate varying number of measurements, measurement errors, and realistic mass functions; for an additional investigation related to the last point, see the text below.

Motivated by the issue behind the significant increase of the reconstruction EOS errors for higher densities and pressures in cases of large measurement uncertainties (merr = 0.1 M and rerr = 1 km) we performed an additional analysis. We understand this as a feature of the non-linearity of the mapping between the observed values of M, R and and the EOS. As shown in Fig. 3, for example, the measurements at high masses probe a significantly larger range of pressure and densities than those at lower masses. In addition, the values of radii R and tidal deformabilities Λ (and hence ) are typically smaller for larger masses: stars are more compact and also less prone to deformation. Sampling the measurements from the high-mass range, where the differences between measurements are small but the errors are comparable to the low-mass measurements, should result in worse reconstruction in the high pressure and density range of the EOS.

In order to study this further also from the point of view of the choice of mass function, we performed additional simulations. Since the double-Gaussian function we initially adopted has its main distribution peak in the low-mass range (around the Chandrasekhar mass), the majority of generated observation points correspond to lower pressures and densities, which are precisely reconstructed by the algorithm. However, the high mass, and therefore the high pressure and density range, is covered sparsely; hence the corresponding high pressures and densities may be reconstructed less precisely. To test this explanation, a new training data using alternative NS mass distribution were prepared. We considered a uniform mass distribution in the range between 1 and 2.2 M. During the training on the uniform mass distribution data set, the ANN reached lower values of MSE with respect to the results presented in Sect. 4 with differences of around one order of magnitude in all considered cases. As a result the EOS reconstruction was characterised by smaller reconstruction errors for pressure and density; see examples of reconstruction in Fig. 7 for 20 observations with σM = 0.1 M and σR = 1 km. Moreover, predicted values probed range of higher values with respect to results from Sect. 4.2. The uniform mass distribution allow to generate observations close to maximum value of 2.2 M (including measurement uncertainties), whereas the previously used double-Gaussian function returned masses rarely higher than 2 M.

thumbnail Fig. 7.

ANN reconstructed EOS using the M(R) data for the SLy4 EOS model (left plot), the APR EOS model (middle plot) and the BSK20 EOS model (right plot) for the uniform NS mass distribution. The presented results were computed for input data consisting of 20 observations with measurement uncertainties equal to σM = 0.1 M, σR = 1 km. Dashed lines correspond to exact values obtained by solving the TOV equations.

Our results suggest that to efficiently probe the high-mass end of the NS distribution, either measurement uncertainties should be significantly decreased with respect to the low-mass range or coverage of masses should be more uniform. The first possibility may be feasible with the 3rd generation GW detectors, such as the Einstein Telescope (Maggiore et al. 2020). On the other hand, the EOS is accurately reconstructed for the low-mass range (low pressure and density regime), which offers the possibility of comparing nuclear parameters with the data from terrestrial experiments.

It is also worth mentioning that a precise reconstruction of EOS using ANN requires training data that is representative of the problem. In order to reconstruct astrophysical EOS models (SLy4, APR, and BSK20), we have selected an appropriate training set. However, ANN tested on different EOS covering different ranges for M, R, Λ, p, and ρ would result in a worse reconstruction. To avoid this problem, it’s necessary to optimise the parameter space of the training set and choose astrophysical models accordingly. It would be straightforward to expand the training dataset with a specific parametric description of dense matter, such as the MIT bag, to describe the de-confined quark matter (Chodos et al. 1974, see Sieniawska et al. 2019 for an example of piecewise relativistic polytrope EOS supplemented by quark EOS approximation of Zdunik 2000). In such a case, the ANN would potentially serve as a tool to discover the presence of exotic phases or signatures of dense-matter phase transitions.

6. Conclusions

We show that the ANN can be successfully applied in the reconstruction of the dense matter EOS from NS observations, either electromagnetic (masses and radii) or based on gravitational-wave measurements (masses and tidal deformabilities). We study the influence of the number of observations and the measurement uncertainties on the EOS reconstruction. The latter factor turned out to have a more significant effect on ANN performance, quantified in terms of the loss function (MSE). Furthermore, we show that the ANN trained on piecewise relativistic polytropes is capable of generalising the EOS reconstruction toward samples it wasn’t previously exposed to: realistic EOSs resulting from microscopic calculations: the SLy4, APR, and BSK20 EOS models.

We also introduce reconstruction errors for ANN: σρ and σp. The presented values vary proportionally to either the uncertainties of measurement with regard to the observables or to the values of pressures and densities. To decrease reconstruction errors, we suggest that either measurement uncertainties should be reduced, which is possible with the new generation of telescopes and detectors (i.e. Einstein Telescope for gravitational observations), or masses should be generated more uniformly. Moreover, we show that ANN can be successfully used in the reconstruction of radius based on the gravitational observables, which can be particularly useful for gravitational astronomy.

Among the many possibilities for further development in studies of NS parameters using ML methods, we plan to focus on the promising direction of variational auto-encoders. The latent space of these algorithms contain features that allow for an in-depth understanding of the distribution of parameters of the input data. Studies of the latent space could be used, for example, to infer information on the nuclear parameters of the EOS or assess the plausibility of the existence of a dense-matter phase transition.


1

Benefiting from the donation via the NVidia GPU seeding grant.

2

Prometheus, Academic Computer Centre CYFRONET AGH, Kraków, Poland.

Acknowledgments

The work was partially supported by the Polish National Science Centre grant no. 2016/22/E/ST9/00037 and the European Cooperation in Science and Technology COST action G2net no. CA17137. The Quadro P6000 used in this research was donated by the NVIDIA Corporation. This research was supported in part by PL-Grid Infrastructure (the Prometheus cluster).

References

  1. Abadi, M., Agarwal, A., Barham, P., et al. 2015, TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems, Software available from: https://tensorflow.org [Google Scholar]
  2. Abbott, B. P., Abbott, R., Abbott, T. D., et al. 2017a, Phys. Rev. Lett., 119, 161101 [NASA ADS] [CrossRef] [PubMed] [Google Scholar]
  3. Abbott, B. P., Abbott, R., Abbott, T. D., et al. 2017b, ApJ, 848, L13 [NASA ADS] [CrossRef] [Google Scholar]
  4. Abbott, B. P., Abbott, R., Abbott, T. D., et al. 2018, Phys. Rev. Lett., 121, 161101 [Google Scholar]
  5. Abbott, B. P., Abbott, R., Abbott, T. D., et al. 2019, Phys. Rev. X, 9, 011001 [Google Scholar]
  6. Abbott, B. P., Abbott, R., Abbott, T. D., et al. 2020, ApJ, 892, L3 [Google Scholar]
  7. Akmal, A., Pandharipande, V. R., & Ravenhall, D. G. 1998, Phys. Rev. C, 58, 1804 [NASA ADS] [CrossRef] [Google Scholar]
  8. Alsing, J., Silva, H. O., & Berti, E. 2018, MNRAS, 478, 1377 [NASA ADS] [CrossRef] [Google Scholar]
  9. Antoniadis, J., Freire, P. C. C., Wex, N., et al. 2013, Science, 340, 448 [Google Scholar]
  10. Chetlur, S., Woolley, C., Vandermersch, P., et al. 2014, ArXiv e-prints [arXiv:1410.0759] [Google Scholar]
  11. Chodos, A., Jaffe, R. L., Johnson, K., Thorn, C. B., & Weisskopf, V. F. 1974, Phys. Rev. D, 9, 3471 [NASA ADS] [CrossRef] [EDP Sciences] [MathSciNet] [Google Scholar]
  12. Chollet, F., et al. 2015, Keras, https://keras.io [Google Scholar]
  13. Cromartie, H. T., Fonseca, E., Ransom, S. M., et al. 2020, Nat. Astron., 4, 72 [NASA ADS] [CrossRef] [Google Scholar]
  14. De, S., Finstad, D., Lattimer, J. M., et al. 2018, Phys. Rev. Lett., 121, 091102 [Google Scholar]
  15. Demorest, P., Pennucci, T., Ransom, S., Roberts, M., & Hessels, J. 2010, Nature, 467, 1081 [NASA ADS] [CrossRef] [PubMed] [Google Scholar]
  16. Douchin, F., & Haensel, P. 2001, A&A, 380, 151 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  17. Fasano, M., Abdelsalhin, T., Maselli, A., & Ferrari, V. 2019, Phys. Rev. Lett., 123, 141101 [NASA ADS] [CrossRef] [Google Scholar]
  18. Ferreira, M., & Providência, C. 2019, ArXiv e-prints [arXiv:1910.05554] [Google Scholar]
  19. Flanagan, E. E., & Hinderer, T. 2008, Phys. Rev. D, 77, 021502 [NASA ADS] [CrossRef] [Google Scholar]
  20. Fonseca, E., Pennucci, T. T., Ellis, J. A., et al. 2016, ApJ, 832, 167 [Google Scholar]
  21. Fujimoto, Y., Fukushima, K., & Murase, K. 2018, Phys. Rev. D, 98, 023019 [NASA ADS] [CrossRef] [Google Scholar]
  22. Fujimoto, Y., Fukushima, K., & Murase, K. 2020, Phys. Rev. D, 101, 054016 [CrossRef] [Google Scholar]
  23. Goodfellow, I., Bengio, Y., & Courville, A. 2016, Deep Learning (The MIT Press) [Google Scholar]
  24. Goriely, S., Chamel, N., & Pearson, J. M. 2010, Phys. Rev. C, 82, 035804 [NASA ADS] [CrossRef] [Google Scholar]
  25. Haegel, L., & Husa, S. 2019, CQG, 37, 135005 [Google Scholar]
  26. Haensel, P., & Pichon, B. 1994, A&A, 283, 313 [NASA ADS] [Google Scholar]
  27. Haensel, P., Potekhin, A. Y., & Yakovlev, D. G. 2007, Neutron Stars 1 : Equation of State and Structure (New York, USA: Springer), 326, 1 [NASA ADS] [Google Scholar]
  28. Hernandez Vivanco, F., Smith, R., Thrane, E., et al. 2019, Phys. Rev. D, 100, 103009 [NASA ADS] [CrossRef] [Google Scholar]
  29. Hinton, G. E., & Zemel, R. S. 1993, Proceedings of the 6th International Conference on Neural Information Processing Systems, NIPS’93 (San Francisco, CA, USA: Morgan Kaufmann Publishers Inc.), 3 [Google Scholar]
  30. Holt, J. W., & Lim, Y. 2019, AIP Conf. Proc., 2127, 020019 [CrossRef] [Google Scholar]
  31. Kingma, D. P., & Ba, J. 2014, ArXiv e-prints [arXiv:1412.6980] [Google Scholar]
  32. Kramer, M. A. 1991, AIChE J., 37, 233 [CrossRef] [Google Scholar]
  33. Love, A. E. H. 1911, Some Problems of Geodynamics (Cambridge Univ. Press) [Google Scholar]
  34. Maggiore, M., Van Den Broeck, C., Bartolo, N., et al. 2020, J. Cosmol. Astropart. Phys., 2020, 050 [CrossRef] [Google Scholar]
  35. Miller, M. C., Lamb, F. K., Dittmann, A. J., et al. 2019, ApJ, 887, L24 [Google Scholar]
  36. Nickolls, J., Buck, I., Garland, M., & Skadron, K. 2008, Queue, 6, 40 [CrossRef] [Google Scholar]
  37. Oppenheimer, J. R., & Volkoff, G. M. 1939, Phys. Rev., 55, 374 [NASA ADS] [CrossRef] [Google Scholar]
  38. Press, W. H., Teukolsky, S. A., Vetterling, W. T., & Flannery, B. P. 1992, Numerical recipes in FORTRAN. The art of scientific computing, 2nd edn. (Cambridge University Press) [Google Scholar]
  39. Raithel, C. A., Özel, F., & Psaltis, D. 2016, ApJ, 831, 44 [NASA ADS] [CrossRef] [Google Scholar]
  40. Riley, T. E., Watts, A. L., Bogdanov, S., et al. 2019, ApJ, 887, L21 [NASA ADS] [CrossRef] [Google Scholar]
  41. Samuel, A. L. 1959, IBM J. Res. Dev., 3, 210 [CrossRef] [Google Scholar]
  42. Sieniawska, M., Turczański, W., Bejger, M., & Zdunik, J. L. 2019, A&A, 622, A174 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  43. Steiner, A. W., Lattimer, J. M., & Brown, E. F. 2010, ApJ, 722, 33 [Google Scholar]
  44. Steiner, A. W., Lattimer, J. M., & Brown, E. F. 2013, ApJ, 765, L5 [Google Scholar]
  45. Tolman, R. C. 1939, Phys. Rev., 55, 364 [NASA ADS] [CrossRef] [Google Scholar]
  46. Tooper, R. F. 1965, ApJ, 142, 1541 [NASA ADS] [CrossRef] [Google Scholar]
  47. Traversi, S., Char, P., & Pagliara, G. 2020, ApJ, 897, 165 [CrossRef] [Google Scholar]
  48. Van Oeveren, E. D., & Friedman, J. L. 2017, Phys. Rev. D, 95, 083014 [NASA ADS] [CrossRef] [Google Scholar]
  49. Wade, L., Creighton, J. D. E., Ochsner, E., et al. 2014, Phys. Rev. D, 89, 103012 [NASA ADS] [CrossRef] [Google Scholar]
  50. Zdunik, J. L. 2000, A&A, 359, 311 [NASA ADS] [Google Scholar]
  51. Zhang, N.-B., Qi, B., & Wang, S.-Y. 2020, Chinese Phys. C, 44, 064103 [NASA ADS] [CrossRef] [Google Scholar]

All Tables

Table 1.

Ranges of piecewise polytrope EOS parameters used in the study.

Table 2.

Reconstruction error ranges for σp and σρ of the ANN for studied measurement uncertainties in case of EOS reconstruction for the M(R) data (upper table) and M() data (lower table).

All Figures

thumbnail Fig. 1.

Top panels: left – sample M(R) relations; right – corresponding M() relations. Red curve and corresponding observations: example of a training datum, consisting on 20 points from the M(R) and M(), selected assuming random normal distribution with the mean value equal to true values and standard deviations σM = 0.1 M, σR = 1.0 km and σ = 1.0 km. The configurations with the same M and R have, in general, different . Bottom panels: left – schematic of a model EOS, composed in the low-density part from the realistic crust of the SLy4 EOS (Haensel & Pichon 1994; Douchin & Haensel 2001) and piecewise relativistic polytropes (Tooper 1965), right – mass-radius M(R) relations generated using the piecewise relativistic polytrope model (thin solid grey curves) and astrophysical models of NS sequences, based on the following EOSs: the SLy4 EOS (solid blue curve), the APR EOS (dashed red curve) and the BSK20 EOS (dash-dotted green curve).

In the text
thumbnail Fig. 2.

Dependence of the MSE (ANN loss function described in Sect. 3.2) as a function of number of observations N and measurement uncertainties in the case of EOS estimation based on M(R) (left figure) and M() (right figure) observations.

In the text
thumbnail Fig. 3.

Top panels: example of the input data (M(R) measurements with errors, left plot), and corresponding output data from ANN (p(ρ) relation, right plot) for the estimation of EOS from M(R). Both input samples consist of 20 observations with masses randomly selected from a mass distribution (Sect. 3.2) and measurement uncertainties equal to σM = 0.1 M, σR = 1 km (blue sample), σM = 0.01 M, σR = 0.01 km (red sample). For the description of the uncertainties on the output, see the text. Bottom panels: example of the input data (M() measurements with errors, left plot), and corresponding output data from ANN (p(ρ) relation, right plot) for the estimation of EOS from M(). Both input samples consist of 20 observations with masses randomly selected from a mass distribution (Sect. 3.2) and measurement uncertainties equal to σM = 0.1 M, σR = 1 km (blue sample), σM = 0.01 M, σ = 0.01 km (red sample). Dashed curves correspond to original (ground-truth, error-free) sequences of input and output of the TOV equations. Presented examples correspond to different EOSs.

In the text
thumbnail Fig. 4.

Top panels: ANN-reconstructed EOS from the M(R) data for the SLy4 EOS model (left plot), the APR EOS model (middle plot) and the BSK20 EOS model (right plot). Results are computed for the input M(R) data consisting of 20 observations with measurement uncertainties equal to merr = 0.1 M, err = 1 km. Bottom panels: ANN-reconstructed EOS from the M() data for the same EOS as in the top panels. Results for the input M() data consisting of 20 observations with measurement uncertainties equal to merr = 0.1 M, err = 1 km. Dashed lines correspond to the exact EOS relations.

In the text
thumbnail Fig. 5.

Evolution of MSE (ANN loss) in the function of number of observations and measurement uncertainties in case of R computation based on M().

In the text
thumbnail Fig. 6.

Top panels: example of the input data (left plot) and corresponding output data from ANN (right plot) in the case of R computation based on M() for the piecewise relativistic polytropes. Bottom panels: radius reconstructed from the M() data for the SLy4 EOS model (left plot), the APR EOS model (middle plot), and the BSK20 EOS model (right plot). All results were computed for the input M(R) data consisting of 20 observations with measurement uncertainties equal to σM = 0.1 M, σ = 1 km. Dashed lines correspond to exact values obtained by solving the TOV equations.

In the text
thumbnail Fig. 7.

ANN reconstructed EOS using the M(R) data for the SLy4 EOS model (left plot), the APR EOS model (middle plot) and the BSK20 EOS model (right plot) for the uniform NS mass distribution. The presented results were computed for input data consisting of 20 observations with measurement uncertainties equal to σM = 0.1 M, σR = 1 km. Dashed lines correspond to exact values obtained by solving the TOV equations.

In the text

Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.

Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.

Initial download of the metrics may take a while.