This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

The BACCO Simulation Project: Exploiting the full power of large-scale structure for cosmology

Raul E. Angulo,1,2 Matteo Zennaro1, Sergio Contreras1, Giovanni Aricò,1 E-mail:[email protected]    Marcos Pellejero-Ibañez1, & Jens Stücker1.
1Donostia International Physics Center (DIPC), Paseo Manuel de Lardizabal, 4, 20018, Donostia-San Sebastián, Guipuzkoa, Spain.
2IKERBASQUE, Basque Foundation for Science, 48013, Bilbao, Spain.
(Accepted XXX. Received YYY; in original form ZZZ)
Abstract

We present the BACCO project, a simulation framework specially designed to provide highly-accurate predictions for the distribution of mass, galaxies, and gas as a function of cosmological parameters. In this paper, we describe our main suite of gravity-only simulations (L2L\sim 2\,Gpc and 432034320^{3} particles) and present various validation tests. Using a cosmology-rescaling technique, we predict the nonlinear mass power spectrum over the redshift range 0<z<1.50<z<1.5 and over scales 102<k/(hMpc1)<510^{-2}<k/(h\,{\rm Mpc}^{-1})<5 for 800800 points in an 88-dimensional cosmological parameter space. For an efficient interpolation of the results, we build an emulator and compare its predictions against several widely-used methods. Over the whole range of scales considered, we expect our predictions to be accurate at the 2%2\% level for parameters in the minimal Λ\LambdaCDM model and to 3%3\% when extended to dynamical dark energy and massive neutrinos. We make our emulator publicly available under http://www.dipc.org/bacco

keywords:
large-scale structure – numerical methods – cosmological parameters
pubyear: 2019pagerange: The BACCO Simulation Project: Exploiting the full power of large-scale structure for cosmologyThe BACCO Simulation Project: Exploiting the full power of large-scale structure for cosmology

1 Introduction

Over the last two decades, our understanding of the Universe has grown tremendously: the accelerated expansion of the Universe and the existence of dark matter are firmly supported; there are strong constraints on the micro-physical properties of dark matter and neutrinos; and the statistics of the primordial cosmic fluctuations are measured with increasing precision (e.g. Alam et al., 2017; Planck Collaboration et al., 2020; Gilman et al., 2020).

Despite the progress, there are several tensions among current data when interpreted within the context of the Λ\LambdaCDM model. For instance, the value of the Hubble parameter inferred from local supernovae is significantly larger than that inferred from the analysis of the CMB (e.g. Riess, 2019; Freedman et al., 2019). Another example is that the amplitude of fluctuations, S8=Ωm/0.3σ8S_{8}=\sqrt{\Omega_{m}/0.3}\,\sigma_{8}, as determined from low-redshift lensing measurements, appears smaller than that inferred from CMB (e.g. Asgari et al., 2020).

In the current era of precision cosmology, the large amount of available large-scale structure (LSS) data promises accurate measurements of cosmological parameters with small systematic errors. These future measurements could shed light on the aforementioned tensions, confirming or ruling out the Λ\LambdaCDM paradigm (e.g. Weinberg et al., 2013). Many observational campaigns that seek to obtain this data are under construction or with an imminent start (e.g. Euclid111https://sci.esa.int/web/euclid, DESI222https://www.desi.lbl.gov/, J-PAS333http://j-pas.org).

To fully exploit the upcoming LSS data and obtain these cosmological constraints, extremely accurate theoretical models are required. This is an active area of research with different approaches being adopted in the literature. On the one hand, recent advances in perturbation theory have increased considerably the range of scales that can be treated analytically (e.g. Desjacques et al., 2018). These scales are, however, still in the quasi-linear regime. On the other hand, cosmological numerical simulations have also steadily increased their robustness and accuracy (e.g. Kuhlen et al., 2012). Currently, simulations are the most accurate method to model smaller and non-linear scales (where, in principle, much more additional constraining power resides).

Traditionally, numerical simulations were very expensive computationally and suffered from large cosmic-variance errors, thus they were only used to calibrate fitting functions or combined with perturbation theory to provide predictions for nonlinear structure as a function of cosmology (e.g. Smith et al., 2003; Takahashi et al., 2012). This has changed recently, as the available computational power keeps increasing and algorithms become more efficient, it is now possible carrying out large suites of simulations spanning different cosmological models. The so-called emulators have become more popular, which interpolate simulation results to provide predictions in the nonlinear regime and for biased tracers (e.g. Heitmann et al., 2014; Nishimichi et al., 2019; Liu et al., 2018; DeRose et al., 2019; Giblin et al., 2019; Wibking et al., 2019).

To keep the computational cost affordable, emulators are typically restricted to a small region in parameter space which is sampled with a small number of simulations (50100\sim 50-100). Additionally, each individual simulation is low resolution or simulates a relatively small cosmic volume. This limits their usability in actual data analyses and/or add a significant source of uncertainty.

Here we take advantage of several recent advances to solve these limitations. First, we employ a very efficient NN-body code to carry out a suite of 6 large-volume high-resolution simulations, which allows to resolve all halos and subhalos with mass >5×1010h1M>5\times 10^{10}h^{-1}{\rm M_{\odot}}, together with their merger histories. Second, we employ initial conditions with suppressed variance, which allows to predict robustly even scales comparable to our simulated volume. Third, these simulations are combined with cosmology-rescaling algorithms, so that predictions can be obtained for any arbitrary set of cosmological parameters. As a result, this approach allows us to make highly accurate predictions for the large-scale phase-space structure of dark matter, galaxies, and baryons.

Our approach has many advantages over others in the literature. Firstly, we can predict the matter distribution over a broad range of scales, with high force accuracy and over a large cosmic volume. This allows for detailed modellings of the distribution of gas and the impact of “baryonic effects” (van Daalen et al., 2011; Schneider et al., 2016; Chisari et al., 2018, 2019; Aricò et al., 2020b; van Daalen et al., 2020). We can also resolve collapsed dark matter structures and their formation history, which enables sophisticated modelling of the galaxies that they are expected to host (e.g. Henriques et al., 2020; Moster et al., 2018; Chaves-Montero et al., 2016). In addition, the cosmological parameter space is large and densely sampled, so that emulator uncertainties are kept under a desired level. Finally, the parameter space includes non-standard Λ\LambdaCDM parameters, dynamical dark energy and massive neutrinos.

As an initial application of our framework, we have used our suite of specially-designed simulations to predict the nonlinear mass power spectrum over the range 0<z<1.50<z<1.5 for 800800 different cosmologies within an 8-dimensional parameter space defined by a 10σ\sim 10\sigma volume around Planck’s best-fit values. From these, we construct and present an emulator so that these predictions are easily accessible to other researchers. Overall, we reach a few percent accuracy over the whole range of parameters considered.

Our paper is structured as follows. Section 2 is devoted to presenting our numerical simulations and to the description and validation of numerical methods. In Section 3 we describe the construction of an emulator for the nonlinear power spectrum. We highlight Section 3.1.1 where we discuss our strategy for selecting training cosmologies, Section 3.1.4 which discusses our power spectra measurements, and Section 3.2.3 that compares our predictions to other approaches. We summarise our results and discuss the implications of our work in Section 4.

2 Numerical methods

2.1 The BACCO Simulations

Refer to caption
Figure 1: The projected mass density field in nenya, one of our six BACCO simulations, at z=0z=0. Each image corresponds to a 25h1Mpc25~{}h^{-1}{\rm Mpc} deep projection employing a tri-cubic Lagrangian interpolation method. Top, middle, and bottom panels progressively zoom into a 1440h1Mpc1440\,h^{-1}{\rm Mpc}, 360h1Mpc360\,h^{-1}{\rm Mpc}, and 90h1Mpc90\,h^{-1}{\rm Mpc}-wide region of the simulation.

The “BACCO simulations” is a suite of 6 NN-body simulations that follow the nonlinear growth of dark matter structure within a cubic region of L=1440h1Mpc1440h^{-1}{\rm Mpc} on a side. These calculations are performed for three different sets of cosmological parameters with two realizations each. The matter distribution is represented by 432034320^{3} (80\sim 80 billion) particles.

The three cosmologies adopted by our BACCO simulations are provided in Table 1. We dub these cosmologies narya, nenya, and vilya444narya, vilya, and nenya are the most powerful rings after Sauron’s “One Ring” in “The Lord of the Rings” mythology.. These sets are inconsistent with the latest observational constraints, but they were intentionally chosen to so that they can be efficiently combined with cosmology-rescaling algorithms (Angulo & White, 2010). Specifically, Contreras et al. (2020) showed that these 3 cosmologies are optimal, in terms of accuracy and computational cost, to cover all the cosmologies within a region of 10σ\sim 10\sigma around the best values found by a recent analysis of the Planck satellite (Planck Collaboration et al., 2014a).

For narya and vilya we stored 5050 snapshots, equally log-spaced in expansion factor, aa, and adopt a Plummer-equivalent softening length of ϵ=6.7h1kpc\epsilon=6.7h^{-1}{\rm kpc}. As pointed out by Contreras et al. (2020), most of the cosmological parameter volume is covered by rescaling nenya, thus we have increased the force resolution and frequency of its outputs to ϵ=5h1kpc\epsilon=5h^{-1}{\rm kpc} and 100100 snapshots, respectively. All simulations were started at a=0.02a=0.02 using 2nd-order Lagrangian Perturbation theory, and were evolved up to a=1.25a=1.25 so that they can be accurately scaled to cosmologies with large amplitude of fluctuations.

For each of the three cosmologies, we carry out two realizations with an initial mode amplitude fixed to the ensemble mean, and opposite Fourier phases (Angulo & Pontzen, 2016; Pontzen et al., 2016). These “Paired-&-Fixed”  initial conditions allow for a significant reduction of cosmic variance in the resulting power spectrum in the linear and quasi-linear scales (which are the most affected by cosmic variance), as it has been tested extensively in recent works (Chuang et al., 2019; Villaescusa-Navarro et al., 2018; Klypin et al., 2020).

The BACCO simulations were carried out in the Summer of 2019 at Marenostrum-IV at the Barcelona Supercomputing Center (BSC) in Spain. We ran our NN-body code in a hybrid distributed/shared memory setup employing 81928192 cores using 40964096 MPI tasks. The run time was 7.2 million CPU hours, equivalent of 3535 days of wall-clock time. The total storage required for all data products is about 8080 TB.

To compare against recent emulator projects, we notice these simulations have 1010 times better mass resolution, 44 times better force resolution, and 2×32\times 3 the volume of those used by the AEMULUS project (DeRose et al., 2019); 50 times the volume and 3 times better mass resolution than MassiveNuS (Liu et al., 2018); 66 times better mass resolution, 50% larger volume, and twice the spatial resolution than those in the EUCLID emulator project (Euclid Collaboration et al., 2019); 1010 times more particles and slightly better force resolution than the runs of the DarkEmulator (Nishimichi et al., 2019); 22 times better mass resolution and similar force resolution and volume to the simulations in the Mira-Titan Universe project (Heitmann et al., 2014). All this, of course, is mostly a result of our project being able to focus computational resources on only 6 simulations, but that is precisely the advantage of the cosmology rescaling approach.

In Fig. 1 we display the projected matter density field of one of our simulations, nenya  at z=0z=0. Each panel shows a region of the simulated volume, the top panel a region 1440h1Mpc1440\,h^{-1}{\rm Mpc} wide – the full simulation side-length –, whereas middle and bottom panels zoom in regions 960960 and 360h1Mpc360\,h^{-1}{\rm Mpc} wide, respectively. We display the density field as estimated via a tri-cubic Lagrangian tessellation using only 108031080^{3} particles. Note that thanks to this interpolation, no particle discreteness is visible and filaments and voids become easily distinguishable.

Table 1: The cosmological parameters of the three “BACCO” simulations presented in this paper: Vilya, Nenya and Narya
σ8\rm\sigma_{8} Ωm\rm\Omega_{\rm m} Ωb\rm\Omega_{b} ns\rm n_{s} h\rm h MνM_{\nu} w0w_{0} waw_{a} LL [h1Mpch^{-1}{\rm Mpc}] mpm_{p} [h1Mh^{-1}{\rm M_{\odot}}] ϵ\epsilon [h1kpch^{-1}{\rm kpc}] NpN_{p}
Vilya 0.9 0.27 0.06 0.92 0.65 0.0 -1.0 0.0 1440 2.77×1092.77\times 10^{9} 6.7 432034320^{3}
Nenya 0.9 0.315 0.05 1.01 0.60 0.0 -1.0 0.0 1440 3.2×1093.2\times 10^{9} 5 432034320^{3}
Narya 0.9 0.36 0.05 1.01 0.70 0.0 -1.0 0.0 1440 3.7×1093.7\times 10^{9} 6.7 432034320^{3}

2.2 The simulation code

The NN-body code we employ is an updated version of L-Gadget3 (Springel, 2005; Angulo et al., 2012). This code was originally developed for the Millennium-XXL project and was successfully executed with more than 10,000 CPUs employing 30Tb of RAM. Compared with previous versions of Gadget, L-Gadget3 features an hybrid OpenMP/P-thread/MPI parallelisation strategy and improved domain decomposition algorithms.

In addition to these improvements, our updated L-Gadget3 version stores all outputs in the HDF5 format, implements the possibility of simulating massive neutrinos via the linear response approach (Ali-Haïmoud & Bird, 2013), and features an improvement in the Tree-PM force split and in Kick-Drift operators. The output data structure is such to allow a straightforward reconstruction of the full phase-space distribution via tri-cubic Lagrangian interpolation (Hahn & Angulo, 2016; Stücker et al., 2020).

The code carries out a large fraction of the required post-processing on the fly. Specifically, this includes the 2LPT initial conditions generator, group finding via Friends-of-Friends algorithms and an improved version of SUBFIND (Springel et al., 2001) that is also able to track tidally-disrupted structures, and a descendant finder and merger tree construction (c.f. §2.2.1).

Thanks to the in-lining of these tools, it is not necessary to store the full particle load at every snapshot, which significantly reduces the I/O and long-term storage requirements. The dark matter distribution is, however, very useful in many applications, thus L-Gadget3 is able to store a subset of particles sampling homogeneously the initial Lagrangian distribution.

2.2.1 SubFind and Group finders

The identification of bound structures is a key aspect of NN-body simulations, thus L-Gadget3 features a version of the SUBFIND algorithm with several improvements.

The first improvement is the ability to track subhalos on-the-fly across snapshots – defining progenitor and descendants –, computing various additional quantities such as peak halo mass, peak maximum circular velocity, infall subhalo mass, and mass accretion rate, among others. These properties become useful when modelling galaxy formation within gravity-only simulations (e.g. Chaves-Montero et al., 2016; Moster et al., 2018).

The second improvement of our updated version of SUBFIND is the use of the subhalo catalogue in the previous snapshot to better identify structures. In the original algorithm, particles are first sorted according to the local density, then when a saddle point is detected, the most massive group at that point is considered as the primary structure. This, however, can cause inconsistencies across time, as small changes can lead to fluctuations in the structure considered as primary. In our version of SUBFIND instead, when a saddle point is detected, we consider as primary the substructure whose main progenitor is the most massive. This has proven to return more stable merger trees and quantities that are not local in time (e.g. peak maximum circular velocity).

Finally, during their evolution, substructures can disappear from a simulation due to artifacts in the structure finder, finite numerical resolution, or because its mass falls below the resolution limit owing to tidal stripping. The last of our improvements to SUBFIND is the ability to track all subhalos with no recognizable descendant, keeping the position and velocity of their most bound particle. This is a indispensable feature to correctly model the small-scale galaxy clustering in dark matter simulations (Guo & White, 2014).

At every output time, we store FoF groups and SUBFIND subhalos with more than 10 particles. In total, there are approximately 129129 billion groups and 214214 billion subhalos in our outputs. This means our simulations are able to resolve halos with mass 5×1010h1M5\times 10^{10}h^{-1}{\rm M_{\odot}}, and subhalos with a number density of 0.1h3Mpc30.1\,h^{3}{\rm Mpc}^{-3} at z=0z=0.

2.3 Validation

2.3.1 Time and force resolution

Refer to caption
Figure 2: The impact of numerical parameters in our simulated nonlinear mass power spectra at z=0z=0 and z=1z=1. We display fractional differences with respect to the measurements in a simulation that adopts the same accuracy parameters as our main BACCO simulations. The grey bands indicate a region of ±1%\pm 1\%.

In order to quantify the accuracy of our results, we have carried out a suite of small, L=64h1MpcL=64\,h^{-1}{\rm Mpc} box, simulations where we systematically vary the numerical parameters around those adopted in the BACCO simulations. Specifically, we consider the time-integration accuracy, force calculation accuracy, softening length, and mass resolution. We adopt the parameters of the nenya simulation as the reference point for these tests.

In Fig. 2 we compare the nonlinear power spectrum between these test runs and one that adopts the same numerical parameters and mean interparticle separation as our main BACCO simulations. Solid and dashed lines denote the results at z=0z=0 and z=1z=1, respectively. In all panels the grey shaded region indicates ±1%\pm 1\% agreement.

In the top panel we display simulations with different mass resolutions, N1/3/L=[2,3,4]N^{1/3}/L=[2,3,4]. The main effect of mass resolution is how well small non-linear structures are resolved. We see no significant effect up to k5hMpc1k\sim 5h\,{\rm Mpc}^{-1} and a mild increase, of about 1% in the power at k10hMpc1k\sim 10h\,{\rm Mpc}^{-1} when improving the mass resolution.

The main source of inaccuracies in the force calculation are the terms neglected in the oct-tree multipole expansion. Specifically, Gadget considers only the monopole contribution down to tree-nodes of mass MM and size \ell that fulfill GM2r2(r)>α|a|\frac{GM^{2}}{r^{2}}\left(\frac{\ell}{r}\right)>\alpha|\vec{a}| for a particle at a distance rr and acceleration |a||\vec{a}|. Thus, the accuracy in the force calculation is controlled by the parameter α\alpha. In the second panel we vary this parameter and see that the power spectrum is converged at a sub-percent level.

As in previous versions of Gadget, time-steps in L-Gadget3 are computed individually for each particle as 2ηϵ/|a|\sqrt{2\eta\epsilon/|\vec{a}|}, where ϵ\epsilon is the softening and |a||\vec{a}| is the magnitude of the acceleration in the previous timestep. The parameter η\eta therefore controls how accurately orbits are integrated. In the third panel we vary this parameter increasing/decreasing it by a factor of 4/2 with respect to our fiducial value, η=0.05\eta=0.05. We see that the power spectrum varies almost negligibly with less than a 1%1\% impact up to k10hMpc1k\sim 10h\,{\rm Mpc}^{-1}.

Perhaps the most important degree of freedom in a numerical simulation is the softening length, ϵ\epsilon, a parameter that smooths two-body gravitational interactions (note that in Gadget, forces become Newtonian at a distance 2.7ϵ2.7\epsilon). In the fourth panel we compare three simulations with 50% higher and lower values of ϵ\epsilon to values equal to 1/1401/140 and 1/35th1/35-th of the mean interparticle separation. On small scales we see that the amount of power increases systematically the lower the value of the softening length. In particular, our fiducial configuration underestimates the power by 1(2.5)1(2.5)% at k5(10)hMpc1k\sim 5(10)\,h\,{\rm Mpc}^{-1}.

In summary, our results appear converged to better than 1% up to k5hMpc1k\sim 5h\,{\rm Mpc}^{-1}, and to 3%\sim 3\% up to k10hMpc1k\sim 10\,h\,{\rm Mpc}^{-1}. The main numerical parameter preventing better convergence is the softening length. To mitigate its impact, we have adopted the following empirical correction to our power spectrum results

P(k)P(k)[1+1.62×102(erfc(ln(k/ϵ)3.1)2)]P(k)\rightarrow P(k)\,\left[1+1.62\times 10^{-2}({\rm erfc}(\ln(k/\epsilon)-3.1)-2)\right] (1)

where ϵ\epsilon is the softening length in units of hMpc1h\,{\rm Mpc}^{-1}, and erfc{\rm erfc} is the complementary error function. We found this expression by fitting the effect seen in the bottom panel of Fig. 2. We will apply this correction by default which brings the expected nominal accuracy of our power spectrum predictions to 1%\sim 1\% on all scales considered.

2.3.2 The EUCLID comparison project

Refer to caption
Figure 3: The nonlinear mass power spectrum at z=0z=0 of the “Euclid code comparison project”. Each coloured curve displays the predictions of a different NN-body code, as indicated by the legend, and we display ratios relative to our simulation result corrected by finite numerical resolution. Note that all NN-body codes agree to better than 11% precision up to k5hMpc1k\sim 5h\,{\rm Mpc}^{-1}, with the exception of the original Gadget3 run presented in Schneider et al (2016).
Refer to caption
Figure 4: Predictions from the BACCO simulations, vilya, nenya, and narya, at z=0z=0. Left panel shows the nonlinear mass power spectrum as solid lines, and linear perturbation theory as dotted lines. The middle panel shows the redshift-space correlation function for subhalos with a number density of 103h3Mpc310^{-3}\,h^{3}\,{\rm Mpc}^{-3}, where solid and dashed lines show the results for each of the two opposite-phase simulations. The right panel shows the Friends-of-Friends halo mass function, with vertical dashed lines indicating the mass limit resolved with 10 and 100 particles.

To further validate our NN-body code and quantify the accuracy of our numerical simulations, we have carried out the main simulation of the “Euclid code comparison project”, presented in Schneider et al. (2016). This simulation consists of 204832048^{3} particles of mass 1.2×109h1M1.2\times 10^{9}\,h^{-1}{\rm M_{\odot}} in a 500h1Mpc500\,h^{-1}{\rm Mpc} box, and has been carried out with several NN-body codes: RAMSES (Teyssier, 2002), PkdGrav3 (Potter et al., 2017), Gagdet3 , and recently with ABACUS by Garrison et al. (2019).

Our realisation of this simulation adopts the same force and time-integration accuracy parameters as of main BACCO simulations, and the same softening length as nenya, ϵ=5h1kpc\epsilon=5\,h^{-1}{\rm kpc}. The full calculation required 62306230 timesteps and took 3×1053\times 10^{5} CPU hours employing 10241024 MPI Tasks.

In Fig. 3 we compare the resulting power spectra at z=0z=0. We display the ratio with respect to our L-Gadget3 results including the correction provided in Eq. 1. For comparison, we display the uncorrected measurement as a blue dashed line. Note the spikes on large scales are caused by noise due to a slightly different kk-binning in the spectra.

Our results, PkdGrav, RAMSES, and that of ABACUS agree to a remarkable level – they differ by less than 1% up to k5hMpc1k\sim 5h\,{\rm Mpc}^{-1}. This is an important verification of the absolute accuracy of our results, but it is also an important cross-validation of all these 4 NN-body codes. Interestingly, the Gadget3 result presented in Schneider et al. (2016) is clearly in tension with the other 44 codes. Since the underlying core algorithms in our code and in Gadget3 are the same, the difference is likely a consequence of numerical parameters adopted by Schneider et al. (2016) not being as accurate as those for the other runs. In the future, it will be important to conduct code comparison projects were numerical parameters are chosen so that each code provides results converged to a given degree.

2.3.3 The BACCO simulations

Having validated our numerical setup, we now present an overview of the results of our BACCO simulations at z=0z=0 in Fig. 4.

In the left panel we show the nonlinear power spectrum in real space. Firstly, we see the low level of random noise in our predictions owing to the “Paired-&-Fixed”  method. On large scales, we have checked that our results agree at the 0.5%0.5\% level with respect to the linear theory solution, which we compute for each cosmology using the Boltzmann code CLASS (Lesgourgues, 2011). Also on large scales, we see that the three simulations display significantly different power spectra, despite the three of them having identical values for σ8\sigma_{8}. This is mostly a consequence of their different primordial spectral index, nsn_{s}. In contrast, on scales smaller than k0.2hMpc1k\sim 0.2h\,{\rm Mpc}^{-1}, the spectra become much more similar as their linear spectra also do (shown by dotted lines).

In the middle panel we show the monopole of the redshift-space correlation functions for subhalos with a spatial number density equal to 103h3Mpc310^{-3}\,h^{3}{\rm Mpc}^{-3} selected according to their peak maximum circular velocity. This is roughly analogous to a stellar mass selection above 5×1010h1M5\times 10^{10}h^{-1}{\rm M_{\odot}}. We see a significant difference in the correlation amplitude among simulations. This hints at the potential constraining power of LSS if a predictive model for the galaxy bias is available. We display each of the two “Paired-&-Fixed” simulations in each cosmology as solid and dashed lines. Unlike in the power spectrum plot, the effect of pairing the initial phase fields is visible.

Finally, in the right panel of Fig. 4 we display the mass function of halos identified by the FoF algorithm. As in previous cases, there are clear differences among cosmologies. The minimum halo mass in our simulations is 34×1010h1M\sim 3-4\times 10^{10}\,h^{-1}{\rm M_{\odot}}, which should suffice to model galaxies with star formation rates above 10h1M10\,h^{-1}{\rm M_{\odot}}/year at z1z\sim 1 (Orsi & Angulo, 2018), as expected to be observed by surveys like EUCLID, DESI, or J-PAS (Favole et al., 2017).

In the next section we will employ our BACCO simulations to predict the nonlinear mass power spectrum as a function of cosmology.

3 Nonlinear mass power spectrum Emulator

Our aim is to make fast predictions for the nonlinear power spectrum over the whole region of currently-viable cosmologies. In this section we describe how we achieve this by building (§3.1) and testing of a matter power spectrum emulator (§3.2).

Our basic strategy is the following:

  • First, we define a target region in cosmological parameter space (§3.1.1) and iteratively select a set of training points that minimise the emulator uncertainty (§3.1.2).

  • Second, we use the cosmology-rescaling approach over the outputs of our BACCO simulations (§3.1.3) to predict the power spectra in those training cosmologies (§3.1.4).

  • Finally, employ those measurements to build an emulator for the power spectra based on either Gaussian Processes or Neural Networks (§3.1.7).

We test the performance and accuracy of our emulator (§3.2.2) and compare it against widely-used methods to predict the nonlinear power spectrum (§3.2.3).

3.1 Building the emulator

Refer to caption
Figure 5: Distribution of cosmologies employed to build our BACCO emulator. Blue symbols display the location of cosmologies in our initial training set, whereas orange and green symbols display those subsequently selected by our iterative method. In each plot, the limits coincide with the full range of values considered (c.f. Eq 3.1.1).

3.1.1 The parameter space

In this paper we aim to cover 8 parameters of the Λ\LambdaCDM model extended with massive neutrinos and dynamical dark energy. Specifically, we consider the parameter range:

σ8\displaystyle\sigma_{8} \displaystyle\in [0.73,0.9]\displaystyle[0.73,0.9]
Ωm\displaystyle\Omega_{\rm m} \displaystyle\in [0.23,0.4]\displaystyle[0.23,0.4]
Ωb\displaystyle\Omega_{b} \displaystyle\in [0.04,0.06]\displaystyle[0.04,0.06]
ns\displaystyle n_{s} \displaystyle\in [0.92,1.01]\displaystyle[0.92,1.01] (2)
h[100kms1Mpc1]\displaystyle h\,[100\,{\rm km}\,{\rm s^{-1}}{\rm Mpc^{-1}}] \displaystyle\in [0.6,0.8]\displaystyle[0.6,0.8]
Mν[eV]\displaystyle M_{\nu}\,[{\rm eV}] \displaystyle\in [0.0,0.4]\displaystyle[0.0,0.4]
w0\displaystyle w_{0} \displaystyle\in [1.15,0.85]\displaystyle[-1.15,-0.85]
wa\displaystyle w_{a} \displaystyle\in [0.3,0.3]\displaystyle[-0.3,0.3]

where MνM_{\nu} is the total mass in neutrinos, σ8\sigma_{8} is the r.m.s. linear cold mass (dark matter plus baryons) variance in 8h1Mpc8h^{-1}{\rm Mpc} spheres, and w0w_{0} and waw_{a} define the time evolution of the dark energy equation of state: w(z)=w0+(1a)waw(z)=w_{0}+(1-a)w_{a}. The parameter range we consider for (σ8\sigma_{8}, Ωm\Omega_{m}, Ωb\Omega_{b}, nsn_{s}) corresponds to a 10σ\sim 10\sigma region around the best-fit parameters of the analysis of Planck Collaboration et al. (2014b). For the dimensionless Hubble parameter, hh, we expand the range to cover a 4σ4\sigma region around current low-redshift measurements from supernovae data (Riess, 2019).

We assume a flat geometry, i.e. Ωk=0\Omega_{k}=0 and Ωm+Ωw+Ων=1\Omega_{m}+\Omega_{w}+\Omega_{\nu}=1. We keep fixed the effective number of relativistic species to Neff=3.046N_{\rm eff}=3.046, and the temperature of the CMB TCMB=2.7255T_{\rm CMB}=2.7255 K and neglect the impact of radiation (i.e. Ωr=0\Omega_{r}=0). Note, however, that it is relatively straightforward to relax these assumptions and include in our framework varying curvature, relativistic degrees of freedom, or other ingredients.

For comparison, we notice that the range of parameters covered by our emulator is approximately twice as large as that of the Euclid emulator (with the exception of w0w_{0}, which is similar), which implies a parameter space volume 200\sim 200 times larger, and that it covers cosmological parameters beyond the minimal Λ\LambdaCDM: dynamical dark energy and massive neutrinos. In contrast, our parameter space is similar to that considered by the Mira-Titan Universe project (Heitmann et al., 2016; Lawrence et al., 2017), however, as we will discuss next, we cover the space with approximately 20 times the number of sampling points.

3.1.2 Training Cosmology Set

We now define the cosmologies with which we will train our emulator. This is usually done by sampling the desired space with a Latin-Hypercube (e.g. Heitmann et al., 2006). We adopt a slightly different strategy based on the idea of iterative emulation of Pellejero-Ibañez et al. (2020) (see also Rogers et al., 2019), where we preferentially select training points in regions of high emulator uncertainty.

Let us first define the quantity we will emulate:

Q(k,z)log[P(k,z)/PlinearsmearedBAO(k,z)],Q(k,z)\equiv\log[P(k,z)/P_{\rm linear}^{\rm smeared-BAO}(k,z)], (3)

where P(k)P(k) is the measured nonlinear power spectrum of cold matter (i.e. excluding neutrinos). PlinearsmearedBAOP_{\rm linear}^{\rm smeared-BAO} is the linear theory power spectrum where the BAO have been smeared out according to the expectations of perturbation theory. Specifically, we define:

PlinearsmearedBAOPlinearG(k)+PlinearnoBAO[1G(k)]P_{\rm linear}^{\rm smeared-BAO}\equiv P_{\rm linear}\,G(k)+P_{\rm linear}^{\rm no-BAO}\,[1-G(k)] (4)

where PlinearnoBAOP_{\rm linear}^{\rm no-BAO} is a version of the linear theory power spectrum without any BAO signal. Operationally, this is obtained by performing a discrete sine transform, smooth the result, and return to Fourier space by an inverse transform (Baumann et al., 2018). G(k)exp[0.5k2/k]G(k)\equiv\exp[-0.5\,k^{2}/k_{*}], with k1(3π2)1dkPlineark_{*}^{-1}\equiv(3\pi^{2})^{-1}\int{\rm d}k\,P_{\rm linear}.

Emulating the quantity Q(k)Q(k), rather than the full nonlinear power spectra, reduces significantly the dynamical range of the emulation, simplifying the problem and thus delivering more accurate results (see e.g. Euclid Collaboration et al., 2019; Giblin et al., 2019). We emphasise, however, that these transformations are simply designed to improve the numerical stability of our predictions – our results regarding large scales and BAO are still uniquely provided by our simulation results.

To build our training set, we first construct a Latin-Hypercube with 400 points. We then build a Gaussian Process (GP) emulator for QQ, which provides the expectation value and variance for the emulated quantity (c.f. §3.1.7 for more details). We then evaluate the GP emulator over a Latin-Hypercube built with 2,000 points, and select the 100 points expected to have the largest uncertainty in their predictions. We add these points to our training set and re-build the GP emulator. We repeat this procedure 4 times.

We display the final set of training cosmologies, comprised of 800 points, in Fig. 5. Blue symbols indicate the initial training cosmologies, whereas orange and green symbols do so for the 2nd and 4th iteration, respectively. We can appreciate that most points in the iterations are located near the boundaries of the cosmological space, which minimise extrapolation.

It is worth mentioning that we sample our space significantly better than typical emulators. For instance, the Mira-Titan and Euclid Emulator projects employ 3636 and 100100 simulations, respectively. This implies that we can keep the emulator errors under any desired level, which could be problematic otherwise, as parameter-dependent uncertainties in models, could in principle bias parameter estimates. On the other hand, sampling points can be designed optimally so that emulation errors are much smaller than a naive Latin Hypercube (Heitmann et al., 2014; Rogers et al., 2019), however, this can be only be done optimally for a single summary statistic: what could be optimal for the matter power spectrum, is not necessarily the best for, e.g, the quadrupole of the galaxy power spectrum.

3.1.3 Cosmology rescaling

Our next step is to predict nonlinear structure for each of our training cosmologies. For this, we employ the latest incarnation of the cosmology-rescaling algorithm, originally introduced by Angulo & White (2010).

In short, the cosmology-rescaling algorithm seeks a coordinate and time transformation such that the linear rms variance coincides in the original and target cosmologies. This transformation is motivated by extended Press-Schechter arguments, and returns highly accurate predictions for the mass function and properties of collapsed objects. On large scales, the algorithm uses 2nd-order Lagrangian Perturbation Theory to modify the amplitude of Fourier modes as consistent with the change of cosmology. On small scales, the internal structure of halos is modified using physically-motivated models for the concentration-mass-redshift relation (e.g. Ludlow et al., 2016).

The accuracy of the cosmology algorithm has been tested by multiple authors (Ruiz et al., 2011; Angulo & Hilbert, 2015; Mead & Peacock, 2014; Mead et al., 2015; Renneby et al., 2018; Contreras et al., 2020), and it has been recently extended to the case of massive neutrinos by Zennaro et al. (2019). Specifically, by comparing against a suite of NN-body simulations, Contreras et al. (2020) explicitly showed that the cosmology rescaling achieves an accuracy of 3%\lesssim 3\% up to k=5hMpc1k=5h\,{\rm Mpc}^{-1} over the same range of cosmological parameter values we consider here (c.f. Eq. 3.1.1). The largest errors appear on small scales and for dynamical dark energy parameters but, when restricted to the 6 parameters of the minimal Λ\LambdaCDM model, the rescaling returns 1%1\% accurate predictions. Note that the level of accuracy is set by the performance of current models for the concentration-mass-redshift relation (which usually are not calibrated for beyond Λ\LambdaCDM parameters), and future progress along those lines should feedback into higher accuracy for our emulator.

For our task at hand, we first split the parameter space into three disjoint regions where nenya, narya, or vilya will be employed (Contreras et al., 2020). Then, we load in memory a given snapshot of a given simulation and then rescale it to the corresponding subset of the 800 cosmologies. We employ 10 snapshots per simulation.

The full rescaling algorithm takes approximately 2 minutes (on 12 threads using OpenMP parallelization) per cosmology and redshift. Thus, all the required computations for building the emulator required approximately 10,00010,000 CPU hours – a negligible amount compared to that employed to run a single state-of-the-art NN-body simulation.

3.1.4 Power Spectrum Measurements

We estimate the power spectrum in our rescaled simulation outputs using Fast Fourier Transforms. We employ a cloud-in-cells assignment scheme over two interlaced grids (Sefusatti et al., 2016) of N=10243N=1024^{3} points. To achieve efficient measurements at higher wavenumbers, we repeat this procedure after folding the simulation box 44 times in each dimension. We measure the power spectrum in 160160 evenly-spaced bins in logk\log{k}, from k=0.01k=0.01 to 5hMpc15h\,{\rm Mpc}^{-1}, and transition from the full to the folded measurement at half the Nyquist frequency. We have explicitly checked that this procedure returns sub-percent accurate power spectrum measurements over the full range of scales considered.

Although our BACCO simulations have high mass resolution, for computational efficiency, hereafter we will consider a subset of 108031080^{3} particles (uniformly selected in Lagrangian space) as our dark matter catalogue. This catalogue, however, is affected by discreteness noise. For a Poisson sampling of NN points over a box of side-length LL, the power spectrum receives a contribution equal to (L/N)3(L/N)^{3}. However, this might not be an accurate estimate for our discreteness noise since our sampling is homogeneous in Lagrangian coordinates. We have estimated its actual contribution as a third-order polynomial by comparing the spectra of the full and diluted samples at different redshifts. We found that the amplitude is proportional to 0.6\sim 0.6 times the Poisson noise at z=0z=0, and that it progressively decreases in amplitude at higher redshifts, to reach 0.2\sim 0.2 times the Poisson noise at z=2z=2.

All our power spectrum measurements will be corrected by discreteness noise by subtracting the term described above, and further corrected for finite numerical accuracy following section §2.3.1. However, this procedure is not perfect, and we still detect 2\sim 2% residuals at z1z\sim 1 at k5hMpc1k\sim 5h\,{\rm Mpc}^{-1}. This will contribute to uncertainties in our emulator, which, however, are smaller than systematic uncertainties induced by the cosmology scaling.

3.1.5 Emulator Data

Refer to caption
Figure 6: Nonlinear power spectra at z=0z=0 and z=1z=1 over the linear theory expectations, as predicted by rescaled BACCO simulations. Grey regions indicate the range covered by the training set employed to construct our emulator. Blue and red lines display two particular, randomly-chosen, measurements.

In total, we employ more than 16,000 power spectrum measurements; 400+4×100400+4\times 100 training cosmologies at approximately 10 different cosmic times for two paired simulations. Grey shaded regions in Fig. 6 show the range covered by this data at z=0z=0 and z=1z=1, with the inset focusing on the range of wavemodes where baryonic acoustic oscillations are found. Blue and red lines show two randomly-chosen cosmologies.

On the largest scales, we can see that our rescaled simulations agree almost perfectly with linear theory. Although expected, it provides further validation of the dataset. On intermediate scales, we can see a lack of any oscillatory behaviour – better appreciated in the figure insets. This is a consequence of our PlinearsmearedBAOP_{\rm linear}^{\rm smeared-BAO} correctly capturing the nonlinear smearing of baryonic acoustic oscillations caused mostly by large-scale flows. This is true at both z=0z=0 and z=1z=1, which confirms the correct redshift dependence of PlinearsmearedBAOP_{\rm linear}^{\rm smeared-BAO}. On small and nonlinear scales, we see an increase of more than one order of magnitude with different cosmologies differing by even factors of 3-4.

3.1.6 Principal Component Analysis

Refer to caption
Figure 7: Top: The amplitude of the first 66 vectors of a principal component analysis of our power spectrum data. Bottom: Ratio between the original and PCA-reconstructed Qlog(P/PlinearsmearedBAO)Q\equiv\log(P/P_{\rm linear}^{\rm smeared-BAO}) for a random 2%2\% of our data at z=0z=0. The grey band indicates a region of ±0.5%\pm 0.5\%, which can be considered as an indication of the statistical noise in our data.
Refer to caption
Figure 8: Comparison between the nonlinear power spectrum at z=0z=0 as predicted by a full NN-body simulation and our BACCO emulator for a set of cosmological parameters consistent with recent analyses of the Planck satellite data. The left panel shows the full power spectrum, whereas middle and rightmost panel show the same data relative to the expectations of linear theory or halofit, respectively. The emulator and simulation predictions agree to better than 2\sim 2% over the full range of scales considered.

To reduce the dimensionality of our power spectrum measurements, we have performed a principal components (PC) analysis over our whole dataset, after subtracting the mean. We have kept the 6 kk-vectors with the highest eigenvalues, which together can explain all but 10310^{-3} of the data variance.

In the top panel of Fig. 7 we show the amplitude of these PCs, as indicated by the legend. The most important vector is a smooth function of wavenumber and roughly captures the overall impact of nonlinear evolution. Subsequent vectors describe further modifications, but none of them shows significant oscillations, which indicates that PlinearsmearedBAOP_{\rm linear}^{\rm smeared-BAO} accurately models the BAO as a function of cosmology and redshift. It is worth noting that only the 6th vector displays any noticeable noise, owing to the highly precise input dataset.

In the bottom panel of the Fig. 7 we display the ratio of the full Qlog(P/Plinear)Q\equiv\log(P/P_{\rm linear}) over that reconstructed using the first PCs aforementioned. We show the results for a random 2% of the power spectrum in our training data. We can see that almost all of them are recovered to better than 0.5%0.5\%. It is interesting to note that the residuals, although increase on intermediate scales, are mostly devoid of structure, which suggests that including additional PCs will simply recover more accurately the intrinsic noise in our dataset rather than systematic dependencies on cosmology.

To confirm this idea, we have repeated our emulation but keeping twice as many PCs. Although the description of the values of Q in our dataset gained accuracy (decreasing the differences down to 0.25%\sim 0.25\%), the performance of the emulator at predicting other cosmologies did not increase. This supports the idea that PCs beyond the 6-th are simply capturing particular statistical fluctuations in the training set, rather than cosmology-induced features. It is interesting to note that this also means that all nonlinear information of the power spectrum beyond the BAO can be described by only 66 numbers.

3.1.7 Emulation

Refer to caption
Figure 9: Comparison between Qlog(P/Plinear)Q\equiv\log(P/P_{\rm linear}) predicted by our emulator and that computed directly by scaling our BACCO simulations to the desired cosmology. Top and bottom panels display results at z=1z=1 and z=0z=0, respectively, for 10 cosmologies that span our whole parameter space.

To interpolate between our training data, and thus predict Q(k,z)Q(k,z) for any cosmology, we will construct two emulators: one built by employing Gaussian Processes Regression, and another one built by training a Neural Network.

Gaussian Processes

In short, Gaussian processes assume that every subset of points in a given space is jointly Gaussian distributed with zero mean and covariance KK. The covariance is a priori unknown but it can be estimated from a set of observations (e.g. our training set). Once the covariance is specified, the Gaussian process can predict the full probability distribution function anywhere in the parameter space.

In our case, we measure the amplitude associated to each PC in each training set. We then build a separate Gaussian Process for each of our PCs, using the package GPy (GPy, 2012). We assume the covariance kernel to be described by a squared exponential, with correlation length and variance set to the values that best describe the correlation among our data, found by maximising the marginal likelihood.

Neural Network

For regression tasks, neural networks perform remarkably well, serving as a highly-precise approximation to almost any continuous function. Thus, they provide an alternative for emulating Q(k,z)Q(k,z) (see e.g. Agarwal et al., 2014; Kobayashi et al., 2020).

In our case, we found that a feed-forward neural network with a simple architecture allowed us to obtain an accurate emulation. Specifically, we employed a fully-connected network with 2 hidden layers, each with 2000 neurons. For the activation function – which takes our input vectors and transforms them non-linearly – we use of Rectified Linear Units (ReLUs) which are commonly used in machine learning. The training of the network is done by using the adaptive stochastic optimization algorithm Adam (Kingma & Ba, 2014), with a default learning rate of 10310^{-3}. For the implementation of the neural network we made use of the open-source neural-network python library, Keras (Chollet et al., 2015).

We employ the PCA-reconstructed Q(k,z)Q(k,z) as our training data for the neural network (NN), so we can compare its results to those of the Gaussian process in equal terms. We randomly selected 5% of our power spectra as our validation set, and employ the remaining 95% as our training data. We define the loss function as the mean squared error function (MSE), and employed 2000 epochs for the training, after which our results appeared converged. At this point, the value of the loss function evaluated in the training and validation sets differed by less than 50%\sim 50\%, which suggest no overfitting. We note that we also tried the so-called batch-normalization and data-dropouts, without finding significant improvements in our predictions.

For a given target cosmology in our parameter space, we can predict the amplitude associated to each PC and then reconstruct the full Q(k,z)Q(k,z) vector. We found both emulators displaying similar performances: we compared their predictions at 1000 random locations within our parameter space, and differences were typically smaller than 121-2 percent. Both emulators display similar computational performances, about 200200 milliseconds for predicting QQ on any cosmology.

Refer to caption
Figure 10: Comparison between nonlinear power spectrum predicted by our emulator and by four other methods. From left to right, panels display a comparison against Halofit, Euclid Emulator, NGenHaloFit, and MiraTitan. Top and bottom rows show results for z=1z=1 and z=0z=0, respectively. Each coloured curve represents a different cosmology within the parameter space where the Euclid Emulator has been calibrated. We highlight the “Euclid reference” cosmology as a thick line. Grey bands indicate ±1\pm 1 and ±2\pm 2% regions.
Refer to caption
Figure 11: The ratio between the cold matter power spectra in massive neutrino cosmologies, PcνP^{\nu}_{c}, and the case without any massive neutrino, Pcν=0P^{\nu=0}_{c}. We show the predictions of linear theory (dotted line) and of three models: HaloFit, Mira-Titan emulator and our BACCO emulator.

3.2 Testing the emulator

3.2.1 A first example

We present a first look at our emulator results in Fig. 8, where we show the nonlinear power spectrum at z=0z=0 for a cosmology consistent with a recent analysis of Planck data. Specifically, we consider: σ8=0.8102\sigma_{8}=0.8102, Ωm=0.30964\Omega_{m}=0.30964, Ωb=0.04897\Omega_{b}=0.04897, ns=0.9665n_{s}=0.9665, h=0.6766h=0.6766, Mν=0.0M_{\nu}=0.0, w0=1.0w_{0}=-1.0, wa=0.0w_{a}=0.0. The left panel shows the full power spectrum, the middle panel do so relative to linear theory, and the rightmost panel relative to HaloFit calibrated by Takahashi et al. (2012).

For comparison, we include as orange circles, the results of a full NN-body simulation with the same mass resolution and numerical parameters as those of our main BACCO suite, but in a smaller box, L=512h1MpcL=512\,h^{-1}{\rm Mpc}. Also as in our BACCO suite, the initial conditions have been “Paired-&-Fixed” for this simulation.

Overall, we can see that the BACCO emulator and the NN-body simulation agree to a remarkable level, being indistinguishable by eye in the leftmost and middle panels. In particular, on large scales, both agree with linear theory (which only can be appreciated thanks to the “Paired-&-Fixed” initial conditions); on intermediate scales, both also predict a BAO featured smeared out compared to linear theory.

In the rightmost panel we can see these aspects in more detail. Firstly, we note that the simulation and emulator results agree to about 1% on all the scales considered. This is consistent with the expected accuracy of the cosmology rescaling method, but also note that due to its somewhat small volume, the cosmic variance in the NN-body simulation results are not negligible. It is also interesting to note the systematic disagreement with the predictions from halofit. In following subsections we will explore these differences further.

3.2.2 Accuracy

To start testing the accuracy of our emulator, we have defined 10 cosmologies distributed over the target parameter space (c.f.§3.1.1) using a Latin hypercube. We then rescale our BACCO outputs to those parameters and compare the results against our emulation predictions. This essentially tests how accurate the PCA decomposition and emulation via Gaussian Process Regression are.

In Fig. 9 we show the ratio of the emulated to the rescaled nonlinear power spectrum at z=0z=0 and z=1z=1 for the 10 cosmologies mentioned before (we recall that neither z=0z=0 nor z=1z=1 were explicitly included in the emulator). We can see that on all scales our results are better than ±2%\pm 2\% percent, and that most of the cosmologies are predicted to better that 1% at z=0z=0. Although this is already a high accuracy, we note that more and more rescaled results can be added over time to progressively improve the quality of the emulation. We recall that this level of uncertainty is comparable to the accuracy expected by the cosmology rescaling algorithm: 1%1\% for parameters of the minimal Λ\LambdaCDM and 3%\sim 3\% when considering massive neutrinos and dynamical dark energy.

3.2.3 Comparison with HaloFit, EuclidEmulator, and NgenHaloFit

We now compare our emulation results against four widely-used methods to predict the nonlinear evolution: HaloFit (Takahashi et al., 2012), the Euclid Emulator (Euclid Collaboration et al., 2019), NGenHalofit (Smith & Angulo, 2019), and Mira-Titan (Lawrence et al., 2017). Since not all of them have been calibrated over the whole parameter space covered by our emulator, we have restricted the comparison to the volume covered by the Euclid Emulator. We note that of our 800 training cosmologies, only 2 of them fall within this parameter volume.

In Fig. 10 we display the ratio of those predictions to that of our BACCO emulator. Coloured lines show 10 cosmologies set by a latin hypercube inside the Euclid Emulator parameter space. In addition, we show as a heavy line the Euclid reference cosmology: Ωcdm=0.26067\Omega_{\rm cdm}=0.26067, σ8=0.8102\sigma_{8}=0.8102, Ωb=0.04897\Omega_{\rm b}=0.04897, ns=0.9665n_{s}=0.9665, h=0.6766h=0.6766, Mν=0M_{\nu}=0, w0=1w_{0}=-1, wa=0w_{a}=0, employed by (Euclid Collaboration et al., 2019).

On large scales, k<0.08hMpc1k<0.08h\,{\rm Mpc}^{-1}, our emulator agrees almost perfectly with NGenHalofit and the Euclid Emulator. Halofit, on the other hand shows a small constant power deficit, whereas the Mira-Titan displays a weakly scale-dependent offset. Over this range of scales, Mira-Titan is given by TimeRG perturbation theory (Pietroni, 2008; Upadhye et al., 2014), which might indicate inaccuracies in that approach.

On intermediate scales, 0.08<k/[hMpc1]<0.50.08<k/[h\,{\rm Mpc}^{-1}]<0.5, our results agree very well with those of NGenHalofit; for almost all cosmologies the differences are within the expected accuracy of our emulator. In contrast, we see that Halofit overestimates the amount of power by about 2% at z=0z=0 and it does not correctly captures the BAO nonlinear smearing (as was already noted by Euclid Collaboration et al., 2019). On the other hand, MiraTitan underestimates the power by up to 2.5% at z=0z=0, and displays a strong feature at k0.2hMpc1k\sim 0.2h\,{\rm Mpc}^{-1} at z=1z=1. Over this range of scales MiraTitan employs a suite of low-resolution, large volume simulations, thus the disagreement could be due to an insufficient numerical accuracy in their simulations.

On small scales, k>0.5hMpc1k>0.5h\,{\rm Mpc}^{-1}, we continue to see differences among the three methods. For halofit, NGenHaloFit and Mira-Titan, they are systematic, roughly independent of cosmology, and decrease at higher redshift. A possibility is that this could be originated by the different numerical accuracy of the underlying simulations. Specifically, the simulations used by Smith & Angulo (2019) employ a softening length ϵ=8h1kpc\epsilon=8h^{-1}{\rm kpc}, which according to Eq. 1 is expected to produce an underestimation of 1.5% at k=5hMpc1k=5h\,{\rm Mpc}^{-1}. In addition, the transition between their high and low-resolution runs occurs at k0.6hMpc1k\sim 0.6h\,{\rm Mpc}^{-1} which might be related to the deficit we observe at k0.70.7hMpc1k\sim 0.7-0.7h\,{\rm Mpc}^{-1} at z=0z=0.

In contrast, the differences with respect to Euclid Emulator vary significantly for different cosmologies. Specifically, for the Euclid Reference Cosmology, the agreement is sub-percent on all scales. However, for our other test cosmologies, differences have a spread of 5\sim 5%, even at k<1hMpc1k<1h\,{\rm Mpc}^{-1}. We do not see this behavior with other methods, which might suggest that there are significant uncertainties in the emulated power spectra of the Euclid Emulator beyond their quoted precision.

Finally, we note that there is a 1% “bump” at k3hMpc1k\sim 3h\,{\rm Mpc}^{-1} in all z=1z=1 panels, which is originated by our imperfect shot-noise correction.

The previous comparison was done in a rather restricted cosmological parameter volume, which served as a strong test of our accuracy. However, one of the biggest advantages of our method is the ability to predict much more extreme cosmologies even with non-standard ingredients. We provide an example of this next.

In Fig. 11 we show the predictions for the effects of massive neutrinos on the z=0z=0 cold matter (baryons plus dark matter) power spectrum. We display the ratio between cases with various neutrino masses, PcνP_{c}^{\nu}, relative to that without massive neutrinos PcνP_{c}^{\nu}. In all cases we use As=2.1×109A_{s}=2.1\times 10^{-9} and fix all the other cosmological parameters to those of the “Euclid reference cosmology”. We only display the predictions of HaloFit and Mira-Titan, since NGenHaloFit nor the Euclid Emulator have been calibrated in the case of massive neutrinos.

We can see that on large scales, our predictions agree with both linear theory and HaloFit. For very massive neutrino cases, there is also a good agreement with MiraTitan, but there is a significant disagreement for Mν=0.1M_{\nu}=0.1, this might be caused by the scale-dependent features on large scales seen in the previous figure. On intermediate scales, all predictions also agree on the broadband shape of the neutrino-induced suppression, but our emulator is also able to capture the slightly different BAO suppression expected when massive neutrinos are present. On small scales, all three methods describe the well-known neutrino-induced spoon-like suppression, disagreeing slightly on the magnitude of the maximum suppression. Note however, we expect our emulator to predict more precisely the full shape of the nonlinear power spectrum, owing to the systematic uncertainties in HaloFit and Mira-Titan discussed before.

4 Summary

In this paper we have presented the BACCO simulation project: a framework that aims at delivering high-accuracy predictions for the distribution of dark matter, galaxies, and gas as a function of cosmological parameters.

The basic idea consists in combining recent developments in numerical cosmology – NN-body simulations, initial conditions with suppressed variance, and cosmology-rescaling methods – to quickly predict the nonlinear distribution of matter in a cosmological volume. The main advantage of our approach is that it requires only a small number of NN-body simulations, thus they can be of high resolution and volume. This in turn allows sophisticated modelling of the galaxy population (for instance in terms of subhalo abundance matching, semi-empirical or semi-analytic galaxy formation model), and of baryons (including the effects of cooling, star formation and feedback) in the mass distribution.

In this paper we have presented the main suite of gravity-only simulations of the BACCO project. These consist in 3 sets of “Paired-&-Fixed” simulations, each of them of a size L=1440h1MpcL=1440\,h^{-1}{\rm Mpc} and with 8080 billion particles. Their cosmologies were carefully chosen so that they maximise the accuracy of our predictions (Fig. 1 and Table 1) while minimising computational resources. We have validated the accuracy of our numerical setup with a suite of small NN-body simulations (Fig. 2) and by presenting a realization of the Euclid comparison project (Fig. 3). These tests indicate our simulations have an accuracy of 1% up to k5hMpc1k\sim 5h\,{\rm Mpc}^{-1}.

We have employed our BACCO simulations to predict more than 16,000 nonlinear power spectra at various redshifts and for 800 different cosmologies (Figs. 5 and 6). These cosmologies span essentially all the currently allowed region of parameter space of Λ\LambdaCDM extended to massive neutrinos and dynamical dark energy. Using these results, we built an emulator for the 6 most important principal components of the ratio of the nonlinear power spectrum over the linear expectation (Fig. 7). We show our emulation procedure to be accurate at the 12%1-2\% level over 0<z<1.50<z<1.5 and 102<k/(hMpc1)<510^{-2}<k/(h\,{\rm Mpc}^{-1})<5 (Figs. 8 and 9). Therefore, our accuracy is currently limited by that of cosmology rescaling methods. We compared our predictions against four popular methods to quickly predict the power spectrum in the minimal Λ\LambdaCDM scenario (Fig. 10) and in the presence of massive neutrinos (Fig. 11).

Since predicting a given cosmology requires an almost negligible amount of CPU time in our BACCO framework, we foresee the accuracy of our emulator to continuously improve as we include more cosmologies in the training set. Extensions to more parameters should also be possible, as, for instance, the number of relativistic degrees of freedom or curvature, can be easily incorporated in cosmology-rescaling methods. Additionally, there are several aspects of such methods that are likely to improve in the future, which should feedback into more accurate predictions and emulated power spectra.

On the other hand, effects induced by baryons on the shape of the nonlinear mass power spectrum can be of 10-30% (e.g. Chisari et al., 2019). Thus, they are much larger than current uncertainties in our emulation, cosmology-rescaling, or even shotnoise. These effects of star formation, gas cooling, and feeback from supermassive black holes are quite uncertain and differ significantly between different hydrodynamical simulations. However, they can be accurately modelled in postprocessing using dark-matter only simulations (Schneider & Teyssier, 2015). Specifically, Aricò et al. (2020b) showed that the effects of 7 different state-of-the-art hydrodynamical simulations could all be modelled to better than 1% within simple but physically-motivated models. In the future, we will implement such models and extend our matter emulation to simultaneously include cosmological and astrophysical parameters (Aricò et al., 2020a).

Acknowledgments

The authors acknowledge the support of the ERC-StG number 716151 (BACCO). SC acknowledges the support of the “Juan de la Cierva Formación” fellowship (FJCI-2017-33816). We thank Simon White, Auriel Schneider, Volker Springel for useful discussions. We thank Aurel Schneider also for making public the initial conditions of the EUCLID comparison project, and to Lehman Garrison for providing us with the respective power spectra of the ABACUS, Ramses, PkdGrav, and Gadget simulations. The authors acknowledge the computer resources at MareNostrum and the technical support provided by Barcelona Supercomputing Center (RES-AECT-2019-2-0012).

Data Availability

The data underlying this article will be shared on reasonable request to the corresponding author.

References

  • Agarwal et al. (2014) Agarwal S., Abdalla F. B., Feldman H. A., Lahav O., Thomas S. A., 2014, MNRAS, 439, 2102
  • Alam et al. (2017) Alam S., et al., 2017, MNRAS, 470, 2617
  • Ali-Haïmoud & Bird (2013) Ali-Haïmoud Y., Bird S., 2013, MNRAS, 428, 3375
  • Angulo & Hilbert (2015) Angulo R. E., Hilbert S., 2015, MNRAS, 448, 364
  • Angulo & Pontzen (2016) Angulo R. E., Pontzen A., 2016, MNRAS, 462, L1
  • Angulo & White (2010) Angulo R. E., White S. D. M., 2010, MNRAS, 405, 143
  • Angulo et al. (2012) Angulo R. E., Springel V., White S. D. M., Jenkins A., Baugh C. M., Frenk C. S., 2012, MNRAS, 426, 2046
  • Aricò et al. (2020a) Aricò G., Angulo R. E., Contreras S., Ondaro-Mallea L., Pellejero-Ibañez M., Zennaro M., 2020a, arXiv e-prints, p. arXiv:2011.15018
  • Aricò et al. (2020b) Aricò G., Angulo R. E., Hernández-Monteagudo C., Contreras S., Zennaro M., Pellejero-Ibañez M., Rosas-Guevara Y., 2020b, MNRAS, 495, 4800
  • Asgari et al. (2020) Asgari M., et al., 2020, A&A, 634, A127
  • Baumann et al. (2018) Baumann D., Green D., Wallisch B., 2018, J. Cosmology Astropart. Phys., 2018, 029
  • Chaves-Montero et al. (2016) Chaves-Montero J., Angulo R. E., Schaye J., Schaller M., Crain R. A., Furlong M., Theuns T., 2016, MNRAS, 460, 3100
  • Chisari et al. (2018) Chisari N. E., et al., 2018, MNRAS, 480, 3962
  • Chisari et al. (2019) Chisari N. E., et al., 2019, The Open Journal of Astrophysics, 2, 4
  • Chollet et al. (2015) Chollet F., et al., 2015, Keras, https://github.com/fchollet/keras
  • Chuang et al. (2019) Chuang C.-H., et al., 2019, MNRAS, 487, 48
  • Contreras et al. (2020) Contreras S., Angulo R. E., Zennaro M., Aricò G., Pellejero-Ibañez M., 2020, MNRAS, 499, 4905
  • DeRose et al. (2019) DeRose J., et al., 2019, ApJ, 875, 69
  • Desjacques et al. (2018) Desjacques V., Jeong D., Schmidt F., 2018, Phys. Rep., 733, 1
  • Euclid Collaboration et al. (2019) Euclid Collaboration et al., 2019, MNRAS, 484, 5509
  • Favole et al. (2017) Favole G., Rodríguez-Torres S. A., Comparat J., Prada F., Guo H., Klypin A., Montero-Dorta A. D., 2017, MNRAS, 472, 550
  • Freedman et al. (2019) Freedman W. L., et al., 2019, ApJ, 882, 34
  • GPy (2012) GPy since 2012, GPy: A Gaussian process framework in python, http://github.com/SheffieldML/GPy
  • Garrison et al. (2019) Garrison L. H., Eisenstein D. J., Pinto P. A., 2019, MNRAS, 485, 3370
  • Giblin et al. (2019) Giblin B., Cataneo M., Moews B., Heymans C., 2019, MNRAS, 490, 4826
  • Gilman et al. (2020) Gilman D., Birrer S., Nierenberg A., Treu T., Du X., Benson A., 2020, MNRAS, 491, 6077
  • Guo & White (2014) Guo Q., White S., 2014, MNRAS, 437, 3228
  • Hahn & Angulo (2016) Hahn O., Angulo R. E., 2016, MNRAS, 455, 1115
  • Heitmann et al. (2006) Heitmann K., Higdon D., Nakhleh C., Habib S., 2006, ApJ, 646, L1
  • Heitmann et al. (2014) Heitmann K., Lawrence E., Kwan J., Habib S., Higdon D., 2014, ApJ, 780, 111
  • Heitmann et al. (2016) Heitmann K., et al., 2016, ApJ, 820, 108
  • Henriques et al. (2020) Henriques B. M. B., Yates R. M., Fu J., Guo Q., Kauffmann G., Srisawat C., Thomas P. A., White S. D. M., 2020, MNRAS, 491, 5795
  • Kingma & Ba (2014) Kingma D. P., Ba J., 2014, Adam: A Method for Stochastic Optimization (arXiv:1412.6980)
  • Klypin et al. (2020) Klypin A., Prada F., Byun J., 2020, MNRAS,
  • Kobayashi et al. (2020) Kobayashi Y., Nishimichi T., Takada M., Takahashi R., Osato K., 2020, Phys. Rev. D, 102, 063504
  • Kuhlen et al. (2012) Kuhlen M., Vogelsberger M., Angulo R., 2012, Physics of the Dark Universe, 1, 50
  • Lawrence et al. (2017) Lawrence E., et al., 2017, ApJ, 847, 50
  • Lesgourgues (2011) Lesgourgues J., 2011, ArXiv e-prints 1104.2932
  • Liu et al. (2018) Liu J., Bird S., Zorrilla Matilla J. M., Hill J. C., Haiman Z., Madhavacheril M. S., Petri A., Spergel D. N., 2018, J. Cosmology Astropart. Phys., 2018, 049
  • Ludlow et al. (2016) Ludlow A. D., Bose S., Angulo R. E., Wang L., Hellwing W. A., Navarro J. F., Cole S., Frenk C. S., 2016, MNRAS, 460, 1214
  • Mead & Peacock (2014) Mead A. J., Peacock J. A., 2014, MNRAS, 440, 1233
  • Mead et al. (2015) Mead A. J., Peacock J. A., Lombriser L., Li B., 2015, MNRAS, 452, 4203
  • Moster et al. (2018) Moster B. P., Naab T., White S. D. M., 2018, MNRAS, 477, 1822
  • Nishimichi et al. (2019) Nishimichi T., et al., 2019, ApJ, 884, 29
  • Orsi & Angulo (2018) Orsi Á. A., Angulo R. E., 2018, MNRAS, 475, 2530
  • Pellejero-Ibañez et al. (2020) Pellejero-Ibañez M., Angulo R. E., Aricó G., Zennaro M., Contreras S., Stücker J., 2020, MNRAS, 499, 5257
  • Pietroni (2008) Pietroni M., 2008, J. Cosmology Astropart. Phys., 2008, 036
  • Planck Collaboration et al. (2014a) Planck Collaboration et al., 2014a, A&A, 571, A16
  • Planck Collaboration et al. (2014b) Planck Collaboration et al., 2014b, A&A, 571, A16
  • Planck Collaboration et al. (2020) Planck Collaboration et al., 2020, A&A, 641, A6
  • Pontzen et al. (2016) Pontzen A., Slosar A., Roth N., Peiris H. V., 2016, Phys. Rev. D, 93, 103519
  • Potter et al. (2017) Potter D., Stadel J., Teyssier R., 2017, Computational Astrophysics and Cosmology, 4, 2
  • Renneby et al. (2018) Renneby M., Hilbert S., Angulo R. E., 2018, MNRAS, 479, 1100
  • Riess (2019) Riess A. G., 2019, Nature Reviews Physics, 2, 10
  • Rogers et al. (2019) Rogers K. K., Peiris H. V., Pontzen A., Bird S., Verde L., Font-Ribera A., 2019, J. Cosmology Astropart. Phys., 2019, 031
  • Ruiz et al. (2011) Ruiz A. N., Padilla N. D., Domínguez M. J., Cora S. A., 2011, MNRAS, 418, 2422
  • Schneider & Teyssier (2015) Schneider A., Teyssier R., 2015, J. Cosmology Astropart. Phys., 12, 049
  • Schneider et al. (2016) Schneider A., et al., 2016, J. Cosmology Astropart. Phys., 4, 047
  • Sefusatti et al. (2016) Sefusatti E., Crocce M., Scoccimarro R., Couchman H. M. P., 2016, MNRAS, 460, 3624
  • Smith & Angulo (2019) Smith R. E., Angulo R. E., 2019, MNRAS, 486, 1448
  • Smith et al. (2003) Smith R. E., et al., 2003, MNRAS, 341, 1311
  • Springel (2005) Springel V., 2005, MNRAS, 364, 1105
  • Springel et al. (2001) Springel V., White S. D. M., Tormen G., Kauffmann G., 2001, MNRAS, 328, 726
  • Stücker et al. (2020) Stücker J., Hahn O., Angulo R. E., White S. D. M., 2020, MNRAS, 495, 4943
  • Takahashi et al. (2012) Takahashi R., Sato M., Nishimichi T., Taruya A., Oguri M., 2012, ApJ, 761, 152
  • Teyssier (2002) Teyssier R., 2002, A&A, 385, 337
  • Upadhye et al. (2014) Upadhye A., Biswas R., Pope A., Heitmann K., Habib S., Finkel H., Frontiere N., 2014, Phys. Rev. D, 89, 103515
  • Villaescusa-Navarro et al. (2018) Villaescusa-Navarro F., et al., 2018, ApJ, 867, 137
  • Weinberg et al. (2013) Weinberg D. H., Mortonson M. J., Eisenstein D. J., Hirata C., Riess A. G., Rozo E., 2013, Physics Reports, 530, 87
  • Wibking et al. (2019) Wibking B. D., et al., 2019, MNRAS, 484, 989
  • Zennaro et al. (2019) Zennaro M., Angulo R. E., Aricò G., Contreras S., Pellejero-Ibáñez M., 2019, MNRAS, 489, 5938
  • van Daalen et al. (2011) van Daalen M. P., Schaye J., Booth C. M., Dalla Vecchia C., 2011, MNRAS, 415, 3649
  • van Daalen et al. (2020) van Daalen M. P., McCarthy I. G., Schaye J., 2020, MNRAS, 491, 2424