This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Accelerating Eigenvalue Computation for
Nuclear Structure Calculations via Perturbative Corrections

Dong Min Roh Applied Mathematics and Computational Research Division, Lawrence Berkeley National Laboratory    Dean Lee Facility for Rare Isotope Beams and Department of Physics and Astronomy, Michigan State University, East Lansing, Michigan 48824    Pieter Maris Department of Physics and Astronomy, Iowa State University, Ames, Iowa 50011    Esmond Ng Applied Mathematics and Computational Research Division, Lawrence Berkeley National Laboratory    James P. Vary Department of Physics and Astronomy, Iowa State University, Ames, Iowa 50011    Chao Yang Applied Mathematics and Computational Research Division, Lawrence Berkeley National Laboratory
Abstract

We present a new method for computing the lowest few eigenvalues and the corresponding eigenvectors of a nuclear many-body Hamiltonian represented in a truncated configuration interaction subspace, i.e., the no-core shell model (NCSM). The method uses the hierarchical structure of the NCSM Hamiltonian to partition the Hamiltonian as the sum of two matrices. The first matrix corresponds to the Hamiltonian represented in a small configuration space, whereas the second is viewed as the perturbation to the first matrix. Eigenvalues and eigenvectors of the first matrix can be computed efficiently. Perturbative corrections to the eigenvectors of the first matrix can be obtained from the solutions of a sequence of linear systems of equations defined in the small configuration space. These correction vectors can be combined with the approximate eigenvectors of the first matrix to construct a subspace from which more accurate approximations of the desired eigenpairs can be obtained. We call this method a Subspace Projection with Perturbative Corrections (SPPC) method. We show by numerical examples that the SPPC method can be more efficient than conventional iterative methods for solving large-scale eigenvalue problems such as the Lanczos, block Lanczos and the locally optimal block preconditioned conjugate gradient (LOBPCG) method. The method can also be combined with other methods to avoid convergence stagnation.

preprint: APS/123-QED

I Introduction

Nuclear structure calculations require solving AA-body Schrödinger equations where A=Z+NA=Z+N is the number of nucleons consisting of ZZ protons and NN neutrons. The Configuration Interaction (CI) method or no-core shell method (NCSM) [1], which represents the solution to the Schrödinger’s equation by a linear combination of AA-body basis functions, reduces the problem to an algebraic eigenvalue problem:

HΨk=EkΨk,\displaystyle H\Psi_{k}=E_{k}\Psi_{k}, (1)

where Hn×nH\in\mathbb{R}^{n\times n} is the matrix representation of the AA-body nuclear Hamiltonian operator in a configuration space (spanned by a set of nn AA-body basis functions), EkE_{k} is the kkth eigenvalue of HH representing an approximation to a discrete energy level, and Ψk\Psi_{k} is the corresponding eigenvector that contains the coefficients of the AA-body basis function in the expansion of the approximate eigenfunction in the AA-body basis.

The dimension (nn) of the matrix HH depends on the number of nucleons AA and the size of the CI model space (determined by a truncation parameter NmaxN_{\max}). Although nn can be quite large, HH is very sparse, and often only a few of its eigenpairs at the low end of the spectrum are of interest, making iterative methods suitable for solving (1).

The construction of the matrix HH by the CI method is typically done in a hierarchical fashion where the leading submatrix of HH corresponds to a matrix constructed from a smaller configuration space. Because the AA-body basis functions that form the lower dimension CI space associated with a small NmaxN_{\max} are typically more important than basis functions outside of such a configuration space, the eigenvectors of HH tend to be localized; i.e., the leading components of the Ψk\Psi_{k} tend to be the larger in magnitude, while the tailing components are relatively small in magnitude. We illustrate these properties of a nuclear Hamiltonian and its wavefunction using nucleus 12C as an example. The Hamiltonian is constructed with a nucleon-nucleon interaction Daejeon16 [2] where hbar-omega value of 2020 describes the harmonic oscillator basis functions. Figure 1(a) shows that the leading submatrix of HH that is 6363 times smaller corresponds to a matrix constructed from a smaller configuration space. Figure 1(b) shows that the eigenvector corresponding to the lowest eigenvalue is localized in its first few components. Therefore, the vector formed by padding the eigenvector of the leading submatrix with zeros can serve as a good initial guess for many iterative methods used to solve large-scale eigenvalue problems (1). Previous works [3, 4] select such initial guesses for algorithms like the Lanczos algorithm [5], the block Lanczos algorithm [6], the Locally Optimal Block Preconditioned Conjugate Gradient (LOBPCG) algorithm [7], and the Residual Minimization Method with Direct Inversion of Iterative Subspace (RMM-DIIS) correction [8, 9, 10]. Another study [11] uses greedy algorithms to incrementally enlarge the submatrix and use the eigenvector of the enlarged submatrix as an improved starting guess.

Refer to caption
Refer to caption
Figure 1: (a) The Hamiltonian matrix HH of 12C constructed by the CI method with a truncation parameter Nmax=4N_{\max}=4, using the nucleon-nucleon interaction. The leading submatrix H^0\hat{H}_{0} of HH that is around 6363 times smaller is equivalent to the Hamiltonian matrix constructed by the CI method with a truncation parameter Nmax=2N_{\max}=2. (b) Illustration of the eigenvector localization observed in the Hamiltonian matrix. The leading components of the eigenvector corresponding to the lowest eigenvalue of HH are several magnitudes larger than the tailing components.

In this paper, we propose solving the eigenvalue problem (1) with an iterative subspace projection method constructed from basis vectors obtained from eigenvalue and eigenvector perturbation theory. We call this method a Subspace Projection with Perturbative Corrections (SPPC) method. In this approach, the Hamiltonian matrix to be partially diagonalized is viewed as the sum of two matrices, i.e., H=H0+VH=H_{0}+V, where the eigenpairs of H0H_{0} are relatively easy and inexpensive to compute, and VV is viewed as a perturbation to H0H_{0}. Perturbative corrections to eigenpairs of H0H_{0} in successively higher order can be computed by solving a sequence of linear systems of equations. Together with the initial approximation to the desired eigenvector obtained from H0H_{0}, these correction vectors form a low dimensional subspace from which approximate eigenpairs of HH are extracted through the Rayleigh-Ritz procedure. When H0H_{0} is chosen to be a block diagonal matrix diag(H^0\hat{H}_{0},0), where H^0\hat{H}_{0} is the matrix representation of the AA-body nuclear Hamiltonian in a smaller configuration space associated with a smaller NmaxN_{\max} value, these linear systems can be solved efficiently in the small configuration space. The overall computational cost of the SPPC method grows linearly with respect to the highest order of perturbation included in the correction subspace. Adding each perturbative correction vector and performing the Rayleigh-Ritz calculation requires multiplying the sparse matrix HH with a single vector. We show numerically that the low dimensional subspace constructed in a low order SPPC method provides more accurate approximation to a few lowest eigenvalues of HH than a Krylov subspace of the same dimension constructed from the same starting guess. The method also appears to be more efficient than the locally optimal block preconditioned conjugate gradient (LOBPCG) method in early iterations even when a good preconditioner is available for the LOBPCG method. Although the SPPC method can stagnate as higher perturbative corrections are included, convergence stagnation can be mitigated by combining SPPC with other iterative algorithms for solving large scale eigenvalue problems using the SPPC’s eigenvector approximation as the starting guess for secondary algorithms.

The rest of this paper is organized as follows. In Section II we describe the basic Subspace Projection with Perturbative Corrections (SPPC) method for computing one eigenpair of HH, and show how successively higher order perturbative corrections can be obtained from solutions of a set of linear systems of equations. We draw the connection between SPPC and previously developed eigenvector continuation methods. In Section III, we present a version of the SPPC algorithm that can be used to compute a few eigenpairs. A few practical implementation details of the SPPC method are discussed in Section IV. Numerical examples that demonstrate the efficiency of the SPPC method relative to other conventional large-scale eigenvalue computation methods are presented in Section V. We also show the effectiveness of combining SPPC with a conventional eigensolver for computing a few lowest eigenpairs of several nuclei.

II Subspace Projection with Perturbative Corrections (SPPC)

We split the matrix HH into the sum of two matrices H0H_{0} and VV, i.e.,

H=H0+V,\displaystyle H=H_{0}+V, (2)

where

H0:=[H^0000],V:=[0V12V21V22],\displaystyle H_{0}:=\begin{bmatrix}\hat{H}_{0}&0\\ 0&0\end{bmatrix},\quad V:=\begin{bmatrix}0&V_{12}\\ V_{21}&V_{22}\end{bmatrix}, (3)

with the matrix H^0n0×n0\hat{H}_{0}\in\mathbb{R}^{n_{0}\times n_{0}} (where n0nn_{0}\ll n) being the leading submatrix of HH that corresponds to the representation of the Hamiltonian within a smaller configuration space.

The eigenvectors of H0H_{0}, which can be obtained from the eigenvectors of H^0\hat{H}_{0}, are computed with a much lower computational cost compared to those of the full matrix HH. They can be used as good initial guesses for conventional algorithms, such as the Lanczos and the LOBPCG algorithms, for computing the eigenpairs of HH.

The SPPC method uses the eigenvectors of H0H_{0} to initiate a subspace construction procedure to produce a subspace from which improved approximations of the desired eigenpairs can be obtained.

Instead of using the Lanczos algorithm or the LOBPCG method to construct the subspace, we use perturbative corrections to the initial eigenvector approximation to construct the subspace. In this approach, we view HH as a matrix obtained from perturbing H0H_{0} by cVcV with c=1c=1, i.e., we can write HH as

H(c)=H0+cV.\displaystyle H(c)=H_{0}+cV. (4)

It follows from the Rayleigh-Schrödinger perturbation theory [12] that the eigenvalues and eigenvectors of H(c)H(c), which we denote by Ek(c)E_{k}(c) and Ψk(c)\Psi_{k}(c) respectively, can be written in terms of perturbative corrections to the eigenvalues and eigenvectors of H0H_{0}, which we denote by Ek(0)E_{k}^{(0)} and Ψk(0)\Psi_{k}^{(0)}, i.e.,

{Ek(c)=Ek(0)+cEk(1)+c2Ek(2)+,Ψk(c)=Ψk(0)+cΨk(1)+c2Ψk(2)+.\displaystyle\begin{cases}E_{k}(c)&=E_{k}^{(0)}+cE_{k}^{(1)}+c^{2}E_{k}^{(2)}+\cdots,\\ \Psi_{k}(c)&=\Psi_{k}^{(0)}+c\Psi_{k}^{(1)}+c^{2}\Psi_{k}^{(2)}+\cdots.\end{cases} (5)

Here, (Ek(p),Ψk(p))p1{(E_{k}^{(p)},\Psi_{k}^{(p)})}_{p\geq 1} represent the ppth order perturbative corrections to the eigenpair (Ek(0),Ψk(0))(E_{k}^{(0)},\Psi_{k}^{(0)}), and are independent of the parameter cc.

Substituting (5) into the equation

H(c)Ψk(c)=Ek(c)Ψk(c)\displaystyle H(c)\Psi_{k}(c)=E_{k}(c)\Psi_{k}(c) (6)

and matching coefficients of the same degree order yields the following set of equations

{Ek(p)=(Ψk(p1))TVΨk(0)(H0Ek(0))Ψk(p)=(Ek(1)V)Ψk(p1)+l=0p2Ek(pl)Ψk(l)\displaystyle\begin{cases}E_{k}^{(p)}=(\Psi_{k}^{(p-1)})^{T}V\Psi_{k}^{(0)}\\ (H_{0}-E_{k}^{(0)})\Psi_{k}^{(p)}=(E_{k}^{(1)}-V)\Psi_{k}^{(p-1)}+\sum_{l=0}^{p-2}E_{k}^{(p-l)}\Psi_{k}^{(l)}\end{cases} (7)

that allow us to compute Ek(p)E_{k}^{(p)} and Ψk(p)\Psi_{k}^{(p)} in a recursive fashion.

The asymptotic expansion used in (5) assumes that cc is a small parameter. As a result, the expansion serves as a good approximation to the desired eigenpair only when cc is sufficiently small, i.e., when cc falls within the radius of convergence for (5), which is generally much smaller than 1. As a result, (5) cannot be used directly in general to approximate the kkth eigenpair of HH [13, 14]. However, the perturbative vectors Ψk(p)\Psi_{k}^{(p)} can be used to construct a subspace

k(P):=span{Ψk(p):p=0,1,,P},\displaystyle\mathcal{M}_{k}^{(P)}:=span\{\Psi_{k}^{(p)}:p=0,1,\ldots,P\}, (8)

from which approximation to EkE_{k} and Ψk\Psi_{k} can be obtained.

The idea of using perturbative corrections to construct an approximating subspace was proposed in [13, 15] in the context of an eigenvector continuation (EC) method [16, 17]. In an EC method, the eigenvectors of H(c)H(c) for some choices of cc’s are used to construct a subspace from which approximations to the eigenvectors of H(c)H(c^{\prime}) for ccc^{\prime}\neq c are obtained from the projection of H(c)H(c^{\prime}) into such a subspace.

It was found in  [13, 15] that instead of using eigenvectors of H(c)H(c) for several choices of c1c\neq 1 to construct a subspace from which approximate eigenvectors of H(1)H(1) are extracted through the standard Rayleigh-Ritz procedure, more accurate approximations to the desired eigenpairs of H(1)H(1) can be obtained from the subspace constructed from the eigenvector of H0H_{0} as well as perturbative eigenvector corrections as discussed above.

In a Rayleigh-Ritz procedure, we compute an orthonormal basis matrix Qk(P)Q_{k}^{(P)} of k(P)\mathcal{M}_{k}^{(P)} and form the projected matrix

H~=(Qk(P))THQk(P).\tilde{H}=\left(Q_{k}^{(P)}\right)^{T}HQ_{k}^{(P)}. (9)

If (θ,v)(\theta,v) is an eigenpair of H~\tilde{H}, then (θ,z)(\theta,z) where z=Qk(P)vz=Q_{k}^{(P)}v, yields an approximate eigenpair of HH. We consider an approximate eigenpair to have converged if its relative residual norm

Hzθz2|θ|\displaystyle\frac{\|Hz-\theta z\|_{2}}{|\theta|} (10)

is smaller than a preset tolerance value.

To obtain an approximate kkth eigenpair (E~k,Ψ~k)(\tilde{E}_{k},\tilde{\Psi}_{k}) of HH, we choose (θ,v)(\theta,v) as the lowest eigenpair of H~\tilde{H} in the above mentioned Rayleigh-Ritz procedure. We present the SPPC method in Algorithm 1.

1
Input: A nuclear CI Hamiltonian Hn×nH\in\mathbb{R}^{n\times n} partitioned as H=H0+VH=H_{0}+V, where H0=diag(H^0,0)H_{0}=diag(\hat{H}_{0},0) with H^0\hat{H}_{0} constructed from a small configuration space (of dimension n0n_{0}); convergence tolerance (toltol); and maximum order of perturbation allowed (maxitermaxiter)
Output: An approximate kkth eigenpair (E~k,Ψ~k)(\tilde{E}_{k},\tilde{\Psi}_{k}) of HH
2
3Compute the kkth nonzero eigenpair (Ek(0),Ψk(0))(E_{k}^{(0)},\Psi_{k}^{(0)}) of H0H_{0}.
4Set θ=(Ψk(0))THΨk(0)\theta=(\Psi_{k}^{(0)})^{T}H\Psi_{k}^{(0)} and z=Ψk(0)z=\Psi_{k}^{(0)}.
5Return (E~k=θ,Ψ~k=z)(\tilde{E}_{k}=\theta,\tilde{\Psi}_{k}=z) if the relative residual norm (10) is less than toltol.
6for p=1,,maxiterp=1,\ldots,maxiter do
7      
8      Compute the correction energy Ek(p)E_{k}^{(p)} and correction vector Ψk(p)\Psi_{k}^{(p)}.
9      Compute an orthonormal basis matrix Qk(p)Q_{k}^{(p)} of k(p)\mathcal{M}_{k}^{(p)} and form a projected matrix H~\tilde{H}.
10      Compute the lowest eigenpair (θ,q)(\theta,q) of H~\tilde{H} and set z=Qk(p)qz=Q_{k}^{(p)}q.
11      Return (E~k=θ,Ψ~k=z)(\tilde{E}_{k}=\theta,\tilde{\Psi}_{k}=z) if the relative residual norm (10) is less than toltol.
12 end for
13
14
Algorithm 1 The SPPC for the kkth eigenpair

We should point out that the main distinction between the SPPC method proposed here and the EC method lies in the subspaces constructed in these methods and the cost of construction. Because EC projects HH onto a subspace constructed from the eigenvectors of H(c)H(c) for several (nonzero) cc’s, the cost of subspace construction may be just as expensive as the cost of solving the target eigenvalue problem with a particular choice of cc. On the other hand, because the subspace constructed in the SPPC method uses perturbative corrections that can be obtained by solving much smaller linear systems, the cost of subspace construction is significantly lower.

Although the basic idea of SPPC was presented in  [13, 15], the computational cost of this method was not carefully analyzed and compared with those of the state-of-the-art large-scale eigensolvers. In [13], the eigenvectors of H0H_{0} can be computed analytically for the one-dimensional quartic anharmonic oscillator. However, in general, identifying a H0H_{0} that can be diagonalized analytically is not possible. In [15], SPPC is used to perform a AA-body nuclear structure calculation. However, a different H=H0+VH=H_{0}+V splitting scheme is used, and the eigenvectors of H0H_{0} are not easier to compute than those of HH.

III Targeting First Few Eigenpairs

Although we can use Algorithm 1 to compute each of the first kevk_{ev} eigenpairs one by one (or in parallel), approximations to larger eigenvalues and the corresponding eigenvectors appear to converge slowly, and sometimes to wrong values, as we will show in section VI. A more effective way to obtain approximations to the first kevk_{ev} eigenpairs is to combine the SPPC subspace k(P)\mathcal{M}_{k}^{(P)} (8) constructed for each eigenpair to create a larger subspace

(P):=k=1kevk(P),\displaystyle\mathcal{M}^{(P)}:=\bigcup_{k=1}^{k_{ev}}\mathcal{M}_{k}^{(P)}, (11)

from which kevk_{ev} approximate eigenpairs can be extracted simultaneously through the Rayleigh-Ritz procedure, by targeting the kevk_{ev} lowest eigenpairs of the projected matrix. Algorithm 2 shows how this approach works.

1
Input: A nuclear CI Hamiltonian Hn×nH\in\mathbb{R}^{n\times n} partitioned as H=H0+VH=H_{0}+V, where H0=diag(H^0,0)H_{0}=diag(\hat{H}_{0},0) with H^0\hat{H}_{0} constructed from a small configuration space (of dimension n0n_{0}); number of desired eigenpairs (kevk_{ev}); convergence tolerance (toltol); and maximum order of perturbation allowed (maxitermaxiter)
2
Output: Approximate kevk_{ev} lowest eigenpairs {(E~k,Ψ~k)}k=1kev\{(\tilde{E}_{k},\tilde{\Psi}_{k})\}_{k=1}^{k_{ev}} of HH
3
4Compute the eigenpairs {(Ek(0),Ψk(0)}k=1kev\{(E_{k}^{(0)},\Psi_{k}^{(0)}\}_{k=1}^{k_{ev}} of H0H_{0}.
5Compute an orthonormal basis matrix Q(0)Q^{(0)} of (0)\mathcal{M}^{(0)} and form the projected matrix H~\tilde{H}.
6Compute the kevk_{ev} lowest eigenpairs {(θk,qk)}k=1kev\{(\theta_{k},q_{k})\}_{k=1}^{k_{ev}} of H~\tilde{H} and set {zk=Q(0)qk}k=1kev\{z_{k}=Q^{(0)}q_{k}\}_{k=1}^{k_{ev}}.
7Return {(E~k=θk,Ψ~k=zk)}k=1kev\{(\tilde{E}_{k}=\theta_{k},\tilde{\Psi}_{k}=z_{k})\}_{k=1}^{k_{ev}} if the relative residual norm (10) is less than toltol for all k=1,,kevk=1,\ldots,k_{ev}.
8for p=1,,maxiterp=1,\ldots,maxiter do
9      
10      Compute the correction energy Ek(p)E_{k}^{(p)} and correction vector Ψk(p)\Psi_{k}^{(p)} for k=1,,kevk=1,\ldots,k_{ev}.
11      Compute an orthonormal basis matrix Q(p)Q^{(p)} of (p)\mathcal{M}^{(p)} and form a projected matrix H~\tilde{H}.
12      Compute the kevk_{ev} lowest eigenpairs {(θk,qk)}k=1kev\{(\theta_{k},q_{k})\}_{k=1}^{k_{ev}} of H~\tilde{H} and set {zk=Q(p)qk}k=1kev\{z_{k}=Q^{(p)}q_{k}\}_{k=1}^{k_{ev}}.
13      Return {(E~k=θk,Ψ~k=zk)}k=1kev\{(\tilde{E}_{k}=\theta_{k},\tilde{\Psi}_{k}=z_{k})\}_{k=1}^{k_{ev}} if the relative residual norm (10) is less than toltol for all k=1,,kevk=1,\ldots,k_{ev}.
14 end for
15
Algorithm 2 The SPPC for the first few eigenpairs

IV Practical considerations

In this section, we describe a few practical implementation details of the SPPC algorithm.

IV.1 Computing the Correction Vectors

We refer to the kkth eigenvector of H0H_{0} denoted by Ψk(0)\Psi_{k}^{(0)} as the zero-th order correction to the kkth eigenvector of HH.

When H0H_{0} is of the form given in (3), Ψk(0)\Psi_{k}^{(0)} can be obtained by computing the kkth eigenvector of H^0\hat{H}_{0}, denoted by Ψ^k(0)\hat{\Psi}_{k}^{(0)}, and appending it with zeros to yield

Ψk(0)=[Ψ^k(0)0].\Psi_{k}^{(0)}=\begin{bmatrix}\hat{\Psi}_{k}^{(0)}\\ 0\end{bmatrix}. (12)

It follows from (7) and the block structures of H0H_{0}, VV, and Ψk(0){\Psi}_{k}^{(0)} that the first order corrections to the kkth eigenvalue and eigenvector of HH are

Ek(1)=0,Ψk(1)=1Ek(0)VΨk(0).\displaystyle E_{k}^{(1)}=0,\quad\Psi_{k}^{(1)}=\frac{1}{E_{k}^{(0)}}V\Psi_{k}^{(0)}. (13)

It is easy to verify that

Ψk(1)=HΨk(0)Ek(0)Ψk(0)Ek(0),\Psi_{k}^{(1)}=\frac{H\Psi_{k}^{(0)}-E_{k}^{(0)}\Psi_{k}^{(0)}}{E_{k}^{(0)}},

i.e., the first order correction to the eigenvector is simply the residual associated with the zero-th order approximation.

Because Ek(1)=0E_{k}^{(1)}=0, we can simplify the linear system in (7) to

(H0Ek(0))Ψk(p)=VΨk(p1)+l=0p2Ek(pl)Ψk(l).\displaystyle(H_{0}-E_{k}^{(0)})\Psi_{k}^{(p)}=-V\Psi_{k}^{(p-1)}+\sum_{l=0}^{p-2}E_{k}^{(p-l)}\Psi_{k}^{(l)}. (14)

The matrix H0Ek(0)H_{0}-E_{k}^{(0)} is in a block diagonal form

H0Ek(0)=[H^0Ek(0)00Ek(0)I]\displaystyle H_{0}-E_{k}^{(0)}=\begin{bmatrix}\hat{H}_{0}-E_{k}^{(0)}&0\\ 0&-E_{k}^{(0)}I\end{bmatrix} (15)

consisting of two blocks where the first block H^0Ek(0)n0×n0\hat{H}_{0}-E_{k}^{(0)}\in\mathbb{R}^{n_{0}\times n_{0}} is relatively small and the second block Ek(0)I(nn0)×(nn0)-E_{k}^{(0)}I\in\mathbb{R}^{(n-n_{0})\times(n-n_{0})} is a scalar multiple of the identity matrix. As a result, the solution of the linear system in (14) essentially reduces to the solution of a much smaller linear system with H^0Ek(0)\hat{H}_{0}-E_{k}^{(0)} being the coefficient matrix. If we partition Ψk(p)\Psi_{k}^{(p)} conformally with the blocks in (15) as Ψk(p)=[x1,x2]T\Psi_{k}^{(p)}=[x_{1},x_{2}]^{T} and the right-hand side of (14) as b=[b1,b2]Tb=[b_{1},b_{2}]^{T} such that x1,b1n0x_{1},b_{1}\in\mathbb{R}^{n_{0}} and x2,b2nn0x_{2},b_{2}\in\mathbb{R}^{n-n_{0}}, x2x_{2} can be easily computed as

x2=1Ek(0)b2,\displaystyle x_{2}=-\frac{1}{E_{k}^{(0)}}b_{2}, (16)

and x1x_{1} can be obtained by solving

(H^0Ek(0))x1=b1.\displaystyle(\hat{H}_{0}-E_{k}^{(0)})x_{1}=b_{1}. (17)

Note that equation (14) is singular because Ek(0)E_{k}^{(0)} is an eigenvalue of H0H_{0}. However, since k(P)\mathcal{M}_{k}^{(P)} already includes Ψk(0)\Psi_{k}^{(0)} and we are only interested in contributions in the orthogonal complement of Ψk(0)\Psi_{k}^{(0)} from the solution of (14), we can project out Ψk(0)\Psi_{k}^{(0)} from the right-hand side of (14) before solving this equation. This is equivalent to projecting out Ψ^k(0)\hat{\Psi}_{k}^{(0)} from the right-hand side of (17), i.e. we solve

(H^0Ek(0))x1=b1(b1TΨ^k(0))Ψ^k(0).\displaystyle(\hat{H}_{0}-E_{k}^{(0)})x_{1}=b_{1}-(b_{1}^{T}\hat{\Psi}_{k}^{(0)})\hat{\Psi}_{k}^{(0)}. (18)

IV.2 Rayleigh-Ritz Calculation

After obtaining the correction vectors, we generate an orthonormal basis matrix of the subspace spanned by these vectors and then perform the Rayleigh-Ritz procedure. We lay out these steps for approximating a single eigenpair and the first few eigenpairs.

Targeting the kkth eigenpair.

We use the Gram-Schmidt process to obtain an orthonormal basis of the subspace k(p)\mathcal{M}_{k}^{(p)}. The orthonormal basis forms the columns of the matrix Qk(p)n×(p+1)Q_{k}^{(p)}\in\mathbb{R}^{n\times(p+1)}. The (p+1)(p+1)th column, denoted by qkq_{k}, is generated as follows.

Φk(p)=[IQk(p1)(Qk(p1))T]Ψk(p),qk(p)=Φk(p)Φk(p)2.\begin{split}\Phi_{k}^{(p)}&=\left[I-Q_{k}^{(p-1)}(Q_{k}^{(p-1)})^{T}\right]\Psi_{k}^{(p)},\\ q_{k}^{(p)}&=\frac{\Phi_{k}^{(p)}}{\|\Phi_{k}^{(p)}\|_{2}}.\end{split} (19)

We append qk(p)nq_{k}^{(p)}\in\mathbb{R}^{n} to Qk(p1)Q_{k}^{(p-1)} such that

Qk(p)=[Qk(p1),qk(p)].\displaystyle Q_{k}^{(p)}=[Q_{k}^{(p-1)},q_{k}^{(p)}]. (20)

Note that the projected matrix (Qk(p))THQk(p)(Q_{k}^{(p)})^{T}HQ_{k}^{(p)} can be constructed recursively. Assuming (Qk(p1))THQk(p1)(Q_{k}^{(p-1)})^{T}HQ_{k}^{(p-1)} has been computed in the previous step, we just need to compute Hqk(p)Hq_{k}^{(p)} and append an additional row and column to (Qk(p1))THQk(p1)(Q_{k}^{(p-1)})^{T}HQ_{k}^{(p-1)} as shown below.

[(Qk(p1))THQk(p1)(Qk(p1))THqk(p)(qk(p))THQk(p1)(qk(p))THqk(p)].\begin{bmatrix}(Q_{k}^{(p-1)})^{T}HQ_{k}^{(p-1)}&(Q_{k}^{(p-1)})^{T}Hq_{k}^{(p)}\\ (q_{k}^{(p)})^{T}HQ_{k}^{(p-1)}&(q_{k}^{(p)})^{T}Hq_{k}^{(p)}\end{bmatrix}. (21)

Therefore, the major cost for constructing the projected matrix in each step of the SPPC method is in performing a single SpMV in Hqk(p)Hq_{k}^{(p)}.

Targeting the first few eigenpairs.

To obtain an orthonormal basis Q(p)n×kev(p+1)Q^{(p)}\in\mathbb{R}^{n\times k_{ev}(p+1)} for the combined subspace (p)\mathcal{M}^{(p)}, we replace the Gram-Schmidt process (19) with the block Gram-Schmidt process. If Ψ(p)=[Ψ1(p),Ψ2(p),,Ψkev(p)]\Psi^{(p)}=[\Psi_{1}^{(p)},\Psi_{2}^{(p)},\ldots,\Psi_{k_{ev}}^{(p)}], the block Gram-Schmidt procedure yields

Φ(p)=[IQ(p1)(Q(p1))T]Ψ(p).\Phi^{(p)}=\left[I-Q^{(p-1)}(Q^{(p-1)})^{T}\right]\Psi^{(p)}. (22)

We then perform a QR factorization of Φ(p)\Phi^{(p)}, i.e.,

Φ(p)=q(p)R(p),\Phi^{(p)}=q^{(p)}R^{(p)}, (23)

to generate an orthonormal basis q(p)q^{(p)} for Φ(p)\Phi^{(p)}. We append q(p)n×kevq^{(p)}\in\mathbb{R}^{n\times k_{ev}} to Q(p1)Q^{(p-1)} such that

Q(p)=[Q(p1),q(p)]\displaystyle Q^{(p)}=[Q^{(p-1)},q^{(p)}] (24)

is an orthonormal basis for (p)\mathcal{M}^{(p)}.

Again, the projected matrix (Q(p))THQ(p)(Q^{(p)})^{T}HQ^{(p)} can be constructed recursively, and the computational cost is dominated by the cost for computing kevk_{ev} SpMVs in Hq(p)Hq^{(p)}.

IV.3 Computational Cost

We now discuss the overall computational cost of the SPPC method. We can see from Algorithm 1 and Algorithm 2 that the two major components of the SPPC algorithm are: (1) Solving the linear system (14); (2) forming the projected matrix (Q(p))THQ(p)(Q^{(p)})^{T}HQ^{(p)}. We have already shown that the projected matrix can be computed recursively using kevk_{ev} SpMVs in each step of the SPPC algorithm. The reduced linear system (17) of the correction equation (14) can be solved iteratively using, for example, the MINRES algorithm. Because it has a much smaller dimension, the SpMVs performed in each MINRES iteration are relatively cheap. However, forming the right-hand side of the equation (14) requires multiplying VV with Ψk(p1)\Psi_{k}^{(p-1)} for p>1p>1 which has nearly the same complexity as multiplying HH with Ψk(p1)\Psi_{k}^{(p-1)}. Therefore, it may appear that each SPPC step requires performing 2kev2k_{ev} SpMVs. We will show below that this is not the case. Both the right-hand side of (14) and the projected matrix can be obtained from the same Hq(p1)Hq^{(p-1)} product. As a result, each step of the SPPC algorithm only requires performing kevk_{ev} SpMVs.

Using the matrix splitting H=H0+VH=H_{0}+V, we can rewrite VΨ(p1)V\Psi^{(p-1)} as

VΨ(p1)=HΨ(p1)H0Ψ(p1)V\Psi^{(p-1)}=H\Psi^{(p-1)}-H_{0}\Psi^{(p-1)} (25)

where Ψ(p1)=[Ψ1(p1),,Ψkev(p1)]\Psi^{(p-1)}=\left[\Psi_{1}^{(p-1)},\ldots,\Psi_{k_{ev}}^{(p-1)}\right]. Therefore, VΨ(p1)V\Psi^{(p-1)} can be obtained by subtracting H0Ψ(p1)H_{0}\Psi^{(p-1)}, a much lower computational cost, from HΨ(p1)H\Psi^{(p-1)}.

We now show that HΨ(p1)H\Psi^{(p-1)} can be easily obtained from Hq(p1)Hq^{(p-1)}. It follows from (22) and (23) that

H[IQ(p2)(Q(p2))T]Ψ(p1)=Hq(p1)R(p1).H\left[I-Q^{(p-2)}(Q^{(p-2)})^{T}\right]\Psi^{(p-1)}=Hq^{(p-1)}R^{(p-1)}. (26)

As a result, we can obtain HΨ(p1)H\Psi^{(p-1)} from Hq(p1)Hq^{(p-1)} by using following identity

HΨ(p1)=HQ(p2)(Q(p2))TΨ(p1)+Hq(p1)R(p1).H\Psi^{(p-1)}=HQ^{(p-2)}(Q^{(p-2)})^{T}\Psi^{(p-1)}+Hq^{(p-1)}R^{(p-1)}. (27)

Note that the analysis of the computational cost assumes HQ(p2)HQ^{(p-2)}, which contains Hq(j)Hq^{(j)} as its columns for j=0,1,,p2j=0,1,...,p-2, has been stored in memory.

As we will show in section VI, the highest order perturbation is often limited to 15, beyond which no significant improvement in the approximate eigenpair can be observed. Therefore, the dense linear algebra operations such as computing (Q(p2))TΨ(p1)(Q^{(p-2)})^{T}\Psi^{(p-1)} and diagonalizing the projected matrix can be performed with a relatively low cost compared to the cost of multiplying HH with q(p1)q^{(p-1)}.

V Combining SPPC with other algorithms

As we will show in the next section, the perturbative correction is typically effective when the order of perturbation pp is relatively low. The convergence of SPPC can stagnate when pp increases, i.e., adding higher order perturbative correction may not help because they may be linearly dependent with respect to the basis vectors included in (p)\mathcal{M}^{(p)} already. In this case, it is useful to combine SPPC with another algorithm that can take the eigenvector approximation produced by SPPC as the starting guess.

Algorithms that can be combined with the SPPC method in a hybrid algorithm include but are not limited to the following:

  • The Lanczos algorithm, which is a classical algorithm that generates an orthonormal basis of a Krylov subspace using the Gram-Schmidt procedure. It is initialized with an approximate eigenvector produced from SPPC or a linear combination of kevk_{ev} approximate eigenvectors. It uses one SpMV per iteration to compute the next basis vector. The eigenpairs are approximated using Ritz pairs obtained from the Rayleigh-Ritz procedure with the basis vectors of the Krylov subspace.

  • The block Lanczos algorithm, which is a variation of the Lanczos algorithm that operates in blocks. It is initialized with a block of vectors approximating several eigenvectors of HH, and builds a Krylov subspace in blocks, with each iteration performing kevk_{ev} SpMVs. The main advantage of the block Lanczos algorithm over the Lanczos algorithm is that it can make use of approximations to several eigenvectors more effectively and most dense linear algebra operations can take advantage of level 3 BLAS.

  • The LOBPCG algorithm, which is an iterative method that solves the equivalent trace minimization formulation of the eigenvalue problem. Similar to the block Lanczos algorithm, it can be initialized with approximations to several eigenvectors and each iteration performs kevk_{ev} SpMVs. One advantage of the LOBPCG algorithm is that it can utilize a preconditioner if it is available. The use of a preconditioner can accelerate convergence. A common choice for the preconditioner is a block diagonal part of the Hamiltonian matrix or a shifted matrix of the Hamiltonian for some appropriately chosen shift [7, 4]. For our numerical experiments, we use a shifted preconditioner that involves a specific block diagonal part of the Hamiltonian matrix and a constant shift that approximates the lowest eigenvalue.

  • The residual minimization method (RMM) combined with the Direct Inversion of Iterative Subspace (DIIS) refinement algorithm (RMM-DIIS), which is a quasi-Newton algorithm for improving a specific eigenpair without computing other eigenpairs, provided that the initial approximation to the desired eigenpairs is sufficiently accurate. Each RMM-DIIS iteration performs one SpMV. The RMM-DIIS can also incorporate a preconditioner to accelerate convergence. For our numerical experiments, we choose a preconditioner that is a specific diagonal part of the Hamiltonian matrix.

VI Numerical Results

In this section, we demonstrate the effectiveness of SPPC algorithm and compare it with other existing algorithms for computing the ground and a few low excited states of several light nuclei. We also show that SPPC can be effectively combined with RMM-DIIS to yield an efficient and accurate hybrid eigensolver for nuclear configuration interaction calculations. We call this hybrid eigensolver the SPPC+RMM-DIIS method.

For all algorithms tested in this section, we use the relative residual norm (10) as the stopping criteria and set the convergence tolerance to be 10610^{-6}. To ensure a fair comparison with the SPPC, we use the eigenvectors of the zero-order part H0H_{0} as the initial guesses for the algorithms. All experiments were conducted using MATLAB.

VI.1 Test Matrices

We use the AA-body Hamiltonian matrices corresponding to the nuclei 6Li, 7Li, 11B, and 12C in the following numerical experiments. The superscripts indicate the number of nucleons in the nuclei; for example, 7Li indicates Lithium with 3 protons plus 4 neutrons. The Hamiltonian matrices HH are constructed in a truncated CI space defined by a truncation parameter NmaxN_{\max}, using the nucleon-nucleon interaction Daejeon16. For the same nucleus, a larger NmaxN_{\max} results in a larger matrix HH, but the size of the matrix is independent of the interaction. Note that with three-nucleon interactions, the number of nonzero matrix elements is an order of magnitude larger than the one for the nucleon-nucleon interactions, and the number of iterations and the actual eigenvalues of any eigensolver will be different. As we indicated earlier, the construction of HH can be done in a hierarchical fashion so that a leading submatrix H^0\hat{H}_{0} of HH corresponds to the same nuclear AA-body Hamiltonian represented in a lower dimensional configuration space associated with a smaller NmaxN_{\max}. In this section, if HH is the matrix representation of a nuclear AA-body Hamiltonian represented in a configuration space associated with Nmax=ncN_{\max}=n_{c}, the submatrix H^0\hat{H}_{0} corresponds to the representation of the same Hamiltonian in a configuration space associated with Nmax=nc2N_{\max}=n_{c}-2. We list the dimension of HH (denoted by dim(HH)), the dimension of H^0\hat{H}_{0} (denoted by dim(H^0\hat{H}_{0})), the number of non-zero elements in HH (denoted by nnz(HH)), H^0\hat{H}_{0} (denoted by nnz(H^0\hat{H}_{0})), and the NmaxN_{\max} value associated with HH in Table 1.

Table 1: Properties of the test matrices HH.
Nucleus NmaxN_{\max} dim(HH) dim(H^0\hat{H}_{0}) nnz(HH) nnz(H^0)\hat{H}_{0})
6Li 6 197,822 17,040 106,738,802 4,122,448
7Li 6 663,527 48,917 421,938,629 14,664,723
11B 4 814,092 16,097 389,033,682 2,977,735
12C 4 1,118,926 17,725 555,151,572 3,365,009

VI.2 Targeting the Lowest Eigenpair

We report the performance of the SPPC, the Lanczos algorithm, the LOBPCG algorithm, and the SPPC+RMM-DIIS for targeting the lowest eigenpair of the matrix HH. As mentioned earlier, we use a preconditioner for the LOBPCG and the RMM-DIIS algorithms. The primary cost of all these algorithms is the number of SpMVs they perform before reaching convergence. Because each algorithm performs one SpMV per iteration, we can directly compare them by the number of iterations required to reach convergence.

The left plot of Figure 2 shows the convergence history of the algorithms chosen for comparison for 12C with respect to the iteration number. We observe that the SPPC algorithm converges in 1616 iterations, which is the least among all methods, while the Lanczos algorithm and the LOBPCG algorithm converge in 2222 iterations.

Refer to caption
Refer to caption
Figure 2: The convergence of the algorithms for computing the lowest eigenpair of the Hamiltonians matrices (12C on the left and 6Li on the right). One iteration equals one SpMV for all algorithms.

The result shown in the right plot of Figure 2 is for the Hamiltonian associated with 6Li. Several features of the SPPC algorithm are observed. The first observation is that the SPPC method converges more rapidly in the early iterations (up to the 15th iteration), and can be up to at most two orders of magnitude more accurate than other algorithms in these early iterations. However, the SPPC approximation appears to stagnate in subsequent iterations. This suggests that higher order correction vectors produced in later iterations (after iteration 15) do not contribute to improving the subspace constructed by the correction vectors produced in the early iterations. To verify this conjecture, we plot the angle between the current correction vector and the subspace spanned by the previous correction vectors, denoted by (Ψ(p),(p1))\angle(\Psi^{(p)},\mathcal{M}^{(p-1)}), with respect to the iteration number pp in Figure 3. We observe that (Ψ(p),(p1))\angle(\Psi^{(p)},\mathcal{M}^{(p-1)}) is relatively large in the first few SPPC iterations, and gradually decreases to the level of 10510^{-5}. This is the point at which the new correction vector contributes minimally to the expansion of the subspace.

Refer to caption
Figure 3: The subspace angle between the correction vector at iteration pp and the subspace spanned by the previous correction vectors from iteration 11 to p1p-1.

To overcome this stagnation, we consider a hybrid approach, the SPPC+RMM-DIIS, where we use the SPPC until the point of stagnation, and then switch to the RMM-DIIS. This hybrid approach takes advantage of the fast convergence of the SPPC for the first few iterations and the fast convergence of the RMM-DIIS when initialized with a good initial guess to the desired eigenvector. Specifically, we choose the initial guess in the RMM-DIIS as the Ritz vector produced from the SPPC method at the point of the switch. For 6Li, we use the RMM-DIIS after the 1515th iteration of the SPPC. We observe that the SPPC+RMM-DIIS breaks the stagnation of the SPPC and converges in 2323 iterations, while the Lanczos algorithm and the LOBPCG algorithm converge in 3131 and 2525 iterations, respectively.

Table 2 gives a comparison of the SpMV counts used by several algorithms tested in this section for all four Hamiltonian matrices. With the exception of the Hamiltonian of the nucleus 12C, the SPPC stagnates around iteration number 15. For these cases, we also consider the SPPC+RMM-DIIS. We observe that this hybrid approach converges the fastest with the fewest SpMVs performed.

Table 2: SpMV count of the algorithms for computing the lowest eigenpair. The convergence of the SPPC stagnates around 15 iterations for the Hamiltonians 6Li, 7Li, and 11B. For these three Hamiltonians, the hybrid method, the SPPC+RMM-DIIS, is also considered where the RMM-DIIS switches with the SPPC after iteration 15.
Nucleus SPPC Lanczos LOBPCG SPPC+RMM-DIIS
6Li >30 31 25 23
7Li >30 31 25 24
11B >30 30 27 18
12C 16 22 22

VI.3 Targeting the Five Lowest Eigenpairs

We can use either Algorithm 1 or Algorithm 2 to compute the lowest kevk_{ev} eigenvalues. In the left plots of Figure 4 and Figure 5, we show the relative residual norms of the approximation to the 55 lowest eigenpairs of the 6Li Hamiltonian at each iteration of Algorithm 1 and Algorithm 2, respectively. The relative residual norms associated with the first three eigenpairs drop below 10410^{-4} by the 1515th iteration. However, for the fourth and the fifth eigenpairs, the relative residuals obtained by Algorithm 1 jump at a certain point and never become small in subsequent iterations. The convergence failure for these eigenpairs can also be seen from the change of Ritz values at each iteration and how they compare with the true eigenvalues of the Hamiltonian as shown in the right plot of Figure 4. It appears that the 4th and 5th Ritz values move towards a different eigenvalue. This is due to a significant round off error in the orthogonalization process where the new basis contains contribution from the other eigenvector. In contrast, as shown in the right plot of Figure 5, by constructing a larger subspace consisting of perturbative corrections to several eigenvectors, Algorithm 2 computes approximate eigenvalues that do not deviate from the true eigenvalues, and as a result, all of its computed eigenpairs converge within a reasonable accuracy. Similar behaviors are observed for the other three Hamiltonians: 7Li, 11B, and 12C.

Refer to caption
Refer to caption
Figure 4: Algorithm 1 to compute the first 55 eigenpairs of the 6Li Hamiltonian, one by one. The left plot shows the relative residual norm, and the right plot shows the approximated eigenvalues in comparison with the true eigenvalues.
Refer to caption
Refer to caption
Figure 5: Algorithm 2 to compute the first 55 eigenpairs of the 6Li Hamiltonian. The left plot shows the relative residual norm, and the right plot shows the approximated eigenvalues in comparison with the true eigenvalues.

We now show that the stagnation in the convergence of the first Ritz pairs in the SPPC method can be eliminated by using a hybrid SPPC+RMM-DIIS. We switch to the RMM-DIIS method from the SPPC method when the relative residual norm of any of the eigenpairs starts to stagnate. In this hybrid approach, a separate RMM-DIIS run is used to refine each approximate eigenpair. It is initialized with the corresponding Ritz vector returned from the SPPC method at the point of the switch.

Figure 6 illustrates the convergence of the SPPC+RMM-DIIS for the 12C Hamiltonian. We choose to switch to the RMM-DIIS from the SPPC at iteration 1515, which is a point of stagnation for most eigenpairs. We observe that the SPPC+RMM-DIIS breaks the stagnation of the SPPC and that all 55 eigenpairs converge rapidly.

Refer to caption
Figure 6: The convergence of the SPPC+RMM-DIIS for computing the 5 lowest eigenapirs of the Hamiltonian 12C. The RMM-DIIS switches with the SPPC after iteration number 15. One iteration of an individual RMM-DIIS costs one SpMVs.

In Table 3, we present the SpMV counts of the block Lanczos algorithm, the LOBPCG algorithm, and the SPPC+RMM-DIIS for computing the 55 lowest eigenpairs of the four Hamiltonians. The switch to the RMM-DIIS algorithm occurs for the SPPC+RMM-DIIS after the 15th iteration for all four Hamiltonian matrices. The block Lanczos algorithm and the LOBPCG algorithm are block methods, thus requiring continued iterations until every eigenpair converges and consequently a full kevk_{ev} SpMVs at each iteration. In contrast, the SPPC+RMM-DIIS targets each eigenpair individually after it switches to RMM-DIIS, so it does not incur more SpMVs for the converged eigenpair as it targets the non-converged eigenpairs. Due to this advantage, we observe that the SPPC+RMM-DIIS converges the fastest.

Table 3: SpMV count of the block Lanczos algorithm, the LOBPCG algorithm, and the SPPC+RMM-DIIS for computing the 5 lowest eigenpairs. The convergence of the SPPC tends to stagnate for some eigenpairs so the hybrid approach, the SPPC+RMM-DIIS, is considered. The RMM-DIIS switches from the SPPC after the 15th iteration.
Nucleus Block Lanczos LOBPCG SPPC+RMM-DIIS
6Li 155 140 127
7Li 175 150 129
11B 170 125 94
12C 155 165 90

VII Conclusion

In conclusion, the Subspace Projection with Perturbative Corrections (SPPC) method, combined with the Residual Minimization Method with Direct Inversion of Iterative Subspace (RMM-DIIS), presents a significant advancement in the efficient computation of eigenpairs for large Hamiltonian matrices in nuclear structure calculations. The SPPC method leverages perturbative correction vectors to enhance the accuracy of eigenpair approximations in the initial iterations, substantially reducing the number of sparse matrix-vector multiplications (SpMVs) required for convergence. Although the SPPC may experience stagnation in subsequent iterations, this challenge is effectively mitigated by integrating it with the RMM-DIIS algorithm, which provides robust refinement of eigenvector approximations. Our numerical experiments across several nuclear Hamiltonians demonstrate that the SPPC+RMM-DIIS hybrid approach outperforms traditional methods in terms of SpMVs. This hybrid method offers a promising solution for large-scale nuclear structure calculations, providing a reliable and efficient approach to solving the AA-body Schrödinger equation.

While the preliminary results of the SPPC are promising, we have not provided a theoretical background explaining why it works. We plan to address this in our future work, discussing the convergence behavior in detail. From a practical standpoint, w e aim to develop a method to automatically detect stagnation, eliminating the need for manual decisions on when to switch to the RMM-DIIS. Additionally, we are interested in implementing the SPPC in a hybrid MPI/OpenMPI code, such as the software MFDn (Many-Fermion Dynamics for nuclear structure) [18, 19, 20], to be run at high-performance computing centers. We also want to explore further optimizations and applications of the SPPC to other large-scale eigenvalue problems in nuclear physics and beyond.

Acknowledgements.
This material is based upon work supported by the Scientific Discovery through Advanced Computing (SciDAC) Program at the U.S. Department of Energy (DOE), Office of Science under Grants DE-SC0023495 and DE-SC0023175 from the Office of Nuclear Physics, and also under funding for the FASTMath Institute from the Office of Advanced Scientific Computing Research through Contract No. DE-AC02-05CH11231. This work used resources at the National Energy Research Scientific Computing Center (NERSC), which is funded by DOE under Contract No. DE-AC02-05CH11231. In addition, D.L. acknowledges partial support from DOE grants DE-SC0024586, DE-SC0023658, DE-SC0013365, and NSF grant PHY-2310620. J.P.V. acknowledges partial support from DOE grants DE-SC0023692 and DE-0023707.

References

  • Barrett et al. [2013] B. R. Barrett, P. Navrátil, and J. P. Vary, Ab initio no core shell model, Progress in Particle and Nuclear Physics 69, 131 (2013).
  • Shirokov et al. [2016] A. Shirokov, I. Shin, Y. Kim, M. Sosonkina, P. Maris, and J. Vary, N3lo nn interaction adjusted to light nuclei in ab exitu approach, Physics Letters B 761, 87 (2016).
  • Shao et al. [2018] M. Shao, H. M. Aktulga, C. Yang, E. G. Ng, P. Maris, and J. P. Vary, Accelerating nuclear configuration interaction calculations through a preconditioned block iterative eigensolver, Computer Physics Communications 222, 1 (2018).
  • Alperen et al. [2023] A. Alperen, H. M. Aktulga, P. Maris, and C. Yang, Hybrid eigensolvers for nuclear configuration interaction calculations, Computer Physics Communications 292, 108888 (2023).
  • Lanczos [1950] C. Lanczos, An iteration method for the solution of the eigenvalue problem of linear differential and integral operators,   (1950).
  • Golub and Underwood [1977] G. H. Golub and R. Underwood, The block lanczos method for computing eigenvalues, in Mathematical software (Elsevier, 1977) pp. 361–377.
  • Knyazev [2001] A. V. Knyazev, Toward the optimal preconditioned eigensolver: Locally optimal block preconditioned conjugate gradient method, SIAM journal on scientific computing 23, 517 (2001).
  • Wood and Zunger [1985] D. Wood and A. Zunger, A new method for diagonalising large matrices, Journal of Physics A: Mathematical and General 18, 1343 (1985).
  • Jia and Stewart [2001] Z. Jia and G. Stewart, An analysis of the rayleigh–ritz method for approximating eigenspaces, Mathematics of computation 70, 637 (2001).
  • Pulay [1980] P. Pulay, Convergence acceleration of iterative sequences. the case of scf iteration, Chemical Physics Letters 73, 393 (1980).
  • Hernandez et al. [2021] T. M. Hernandez, R. Van Beeumen, M. A. Caprio, and C. Yang, A greedy algorithm for computing eigenvalues of a symmetric matrix with localized eigenvectors, Numerical Linear Algebra with Applications 28, e2341 (2021).
  • Shavitt and Bartlett [2009] I. Shavitt and R. J. Bartlett, Many-body methods in chemistry and physics: MBPT and coupled-cluster theory (Cambridge university press, 2009).
  • Franzke et al. [2022] M. C. Franzke, A. Tichai, K. Hebeler, and A. Schwenk, Excited states from eigenvector continuation: The anharmonic oscillator, Physics Letters B 830, 137101 (2022).
  • Duguet et al. [2023] T. Duguet, A. Ekström, R. J. Furnstahl, S. König, and D. Lee, Eigenvector continuation and projection-based emulators, arXiv preprint arXiv:2310.19419  (2023).
  • Demol et al. [2020] P. Demol, T. Duguet, A. Ekström, M. Frosini, K. Hebeler, S. König, D. Lee, A. Schwenk, V. Somà, and A. Tichai, Improved many-body expansions from eigenvector continuation, Physical Review C 101, 041302 (2020).
  • Frame et al. [2018] D. Frame, R. He, I. Ipsen, D. Lee, D. Lee, and E. Rrapaj, Eigenvector continuation with subspace learning, Physical review letters 121, 032501 (2018).
  • Sarkar and Lee [2021] A. Sarkar and D. Lee, Convergence of eigenvector continuation, Physical review letters 126, 032501 (2021).
  • Sternberg et al. [2008] P. Sternberg, E. G. Ng, C. Yang, P. Maris, J. P. Vary, M. Sosonkina, and H. V. Le, Accelerating configuration interaction calculations for nuclear structure, in SC’08: Proceedings of the 2008 ACM/IEEE Conference on Supercomputing (IEEE, 2008) pp. 1–12.
  • Maris et al. [2010] P. Maris, M. Sosonkina, J. P. Vary, E. Ng, and C. Yang, Scaling of ab-initio nuclear physics calculations on multicore computer architectures, Procedia Computer Science 1, 97 (2010).
  • Aktulga et al. [2014] H. M. Aktulga, C. Yang, E. G. Ng, P. Maris, and J. P. Vary, Improving the scalability of a symmetric iterative eigensolver for multi-core platforms, Concurrency and Computation: Practice and Experience 26, 2631 (2014).