Continuous-variable neural-network quantum states and the quantum rotor model
Abstract.
We initiate the study of neural-network quantum state algorithms for analyzing continuous-variable lattice quantum systems in first quantization. A simple family of continuous-variable trial wavefunctons is introduced which naturally generalizes the restricted Boltzmann machine (RBM) wavefunction introduced for analyzing quantum spin systems. By virtue of its simplicity, the same variational Monte Carlo training algorithms that have been developed for ground state determination and time evolution of spin systems have natural analogues in the continuum. We offer a proof of principle demonstration in the context of ground state determination of a stoquastic quantum rotor Hamiltonian. Results are compared against those obtained from partial differential equation (PDE) based scalable eigensolvers. This study serves as a benchmark against which future investigation of continuous-variable neural quantum states can be compared, and points to the need to consider deep network architectures and more sophisticated training algorithms.
1. Introduction
Variational Monte Carlo (VMC) approaches to the quantum many-body problem [1] have witnessed a recent resurgence in activity fueled by the realization that when neural networks are exploited as systematically improvable trial wavefunctions, direct attack on otherwise intractable quantum many-body systems becomes possible. The success of so-called neural-network quantum states [2] is closely paralleled by the ability of deep neural networks to overcome a related curse-of-dimensionality in a variety of high-dimensional machine learning tasks. The key features shared by these learning tasks is that they involve iterative fitting of a high-dimensional function approximator using various forms of stochastic approximation to an objective function, for which the learner has incomplete knowledge in the form of samples. Applying this perspective to the VMC for ground state determination, the variational wavefunction acts as a data generating process from which IID samples can be drawn. These samples provide noisy estimates for a Rayleigh quotient to be optimized, whose exact dependence on the variational parameters is unknown.
In this paper, we argue that quantum many-body simulation can leverage the success of geometric deep learning [3, 4] from two different perspectives which are based on group invariances of the Hamiltonian and the space of states, respectively. The first perspective applies whenever the configuration space of the Hamiltonian has a non-Euclidean structure, as in the hyperbolic lattices [5, 6] currently under investigation using circuit quantum electrodynamics [7]. In the second perspective, one is concerned with efficiently describing states of the Hilbert space which are invariant or equivariant to group invariances of the Hamiltonian. In the theory of many-body Schrödinger operators, for example, it can be shown that the unrestricted ground state inherits any group invariances of the parent Hamiltonian.
Motivated by the above simulation prospects, in this paper we initiate the study of continuous-variable VMC on non-Euclidean spaces, proposing a Hamiltonian which can be summarized by a small amount of geometric data, whose invariances are respected by the quantum theory. We limit our investigation to the simplest non-trivial target space geometry corresponding to a system of quantum rotors. Ref. [8] undertook a variational ground state study of the quantum rotor model in one spatial dimension using basis truncation and matrix product states. The neural network approach advocated here, in contrast, does not rely on a choice of basis and could potentially be advantageous in the analysis of more complicated geometries such as hyperbolic manifolds, which are central to the study of quantum chaos [9]. Another possible advantage of the neural network approach compared to [8] is the lack of reliance on the density matrix renormalization group training algorithm [10, 11], which tends to struggle beyond one spatial dimension or in applications to disordered systems without a lattice.
Rather than pursuing the full machinery of geometric deep learning for variational simulation, we choose to focus on the introduction of baselines and illustration of general techniques using minimalist architectures which generalize those originally introduced for quantum spin systems [2]. In particular, experiments are conducted using a rotor variant of the restricted Boltzmann machine applicable to the planar quantum rotor model, which can be understood as a continuous-variable relaxation of the quantum Ising model. The results are compared against scalable PDE-based eigensolvers and the extension of the method to other geometries is discussed.
The paper is structured as follows: we first introduce a geometrically motivated Hamiltonian, highlighting some subtleties that arise in non-Euclidean space. The planar quantum rotor model is identified as the simplest candidate system for illustrating the applicability of geometric machine learning techniques. Baselines are introduced for ground state preparation of the rotor model, which we investigate using a combination of variational techniques inspired by shallow neural networks, as well as techniques based on scalable PDE eigensolvers. We conclude by summarizing future directions.
2. States and Hamiltonian
The quantum systems under consideration describe finitely many particles moving on a Riemannian target space subject to two-body interactions. In particular, given a choice of Riemannian manifold , a finite simple undirected graph decorated by vertex and edge weights functions and respectively, and a choice of potential function , we define the generator of time evolution in the infinite-dimensional Hilbert space of states as follows,
(1) |
where denotes the Laplace-Beltrami operator acting on and denotes the Riemannian distance function. Since the Hamiltonian depends only on the intrinsic geometry of the Riemannian manifold (via the distance and Laplace operator), the associated quantum theory inherits the invariances of the Riemannian space and likewise for the invariances of the interaction graph. Under appropriate assumptions on the potential, these invariances are inherited by the unique positive ground state of [12]. The geometry of the target space has non-trivial implications for the smoothness of the potential energy. In particular, the distance function , viewed as a function of its second argument, will generically suffer from cusp singularities at and at any points admitting multiple minimizing geodesics to (the so-called cut locus).
The primary goals of this paper are the development of scalable variational approaches to solve both the ground state eigenvalue problem and the time evolution problem corresponding to the Schrödinger operator (1). In this paper, however, we only consider the ground state eigenvalue problem, which can be rephrased as the following unconstrained optimization problem over ,
(2) |
The simplest Riemannian manifold after Euclidean space is the -dimensional unit sphere . In the interests of simplicity we focus in this initial work on , corresponding to the unit circle . In terms of the implicit angular parametrization , the Riemannian distance function on the circle is given by
(3) |
which exhibits the expected cusp singularities at antipodal points defined by . In this example the cut locus is a single point. The singularities of the distance function can smoothed out by a suitable choice of potential, which we take to be of cosinusoidal form. The resulting Hamiltonian is that of the quantum rotor model formulated on an arbitrary graph,
(4) |
3. Rotor Restricted Boltzmann Machine
Inspired by the restricted Boltzamnn neural network quantum state introduced in [2], we introduce a class of trial wavefunctions suitable for time evolution and ground state determination for the quantum rotor model (4). Since everything generalizes to rotors of arbitrary dimension, we first consider the target space and subsequently specialize to the circle (). Denote the classical configuration of visible rotors by . Following [2], we assign a probability amplitude to each configuration of rotors by integrating a Boltzmann factor over a space consisting of hidden rotors, whose collective coordinates are denoted . In order to construct a suitable Boltzmann factor, we consider the isometric embedding of the target space into ()-dimensional Euclidean space and choose the exponent to be of restricted Boltzmann form,
(5) |
where denotes the surface measure on (the counting measure for ). In the above expression denotes the dot product for and the variational parameters are given by , and for all . If all the bias terms vanish, then the amplitude is invariant under global transformations of the visible rotors, the proof of which follows by a change of integration variables combined with invariance of the integration measure. For one has and , reproducing the proposal of [2]. The weights and biases can also be promoted to complex numbers, resulting in a holomorphic parametrization suitable for time evolution. In this paper, however, we only consider real parametrizations since the Hamiltonian (4) is known to have a non-negative ground state. The integration over the hidden rotors can be performed for any value of . In the case of relevance to (4) we obtain the following logarithmic probability amplitude
(6) |
where is an affine transformation of the embedded rotor configurations , defined in terms of and , and denotes a modified Besssel function. The model can be trained using an efficient Markov Chain Monte Carlo method generalizing [2], which is summarized in Appendix B.
4. Benchmarks: Jastrow Variational Wavefunction
It is instructive to compare the form of the continuous-variable RBM with that of a -invariant Jastrow wavefunction, which is defined by local interactions dictated by the choice of interaction graph,
(7) |
where denote the variational parameters characterizing the trial wavefunction. Since the number of parameters is dictated by the choice of interaction graph, the Jastrow wavefunction, unlike the RBM, lacks the property of systematic improvability. The Jastrow wavefunction has the advantage that the Rayleigh quotient can be computed analytically for certain interaction graphs. We carry out this calculation in the case of a linear (path) graph in Appendix C
5. Benchmarks: High-Dimensional PDE Solvers
As a second validation of the ground states obtained from VMC simulations, we compare them against those obtained using partial differential equation (PDE) based eigensolvers. Here, we give a brief overview of our implementation of these PDE solvers. We restrict ourselves to the Hamiltonian (4) associated with the unit circle with fixed positive weight on vertices . The associated eigenvalue problem is given by
(8) |
where we denote , for .
One straightforward approach to solving (8) is to discretize the domain using a regular grid on , applying the finite difference approximation to the Schrödinger operator and solving the resulting algebraic eigenvalue problem. This will lead to a solution scheme whose error decays only polynomially depending on the order of the finite difference scheme used to approximate the Hamiltonian [13]. Instead, we can take advantage of the periodicity of the domain and switch to the frequency domain using Fourier series expansion for the wavefunction:
(9) |
As is well known, assuming is smooth, its truncated Fourier approximation will yield spectral accuracy. A numerical solver for (8) can be constructed based on transforming the eigenvalue problem into the Fourier domain; its description is provided in Appendix D, along with details on the eigenvalue algorithms.
One major obstacle to implementing either the finite difference or the Fourier spectral schemes described above is the memory cost: assuming degrees of freedom per dimension (number of grid points for finite difference or number of Fourier modes for spectral schemes), the discretized wavefunction will require storing scalar variables for a full representation. As the dimensionality of the problem (number of rotors) increases, it quickly becomes impossible to store the eigenfunction in the memory of a single computing node. This is a direct consequence of the curse of dimensionality. To alleviate this issue and help scale our benchmark solvers, we adopt the distributed memory computing model, where we split the wavefunction across multiple computing nodes and use the message passing interface (MPI) library to implement necessary communication between the nodes. For our particular form of the Hamiltonian, the discretized operator ends up being sparse (both in the case of finite difference and Fourier spectral schemes), and this allows us to design fast matrix-vector products with minimal communication [14].
6. Numerical Results
We implemented the PDE eigensolvers and VMC algorithm in C++ and using state-of-the-art open source libraries; the PDE eigensolvers were built on top of Trilinos [15] to support distributed computing. The code is available publicly at https://github.com/shravanvn/cnqs. The numerical experiments described in this section were run using version 1.0.0 of the code. All simulations were run on the Great Lakes cluster at the University of Michigan. Each compute node of this cluster is equipped with two 18-core 3.0 GHz Intel Xeon Gold 6154 processors and 192 GB RAM.
6.1. Convergence of the Benchmark PDE Solver
We performed self-convergence analysis of our PDE eigensolver on two four-rotor networks depicted in Figure 1 with vertex weights and edge weights . We ran the Fourier PDE eigensolver for maximum frequencies , and Conjugate Gradient (CG) and inverse power iteration tolerances .


We plot the state amplitude for both networks corresponding to the discretization in Figure 3; as we can see, the amplitudes vanish for approximately . This leads to spectral convergence of our Fourier PDE eigensolver; we demonstrate this in Figure 3, where we plot the error in the ground state energy (using solution as reference) as we increase the maximum frequency. The error decays exponentially until it reaches machine precision.


In Figure 5 we plot the number of inverse power iterations necessary for convergence; we note that the iteration count is largely independent of the maximum frequency discretization parameter. Additionally, from Figure 5 we see that with our preconditioner for the Fourier problem, the number of CG iterations per inverse iteration plateaus very quickly.
6.2. VMC Simulations
We ran VMC simulations with the same four-rotor Hamiltonians. In these simulations, 20 hidden nodes were used to construct the restricted Boltzmann machine and 10000 stochastic gradient descent steps were performed with a learning rate of . During each of these steps 24000 Metropolis-Hastings samples were generated, the first 4000 of which were discarded as burn-in and then every 20th sample were cherry-picked to compute the expected value quantities (e.g. energy and gradient). A stochastic reconfiguration parameter of was selected for the simulations. For details on the stochastic reconfiguration algorithm we refer the reader to the appendices of [2].

In Figure 6, we demonstrate the convergence of the energy estimates for the ground state of the corresponding Hamiltonians over the optimization process.
6.3. Comparison of PDE and VMC Solvers
# Rotors () | PDE () | PDE () | VMC | Jastrow | ||||
---|---|---|---|---|---|---|---|---|
avg | std | |||||||
2 | 1.627 | |||||||
3 | 3.254 | |||||||
4 | 4.882 | |||||||
5 | 6.509 | |||||||
6 | 8.136 | |||||||
7 | 9.763 |
In Table 1, we compare the performances of the PDE and VMC solvers on linear chain networks. For each , we ran inverse power iteration with tolerances and for the eigenproblem in Fourier domain with two discretizations: and . This was done to ensure that the PDE eigensolver converged to the correct ground state energy. We then compared these reference eigenvalues against those obtained using the VMC simulations; as we can clearly see from the table, the eigenvalues match to one significant digit. We also note the mild increase in elapsed time in VMC simulations as the number of rotors is increased, compared to the exponential increase in the PDE eigensolver time. This implies that using VMC, we will be able to solve much larger problems than using the PDE eigensolver.
6.4. VMC State Parametrization
# Hidden nodes | Metric | SGD Step | Average Energy | |||
---|---|---|---|---|---|---|
100 | 500 | 5000 | 9999 | |||
20 | std. dev. | |||||
grad. norm | ||||||
40 | std. dev. | |||||
grad. norm | ||||||
60 | std. dev. | |||||
grad. norm | ||||||
80 | std. dev. | |||||
grad. norm | ||||||
100 | std. dev. | |||||
grad. norm |
The main advantage of the VMC approach over the PDE approach is the ability to parametrize the infinite-dimensional quantum rotor state using a finite-dimensional manifold of parameters. The dimensionality of the manifold is dependent on the number of hidden nodes; as the number of hidden nodes is increased, we are able to capture increasingly complicated quantum states. Note however, this also increases the number of parameters to learn (on top of increasing the computational complexity), and this may lead to poor performance if the model is not trained for long enough. In Table 2, we record the eigenvalue standard deviation and gradient norm of a ten-rotor chain quantum rotor system with various number of hidden RBM nodes during various stages of the training process. We note that after 100 stochastic gradient descent steps, the RBM model with 60 hidden nodes has the smallest standard deviation in the eigenvalue; however as the training continues, all of the models end up with comparable converged eigenvalue.
7. Conclusion and Future Directions
We introduced continuous-variable neural quantum states as a variational ansatz for finding the ground-states of quantum Hamiltonian operators on continuous manifolds. We demonstrated the ability of these neural states to converge to the minimal eigenpair of the rotor Hamiltonian by comparing the obtained eigenvalue against those obtained using a baseline PDE based eigensolver. We observed that the scalability of our variational solver increases far slower when compared to the PDE eigensolver as the number of dimensions increase.
The baseline PDE eigensolver introduced in this paper leverages simple techniques from scalable scientific computing algorithms. While the implementation supports distributed computing, allowing us to scale beyond the memory limits of a single node, it does not address the issues related to the curse of dimensionality. Tensor factorization techniques can be used to compress the quantum state of high-dimensional systems to enable memory efficient computing. In [16, 17], the authors use the canonical polyadic (CP) decomposition to find the ground state of a Hamiltonian similar to the one we consider and in [18], the author uses the tensor-train (TT)/matrix-product state (MPS) decomposition on the same problem. [19, 20] use the TT decomposition to find the vibrational spectra and ground states of molecules.
A line of investigation promising improved scalability of the VMC is the exploitation of symmetries of the interaction graph . Although simple convolutional architectures are likely sufficient for square grid graphs with discrete translational symmetry, a detailed investigation of the interplay between the automorphism group of and the isometry group of the target space would be desirable.
If the first quantized approach of this paper can ultimately be made to scale, then it should become possible to analyze the quantum phase transitions corresponding to the Berezinskii–Kosterlitz–Thouless (BKT) transition via the quantum classical mapping. In two (Euclidean) dimensions the BKT topological phase transition is a well-known consequence of the proliferation of vortices. Can systematically improvable variational wavefunctions cast light on the nature of the corresponding quantum phase transition? Similar techniques adapted to variational real-time dynamics could potentially offer a window into many-body quantum chaos via disordered rotor models [21] or the models with negative curvature [22].
In closing, the approximation properties of continuous-variable neural-network quantum states is poorly understood. Considerable effort has been expended in the search for exact representations of many-body quantum states using restricted Boltzmann machines [23, 24, 25, 26, 27, 28] and it would be very interesting if similiar techniques can be adapted to continuous-variable lattice systems.
8. Acknowledgements
J.S. thanks Di Luo and Tobias Osborne for discussion and encouragement. Authors gratefully acknowledge support from NSF under grant DMS-2038030. This research was supported in part through computational resources and services provided by the Advanced Research Computing (ARC) at the University of Michigan.
Appendix A Nonlinear Sigma Model Regularization
A nonlinear sigma model is a theory of maps from a spacetime manifold to the target Riemannian manifold . In this appendix we focus on Minkowski space which is topologically the product . The following discussion can be easily adapted to replace with a manifold of finite spatial volume such as the -torus. The degrees of freedom of the theory consist of target-space-valued functions of spacetime, which we denote by , where indexes the implicit coordinates for the target manifold . Classical trajectories of the sigma model satisfy a nonlinear wave equation obtained by extremizing the following functional,
(10) |
where is an overall normalization. Define the momentum density as the following functional derivative,
(11) |
in terms of which the energy density is given by
(12) |
where is the Lagrangian density. In the simplest Euler discretization, the total energy at time of a map can be obtained from the limiting procedure ,
(13) |
where is the finite difference operator in the direction of the unit basis vector , and is the momentum at vertex . The double summation runs over the edges of the cubic lattice graph. Observe that the potential corresponding to each edge of the graph is the quadratic approximation to the squared Riemannian distance , which provides an accurate approximation when the involved distances are much less than the radius of curvature of the target space . The form of the energy (13) motivates a quantum Hamiltonian in which the quadratic potential is replaced by its nonlinear completion. It follows from the requirement of self-adjointness of the conjugate momentum operator, that the kinetic term of the quantum Hamiltonian is obtained by replacing with minus the Laplace-Beltrami operator defined as follows
(14) |
where . The resulting Hamiltonian is of the form (1) with and with uniform vertex and edge weights given by and , respectively.
Appendix B Training the VMC
The components of the variational derivatives are given by
(15) | ||||
(16) | ||||
(17) |
where and . Each time a Metropolis update of the th visible rotor occurs of the form , the variable is updated for all according to the following rule:
(18) | ||||
(19) |
In our quantum rotor setup, the vectors are parametrized by as . The Metropolis update is then simply updating this angle parameter
(20) |
while shifting by multiples of to keep it in the range . Here is another hyperparameter for the training.
Appendix C Exact Energy for Jastrow Wavefunction
In this appendix we analytically compute the Rayleigh quotient for the quantum rotor Hamiltonian on a linear graph,
(21) |
using a generalized Jastrow trial wavefunction of the form
(22) |
where for each , the function is even, -periodic and satisfies the following normalization condition,
(23) |
which ensures overall normalization
(24) |
If we define constant functions , then
(25) | ||||
(26) | ||||
(27) |
where in the last equality we used the fact that is odd since is even. Similarly,
(28) |
and thus the total energy in the state is
(29) |
In the particular case of Eq. (7) of the main text we have,
(30) |
and the energy is found to be
(31) |
Appendix D Preconditioned Fourier Eigensolver
Here, we provide details on the Fourier-based spectral method for solving (8). Substituting the Fourier expansion (9) in the eigenvalue problem (8), multiplying both sides by and integrating over the domain , we obtain
(32) |
where is the vector of all zeros except at the -th entry, which is . Clearly, (32) is an eigenvalue equation for the Fourier coefficients
(33) |
where the operator defined as
(34) |
Note that given with periodic weak derivatives, the Fourier coefficients satisfy
(35) |
This condition (35) suggests that the Fourier coefficients decay rapidly as infinity norm of the the frequency vector increases in size. Thus, setting the Fourier coefficients to zero outside a hypercube in the frequency space
(36) |
provides a sufficiently accurate approximation to the full wavefunction as long as the cut-off frequency is not too small. This assumption also allows us to represent the wavefunction as a -dimensional tensor for computational purposes. Next, restriction of the Hamiltonian operator to the frequency range is sparse, with at most non-zero entries per row. This structure leads to efficient matrix-vector operations and minimal inter-node communications in distributed computing setups.
Following this restriction of the eigenstate and the Hamiltonian in the Fourier domain, we need to find the minimal eigenvalue-eigenvector pair corresponding to the linear system
(37) |
This is equivalent to finding the maximal eigenvalue-eigenvector pair of the system
(38) |
where is a lower bound on the eigenvalues of , implying is invertible. Such a lower bound is derived in Appendix E.
Power iteration provides a simple method for computing the maximal eigenpair of (38): starting from an arbitrary initial state , we iterate
(39) |
As long as the minimal eigenpair of is non-degenerate, and the initial state is not orthogonal to the minimal eigenstate, this (inverse) power iteration is guaranteed to converge to the correct answer [29].
Finally, note that computing in the iteration requires solving a linear system
(40) |
This can be done by any iterative solvers such as conjugate gradient (CG) or generalized minimal residual (GMRES) methods. They only require a routine for computing for arbitrary vector [29]. We have already pointed out how the sparsity structure of can be used to compute the matrix-vector product efficiently.
In our implementation, we choose CG as the iterative linear solver. To accelerate the convergence, we utilize the diagonal preconditioner
(41) |
and solve the following linear system
(42) |
Algorithm 1 describes the full inverse power iteration process for determining the ground state of the quantum rotor Hamiltonian.
Appendix E Lower Bound on Minimal Eigenvalue
Here, we derive a simple lower bound on the magnitude of the eigenvalues of the rotor Hamiltonian , and equivalently, the operator in the Fourier basis:
Theorem E.1.
All eigenvalues of the Hamiltonian (8) satisfy with
(43) |
Proof.
Let us denote
(44) |
so that we can write
(45) |
It follows that
(46) |
and we can compute
(47) |
where the integration is performed over the domain and the second equality follows from integration by parts (boundary terms vanish due to the periodic boundary conditions). Further
(48) |
It follows that
(49) |
Clearly,
(50) |
serves as a lower bound for the eigenvalues of the Hamiltonian . ∎
References
- [1] William Lauchlin McMillan. Ground state of liquid he 4. Physical Review, 138(2A):A442, 1965.
- [2] Giuseppe Carleo and Matthias Troyer. Solving the quantum many-body problem with artificial neural networks. Science, 355(6325):602–606, 2017.
- [3] Michael M Bronstein, Joan Bruna, Yann LeCun, Arthur Szlam, and Pierre Vandergheynst. Geometric deep learning: going beyond euclidean data. IEEE Signal Processing Magazine, 34(4):18–42, 2017.
- [4] Michael M Bronstein, Joan Bruna, Taco Cohen, and Petar Veličković. Geometric deep learning: Grids, groups, graphs, geodesics, and gauges. arXiv preprint arXiv:2104.13478, 2021.
- [5] Igor Boettcher, Przemyslaw Bienias, Ron Belyansky, Alicia J Kollár, and Alexey V Gorshkov. Quantum simulation of hyperbolic space with circuit quantum electrodynamics: From graphs to geometry. Physical Review A, 102(3):032208, 2020.
- [6] Xingchuan Zhu, Jiaojiao Guo, Nikolas P Breuckmann, Huaiming Guo, and Shiping Feng. Quantum phase transitions of interacting bosons on hyperbolic lattices. arXiv preprint arXiv:2103.15274, 2021.
- [7] Alicia J Kollár, Mattias Fitzpatrick, and Andrew A Houck. Hyperbolic lattices in circuit quantum electrodynamics. Nature, 571(7763):45–50, 2019.
- [8] S Iblisdir, R Orus, and JI Latorre. Matrix product states algorithms and continuous systems. Physical Review B, 75(10):104305, 2007.
- [9] Martin C Gutzwiller. The geometry of quantum chaos. Physica Scripta, 1985(T9):184, 1985.
- [10] Steven R White. Density matrix formulation for quantum renormalization groups. Physical review letters, 69(19):2863, 1992.
- [11] Steven R White. Density-matrix algorithms for quantum renormalization groups. Physical Review B, 48(14):10345, 1993.
- [12] James Glimm and Arthur Jaffe. Quantum physics: a functional integral point of view. Springer Science & Business Media, 2012.
- [13] Randall J LeVeque. Finite difference methods for ordinary and partial differential equations: steady-state and time-dependent problems. SIAM, 2007.
- [14] Yousef Saad. Iterative methods for sparse linear systems. SIAM, 2003.
- [15] Michael A Heroux, Roscoe A Bartlett, Vicki E Howle, Robert J Hoekstra, Jonathan J Hu, Tamara G Kolda, Richard B Lehoucq, Kevin R Long, Roger P Pawlowski, Eric T Phipps, et al. An overview of the trilinos project. ACM Transactions on Mathematical Software (TOMS), 31(3):397–423, 2005.
- [16] Gregory Beylkin and Martin J Mohlenkamp. Numerical operator calculus in higher dimensions. Proceedings of the National Academy of Sciences, 99(16):10246–10251, 2002.
- [17] Gregory Beylkin and Martin J Mohlenkamp. Algorithms for numerical analysis in high dimensions. SIAM Journal on Scientific Computing, 26(6):2133–2159, 2005.
- [18] Ivan V Oseledets. Tensor-train decomposition. SIAM Journal on Scientific Computing, 33(5):2295–2317, 2011.
- [19] Maxim Rakhuba and Ivan Oseledets. Calculating vibrational spectra of molecules using tensor train decomposition. The Journal of chemical physics, 145(12):124101, 2016.
- [20] Alexander Veit and L Ridgway Scott. Using the tensor-train approach to solve the ground-state eigenproblem for hydrogen molecules. SIAM Journal on Scientific Computing, 39(1):B190–B220, 2017.
- [21] Gong Cheng and Brian Swingle. Chaos in a quantum rotor model. arXiv preprint arXiv:1901.10446, 2019.
- [22] Steven Gubser, Zain H Saleem, Samuel S Schoenholz, Bogdan Stoica, and James Stokes. Nonlinear sigma models with compact hyperbolic target spaces. Journal of High Energy Physics, 2016(6):1–15, 2016.
- [23] Yichen Huang and Joel E Moore. Neural network representation of tensor network and chiral states. arXiv preprint arXiv:1701.06246, 2017.
- [24] Dong-Ling Deng, Xiaopeng Li, and S Das Sarma. Machine learning topological states. Physical Review B, 96(19):195145, 2017.
- [25] Giuseppe Carleo, Yusuke Nomura, and Masatoshi Imada. Constructing exact representations of quantum many-body systems with deep neural networks. Nature communications, 9(1):1–11, 2018.
- [26] Jing Chen, Song Cheng, Haidong Xie, Lei Wang, and Tao Xiang. Equivalence of restricted boltzmann machines and tensor network states. Physical Review B, 97(8):085104, 2018.
- [27] Ermal Rrapaj and Alessandro Roggero. Exact representations of many-body interactions with restricted-boltzmann-machine neural networks. Physical Review E, 103(1):013302, 2021.
- [28] Michael Y Pei and Stephen R Clark. Compact neural-network quantum state representations of jastrow and stabilizer states. arXiv preprint arXiv:2103.09146, 2021.
- [29] Lloyd N Trefethen and David Bau III. Numerical linear algebra, volume 50. SIAM, 1997.