Linear operator theory of phase mixing
Abstract
We study solutions of the collisionless Boltzmann equation (CBE) in a functional Koopman representation. This facilitates the use of linear spectral techniques characteristic of the analysis of Schrödinger-type equations. For illustrative purposes, we consider the classical phase mixing of a non-interacting distribution function in a quartic potential. Solutions are determined perturbatively relative to a harmonic oscillator. We impose a form of coarse-graining by choosing a finite dimensional basis to represent the distribution function and time evolution operators, which sets a minimum length scale on phase space structure. We observe a relationship between the dimension of the representation and the multiplicity of the harmonic oscillator eigenvalues. System dynamics are understood in terms of degenerate subspaces of the linear operator spectra. Each subspace is associated with a mode of the harmonic oscillator, the first two being bending and breathing structures. The quartic potential splits the degenerate eigenvalues within each subspace. This facilitates the formation of spiral structure as deformations from the harmonic oscillator modes. We ultimately argue that this construction provides a promising avenue for study of self-interacting systems experiencing phase mixing, which is an outstanding problem in the context of the Gaia DR2 vertical phase space spirals.
keywords:
Galaxy: disc – Galaxy: kinematics and dynamics – Galaxy: structure.1 Introduction
In this paper we investigate the relaxation of collisionless systems through phase mixing. In a statistical description, the macrostate of a system is specified by the phase space distribution function, . This quantifies the probability that a particle exists within an infinitesimal volume of phase space (Sethna, 2006). Liouville’s theorem requires that is conserved along orbits, and it therefore satisfies the collisionless Boltzmann equation (CBE) (Arnold, 1989). For spatial degrees of freedom, this is equivalent to the incompressible flow of a dimensional fluid in a velocity field specified by Hamilton’s equations. An introductory description of this can be found in Binney & Tremaine (2008), but we summarize as follows. Let us assume an anharmonic potential in which orbital frequency of test particles is dependent on their amplitudes of oscillation. In the fluid analogy, this means that vorticity of the velocity field depends on the spatial coordinate. A distribution out of equilibrium with such a potential will deform with time as packets of density with different energies orbit at varying frequencies. For a fixed conservative Hamiltonian this continues indefinitely, with different energy orbits becoming increasingly out of phase with each other. In this process, the scale of structure in the distribution decreases. Eventually when the scale becomes so small that adjacent wraps of the mixed distribution become indistinguishable, the system has equilibrated.
Phase mixing in general leads to complex structures, especially for the case present in galactic dynamics and cosmology (Tremaine, 1999; Perrett et al., 2003; Abel et al., 2012). Even for , which applies to considerations of the vertical motion in the Galactic disc, this process is not trivial. With astrometric data from Gaia Collaboration et al. (2018), a one armed spiral was observed in the vertical phase space of Solar Neighborhood stars (Antoja et al., 2018). In reality this is a system, as it is unlikely that the vertical dynamics are decoupled from motion in the plane (Hunt et al., 2021), but much attention has been given to the structure in the univariate case (Schönrich & Binney, 2018; Darling & Widrow, 2019a; Bennett & Bovy, 2018, 2021). At present, no models have reproduced the exact form of the Gaia spiral.
In Darling & Widrow (2019b), it was suggested that the phase mixing process can be represented with the discrete spectrum of a linear time-evolution operator of finite dimension. There, eigenfunctions were estimated numerically from the full temporal history of a system by applying Dynamic Mode Decomposition (DMD) (Mezić, 2005; Rowley et al., 2009; Kutz et al., 2016) to -body simulations. This served to investigate the claim that self-gravity should not be ignored in phase mixing (Darling & Widrow, 2019a), as well as to explore the representation of this process with persistent oscillatory structures. Stable oscillations such as bending and breathing modes were observed in a self-interacting system in an anharmonic potential. When modifying the relative dominance of self-interaction and anharmonic forcing, these oscillatory structures were more prominent the closer the system was to purely self-interacting. For stronger anharmonic forcing, they were deformed to include spiral structure.
DMD is closely related to Koopman theory (Koopman, 1931), which supposes that a complex, potentially nonlinear system can be represented as a simpler linear one by studying its evolution in terms of observable functions of its state space. This concept was used to interpret the results in Darling & Widrow (2019b), arguing that the binning of -body simulations constituted a mapping to observables. Because of the numerical nature of that work, it was difficult to study the supposed mechanism of phase mixing from discrete modes, or establish a concrete connection to Koopman theory. The DMD approach also draws comparison to the use of multichannel Singular Spectrum Analysis (mSSA) (Weinberg & Petersen, 2021; Johnson et al., 2023). In both cases, principle component analysis (PCA) based techniques are applied to time-series data. A novel aspect of the mSSA work was the temporal analysis of basis function expansion coefficients. This motivates treating the expansion coefficients as Koopman observables, rather than the distribution function.
In the present work, we aim to investigate a mechanism of phase mixing with a discrete linear operator spectrum, emphasizing the role of representation scale, and properties of the spectrum. The essential premise is that the minimum length scale and degeneracy of evolution operator eigenvalues depend on the dimension of the representation. The degenerate eigenvalues define subspaces of the operator spectrum. Splitting of these eigenvalues by anharmonic forcing causes differential rotation among the subspace eigenvectors, which manifests as phase mixing in . System dynamics are determined in terms of basis function expansion coefficients, which we treat as functionals of , adopting a functional Koopman formalism for our calculations. In doing so, we apply techniques from degenerate perturbation theory of the Schrödinger equation. All of our calculations are carried out symbolically.
This paper is organized as follows. In Section 2 we define our coordinates and Hamiltonian, as well as the vector spaces in which we perform our calculations. In section 3 we define the time evolution operator for functionals of , and then restate the problem in this formalism. In Section 4 we define a matrix representation for our operators, and compute the harmonic oscillator spectrum. Section 5 contains the perturbative treatment of the anharmonic potential, and a description of the eigenvalue splitting mechanism for phase mixing. Sections 6 and 7 contain a discussion of connections to other works, and concluding remarks on foreseeable extensions.
2 Preliminaries
Consider a two-dimensional phase space comprising position and momentum . We concern ourselves with the phase space distribution function, . This is defined such that,
(1) |
is the probability of finding a particle in the region of phase space, . The dynamics of are prescribed by the Hamiltonian density, , according to the CBE,
(2) |
Here denotes the Poisson bracket.
2.1 System Hamiltonian
We begin with the Hamiltonian density for a harmonic oscillator with frequency ,
(3) |
This is to be understood as the standard kinematic term plus the leading term in a Taylor series of any even potential. By Jeans’ theorem, the Hamiltonian defines the coordinate dependence of the equilibrium distribution function. Let us choose
(4) |
Here, is a coldness parameter inversely proportional to the velocity dispersion, and is the partition function. It is convenient to work with scaled dimensionless coordinates, so we let
(5) |
In these coordinates, the Hamiltonian density and distribution function become
(6) |
Now we add an anharmonic correction of the form , representative of the next term in the even Taylor series. Our total Hamiltonian density in scaled coordinates is
(7) |
The factor comes from the coordinate transformation in equation 5. We will use in our calculations. This is chosen to assure stable orbits for a range of initial conditions while introducing vorticity that decreases with position, of which the effects can be seen within a few dynamical times.
2.2 Space of bivariate functions
Any configuration of is an element of a function space defined for the independent variables and . Denote by the space of functions of and on a domain . We treat as a Hilbert space equipped with an inner product, which we denote for any (see equation 59). We further assume that there exists a set of linearly independent functions that form a basis in . Any function can be written in terms of the basis functions by projecting onto them (equation 60).
2.3 Conjugate space of functionals
Given the function space , one may consider another, distinct vector space housing its functionals. Such a space is called dual to , and is denoted (also a Hilbert space). For the purposes of this work, we refer to a functional of as any linear mapping
(8) |
which takes the input function, here the distribution function , and maps it to (the complex numbers) by integration with a given . For every , there is a unique given by equation 8 (Riesz representation theorem, see for example Conway (1994)). In the present context, the input function will always be , and it is treated as a variable relative to a functional in the same sense that and are variables with respect to a function . The functions and the functional are further related by the functional derivative, .
The dual space has an inner product and basis that can be understood in terms of the corresponding constructions in . Given a basis of , there exists a basis for , which we denote . The two are related by the biorthonormal condition, . The inner product of any two is denoted , and defined in equation 61. The subscript asterisk signifies that the operation is in .
Finally we highlight some important quantities that exist in . The total energy, which is the expectation value of the Hamiltonian with respect to , is the functional . Additionally, if we project onto a set of basis functions in as in equation 60, the coefficients are the functionals . The † operation denotes complex conjugation on scalars, and the conjugate transpose on vectors and matrices. If we compute the dynamics of according to in , we obtain the time-evolution of the basis function expansion coefficients for in .
2.4 Particular choice of basis functions
We choose as a basis for products of univariate Gaussian-Hermite functions. That is,
(9) |
where denotes the th Hermite polynomial of either or , and is a normalization constant. The corresponding functionals are denoted . In Appendix A we define the polynomials , the coefficients , and state two recurrence relations for the Hermite polynomials we will make use of in Sections 4.1 and 5.1.
3 Dynamics
In what follows, we will use the notion of a flow map. This is an operator on denoted , which takes an initial state to another state at time , . The particular action of is prescribed by the CBE (equation 2).
3.1 Time evolution operator for functionals
Knowing the flow map is tantamount to solving equation 2. Here, we leverage Koopman (1931) to effectively apply the flow map, without dealing explicitly with the CBE. Discussion of the Koopman formalism in the context of astrophysics can be found in Darling & Widrow (2019b) and Darling & Widrow (2021), but for a more rigorous description of the functional realization we will use here, see Nakao & Mezić (2020). Briefly, the idea is that rather than consider the distribution function directly, one can instead study the dynamics of “observables”. In the case of a partial differential equation like the CBE, the observables take the form of functionals, . Let us define a linear time evolution operator that acts on , which we denote . This is called the Koopman operator, and its action on any is,
(10) |
That is, applies the flow map to the argument function , but does not change the form of the functional . If it is easier to determine the action of than , one can compute the functional , which encodes information about at time . Subsequently, if we know how to infer from a functional or set of functionals, we can compute the future state .
Treating as a function of , we may write its infinitesimal change with respect to as
(11) |
To evaluate the limit, we separate out the (by linearity), and apply equation 10 for . That is,
(12) |
For small , . Replacing in equation 12 with this takes care of the limit, and we are left with
(13) |
Here we have defined as the infinitesimal generator of . This is another operator on , and will be the focus of much of this work. To be clear, the action of on any is
(14) |
The final equality is obtained by expanding the Poisson bracket and performing integration by parts with respect to on the first term and on the second. This final form is called a Morrison bracket, which originates in plasma physics (Morrison, 1980). We denote this multiplicative operation between vectors in by , indicating that it is the bracket between and with respect to . We prefer this form, as it makes clear that a functional acted on by is another vector in . For , maps the expectation value of to that of . An operator of this form has appeared in the gravitational context previously in for example Perez (2005), where the Morrison bracket is used to form a so-called functional Vlasov equation.
Equation 13 is a Schrödinger-type equation. It follows that the operator satisfies an ordinary differential equation, and has the general solution
(15) |
The reader familiar with quantum mechanics (QM) may find that this resembles the Heisenberg representation, with analogous to the Hamiltonian operator. Like in QM, we can construct solutions to equation 13 by solving the associated eigenvalue problem of .
3.2 Time evolution of
Denote the th eigenvalue and eigenfunctional of by and respectively. The eigenvalue problem is then, . Let . For convenience, we call this quantity an eigenfunction of , but really it is an eigenfunction of the Liouville operator, . Suppose that forms a basis of , and of .
In this paper, we focus on the case where , which corresponds to a time-independent potential. This holds for our target Hamiltonian in equation 7, but is not true in general for a self-consistent distribution function satisfying both the CBE and Poisson equation. To study the response to both an external potential and self-consistent density perturbations, one must consider . We will discuss this briefly in Sections 6 and 7, but treatment of that problem is not in the scope of this work.
In the time-independent generator case, equation 15 simplifies to , and we have
(16) |
By conjugate symmetry of the inner product, the expansion coefficient of onto at time is then
(17) |
Applying to equation 60 and using this result, is
(18) |
3.3 Restatement of the problem
Now that we have established the formalism in which we will carry out our calculations, it is advantageous to restate the original problem. Recall that the target Hamiltonian is stated in equation 7. In terms of total energy functionals, this corresponds to , where and . It follows that the generators are , , and . Our goal then is to compute the spectrum of , so that we can use equation 18.
4 Matrix Representation
To carry out calculations, we map the quantities in to finite dimensional matrices and vectors. We begin by choosing a finite dimension, . If we take all bivariate Hermite polynomials up to order , the dimension of our finite basis is . We define a reference vector for the basis as
(19) |
This is a vector-valued function of the phase space coordinates. For any vector , the contraction returns a linear combination of the basis functions.
When we choose the dimension for the matrix representation, we introduce coarse-graining by imposing a minimum length scale. The resultant scale is set by the maximum polynomial order , and the model parameters that appear in the coordinate scaling (equation 5). The minimum scale can be quantified by the distance between adjacent roots in the highest order polynomial appearing in the basis functions, . Note that the roots are not necessarily evenly spaced, so the effective resolution must be understood as a function of the phase space coordinates.
4.1 Matrix Elements of
In order to compute the spectrum of , we construct a matrix, which we denote . Its elements are obtained by computing the inner products,
(20) |
Substituting , and into equation 14, and applying the definition of the inner product in equation 61, we obtain
(21) |
Expanding the Poisson bracket, and applying equation 63 we are left with the set of integrals,
(22) | ||||
The integrals are easier to evaluate if we first apply equation 64. Doing so for in the first term and in the second gives
(23) | ||||
From the orthogonality of the Hermite polynomials with respect to , this reduces to
(24) |
We organize the resulting real numbers into an matrix according to the ordering of the indices in the reference vector (equation 19).
4.2 Spectrum of
With a matrix representation of , we can compute a subset of its spectrum. Let denote the eigendecomposition of , where
(25) |
Here, and . We can transform the eigenvectors of into eigenfunctions of by taking their contraction with the reference vector . Denoting the eigenfunctions of by , we have
(26) |
In general, there can be a combination of real and imaginary eigenvalues. Complex eigenvalues come in conjugate pairs, as do their associated eigenvectors. All eigenvalues are determined by the characteristic polynomial,
(27) |
where denotes the identity matrix. We compute the by finding the roots of equation 27. It is possible that some of these roots are repeated, and it is such repeated roots that we refer to as degenerate eigenvalues of . We will refer to the number of times a given eigenvalue is repeated as its multiplicity, denoted .
The eigenvectors associated with are computed by solving,
(28) |
for . We assume that all of the eigenvectors are linearly independent, including those associated with the degenerate eigenvalues. This means that associated with each eigenvalue , is a subspace of , spanned by the corresponding eigenvectors. For the non-degenerate eigenvalues, these subspaces are 1-dimensional, where as for the degenerate eigenvalues they have dimension equal to the degenerate eigenvalue multiplicity, .
We show in Fig. 1 and 2 the sets of linearly independent eigenfunctions associated with and respectively. The former corresponds to the standard bending mode structure, and the latter the breathing mode. In each case, the presented function is the real part of one member of a conjugate pair. When the full complex functions are taken together with their conjugates, multiplication with the relevant complex exponentials produces real valued clockwise rotating bending and breathing patterns.


4.3 Dynamics of from
The dynamics of an arbitrary initial condition according to are determined by the spectrum of . Explicitly, is mapped to a new function at time by the flow of equation 2, for . We have from equation 18,
(29) |
For the harmonic potential, the dynamics of are a rigid rotation of the initial condition about the origin. It is not immediately apparent from equation 29, but we can verify that no deformation to the distribution occurs with the conserved quantities in the spectrum of . There are several eigenvectors of associated with a zero eigenvalue. Each of these corresponds to a conserved quantity. By solving equation 28 for , we find the sequence:
(30) | ||||
The first two correspond to conservation of phase space density and energy, and are themselves constants of motion in the sense that . The other eigenfunctions are not conserved in this sense. It is rather the corresponding functionals which are conserved. This can be understood in terms of the weighted statistical moments of , defined as . All of the zero-eigenvalue functions in equation 30 contain even polynomials of and . This means that the corresponding functionals are linear combinations of bivariate moments corresponding only to even powers in both variables. Such moments are variance, covariance, kurtosis, cokurtosis and so on. These moments quantify the character of symmetric properties of the distribution, ignoring asymmetries like mean and skew. This means that when the integrals over phase space are carried out, the result is the same regardless of where along the harmonic oscillator orbit the distribution is. Any deformation in will change the values of the integrals in this sequence. Their presence as zero-eigenvalue functionals in the spectrum of restricts the dynamics to rigid rotation.
5 Anharmonic Potential
With the harmonic oscillator spectrum calculated, we can now proceed to estimating the spectrum of . We begin by computing the matrix elements of .
5.1 Matrix elements of
The equivalent expression to equation 21 for the matrix elements is obtained by making the substitutions , and in equation 14. We have,
(31) |
We evaluate the Poisson bracket, applying equation 63 to the Hermite polynomial derivatives, and equation 64 to the derivatives. We are left with
(32) | ||||
To evaluate the these integrals for all indices, it is easiest to re-write strictly in terms of Hermite polynomials. Recursive application of equation 64 yields (for arbitrary ) an expansion of the form
(33) |
where the coefficients are polynomials in of order . Table 1 contains a list of the relevant coefficients for both a quartic and sextic potential. To compute the contribution from , let us define the integral
(34) |
With this definition, the matrix elements of (equation 32) are expressed as
(35) | ||||
where we have used .
(36) |
This is evaluated easily by the orthogonality of the Hermite polynomials, leaving
(37) |
We then substitute the evaluated into equation 35, and compute the remaining momentum space integral. We are left with
(38) | ||||
Again, we organize the resulting values into an matrix according to the ordering of indices in .
0 | ||
---|---|---|
1 | ||
2 | ||
3 | ||
4 | ||
5 |
5.2 Spectrum of
To apply the ordinary differential equation solution for , and compute from equation 18, we need the spectrum of . The matrix
(39) |
does not admit a diagonalization directly, so we adopt a perturbative treatment. We follow a standard procedure for analysis of the Schrödinger equation in QM (for an introductory description, see for example Griffiths & Schroeter (2018)). Suppose that the eigenvalues and eigenvectors of can be written as a power series in the parameter . That is,
(40) | ||||
We assume that these quantities satisfy the eigenvalue problem,
(41) |
Substituting the expressions in equation 40 into 41 and equating terms of equal power in , we obtain a sequence of equations relating the power series coefficients and the operators and . The zero order equation is the eigenvalue problem for , stated in Section 4.2. The first order equation is
(42) |
We assume that the basis formed by the eigenvectors of spans the eigenspace of . That is,
(43) |
Note that indicates a dot product in and the result is scalar. Substituting as in equation 40 into equation 43, we express the first order correction as
(44) |
In order to compute this correction we need an expression for the coefficients, . We proceed by contracting equation 42 with . This yields
(45) |
For , the left hand side is zero, and we have the first order correction to the eigenvalues,
(46) |
For , and we obtain the coefficients we need for equation 44. That is,
(47) |
Summing over these coefficients as in equation 44 yields the correction to the eigenvectors of . Explicitly, the first order correction to the non-degenerate eigenvectors of is
(48) |
We can use the corrections computed here as-is for the non-degenerate subset of the spectrum, where if . The situation is more complicated if . As discussed in Section 4.2, the spectrum of in general contains degenerate subspaces, where there are multiple distinct eigenvectors associated with the same eigenvalue (see for example Fig. 1 and 2). This poses an issue for the correction in equation 48, where we must sum over all of the linearly independent eigenvectors. If and are such that both eigenvectors lie within a degenerate subspace, the denominator here is zero. To proceed, we treat the contribution separately for each degenerate subspace.
5.3 Treatment of degenerate eigenvalues
We begin by adding a second index label to the eigenvalues of . That is, let denote the th repetition of the th unique eigenvalue of . The same indexing scheme is applied to the eigenvectors. As stated in Section 4.2, all eigenvectors corresponding to repeated eigenvalues are themselves linearly independent. So although we have , for the eigenvectors, .
For any fixed , the are linearly independent and span a subspace of the eigenspace corresponding to a single eigenvalue . Any linear combination of these vectors is itself an eigenvector of , with eigenvalue . We can make use of this property in determining the contribution. We begin by looking at the spectrum of when it is projected onto each degenerate subspace of separately. This process is as follows.
For each eigenvalue with multiplicity , we define the subspace basis as
(49) |
We then compute the matrix elements of projected onto the degenerate subspace. Denoting the th subspace projection , we have
(50) |
Note that the projected matrices are rather than . With the projected generator matrix, we compute its eigenvalues, and eigenvectors, . These satisfy
(51) |
5.4 Simultaneous eigenvectors of and
The comprise weights for linear combinations of the degenerate subspace basis vectors . Contracting with the subspace basis in equation 49, we can get back vectors in . Explicitly, we write
(52) |
Each of the are eigenvectors of both the projected , with eigenvalue , and of with eigenvalue . Because of this, we can replace the original eigenvectors of making up the th degenerate subspace with the new .
We need these new eigenvectors to resolve an ambiguous definition of the unperturbed . If we did not do this, the transition between and could constitute a discontinuous change in the eigenvectors. We must avoid this to assure that we have a smooth variation of the eigenvectors with respect to change in the perturbation parameter . Otherwise, the original power series assumption (equation 40) would be invalid. From this point on, we replace all of the degenerate subspace eigenvectors, with the new vectors we just found, . For notational simplicity, we drop the tilde. Any time from this point on that we refer to the degenerate subspace vectors, it should be assumed that they are the ones defined by equation 52 that diagonalize in their respective subspaces. The eigenfunctions are obtained from the eigenvectors in the same way as for the harmonic oscillator, using an analog to equation 26.
5.5 Degenerate Subspace Corrections
To compute the corrections for the degenerate eigenvalues and eigenvectors, we follow a similar procedure to Section 4.2, carrying out the necessary algebra in Appendix B. To summarize, we again assume a power series in for both the eigenvalues and eigenvectors of (equation 68), and then compute the coefficients of the terms in these series.
5.5.1 Eigenvalues of
The expression for the first order corrected eigenvalues is the same as it was in the non-degenerate case, just with the double index on the degenerate subsets. Explicitly,
(53) |
where is given by equation 72, derived in Appendix B.1. This is obtained by diagonalizing in the th degenerate subspace as in Section 5.3. Here, the eigenvalues of are distinct within any given degenerate subspace . That is, for . This means that the degenerate eigenvalues of have been split by . If this is not true, a higher order expansion in is required.
We show in Fig. 3 the corrected eigenvalues corresponding to half of the degenerate subspaces (the values shown are conjugate partners to those in the omitted half). The curves serve to illustrate the splitting of the degenerate subset of the spectrum. The degenerate subspace dimension decreases as increases. The curves corresponding to a single degenerate eigenvalue of converge at , but possess distinct values otherwise. The spread of the for fixed , increases with . This corresponds to the strength of the anharmonic contribution to the potential. It is this distance between eigenvalues of within a given degenerate subspace that drives differential rotation in the distribution function. The vertical line at indicates the chosen value of used in our example calculations.

5.5.2 Eigenvectors of
The eigenvectors of are expressed as the power series,
(54) |
The first order correction is derived in Appendix B.2, and can ultimately be written as
(55) |
As in Section 4.2, we can transform the eigenvectors of into eigenfunctions of by taking their contraction with the reference vector, . We have
(56) |

We show in Fig. 4 an example set of eigenfunctions associated with . As discussed in Section 4.2, this is the eigenvalue that controls the rotation of the simple bending mode in Fig. 1. The eigenfunctions contain more sign changes along a radial path from the origin than those of the harmonic oscillator. This facilitates a segmenting of phase space density with distance from the origin (proportional to the action in this case). Each of these structures possesses a different orbital frequency, corresponding to the four diverging solid red lines starting at in Fig. 3. The relative spacing of these lines determines the relative rates of rotation, and therefore the relative phase of the structures in Fig. 4, as time progresses. The combination of these two factors causes differential rotation in the distribution function, producing the spiral structure characteristic of phase mixing. The same general idea applies to the other degenerate subspace eigenfunctions, although there are more complicated structures present in the larger eigenfunctions.
To illustrate the importance of degenerate subspaces to the mixing process, we consider the sum over eigenvectors within a given subspace. Let us define the th subspace sum as
(57) |
We show this quantity in Fig. 5, with the particular value corresponding to the index of the subspace. The initial condition is , with . At , the split eigenvalues have no effect, and the net contribution from the subspace is the bending mode. As time progresses, the eigenfunctions in Fig. 4 rotate increasingly out of phase with each other, deforming the bending mode structure into interlocked positive and negative one-armed spirals.

We have observed so far that particular types of structures correspond to degenerate subspaces of the harmonic oscillator spectrum. To reiterate, these subspaces are sets of linearly independent eigenfunctions of , all associated with the same eigenvalue. When the anharmonic potential contribution is added, the degenerate eigenvalues split, but the correspondence between particular structures (bending, breathing, etc.) and the degenerate subspaces remains. The subspace dimension determines the complexity of representable structure, and taken in combination with the properties of the contained eigenfunctions (spacing of their roots), sets the minimum length scale of that structure.
5.6 Dynamics of from
With an estimated spectrum of the full generator, , we can compute the distribution function at time . Separating the sum in equation 18 into the distinct contributions from our perturbative analysis we have
(58) |
On the right hand side from left to right these are, the non-degenerate subspace of the spectrum of , and the total contribution from the degenerate subspaces.
In Fig. 6, we show three snapshots of for three values of . The initial condition is again (the same as for Fig. 5). Time progresses down the rows, and basis size increases across the columns. Looking first at (rightmost column), the distribution function winds into a one-armed spiral as increases. Let us compare this to the and columns, focusing on (second row). Relative to , the other two cases appear to have more sophisticated structure, with the complexity increasing as decreases. This apparent structure is however an artifact of the truncated basis representation in a manner similar to the Gibbs phenomenon. That is, a deficiency of terms in the series over the basis functions leads to a poor representation of the true .
The basis size set by the polynomial order determines how well a temporally isolated snapshot of the distribution function can be represented. In a finite dimensional representation, the exact form of is only known at the initial condition, and is projected onto the basis functions. For , every future state is determined from the initial projection according to the eigenvalues of . The manner in which the distribution deforms from some to another time depends on the structure of the eigenfunctions and their associated eigenvalues. In the present context, this hinges on the relative spacing of the eigenvalues within each degenerate subspace. As discussed in Section 5.5.2 and demonstrated in Fig. 5, the mechanism for spiral formation is the out of phase rotation of the degenerate subspace eigenvectors. The temporal characteristics of this process are determined by the split eigenvalues in Fig. 3. This occurs within each degenerate subspace, and the total contribution from the mechanism is encapsulated in the rightmost term in equation 58.
The number of eigenvalues, and their multiplicities in the spectrum increase with . For , we have for . When the degenerate eigenvalues have been split by the anharmonic potential, the multiplicity prescribes the number of distinct rates of rotation for the subspace eigenvectors. This can also be understood as an increase in frequency domain coverage in the sense of a Fourier transform. The length scale set by determines which structures can be well represented by the basis regardless of . The fanning out in frequency shown in Fig. 3 coupled with the form of the determines the resolvability of spatiotemporal structure.

6 Discussion
The resolvability of structure in depends on its representation, or means of observation. When one chooses a number of particles in an -body simulation, a grid resolution in an equation solver, or a finite dimensional basis, a scale is imposed. In this work, we aimed to highlight the relationship between imposed length scale from dimension of representation, and multiplicity of eigenvalues in the spectrum of . Further, that in such a case splitting of the degenerate eigenvalues by drives mixing in .
For the spectrum of , the number of unique degenerate eigenvalues, and their average multiplicity increases with , the size of the basis. In this construction, spiral formation is achieved through a linear combination of structures that span the entire plane (Fig. 4). This is in contrast to the picture described in Section 1, in which essentially every infinitesimal annulus of phase space volume at a different radius from the origin orbits with a different frequency. In Banik et al. (2022), phase mixing is achieved through a linear response term of the form , where are angle-action coordinates, and is the oscillatory frequency of the orbits. Since the frequencies depends on the actions, orbits at different actions will possess varying frequencies, leading to deformation of .
One could suppose that for an infinite dimensional basis, decomposes into an infinite set of delta functions of and , each nonzero at a different point in the plane. In this case, all of the different packets of phase space density may have different orbital frequencies, and we obtain the original picture of the process, essentially moving to a discrete particle representation. Given the premise of increasing dimension corresponding to increasing degeneracy of the eigenvalues, we suppose that in the limit case of an infinite dimensional representation, the discretely split degenerate eigenvalues may become the continuous bands described in Mathur (1990) and Weinberg (1991).
The analysis here omits a self-interaction potential. This means that does not depend on , and the integral in equation 15 is trivial. Were this not the case, the operator solution would take the form of a Dyson series, obtained by iterative solution of the implicit equation (see for example, Sakurai & Napolitano (2017)). In doing this, it would be sensible to separate the self-interaction contribution to the Hamiltonian, and parameterize its strength with for example, . Computing the necessary Dyson series to some order in is technically possible. The main constraint is that the matrix element integrals cannot necessarily be evaluated analytically for a non-polynomial potential in our chosen basis. Using the exact one-dimensional Green’s function of the Poisson equation poses a challenge in this regard. One option is to adopt a Hamiltonian Mean Field (HMF) approach as in Inagaki & Konishi (1993), replacing the absolute value function with a polynomial that preserves some desired properties. In this case, the self-interaction can be expressed in terms of the moments of as in Section 4.3, which work nicely with the functional formalism used here. The simplest case would be a quadratic interaction, . In that case, the self-interaction is a quadratic potential well that tracks the expectation value of position with respect to , which is the moment .
In this work we considered systems with only a single spatial degree of freedom. This is applicable to the vertical motions of Solar Neighborhood stars in the limit that vertical and in-plane motions decouple. The general approach described in this work may be extended to higher dimensional phase spaces in order to incorporate radial and/or azimuthal motions. The general framework requires that CBE solutions reside in a Hilbert space , and that there exists a dual space of the corresponding functionals . We can in principle assume this for any number of spatial degrees of freedom. It is when computing a representation of the Koopman generator that we must handle the problem a little differently. First it should be noted that increasing the dimension of the phase space will require additional degrees of freedom on the basis functions. In the case of a truncated finite dimensional basis, this means that in order to achieve a comparable resolution to the calculations here, a larger number of basis functions is necessary.
Assuming a Milky Way-like geometry, we briefly outline here possible choices for the basis functions, and discuss how one might proceed. In Section 4, we used Gaussian-Hermite functions for the vertical position and momentum, and can similarly use such functions for the radial and azimuthal momenta. For the radial position, we need orthogonal functions on the interval that approach zero at infinity. These criteria are met by Laguerre polynomials weighted by a radially decaying exponential function. For the azimuthal position we could use any set of periodic functions orthogonal on .
If the increased matrix size poses a computational barrier, one may consider studying the system dynamics with expectation values of particular functions rather than the distribution function itself. One approach is to look at the mean vertical position and/or momentum as a function the radial and azimuthal coordinates as in for example Chequers & Widrow (2017). These partial moments have been represented in terms of Koopman generator eigenfunctions computed with DMD in Widrow et al. (2020). The functional formalism used in this work can be extended to various moments by transforming the Morrison bracket definition of (equation 14) into a form prescribed by Kupershmidt & Manin (1978). This yields an equation of motion for moments rather than the distribution function itself. A description of this and its correspondence to Jeans’ equations can be found in Chapter 3 of Darling (2024).
7 Conclusions
We have described a procedure for determining the CBE dynamics of , by mapping the problem into a set of functionals which satisfy a Schrödinger-type equation. To calculate explicit matrix representations of operators, we adopted a finite dimensional set of basis functions that impose a minimum length scale on . For an illustrative example of phase mixing, we treated a quartic potential perturbatively with respect to a harmonic oscillator solution. We observed that the multiplicity of harmonic oscillator eigenvalues increases with the dimension of the finite basis representation. Subsequently, we showed that the anharmonic potential splits the degenerate eigenvalues, leading to the amplitude dependent frequencies characteristic of phase mixing.
In Section 5.5.2, we computed the corrected eigenfunctions of . These structures were segmented into subspaces, with the particular groupings set by the spectrum of according to the degeneracy of its eigenvalues. Different classes of structure (for example bending or breathing) were associated with each subspace. The scenario depicted in Fig. 5 can be understood as the kinematic response to a bending mode perturbation, projected onto the subspace. This demonstrated how the standard bending mode structure deforms to contain spiral structure, similar to the observations in Darling & Widrow (2019b). This is of interest in the context of the Gaia spiral, especially with respect to the hypothesis that the observed structure stems from a bending mode excitation (Darling & Widrow, 2019a). The essential takeaway from this is that the spectrum derived from a solvable model like the harmonic oscillator can be used to model both processes characteristic of self-gravity, and phase mixing. For the approach taken in Darling & Widrow (2019b), one could compute basis functions for the evolution of which were suited to representing the particular dynamics of the simulation they were derived from. The form of the basis functions derived from DMD was heavily dependent on the parameters of the simulation. They were especially disparate between simulations dominated by self-gravity and anharmonic forcing. In the formalism used here, we suggest that one can construct a single basis suited to both cases. The trade-off needed for this is that instead of a single eigenfunction representative of a principal component of the dynamics, one has a subspace of dimension (usually ). The eigenfunctions spanning a subspace would at first glance appear suited only to represent the highly symmetric oscillatory structures typical of the harmonic oscillator spectrum, but eigenvalue splitting in proportion to anharmonic forcing magnitude allows for out of phase rotation. Such rotation facilitates spiral structure formation in the subspace contributions and consequently in the distribution function.
In modeling the Gaia spiral, it is possible that there does not exist a purely kinematic description which yields the observed structure. Entertaining that possibility, one must ask how self-interaction changes the standard mechanism for the formation of phase space spirals. It is feasible to consider self-interaction and comparatively static anharmonic forcing as competing effects, with their relative magnitudes impacting the spatiotemporal structure of the distribution function. This relative dominance has previously been quantified by a “live fraction" parameter, denoted (Darling & Widrow, 2019b; Bennett & Bovy, 2021; Darling & Widrow, 2021). It was observed in Darling & Widrow (2019b) that when evolution is primarily driven by self-interaction (), the dominant contributions to the Koopman spectrum resemble the bending and breathing modes that appear in the harmonic oscillator spectrum (Fig. 1 and 2). As the relative strength of the self-interaction was increased, the structures associated with the harmonic oscillator spectrum were deformed to contain spiral structure, resembling quite closely the sum over the subspace in Fig. 5 for . We suppose that the spiral-bending and spiral-breathing contributions to the Koopman spectrum observed numerically in Darling & Widrow (2019b) can be modeled by the degenerate subspaces of the perturbed harmonic oscillator spectrum associated with their respective structures. For the quartic Hamiltonian, the eigenvalue splitting mechanism represents the effect of the anharmonic forcing. The self-interaction can be treated with time-dependent perturbation theory. Specifically this involves the more general Dyson series for discussed in Section 6. This treatment can be applied to each subspace individually, so one can observe the effect of self-interaction on the spiral-bending and spiral-breathing modes in a similar way to Darling & Widrow (2019b), but with an analytic model.
A crucial assumption here is that one can determine a basis sufficient for the evolution of the distribution function. If that is the case, the problem is reduced to determining the set of coefficients that produce the correct linear combination. It is the relative magnitudes of the coefficients that determine the form of at any particular time, but it is the relative phase of the time-dependent coefficients that prescribe the evolution and mixing. Acknowledging the possibility that there is no fully time-independent mapping from the hypothetically “correct" initial condition to the observed structure, the effect of time-dependent forces from the self-interaction of the distribution is essential for successful modeling. We suggest that a subspace-wise analysis of the interplay between anharmonic forcing and self-interaction is a preferable way forward. Preliminary calculations using the time-dependent Dyson series indicate that addition of a self-interaction term to the quartic Hamiltonian leads to a time-varying rate of mixing. This is corroborated by fully self-consistent (Darling & Widrow, 2019a) and live fraction based (Darling & Widrow, 2019b) numerical experiments. In all cases, it was observed that a self-interacting distribution undergoes mostly kinematic phase mixing at early times to form a spiral, but the winding slows down after the initial formation. A detailed exploration of this is beyond the scope of the present work, and is left to a future article.
It is also of interest to further investigate the scale-dependent mixing process studied here in the context of other work around coarse-grained evolution of the CBE, as in Chavanis et al. (1996) or Chavanis, P. H. & Bouchet, F. (2005). In the latter, one of the definitions of a coarse-grained is a windowed functional. That is, is taken in convolution with some kernel that sets the representation scale. Preliminary work suggests that the equation of motion in that case is equivalent to equation 13 up to a diffusion current term.
Acknowledgements
This work was supported by a Discovery Grant with the Natural Sciences and Engineering Research Council of Canada. We are thankful to Mike Petersen for insightful comments. We also thank Scott Tremaine, Francesco Cellarosi, Aaron Vincent, and Stephen Hughes for helpful discussions.
Data Availability
Calculations were carried out in MATLAB. Custom software used for this paper is available upon reasonable request. Colormaps used in contour plots are thanks to Thyng et al. (2016).
References
- Abel et al. (2012) Abel T., Hahn O., Kaehler R., 2012, MNRAS, 427, 61
- Antoja et al. (2018) Antoja T., et al., 2018, Nature, 561, 360
- Arnold (1989) Arnold V., 1989, Mathematical methods of classical mechanics. Vol. 60, Springer
- Banik et al. (2022) Banik U., Weinberg M. D., van den Bosch F. C., 2022, ApJ, 935, 135
- Bennett & Bovy (2018) Bennett M., Bovy J., 2018, MNRAS, 482, 1417
- Bennett & Bovy (2021) Bennett M., Bovy J., 2021, MNRAS, 503, 376
- Binney & Tremaine (2008) Binney J., Tremaine S., 2008, Galactic Dynamics: Second Edition. Princeton University
- Chavanis, P. H. & Bouchet, F. (2005) Chavanis, P. H. Bouchet, F. 2005, A&A, 430, 771
- Chavanis et al. (1996) Chavanis P. H., Sommeria J., Robert R., 1996, The Astrophysical Journal, 471, 385
- Chequers & Widrow (2017) Chequers M. H., Widrow L. M., 2017, MNRAS, 472, 2751
- Conway (1994) Conway J., 1994, A Course in Functional Analysis. Graduate Texts in Mathematics, Springer New York
- Darling (2024) Darling K., 2024, Linear Operator Theory of Phase Mixing in Collisionless Systems
- Darling & Widrow (2019a) Darling K., Widrow L. M., 2019a, MNRAS, 484, 1050
- Darling & Widrow (2019b) Darling K., Widrow L. M., 2019b, MNRAS, 490, 114
- Darling & Widrow (2021) Darling K., Widrow L. M., 2021, MNRAS, 506, 3098
- Gaia Collaboration et al. (2018) Gaia Collaboration et al., 2018, A&A, 616, A11
- Griffiths & Schroeter (2018) Griffiths D. J., Schroeter D. F., 2018, Introduction to Quantum Mechanics, 3 edn. Cambridge University Press
- Hunt et al. (2021) Hunt J. A. S., Stelea I. A., Johnston K. V., Gandhi S. S., Laporte C. F. P., Bédorf J., 2021, MNRAS, 508, 1459
- Inagaki & Konishi (1993) Inagaki S., Konishi T., 1993, PASJ, 45, 733
- Johnson et al. (2023) Johnson A. C., Petersen M. S., Johnston K. V., Weinberg M. D., 2023, MNRAS, 521, 1757
- Koopman (1931) Koopman B. O., 1931, Proceedings of the National Academy of Sciences, 17, 315
- Kupershmidt & Manin (1978) Kupershmidt B. A., Manin Y. I., 1978, Functional analysis and its applications, 11, 188
- Kutz et al. (2016) Kutz J. N., Brunton S. L., Brunton B. W., Proctor J. L., 2016, Dynamic Mode Decomposition: Data-Driven Modeling of Complex Systems. SIAM
- Mathur (1990) Mathur S. D., 1990, MNRAS, 243, 529
- Mezić (2005) Mezić I., 2005, Nonlinear Dynamics, 41, 309
- Morrison (1980) Morrison P. J., 1980, Physics letters. A, 80, 383
- Nakao & Mezić (2020) Nakao H., Mezić I., 2020, Chaos, 30, 113131
- Perez (2005) Perez J., 2005, Transport Theory and Statistical Physics, 34, 391
- Perrett et al. (2003) Perrett K. M., Stiff D. A., Hanes D. A., Bridges T. J., 2003, ApJ, 589, 790
- Rowley et al. (2009) Rowley C. W., Mezic I., Bagheri S., Schlatter P., Henningson D. S., 2009, Journal of Fluid Mechanics, 641, 115–127
- Sakurai & Napolitano (2017) Sakurai J. J., Napolitano J., 2017, Modern Quantum Mechanics, 2 edn. Cambridge University Press, doi:10.1017/9781108499996
- Schönrich & Binney (2018) Schönrich R., Binney J., 2018, MNRAS, 481, 1501
- Sethna (2006) Sethna J. P., 2006, Statistical Mechanics: Entropy, Order Parameters and Complexity, first edition edn. Oxford University Press, Great Clarendon Street, Oxford OX2 6DP
- Thyng et al. (2016) Thyng K. M., Greene C. A., Hetland R. D., Zimmerle H. M., DiMarco S. F., 2016, Oceanography, 29, 9
- Tremaine (1999) Tremaine S., 1999, MNRAS, 307, 877
- Weinberg (1991) Weinberg M. D., 1991, ApJ, 373, 391
- Weinberg & Petersen (2021) Weinberg M. D., Petersen M. S., 2021, MNRAS, 501, 5408
- Widrow et al. (2020) Widrow L. M., Darling K., Li H., 2020, in Valluri M., Sellwood J. A., eds, Vol. 353, Galactic Dynamics in the Era of Large Surveys. pp 65–70, doi:10.1017/S1743921319009049
Appendix A Preliminary Definitions
A.1 Inner products
The inner product in is
(59) |
where † indicates complex conjugation. This product possesses conjugate symmetry, . For a set of linearly independent basis functions as defined in Section 2, the inner product facilitates projection of an arbitrary function . Such an expansion may be expressed as
(60) |
The inner product in is
(61) |
An expansion analogous to equation 60 applies to arbitrary functionals .
A.2 Hermite polynomials
The Hermite polynomials introduced in Section 2.4 are defined by the Rodrigues formula,
(62) |
Hermite polynomials satisfy the following two recurrence relations:
(63) |
(64) |
The normalization constants for the Gaussian-Hermite basis functions defined in Section 2.4 are
(65) |
Our chosen bivariate functions inherit orthogonality from the Hermite polynomials. Explicitly,
(66) |
This applies to the entire real number line in both variables, so our domain is set to the entire phase space.
Appendix B Perturbative treatment of degenerate eigenvalues
We aim to find eigenvectors and eigenvalues that satisfy the perturbed eigenvalue problem
(67) |
This is achieved by assuming a power series in for both quantities,
(68) | ||||
This requires that both the eigenvectors and eigenvalues vary smoothly with respect to change in the perturbation parameter . Since the eigenvalues are scalar, this is the case by default. We assure a smooth variation in the eigenvectors by the procedure described in Section 5.4.
Substituting the assumed power series into equation 67, and equating terms of equal power in , one finds a set of equations. Those corresponding to the first three powers of are,
(69) | ||||
B.1 Degenerate case: eigenvalues
To determine the first order correction to the eigenvalues, we begin by contracting the case (second line) in equation 69 with . That is
(70) |
Since is a left-eigenvector of , , and the left hand side is zero. We are left with
(71) |
Since , equation 71 then reduces to
(72) |
That is, the corrections to the degenerate eigenvalues are the diagonal entries of the perturbation, , when it is projected onto the degenerate subspace corresponding to . In other words, if we project onto a degenerate subspace, the eigenvalues of the projected matrix are the first order corrections . The first order correction in equation 72 goes into the power series definition for the eigenvalues of in Section 5.5.1 (equation 53).
B.2 Degenerate case: eigenvectors
In general, the eigenvectors of may contain components from the entire space, including vectors from both in and not in the degenerate subspace. That is, we want to know how to project onto the sets of and . It follows that we can do this if we have general expressions for both of the contractions, and .
We will treat both cases separately, first computing the contribution from the subset of the complete basis that excludes the degenerate subspace. We begin by contracting the first order case in equation 69 with . We have
(73) |
Similar to when we computed the eigenvalue correction, we note that . Additionally, we know that and are orthogonal for all , since one of them is within the degenerate subspace and the other is not. From these two arguments, we can simplify equation 73 to
(74) |
For the degenerate subset contribution, we do not get any new information by attempting a contraction of with the first order case in equation 69. Let us instead try the equation. We take
(75) | |||
Note that we use the same index on the first index for the introduced vector, as if we allowed for a different index, we would be considering the case that one degenerate subspace is projected onto another. All of the distinct subspaces are orthogonal, so this can only result in zero.
We can proceed by again noting that . Since , the left hand side is zero. Further, , so we are are left with
(76) |
Of course, we do not know , as that is what we are trying to determine here. We do know however that it can have contributions from and . Let us suppose that it takes the form
(77) |
We have already found what we need to express the first sum explicitly in equation 74. With this, we may write
(78) |
Continuing with equation 76, we have
(79) | ||||
The first term on the right hand side is zero since are orthogonal for . The sum in the term on the right hand side collapses because . Making these simplifications leaves us with,
(80) | ||||
Since the degenerate subspace basis vectors are chosen such that they diagonalize , the only nonzero matrix elements are those along the diagonal. We may therefore simplify this term as
(81) | ||||
Collapsing the sum according to the we are left with
(82) | |||
This equation has two cases. First, if , we obtain the explicit expression for that we have been looking for. That is
(83) |
Taking the case yields the correction to the eigenvalues. That is,
(84) |
These second order corrections are shown in the dotted lines in Fig. 3.
Now that we have explicit expressions for both and we can express the first order corrections to the eigenvectors of as we had outlined in equation 77.
(85) | ||||