Gentlest ascent dynamics on manifolds defined by adaptively sampled point-clouds
Abstract
Finding saddle points of dynamical systems is an important problem in practical applications such as the study of rare events of molecular systems. Gentlest ascent dynamics (GAD) 1 is one of a number of algorithms in existence that attempt to find saddle points in dynamical systems. It works by deriving a new dynamical system in which saddle points of the original system become stable equilibria. GAD has been recently generalized to the study of dynamical systems on manifolds (differential algebraic equations) described by equality constraints 2 and given in an extrinsic formulation. In this paper, we present an extension of GAD to manifolds defined by point-clouds, formulated using the intrinsic viewpoint. These point-clouds are adaptively sampled during an iterative process that drives the system from the initial conformation (typically in the neighborhood of a stable equilibrium) to a saddle point. Our method requires the reactant (initial conformation), does not require the explicit constraint equations to be specified, and is purely data-driven.
JHU]Department of Chemical and Biomolecular Engineering, Whiting School of Engineering, Johns Hopkins University, 3400 North Charles Street, Baltimore, 21218, MD, USA JHU]Department of Chemical and Biomolecular Engineering, Whiting School of Engineering, Johns Hopkins University, 3400 North Charles Street, Baltimore, 21218, MD, USA KULeuven]Department of Computer Science, KU Leuven, Celestijnenlaan 200A, 3001 Leuven JHU]Department of Chemical and Biomolecular Engineering, Whiting School of Engineering, Johns Hopkins University, 3400 North Charles Street, Baltimore, 21218, MD, USA \alsoaffiliationDepartments of Applied Mathematics and Statistics, Johns Hopkins University, 3400 North Charles Street, Baltimore, 21218, MD, USA
![[Uncaptioned image]](https://cdn.awesomepapers.org/papers/87e69b29-6d92-42fc-9673-8a5479feb9db/fig-isd-simple-sphere.png)
1 Introduction
The problem of finding saddle points of dynamical systems has one of its most notable applications in the search for transition states of chemical systems described at the atomistic level, since saddle points coincide with transition states 3 at the zero temperature limit. While finding local stable equilibria (sinks) is a relatively straightforward matter, finding saddle points is a more complicated endeavor for which a number of algorithms have been presented in the literature 4.
Saddle point search methods can be classified according to whether they require one single input state (usually the reactant, located at the minimum of a free energy well) or two states (reactant and product). The Gentlest Ascent Dynamics (GAD) 1 belongs to the class of methods requiring a single reactant state as input and its applicability has been demonstrated with atomistic chemical systems 5, 6. GAD can be regarded as a variant of the dimer method that is formulated as a continuous dynamical system whose integral curves with initial condition at the reactant state can lead to saddle points. Variants of GAD such as high-index saddle dynamics (HiSD) have been the subject of recent research efforts and applications 7, 8, 9, 2, 10, 11, 12, 13.
While many search schemes attempt to find an optimal path between reactant and product (or between reactant state and transition state), it is interesting that there exist continuous curves joining the desired states in a variety of ways: following isoclines 14, 15, gradient extremals 16, 17, and the GAD studied in this paper. In most cases the study of these curves has been carried out in the Euclidean setting with some exceptions on the manifold of internal coordinates 18, 6 and on manifolds defined by the zeros of smooth maps 2 (in these cases, the algorithms are formulated extrinsically on the ambient space). Our contribution is formulated intrinsically and is valid on arbitrary manifolds, not necessarily explicitly defined by an atlas or by the zeros of maps. Importantly, algorithms like the one presented here or our previous work 19 do not rely on a priori knowledge of good collective coordinates, but rather use manifold learning to find them on the fly. In our method, there is a feedback loop of data collection that drives progress towards a saddle point.
In this paper, we study an application of the GAD to manifolds defined by point-clouds. The manifold does not need to be characterized in advance either by the zeros of a smooth function or by an atlas, and it is only assumed that the user is capable of sampling the vicinity of arbitrary points on the manifold (e.g., umbrella sampling based on reduced local coordinates). The algorithm uses dimensionality reduction (namely, diffusion map coordinates 20) to define a dynamical system intrinsically on the reduced coordinates that can lead to a saddle point. Since the saddle point in general is not expected to lie in the vicinity of the reactant, our algorithm works by iteratively sampling the manifold on the fly, resolving the path on the local chart, and repeatedly switching charts until convergence. Our approach shares algorithmic elements with our previous work 19 which, however, instead of GAD dynamics on manifolds, was following isoclines on manifolds.
2 Gentlest Ascent Dynamics and Idealized Saddle Dynamics
Gentlest Ascent Dynamics (GAD) 1 is an algorithm for finding saddle points of dynamical systems. We propose an extension of GAD to manifolds defined by point-clouds that finds saddle points combining nonlinear dimensionality reduction and adaptive sampling.
Let be a smooth potential energy function and consider the associated gradient vector field in given by , where with denoting the gradient (or, equivalently, the Jacobian matrix). We restrict ourselves here, for the sake of simplicity, to the case of gradient systems and index-1 saddle points. The GAD algorithm consists of integrating the equations of motion of the related dynamical system on an extended phase space given by
(1) |
where , is the Householder reflection 21 of across the hyperplane and is the Rayleigh quotient of the Hessian matrix corresponding to the vector .
Remark 1.
The right hand side of the ordinary differential equation for is the term
If is an eigenvector of with eigenvalue , then . Now note that the constrained optimization problem consisting of finding the extrema of along the level sets is precisely given by the Lagrange equation
which says that the extrema of the magnitude of the gradient, , along are attained wherever the gradient field happens to be an eigenvector of the Hessian .
The intuition behind GAD is that the force can be written in a basis determined by the eigenvectors of the Hessian , which gives us the stable and unstable directions in the vicinity of an index-1 saddle point. The second differential equation in (1) acts as a (continuous) eigensolver yielding the unstable direction. Then, we can flip the sign of the component of the force corresponding to the unstable direction to make the resulting vector point towards the saddle point (see Figure 1).

2.1 Idealized saddle-point dynamics
We consider a variant of the GAD algorithm named Idealized Saddle Dynamics (ISD) 22, given by
where is an eigenvector of corresponding to the smallest eigenvalue .
We now leave the Euclidean setting behind and study the Riemannian setting (see the appendix for a summary of the topic and the notation). Let be a -dimensional smooth manifold with Riemannian metric and let be a potential energy function. We consider the gradient field and we write the following equation for the ISD vector field on ,
(2) |
where is the vector field on defined by choosing the eigenvector (normalized so that ) corresponding to the smallest eigenvalue of the Hessian matrix. However, in this case the Hessian must be defined in terms of the covariant derivative on with respect to the Levi-Civita connection 23, 24. To be precise, given two vector fields , we have
which is a tensor of type , where denotes the covariant derivative induced by the Levi-Civita connection (the notation represents the covariant derivative of in the direction of ). We apply the sharp () isomorphism (raising or lowering indices) to turn into a -tensor. Therefore,
The eigenvector corresponding to the lowest eigenvalue of at each point induces a vector field whose integral curves are curves that may join an initial point with a saddle point.
The ISD formulation of GAD is particularly amenable to be coupled with dimensionality reduction approaches because the resulting eigenproblem is often of much lower dimensionality than the ambient space in which the original dynamical system is defined.
Example 1 (Exact solution of a model system).
Let us compute a concrete case of ISD, first with known exact formulas and later on with approximations using diffusion maps and Gaussian processes. Consider the sphere
and the stereographic projection from the North pole onto the tangent plane at the South pole. The system of coordinates is given by
Let . The corresponding parameterization, , is given by
The pullback of the Euclidean metric by gives us the metric
The non-redundant Christoffel symbols that characterize the Levi-Civita connection are
The energy , constrained on is transformed to on ,
The force on is the negative of the gradient
The potential energy and the gradient field are shown in Figure 2(a).


Example 1 required the knowledge of a particular system of coordinates (namely, the stereographic projection from the North pole to the tangent space at the South pole) mapping three-dimensional points on the sphere to two-dimensional coordinates; here this allows us to work with a single chart. In some settings the system of coordinates of the underlying manifold is either unknown or difficult to obtain. It is possible under those circumstances to replicate the steps in Example 1 in the absence of a system of coordinates given as a closed-form expression by resorting to manifold learning / dimensionality reduction techniques. In our case, as we shall discuss next, we use diffusion maps on points sampled from a local neighborhood of the manifold to extract a suitable system of coordinates. Fitting a Gaussian process to the diffusion map coordinates of the point-cloud yields a local system of coordinates that can be evaluated at arbitrary points (not necessarily those in the sampled point-cloud). Once we have a system of coordinates, we can again proceed to estimate the Riemannian metric as well as the Levi-Civita connection, and compute the flow of the ISD vector field given in (2) to find saddle points.
Revisiting Example 1 and approaching it with the aforementioned procedure allows us to obtain trajectories that lead to saddle points, as shown in Figures 3 to 5.
The choice of diffusion map coordinates for dimensionality reduction and Gaussian processes for non-linear regression is inessential and both methods could be replaced by, for instance, neural networks (e.g., a variational autoencoder and a graph neural network, respectively). An important aspect of our approach is the fact that it operates on local neighborhoods of points, as opposed to using data from longtime simulations, and only then applying dimensionality reduction techniques to extract global collective variables. From a geometric viewpoint, this is motivated by the fact that a sufficiently small neighborhood of a manifold can always be transformed onto a subset of its tangent space by a smooth invertible map. From a physical viewpoint, different sets of collective variables govern different stages of a reaction (e.g., the distance between a ligand and a receptor is important when both molecules are far away whereas their relative orientation might be more important to ligand-binding, when both molecules become close).
One interesting aspect of diffusion maps is its relation to the infinitesimal generator of a diffusion process 25, which in turn has connections to the committor function, which is an optimal reaction coordinate 26, 27, 28. We conclude this discussion on collective variables by pointing out that a priori knowledge of good reaction coordinates (some recent examples can be found in 29, 30) can often be put in one-to-one correspondence with diffusion map coordinates 31, 32.
2.2 Mean force
In computational statistical mechanics we are often interested in dynamics on Riemannian manifolds endowed with a probability distribution. For instance, for simulations at constant temperature, we use the Boltzmann distribution with probability density proportional to , where is the inverse temperature. The relevant vector field at in the local chart is the mean force 33, 34, 35, given by
where denotes the ensemble average with respect to the Boltzmann distribution, conditioned on . There are a number of numerical methods that estimate the mean force as a step in their calculations. Adaptive biasing force 36 and the string method 37 are examples of those.
3 Algorithm
Consider the manifold and the gradient dynamical system . We begin by drawing a total of samples from the manifold in the neighborhood of an initial point . This can be done in a variety of ways depending on the application. If is the inertial manifold of a dynamical system , then a reasonable way to approach this problem is to generate distinct perturbations of and propagate them according to the flow of during a (short) time horizon . Doing so, we obtain a data set approximately on the manifold. Alternatively, one may numerically solve a stochastic differential equation such as the Brownian dynamics equation,
(3) |
where the drift is the vector field , is a constant, and is a standard -dimensional Brownian motion. Solving (3) (possibly with an added RMSD-based restraint around the initial conformation) up to a certain time and extracting an uncorrelated subset of the states at different time steps yields a data set .
Next, we apply a dimensionality reduction algorithm on the data set to obtain a set of reduced coordinates . In our case, we use diffusion maps 20, 25 to obtain a set of vectors with but other methods, such as local tangent space alignment 38, may be used as well. It is important to note that our dimensionality reduction method is applied to a local neighborhood of an initial point and, therefore, it is expected to yield a reasonable approximation to a chart on that neighborhood.
We fit a Gaussian process regressor to the pairs of points to obtain a smooth map that will act as a system of coordinates (in particular, ). Proceeding in an analogous fashion, we compute the inverse mapping .
Note that one possible way of estimating the dimensionality is by computing the average of the approximate rank of the Jacobian matrix of at (a subset of) the data points and retaining the components that yield a local chart.
Remark 2.
We can reduce the computational expense of the Gaussian process regression by reusing the kernel matrix with entries (for some ) calculated during the computation of the diffusion map coordinates as the covariance matrix for the Gaussian process (assuming that it is formulated using the squared exponential kernel).
Remark 3.
It is not always possible to obtain a suitable Gaussian process regressor for . An alternative is to add an Ornstein-Uhlenbeck process to the stochastic differential equation (3) in order to obtain
where is a hyperparameter and is a prescribed point not necessarily in . Computing the ensemble average of the solution to the above equation yields a point such that or, equivalently, . This works because the new term added to the drift nudges the system towards a point in ambient space such that its image by is the prescribed point .
We consider the values of the vector field at the points in the data set and map them via the system of coordinates in order to obtain the vector field in the new coordinates, where is the Jacobian matrix of (note that the Jacobian-vector product can be computed either as a closed-form formula or efficiently using automatic differentiation —e.g., using jvp in JAX 39). We fit another Gaussian process to the pushforward vector field in order to be able to evaluate it at arbitrary points in the new coordinates and we abuse notation in what follows by also referring to as .
At this stage, we can readily compute the Riemannian metric as
(4) |
where for . The metric induces an inner product, denoted by the bracket , between tangent vectors such that if and are the expressions in local coordinates of two tangent vectors at a point , then .
Using (4), we obtain the coefficients of the Levi-Civita connection 23,
for , where denotes the entries of the inverse of the matrix with components . This, in turn, allows us to take the covariant derivative and the Hessian, defined by
for arbitrary tangent vectors . Observe that the Hessian is defined on the local chart, not on the ambient space. The eigenvector , normalized with respect to , with smallest eigenvalue of the Hessian then yields a vector field
such that the index-1 saddle points of become stable equilibria 1 of . Consequently, an integral curve of in the vicinity of a saddle point leads to the saddle point.
In general, in order to carry out the computation until convergence, we must frequently switch charts. This is due to the fact that at each step, we sample a point-cloud in a small neighborhood of a given point and the Gaussian process regressor yields an approximation to on the chart that cannot extrapolate far away from the sampled points. Therefore, when we reach the confines of (which can be determined by the density of points), we ought to map the latest point of our integral curve of from the chart back to the ambient space using , as discussed earlier. After that, we start the whole procedure again: sample a new data set , compute , trace an integral curve of , etc.
Remark 4.
The factor that impacts the algorithm the most is the deviation of the sampled points in the data set from the manifold . In other words, the farther away from our data points are, the more noisy is the estimate of , and, consequently, the harder it is to find the saddle points.
3.1 Numerical examples
The computations that follow were carried out using the JAX 39 and Diffrax 43 libraries, and the code to reproduce our results is available at https://github.com/jmbr/gentlest_ascent_dynamics_on_manifolds.
In our two examples below, we automatically set the bandwidth parameter (usually denoted by ) in diffusion maps to be equal to the squared median of the pairwise distances between points. The regularization parameters in the Gaussian processes and are chosen so that their scores are sufficiently close to 1 on a test set.
3.1.1 Sphere
The preceding algorithm applied to the vector field on the sphere introduced in Example 1 produces the iterations shown in Figures 3, 4, and 5. These figures depict the integral curves (highlighted) in the local neighborhoods obtained by sampling and integrating . The full integral curve joining the initial point to a saddle point at the equator of the sphere in Figure 6.
We sampled points per iteration and integrated the ISD vector field using an explicit Euler integrator for a total steps with a time-step length of . The algorithm converges to a saddle point in ten iterations.




3.1.2 Regular surface
Our second example is the Müller-Brown (MB) potential 44 mapped onto a regular surface. Namely, the manifold is the graph of the function (see Figure 7),
where and the coefficients are given in Table 1.
0 | 1 | 0.9490 | 0.8838 |
---|---|---|---|
0 | 2 | 0.4575 | 0.6564 |
1 | 0 | 0.4152 | 0.7449 |
1 | 2 | 0.2911 | 0.3619 |
2 | 0 | 0.4121 | 0.5469 |
3 | 2 | 0.2817 | 0.4719 |
After eight iterations, sampling points and integrating for steps per iteration with a time step length of , our algorithm successfully constructs a path joining the initial point located at a minimum (sink) of the MB potential to the nearest saddle point. The relative errors between the points in the constructed path and the known coordinates of the saddle point are shown in Figure 8.


4 Conclusions
We have presented a formulation of GAD on manifolds defined by point-clouds that are meant to be sampled on-demand. Our formulation is intrinsic and does not require the specification of the manifolds either by a given atlas or by the zeros of a smooth map. The required charts are discovered through a data-driven, iterative process that only requires knowledge of an initial conformation of the reactant. We illustrated our approach with two simple examples and we expect the results to transfer to the high-dimensional dynamical systems of interest in computational statistical mechanics.
This work was supported by the US Air Force Office of Scientific Research (AFOSR) and the US Department of Energy DOE with IIT: SA22-0052-S001 and AFOSR-MURI: FA9550-21-1-0317.
Appendix A Appendix
In this section we review basic notions of Riemannian geometry relevant to this paper. Our aim is to emphasize intuition over formalism as much as possible. We refer the reader to general treatises on the topic such as 23, 45 for a deeper presentation and we especially recommend the relevant material in 46, 47 for the working physical chemist.
A.1 Tensor spaces
Let and be two vector spaces over the real numbers and denote by the vector space generated by all finite linear combinations of elements of the Cartesian product . The tensor product is a vector subspace of with elements of the form , where , , such that:
-
1.
, where .
-
2.
, where .
-
3.
, where .
-
4.
.
The above construction can be recursively extended to tensor spaces of arbitrary numbers of factors (i.e., given vector spaces , with , we construct as and so on and so forth).
Given a vector space , its dual space, denoted by , is the vector space formed by all linear maps . A tensor space of type is a tensor product
(5) |
Elements of are called tensors of type .
Similarly to how all vector spaces of fixed dimension over the real numbers are linearly isomorphic to each other, all tensor spaces of type over the reals are linearly isomorphic to each other and, therefore, we can assume that the factors are arranged as in (5).
Example 2.
The set of linear maps from to is a vector space. If is an orthonormal basis of and is an orthonormal basis of , then is the basis of the dual space determined by
Consequently, the linear map can be written as
where . Additionally, if and are, respectively, the canonical bases of and , then coincides with the outer product and coincides with the real matrix
The wedge product is defined as
where is the group of permutations of and is equal to if the permutation is odd and if it is even. This is equivalent to the projection of the tensor onto the subspace of anti-symmetric tensors of .
Remark 5.
The signature coincides with the Levi-Civita symbol. The latter should not be confused with the Levi-Civita connection, to be introduced later.
A.2 Smooth manifolds and smooth curves
In what follows, we consider a smooth map to be a map that has sufficiently many derivatives and whose derivatives are continuous functions.
A chart is a tuple such that is an open set in a topological space and is a diffeomorphism (i.e., a smooth mapping with a smooth inverse ). A smooth manifold is a topological space together with a family of charts (i.e., an atlas),
such that for every there exists a chart with . Moreover, for any other chart such that , the function is smooth (see Figure 9).
Example 3.
The sets
and
where , are both smooth manifolds. The former is the graph of the function and the latter is implicitly defined by the zeros of the equation of a 2-sphere (i.e., the points in at which the function vanishes).
Differential Geometry is often explained following the intrinsic point of view in which nothing outside of the manifold is considered. Nevertheless, for didactic purposes it is sometimes useful to think of a manifold as being embedded in a sufficiently high-dimensional Euclidean space. This can always be done due to the celebrated embedding theorem of Whitney 48, which guarantees that we can embed any smooth -dimensional manifold in .
A smooth curve on a manifold is a smooth mapping where is an interval of .
A.3 Tangent spaces and tangent bundles
Again, let be a smooth -dimensional manifold and let be some arbitrary point on it. A tangent vector at is a map that assigns a real number to each smooth, real-valued function of such that
-
1.
,
-
2.
,
-
3.
,
For any smooth and .
The set of tangent vectors at is a vector space called the tangent space, denoted by . Indeed, is a vector space because if are tangent vectors at and , then the axioms
-
1.
,
-
2.
,
are satisfied.
In a local chart, the vectors
constitute a basis of .
Example 4.
Let . In this case, the directional derivative of at in the direction is
In our notation, we would instead say that is a tangent vector of at .
We can view each tangent vector at as the velocity vector of a smooth curve such that . In that case,
The geometric interpretation of tangent vectors and tangent spaces is shown in Figure 10.
The collection
of the tangent spaces at each point of is called the tangent bundle and it is a smooth manifold in its own right.
Example 5 (The tangent bundle in classical mechanics).
If we denote the position of a mechanical system by , where is the configurational manifold of the system, and we consider an arbitrary trajectory starting at the point with initial velocity , then we see that the tangent bundle is the collection of all the positions and velocities of the mechanical system under consideration. Moreover, the Lagrangian is a real-valued function of the tangent bundle,
where is the mass matrix of the system.
The momenta are obtained from the velocities by duality (via the Legendre transform 49). Because of that, the phase space of a mechanical system is a manifold that is closely related to the tangent bundle of the configurational manifold . Indeed, the phase space is the co-tangent bundle of the configurational manifold .
The tangent bundle of a manifold is the prototype of a more general construction called a vector bundle that we will introduce next (albeit not fully rigorously). Informally, a vector bundle is a collection of vector spaces that are in correspondence to each point of . Slightly more formally, we say that a vector bundle over is a tuple where:
-
1.
The total space and the base space are smooth manifolds.
-
2.
The projection is a smooth map whose preimages (known as fibers) for are vector spaces (there are additional conditions that must be satisfied but they are beyond the scope of this appendix).
A trivial example of a vector bundle is shown in Figure 11.
In the case of the tangent bundle, the fiber is the tangent space of at . More generally, we can take the fibers to be tensor spaces.
A.4 Vector fields and sections
A vector field is a correspondence of a tangent vector to each point . The set of all vector fields over a manifold is denoted by . An example of a vector field is shown in Figure 12.

In local coordinates,
From now on, we adopt the convention
To any smooth vector field and any , we can associate the flow map, , that satisfies
The smooth curve is also called an integral curve of .
In local coordinates, if , then solves the initial value problem
with the initial condition . We often denote by .
Let be a vector bundle. A section is a smooth map
such that the composition of followed by the projection is the identity in (or, equivalently, ). Indeed, we have the following diagram:
The set of sections of a vector bundle , denoted by , is a vector space with the operations of addition,
and scalar product,
In local coordinates around a point with a basis , the expression of a section can be written as
A.5 Pullbacks and pushforwards
Consider a smooth map between manifolds and with respective local coordinates and . Let and let be a smooth function. The pullback of the expression is defined by
Consider a vector field , the pushforward of by is a vector field in defined by
for all smooth functions and all points .
Remark 6.
The pullback generalizes the transformation of the integrand in the change of variables formula known from integral calculus. The pushforward of a vector field is an expression of the chain rule from differential calculus.
A.6 Riemannian metrics
Let be an -dimensional smooth manifold. A Riemannian metric is a smooth map such that each point is mapped to a symmetric and positive definite matrix . If are vectors from an orthonormal frame in the local chart , the components of the Riemannian metric can be written as
for all and . The inner product between and is defined by linearity. The norm of an arbitrary vector is .
Remark 7.
If we take , then its dual is with orthonormal basis given by . When , as in this case, it is customary to write and . From the previous discussion, we see that the Riemannian metric is a section in because is linear and we can write it as
Example 6 (Euclidean metric in ).
The Euclidean metric in is
(6) |
Consider the polar coordinates in given by
where and . We have
Therefore, the Euclidean metric (6) is pulled back as:
While in Euclidean coordinates the components of the metric tensor are , in polar coordinates they are , , .
Let and let be a smooth curve such that . The arc-length of a curve is
Given an arbitrary smooth manifold , we can regard it as the configurational space of a (constrained) mechanical system. Let be a curve on . If we set the potential energy to be constant (, for simplicity) and we think of the Riemannian metric as a (position-dependent) mass matrix, then it turns out that the action integral is
(7) |
where the Lagrangian contains only a kinetic energy term,
A curve that locally minimizes the action is called a geodesic curve of . By Jensen’s inequality 50, curves that minimize (7) also minimize the arc-length.
A.7 Connections
Consider a vector bundle on the manifold and let denote the sections of . We are interested in studying how a section changes along a given direction . In essence, we want to define what it means to take the limit
(8) |
where is a linear isomorphism between fibers and is a smooth curve on such that and . If , then we can take to be the identity map but in general we need a non-trivial to account for the fact that the fibers and are different vector spaces. To formalize this notion, we define the concept of a connection.
Let , smooth, and . A connection acts on sections as follows:
-
1.
.
-
2.
.
-
3.
.
-
4.
.
-
5.
.
In short, the connection is linear on its first argument (a vector field) and acts as a derivation (i.e., as Leibniz’s rule) on its second argument (a section). The covariant derivative of the section in the direction of is defined by .
We can characterize the covariant derivative by writing it as:
where is known as the flat connection. If we introduce local coordinates around a point in the open set , and consider the coordinate vector fields and a basis of , then where . Therefore,
where are the components of the vector potential. In full generality, we have
Consequently, each component of the covariant derivative of the section along the vector is of the form
Remark 8.
The definition of a covariant derivative is equivalent to (8) (see 51 Chapter 6). In other words, the flat connection and the vector potential fully characterize all the possible ways in which we could define a derivative as a limit of the form (8). We will see next that there is a special type of connection that is determined by the Riemannian metric.
A.8 The Levi-Civita connection
A Riemannian metric induces a unique connection called the Levi-Civita connection. The components of the vector potential of the Levi-Civita connection are called the Christoffel symbols and we shall derive them next using classical mechanics.
Consider a free particle with unit mass moving on the manifold with position and velocity . Its Lagrangian is
The Euler-Lagrange equations for this mechanical system are
This implies that
for . Using the identity we find
for . Finally, multiplying both sides above by , where are the components of the inverse metric tensor (i.e., where is the Kronecker delta), we arrive at the geodesic equation
(9) |
where
is the expression of the Christoffel symbol for .
Remark 9.
Remark 10.
In the Euclidean case, and , leading to the usual expression of the kinetic energy and to the free particle moving in a straight line. When is an arbitrary Riemannian manifold and the free particle moves along a geodesic, we can interpret the Christoffel symbols as the terms giving rise to the corrections to an otherwise straight trajectory so that it remains on the manifold .
A.9 Raising and lowering indices
The inner product between two vectors determined by the Riemannian metric gives rise to a linear isomorphism (called flat) between and by mapping
The inverse linear map (called sharp) is
such that . The isomorphisms introduced above are called the musical isomorphisms.
Denoting by the inverse of the metric tensor , we write the components of and as
Example 7.
The gradient of a function is defined as , where . We have seen that in the case of the plane with polar coordinates, . Consequently, .
The musical isomorphisms can be used with individual factors of a tensor product to map between the primal and dual vector spaces.
References
- E and Zhou 2011 E, W.; Zhou, X. The Gentlest Ascent Dynamics. Nonlinearity 2011, 24, 1831–1842
- Yin et al. 2022 Yin, J.; Huang, Z.; Zhang, L. Constrained High-Index Saddle Dynamics for the Solution Landscape with Equality Constraints. Journal of Scientific Computing 2022, 91, 62
- Hänggi et al. 1990 Hänggi, P.; Talkner, P.; Borkovec, M. Reaction-Rate Theory: Fifty Years after Kramers. Reviews of Modern Physics 1990, 62, 251–341
- Olsen et al. 2004 Olsen, R. A.; Kroes, G. J.; Henkelman, G.; Arnaldsson, A.; Jónsson, H. Comparison of Methods for Finding Saddle Points without Knowledge of the Final States. The Journal of Chemical Physics 2004, 121, 9776–9792, Publisher: American Institute of Physics
- Samanta et al. 2014 Samanta, A.; Chen, M.; Yu, T.-Q.; Tuckerman, M.; E, W. Sampling saddle points on a free energy surface. The Journal of Chemical Physics 2014, 140, 164109–164109
- Quapp and Bofill 2014 Quapp, W.; Bofill, J. M. Locating saddle points of any index on potential energy surfaces by the generalized gentlest ascent dynamics. Theoretical Chemistry Accounts 2014, 133, 1510
- Gu and Zhou 2017 Gu, S.; Zhou, X. Multiscale gentlest ascent dynamics for saddle point in effective dynamics of slow-fast system. Communications in Mathematical Sciences 2017, 15, 2279–2302, Publisher: International Press of Boston
- Gu and Zhou 2018 Gu, S.; Zhou, X. Simplified gentlest ascent dynamics for saddle points in non-gradient systems. Chaos: An Interdisciplinary Journal of Nonlinear Science 2018, 28, 123106, Publisher: American Institute of Physics
- Yin et al. 2019 Yin, J.; Zhang, L.; Zhang, P. High-Index Optimization-Based Shrinking Dimer Method for Finding High-Index Saddle Points. SIAM Journal on Scientific Computing 2019, 41, A3576–A3595, Publisher: Society for Industrial and Applied Mathematics
- Yin et al. 2022 Yin, J.; Zhang, L.; Zhang, P. Solution Landscape of the Onsager Model Identifies Non-Axisymmetric Critical Points. Physica D: Nonlinear Phenomena 2022, 430, 133081
- Gu et al. 2022 Gu, S.; Wang, H.; Zhou, X. Active Learning for Saddle Point Calculation. Journal of Scientific Computing 2022, 93, 78
- Luo et al. 2022 Luo, Y.; Zheng, X.; Cheng, X.; Zhang, L. Convergence Analysis of Discrete High-Index Saddle Dynamics. 2022; arXiv:2204.00171 [cs, math]
- Zhang et al. 2022 Zhang, L.; Zhang, P.; Zheng, X. Error Estimates for Euler Discretization of High-Index Saddle Dynamics. SIAM Journal on Numerical Analysis 2022, 60, 2925–2944, Publisher: Society for Industrial and Applied Mathematics
- Quapp et al. 1998 Quapp, W.; Hirsch, M.; Imig, O.; Heidrich, D. Searching for saddle points of potential energy surfaces by following a reduced gradient. Journal of Computational Chemistry 1998, 19, 1087–1100
- Quapp et al. 1998 Quapp, W.; Hirsch, M.; Heidrich, D. Bifurcation of reaction pathways: the set of valley ridge inflection points of a simple three-dimensional potential energy surface. Theoretical Chemistry Accounts 1998, 100, 285–299
- Basilevsky 1982 Basilevsky, M. V. The topography of potential energy surfaces. Chemical Physics 1982, 67, 337–346
- Hoffman et al. 1986 Hoffman, D. K.; Nord, R. S.; Ruedenberg, K. Gradient extremals. Theoretica chimica acta 1986, 69, 265–279
- Quapp 2004 Quapp, W. Newton Trajectories in the Curvilinear Metric of Internal Coordinates. Journal of Mathematical Chemistry 2004, 36, 365–379
- Bello-Rivas et al. 2022 Bello-Rivas, J. M.; Georgiou, A.; Guckenheimer, J.; Kevrekidis, I. G. Staying the Course: iteratively locating equilibria of dynamical systems on Riemannian manifolds defined by point-clouds. Journal of Mathematical Chemistry 2022,
- Coifman and Lafon 2006 Coifman, R. R.; Lafon, S. Diffusion Maps. Applied and Computational Harmonic Analysis 2006, 21, 5–30
- Golub and Van Loan 2013 Golub, G. H.; Van Loan, C. F. Matrix computations, 4th ed.; Johns Hopkins Studies in the Mathematical Sciences; Johns Hopkins University Press, 2013
- Levitt and Ortner 2017 Levitt, A.; Ortner, C. Convergence and Cycling in Walker-type Saddle Search Algorithms. SIAM Journal on Numerical Analysis 2017, 55, 2204–2227
- do Carmo 1992 do Carmo, M. P. Riemannian geometry; Mathematics: Theory & Applications; Birkhäuser Boston, Inc., Boston, MA, 1992
- Absil et al. 2008 Absil, P.-A.; Mahony, R.; Sepulchre, R. Optimization Algorithms on Matrix Manifolds; Princeton University Press, 2008
- Nadler et al. 2006 Nadler, B.; Lafon, S.; Coifman, R. R.; Kevrekidis, I. G. Diffusion Maps, Spectral Clustering and Reaction Coordinates of Dynamical Systems. Applied and Computational Harmonic Analysis 2006, 21, 113–127
- Peters 2016 Peters, B. Reaction Coordinates and Mechanistic Hypothesis Tests. Annual Review of Physical Chemistry 2016, 67, 669–690
- Berezhkovskii and Szabo 2019 Berezhkovskii, A. M.; Szabo, A. Committors, First-Passage Times, Fluxes, Markov states, Milestones, and All That. The Journal of Chemical Physics 2019, MMMK, 054106
- Roux 2022 Roux, B. Transition Rate Theory, Spectral Analysis, and Reactive Paths. The Journal of Chemical Physics 2022, 156, 134111, Publisher: American Institute of Physics
- Wu et al. 2022 Wu, S.; Li, H.; Ma, A. Exact Reaction Coordinates for Flap Opening in HIV-1 Protease. Proceedings of the National Academy of Sciences 2022, 119, e2214906119
- Manuchehrfar et al. 2022 Manuchehrfar, F.; Li, H.; Ma, A.; Liang, J. Reactive Vortexes in a Naturally Activated Process: Non-Diffusive Rotational Fluxes at Transition State Uncovered by Persistent Homology. The Journal of Physical Chemistry B 2022, 126, 9297–9308
- Chiavazzo et al. 2017 Chiavazzo, E.; Covino, R.; Coifman, R. R.; Gear, C. W.; Georgiou, A. S.; Hummer, G.; Kevrekidis, I. G. Intrinsic Map Dynamics Exploration for Uncharted Effective Free-Energy Landscapes. Proceedings of the National Academy of Sciences 2017, 201621481–201621481
- Fujisaki et al. 2018 Fujisaki, H.; Moritsugu, K.; Mitsutake, A.; Suetani, H. Conformational Change of a Biomolecule Studied by the Weighted Ensemble Method: Use of the Diffusion Map Method to Extract Reaction Coordinates. The Journal of Chemical Physics 2018, 149, 134112
- Fixman 1974 Fixman, M. Classical Statistical Mechanics of Constraints: A Theorem and Application to Polymers. Proceedings of the National Academy of Sciences 1974, 71, 3050–3053
- Carter et al. 1989 Carter, E. A.; Ciccotti, G.; Hynes, J. T.; Kapral, R. Constrained Reaction Coordinate Dynamics for the Simulation of Rare Events. Chemical Physics Letters 1989, 156, 472–477
- den Otter 2000 den Otter, W. K. Thermodynamic Integration of the Free Energy along a Reaction Coordinate in Cartesian Coordinates. The Journal of Chemical Physics 2000, 112, 7283–7292, Publisher: American Institute of Physics
- Darve et al. 2008 Darve, E.; Rodríguez-Gómez, D.; Pohorille, A. Adaptive Biasing Force Method for Scalar and Vector Free Energy Calculations. The Journal of Chemical Physics 2008, 128, 144120–144120
- Maragliano et al. 2006 Maragliano, L.; Fischer, A.; Vanden-Eijnden, E.; Ciccotti, G. String Method in Collective Variables: Minimum Free Energy Paths and Isocommittor Surfaces. The Journal of Chemical Physics 2006, 125, 024106
- Zhang and Zha 2004 Zhang, Z.; Zha, H. Principal Manifolds and Nonlinear Dimensionality Reduction via Tangent Space Alignment. SIAM Journal on Scientific Computing 2004, 26, 313–338, Publisher: Society for Industrial and Applied Mathematics
- Bradbury et al. 2018 Bradbury, J.; Frostig, R.; Hawkins, P.; Johnson, M. J.; Leary, C.; Maclaurin, D.; Necula, G.; Paszke, A.; VanderPlas, J.; Wanderman-Milne, S.; Zhang, Q. JAX: composable transformations of Python+NumPy programs. http://github.com/google/jax (Last accessed 2023-04-20), 2018
- Bofill and Quapp 2016 Bofill, J. M.; Quapp, W. The variational nature of the gentlest ascent dynamics and the relation of a variational minimum of a curve and the minimum energy path. Theoretical Chemistry Accounts 2016, 135, 1–14
- Randers 1941 Randers, G. On an Asymmetrical Metric in the Four-Space of General Relativity. Physical Review 1941, 59, 195–199
- Chern and Shen 2005 Chern, S.-S.; Shen, Z. Riemann-Finsler Geometry; Nankai Tracts in Mathematics; WORLD SCIENTIFIC, 2005; Vol. 6
- Kidger 2021 Kidger, P. On Neural Differential Equations. Ph.D. thesis, University of Oxford, 2021
- Müller and Brown 1979 Müller, K.; Brown, L. D. Location of Saddle Points and Minimum Energy Paths by a Constrained Simplex Optimization procedure. Theoretica chimica acta 1979, 53, 75–93
- Frankel 2011 Frankel, T. The Geometry of Physics: An Introduction, 3rd ed.; Cambridge University Press, 2011
- Mezey 1987 Mezey, P. G. Potential Energy Hypersurfaces; Studies in physical and theoretical chemistry; Elsevier Scientific Publishing Co., Amsterdam, 1987; Vol. 53
- Wales 2001 Wales, D. Energy Landscapes: Applications to Clusters, Biomolecules and Glasses, 1st ed.; Cambridge University Press, 2001
- Whitney 1944 Whitney, H. The Self-Intersections of a Smooth -manifold in -space. The Annals of Mathematics 1944, 45, 220–220
- Arnold 1989 Arnold, V. I. Mathematical Methods of Classical Mechanics; Graduate Texts in Mathematics; Springer New York, 1989; Vol. 60
- Rudin 1987 Rudin, W. Real and Complex Analysis, 3rd ed.; McGraw-Hill Book Co., 1987
- Spivak 1999 Spivak, M. A Comprehensive Introduction to Differential Geometry. Vol. II, 3rd ed.; Publish or Perish, Inc., 1999
- 52 Ref. 51, Chapter 6