Functorial Manifold Learning
Abstract
We adapt previous research on category theory and topological unsupervised learning to develop a functorial perspective on manifold learning, also known as nonlinear dimensionality reduction. We first characterize manifold learning algorithms as functors that map pseudometric spaces to optimization objectives and that factor through hierarchical clustering functors. We then use this characterization to prove refinement bounds on manifold learning loss functions and construct a hierarchy of manifold learning algorithms based on their equivariants. We express several popular manifold learning algorithms as functors at different levels of this hierarchy, including Metric Multidimensional Scaling, IsoMap, and UMAP. Next, we use interleaving distance to study the stability of a broad class of manifold learning algorithms. We present bounds on how closely the embeddings these algorithms produce from noisy data approximate the embeddings they would learn from noiseless data. Finally, we use our framework to derive a set of novel manifold learning algorithms, which we experimentally demonstrate are competitive with the state of the art.
1 Introduction
Suppose we have a finite pseudometric space that we assume has been sampled from some larger space according to some probability measure over . Manifold Learning algorithms like Isomap [23], Metric Multidimensional Scaling [2], and UMAP [19] construct -embeddings for the points in , which we interpret as coordinates for the support of . These techniques are based on the assumption that this support can be well-approximated with a manifold. In this paper we use functoriality, a basic concept from Category Theory, to explore two aspects of manifold learning algorithms:
-
โข
Equivariance: A manifold learning algorithm is equivariant to a transformation if applying that transformation to its inputs results in a corresponding transformation of its outputs. Understanding the equivariance of a transform lets us understand how it will behave on new types of data.
-
โข
Stability: The stability of a manifold learning algorithm captures how well the embeddings it generates on noisy data approximate the embeddings it would generate on noiseless data. Understanding the stability of a transform helps us predict its performance on real-world applications.
1.1 Functoriality
In order for a manifold learning algorithm to be useful, the properties of the embeddings that the algorithm derives from must be somewhat in line with the structure of . One way to formalize this is to cast the algorithm as a functor between categories. A category is a collection of objects and morphisms between them. Morphisms are closed under an associative composition operation, and each object is equipped with an identity morphism. An example of a category is the collection of sets (objects) and functions (morphisms) between them.
A functor is a mapping between categories that preserves identity morphisms and morphism composition. Underlying this straightforward definition is a powerful concept: the functoriality of a transformation is a blueprint for its structure, expressed in terms of the equivariants it preserves. If a given transformation is functorial over some pair of categories, then the transformation preserves the structure represented by those categoriesโ morphisms. By identifying the settings under which an algorithm is functorial, we can derive extensions of the algorithm that preserve functoriality and identify modifications that break it. See โBasic Category Theoryโ [18] or โSeven Sketches in Compositionalityโ [14] for more details on categories and functoriality.
1.2 Summary of Contributions
In Section 2, we demonstrate that manifold learning algorithms can be expressed as optimization problems defined on top of hierarchical overlapping clusterings. That is, we can express these algorithms in terms of the composition of:
-
โข
a strategy for clustering points at different distance scales
-
โข
an order-preserving transformation from a clustering of points to a loss function
We formalize this relationship in terms of the composition of functors between categories of pseudometric spaces, hierarchical overlapping clusterings, and optimization problems. This allows us to formally extend clustering theory into manifold learning theory.
In Section 2.1 we build on clustering theory to demonstrate that every manifold learning objective lies on a spectrum based on the criterion by which embeddings are penalized for being too close together or too far apart. In Section 2.2 we introduce a hierarchy of manifold learning algorithms and categorize algorithms based on the dataset transformations over which they are equivariant. In Section 2.3 we provide several examples of this categorization. We show that UMAP is equivariant to isometries and both IsoMap and Metric Multidimensional Scaling are equivariant to surjective non-expansive maps. Identifying these equivariants is helpful for identifying the circumstances under which we would expect our algorithms to generalize. For example, while adding points to a dataset will modify the IsoMap objective in a predictable way, this will completely disrupt the UMAP objective. This is caused by the fact that UMAP uses a local distance rescaling procedure that is density-sensitive, and is therefore not equivariant to injective or surjective non-expansive maps.
In Section 3 we use interleaving distance to study the stability of a broad class of manifold learning algorithms. We present novel bounds on how well the embeddings that algorithms in this class learn on noisy data approximate the embeddings they learn on noiseless data.
In Section 4 we build on this theory to describe a strategy for deriving novel algorithms from existing manifold learning algorithms. As an example we derive the Single Linkage Scaling algorithm, which projects samples in the same connected component of the Rips complex as close together as possible. We also present experimental results demonstrating the efficacy of this algorithm.
1.3 Related Work
Several authors have explored functorial perspectives on clustering algorithms. Carlsson et al.ย [9] introduce clustering functors that map metric spaces to partitionings, whereas Culbertson et al.ย [12] take a slightly broader scope and also explore overlapping clustering functors that map metric spaces to coverings. Both approaches demonstrate that metric space categories with fewer morphisms permit a richer class of clustering functors. For example, while the single linkage clustering algorithm is functorial over the full category of metric spaces and non-expansive maps, density-sensitive clustering algorithms like robust single linkage are only functorial over the subcategory of metric spaces and injective non-expansive maps. In order to get around the Kleinberg Impossibility Theorem [17], which states that any scale invariant flat clustering must sacrifice either surjectivity or a notion of consistency, several authors [9, 13, 20] also explore hierarchical clustering functors that map metric spaces to multi-scale dataset partitionings or covers. Shiebler [21] builds on this perspective to factor clustering functors through a category of simplicial complexes.
Manifold learning shares structure with hierarchical clustering, and some authors have begun applying categorical ideas to manifold learning. For example, McInnes et al.ย [19] introduce the UMAP manifold learning algorithm in terms of Spivakโs fuzzy simplicial sets [22], which are a categorical analog of simplicial filtrations.
In Section 3 we study the stability of manifold learning algorithms to dataset noise. Due to the importance of this topic, many other authors have researched the stability properties of manifold learning algorithms. For example, Baily [4] explores adaptations of PCA to noisy data, and Gerber et al.ย [16] demonstrate that Laplacian Eigenmaps has nicer stability properties than IsoMap. However, we believe that ours is the first work that uses interleaving distance to formalize a stability property.
1.4 Preliminaries on Functorial Hierarchical Overlapping Clustering
We briefly review some definitions from the functorial perspective on hierarchical overlapping clustering. For more details, see Shiebler [21]. Given a set , a non-nested flag cover of is a cover of such that: (1) if and , then , (2) the simplicial complex with vertices corresponding to the elements of and faces to all finite subsets of the sets in is a flag complex, or a simplicial complex that can be expressed as the clique of its -skeleton. Intuitively, is a non-nested flag cover if there exists a graph whose cliques are the sets in . The category has tuples as objects where is a non-nested flag cover of the finite set . The morphisms between and are functions where for any set in there exists some set in such that .
Next, we will represent datasets with the category of finite pseudometric spaces and non-expansive maps between them. Given a subcategory of , a flat -clustering functor is a functor that is the identity on the underlying set. Intuitively, a flat -clustering functor maps a dataset to a cover of the set in such a way that increasing the distances between points in may cause clusters to separate.
A fuzzy non-nested flag cover is a functor such that for any morphism in , the morphism is the identity on the underlying set. In the category objects are fuzzy non-nested covers and morphisms are natural transformations between them. Given a subcategory of a hierarchical -clustering functor is a functor such that for , is a flat -clustering functor. Intuitively, a hierarchical -clustering functor maps a pair of a dataset and a strength to a cover of the set in such a way that increasing the distances between points in or increasing the strength may cause clusters to separate.
2 Manifold Learning
Given a choice of , a manifold learning algorithm constructs an real valued matrix of embeddings in from a finite pseudometric space with points. In this work we focus on algorithms that operate by solving embedding optimization problems, or tuples where is a loss function, or a penalty term. We call the set of all that minimize the solution set of the embedding optimization problem. In particular, we focus on pairwise embedding optimization problems, or embedding optimization problems where can be expressed as a sum of pairwise terms such that . Intuitively the terms act as a penalty on bad embeddings. We express such a pairwise embedding optimization problem with the tuple .
Formally we define a manifold learning problem to be a function that maps the pseudometric space to a pairwise embedding optimization problem of the form . Note that this definition does not include any specification of how the optimization problem is solved (such as gradient descent or reduction to an eigenproblem). For example, the Metric Multidimensional Scaling manifold learning problem maps the pseudometric space to where . Optimizing this objective involves finding a matrix that minimizes . That is, the Euclidean distance matrix of the rows of the optimal is as close as possible to the distance matrix.
If a manifold learning problem maps isometric pseudometric spaces to embedding optimization problems with the same solution set, we call it isometry-invariant. Intuitively, isometry-invariant manifold learning algorithms do not change their output based the ordering of . A particularly useful property of isometry-invariant manifold learning problems is that they factor through hierarchical clustering algorithms.
Proposition 1.
Given any isometry-invariant manifold learning problem , there exists a manifold learning problem , where is a hierarchical overlapping clustering algorithm (as defined by Shiebler [21]) and is a function that maps the output of to an embedding optimization problem, such that the solution spaces of the images of and on any pseudometric space are identical.
Intuitively, Proposition 1 holds because manifold learning problems generate loss functions by grouping points in the finite pseudometric space together. In order to use this property to adapt clustering theorems into manifold learning theorems we will introduce a target category of pairwise embedding optimization problems and replace functions with functors from into this category. To start, consider the category :
Definition 2.1.
The objects in are tuples where is a natural number and is a real-valued function that satisfies for or . is a preorder where iff for any we have .
Given a choice of we can view the object in as a pairwise embedding optimization problem. However, is not quite expressive enough to serve as our target category. Recall the Metric Multidimensional Scaling manifold learning problem, which maps the pseudometric space to the pairwise embedding optimization problem where . Since does not vary monotonically with , it is clear that this manifold learning problem is not a functor from to . However, note that we can express as the sum of a term that increases monotonically in and a term that decreases monotonically in :
We will see in Section 2.3 that the embedding optimization problems associated with many common manifold learning algorithms decompose similarly. We can build on this insight to create a new category with the following pullback:
where is the forgetful functor that maps to . Intuitively is a subcategory of and we can write the objects in as tuples where iff for any we have and . Given a choice of , each object in corresponds to the pairwise embedding optimization problem where .
Similarly to the representation of hierarchical clustering algorithms as maps into a category of functors , we will represent manifold learning algorithms as mapping into a category of functors . The objects in are functors that commute with the forgetful functor that maps to , and the morphisms are natural transformations. We call the cardinality of . We can define a functor which maps the functor where to the tuple where:
Therefore, each functor with cardinality corresponds to the pairwise embedding optimization problem where . We will call the sum of these terms, , the -loss:
We can now give our definition of a manifold learning functor:
Definition 2.2.
Suppose is the category of pseudometric spaces and non-expansive maps and is the category of fuzzy non-nested flag covers and natural transformations (see Section 1.4). Then given the subcategories , the composition forms a -manifold learning functor if is a hierarchical -clustering functor and is a functor that maps a fuzzy non-nested flag cover with vertex set to some with cardinality .
Intuitively a manifold learning functor factors through a hierarchical clustering functor and sends to where . We will say that is in standard form if maps the one-point metric space to some where and . Each manifold learning functor corresponds to a manifold learning problem that maps to .
2.1 A Spectrum of Manifold Learning Functors
Recall the single and maximal linkage hierarchical overlapping clustering algorithms and . These algorithms map the pseudometric space to the fuzzy non-nested flag cover where is respectively the connected components of the -Vietoris-Rips complex of and the maximally linked sets of the relation in which if [21, 12]. If we apply functoriality to Proposition 6 in Shiebler [21] we see:
Proposition 2.
Intuitively, this proposition states that every manifold learning functor maps to a loss that is both no more interconnected than the loss that does not distinguish points within the same connected component of the Vietoris-Rips complex and no less interconnected than the loss that treats each pair of points independently.
There are many manifold learning functors that lie between these extremes. In particular, for any functor and sequence of clustering functors whose outputs refine each other, we can apply functoriality to form a sequence of manifold learning functors . For example, consider the family of hierarchical overlapping clustering functors from Culbertson et al.ย [12]: for , the cover is the maximal linked sets of the relation where if there is a sequence in where . The functor therefore maps to a loss that distinguishes only between points whose shortest path in the Vietoris-Rips complex is longer than . For this loss is more interconnected than and less interconnected than . This also suggests a recipe for generating new manifold learning algorithms (see Section 4): first express an existing manifold learning problem in the form , and then form , or any of the functors along the spectrum .
2.2 Characterizing Manifold Learning Problems
Similarly to how Carlsson et al.ย [9] characterize clustering algorithms in terms of their functoriality over different subcategories of pseudometric spaces, we can characterize manifold learning algorithms based on the subcategory over which they are functorial.
We have already introduced isometry-invariant manifold learning problems. Any -manifold learning functor is isometry-invariant, and an isometry-invariant manifold learning problem is expansive-contractive if the loss that it aims to minimize decomposes into the sum of an expansive term that decreases as distances increase and a contractive term that increases as distances increase. Intuitively, expansive-contractive manifold learning problems use the term to push together points that are close in the original space and use the term to push apart points that are far in the original space. Any -manifold learning functor is expansive-contractive.
An expansive-contractive manifold learning problem is positive extensible if increases and decreases when we add new points to increase . If instead decreases and increases when we add new points to increase , we say it is negative extensible. Intuitively, many positive-extensible manifold learning problems are minmax problems that aim to simultaneously minimize and maximize . Any -manifold learning functor is positive extensible and any -manifold learning functor is negative extensible.
Proposition 3.
Suppose is a standard form -manifold learning functor and is a standard form -manifold learning functor. Then for any and we have that , are non-positive and , are non-negative.
In the next section we show examples of manifold learning algorithms in each of these categories.
2.3 Examples
2.3.1 Metric Multidimensional Scaling (-Manifold Learning Functor)
The most straightforward strategy for learning embeddings is to minimize the difference between the pairwise distance matrix of the original space and the pairwise Euclidean distance matrix of the learned embeddings. The Metric Multidimensional Scaling algorithm [2] does exactly this. Given a finite pseudometric space , the Metric Multidimensional Scaling embedding optimization problem is where . When the distance matrix of the pseudometric space is double-centered (i.e. mean values of rows/columns are zero) this is the same objective that Principal Components Analysis (PCA) optimizes [3]. Now define to map the fuzzy non-nested flag cover with vertex set to where and:
where:
Proposition 4.
is a functor, and is a -manifold learning functor that maps the finite pseudometric space to the Metric Multidimensional Scaling embedding optimization problem.
2.3.2 IsoMap (-Manifold Learning Functor)
For many real world datasets it is the case that the distances between nearby points are more reliable and less noisy than the distances between far away points. The IsoMap algorithm [23] redefines the distances between far apart points in terms of the distances between near points. Given a finite pseudometric space , the IsoMap embedding optimization problem is where such that is the length of the shortest path between and in the graph in which there exists an edge of length between each pair of points with .
Proposition 5.
For any , there exists a hierarchical -clustering functor such that the -manifold learning functor maps the finite pseudometric space to the IsoMap embedding optimization problem.
2.3.3 -Path Scaling (-Manifold Learning Functor)
Given a finite pseudometric space and , suppose is the smallest such that there exists a path of edges between and in the -Vietoris Rips complex of . Then the -manifold learning functor maps to the -Path Scaling embedding optimization problem , where .
2.3.4 -Vertex-Connected Scaling (-Manifold Learning Functor)
For the hierarchical overlapping clustering functor maps the finite pseudometric space to the fuzzy non-nested flag cover where is the set of maximal -vertex-connected subgraphs of the -Vietoris-Rips complex of . Note that and . Note also that for the map is functorial over but not all of since a non-injective map may split a -vertex-linked subgraph [12].
Now given a finite pseudometric space and , suppose is the smallest such that there exists a -vertex-connected subgraph of the -Vietoris-Rips complex of that contains both and . Then the -manifold learning functor maps to the -Vertex Connected Scaling embedding optimization problem where . Note that unlike , for the map is not functorial over all of since is not functorial over .
2.3.5 UMAP (-Manifold Learning Functor)
The UMAP algorithm builds a local uber-metric space around each point in , converts each local uber-metric space to a fuzzy simplicial complex, and minimizes a loss function based on a fuzzy union of these fuzzy simplicial complexes. Given a finite pseudometric space , the UMAP embedding optimization problem is where is the fuzzy cross-entropy:
and is the weight of the fuzzy union of the -simplices connecting and in the Vietoris-Rips complexes formed from the local uber-metric spaces where:
Proposition 6.
There exists a hierarchical -clustering functor that decomposes into the composition of four functors that:
-
1.
build a local uber-metric space around each point in ;
-
2.
convert each local uber-metric space to a fuzzy simplicial complex;
-
3.
take a fuzzy union of these fuzzy simplicial complexes;
-
4.
convert the resulting fuzzy simplicial complex to a fuzzy non-nested flag cover.
Proposition 7.
There exists a functor where is a -manifold learning functor that maps the pseudometric space to the UMAP embedding optimization problem.
Since the UMAP distance rescaling procedure does not preserve non-expansive maps, even if a map from to is non-expansive, this will not necessarily be the case for all of the local uber-metric spaces that we build from and . For this reason is not functorial over .
3 Stability of Manifold Learning Algorithms
We can use this functorial perspective on manifold learning to reason about the stability of manifold learning algorithms under dataset noise. An -interleaving between the functors is a collection of commuting natural transformations between and [10, 11]. The interleaving distance between such functors is the smallest such that an -interleaving exists. In order to study interleavings between functors in or whose domain is rather than , we will say that the functors are -interleaved when there is an -interleaving between the functors and where . We will also write .
Proposition 8.
Given a subcategory of , a standard form -manifold learning functor and a pair of finite pseudometric spaces such that there exists a pair of morphisms in , we have .
Proposition 8 is similar in spirit to previous results that use the Gromov-Hausdorff distance between metric spaces to bound the bottleneck or homotopy interleaving distances between their corresponding Vietoris-Rips complexes [11, 8, 20, 6]. As a special case, if is an -manifold learning functor and there exists an -isometry between then . We can use this to prove the following:
Proposition 9.
Suppose we have a standard form -manifold learning functor , a pair of -isometric finite pseudometric spaces and the matrices that respectively minimize and . Then if:
and:
we have:
(2) |
If is constant in (such as for any that factors as ) we have:
(3) |
For a very simple example, consider Multidimensional Scaling without dimensionality reduction. In this case and are each finite ordered -element subspaces of with Euclidean distance. If we write the vectors in and as matrices , then these matrices respectively minimize and , and . Since the function that sends the th element of to the th element of must be an -isometry, Proposition 9 bounds the average of the squared distances between the paired rows of two matrices in terms of the largest such distance.
These bounds apply to a very general class of manifold learning algorithms, including topologically unstable algorithms like IsoMap [5]. As an example, consider using IsoMap to project evenly spaced points that lie upon the surface of a radius circle in onto . In this case is a finite ordered -element subspace of with Euclidean distance, and for any matrix that consists of evenly spaced points along the real line such that we have . Now suppose that we instead apply IsoMap to a noisy view of : a finite ordered -element subspace of where is Euclidean distance and . Then for any matrix that minimizes , Proposition 9 bounds the average squared difference between and .
4 Experiments in Functorial Recombination
One benefit of the functorial perspective on manifold learning is that it provides a natural way to produce new manifold learning algorithms by recombining the components of existing algorithms. Suppose we are able to express two existing manifold learning algorithms in this framework such that and where are hierarchical clustering functors. Then we can use the compositionality of functors to define the manifold learning algorithms or . We can use this procedure to combine the strengths of multiple algorithms in a way that preserves functoriality (and therefore also stability by Proposition 9). For example, if we compose the functor from Proposition 6 with we form the -manifold learning functor which maps to the embedding optimization problem where and is the strength of the fuzzy simplex that UMAP forms between and .
For a more illustrative example, consider a DNA recombination task in which we attempt to match a string of DNA that has been repeatedly mutated back to the original string. One way to solve this task is to generate embeddings for each string of DNA and look at the nearest neighbors of the mutated string. We can simulate this task as follows
-
1.
Generate original random sequences of DNA of length (strings of โAโ, โCโ, โGโ, โTโ).
-
2.
For each sequence, mutate the sequence times to produce a mutation list, or a list of sequences which each start with an original DNA sequence and end with a final DNA sequence.
-
3.
Collect each of the sequences in each of these mutation lists into an element finite pseudometric space with Hamming distance.
-
4.
Build embeddings from this pseudometric space and compute the percentage of mutation lists for which the nearest neighbor of the last DNA sequence in that list among the set of all original sequences is the first sequence in that list (the accuracy).
A manifold learning algorithm that performs well on this task will need to take advantage of the intermediate mutation states to recognize that the first state and final state in a mutation list should be embedded as close together as possible. We can follow the procedure in Section 2.1 to adapt the Metric Multidimensional Scaling algorithm (Section 2.3.1) into such an algorithm by forming the maximally interconnected functor . Intuitively, this functor maps to a loss function that corresponds to the optimization objective for Metric Multidimensional Scaling where Euclidean distance is replaced with:
We call this the Single Linkage Scaling algorithm (Algorithm 1). Since this algorithm embeds points that are connected by a sequence in the original space as close together as possible, we expect Single Linkage Scaling to outperform Metric Multidimensional Scaling on this task. This is exactly what we see in Table 4. We also show the embeddings for each sequence in each list in Figure 1.

Algorithm | Accuracy | |||
---|---|---|---|---|
Accuracy | ||||
Accuracy | ||||
Accuracy | ||||
Metric Multidimensional Scaling | ||||
Embedding Size 2 | 0.21 ( 0.05) | 0.01 ( 0.02) | 0.29 ( 0.02) | 0.01 ( 0.00) |
Single Linkage Scaling | ||||
Embedding Size 2 | 0.61 ( 0.02) | 0.68 ( 0.05) | 0.76 ( 0.01) | 0.32 ( 0.02) |
Metric Multidimensional Scaling | ||||
Embedding Size 5 | 0.74 ( 0.01) | 0.13 ( 0.02) | 0.84 ( 0.01) | 0.04 ( 0.01) |
Single Linkage Scaling | ||||
Embedding Size 5 | 0.93 ( 0.05) | 0.91 ( 0.02) | 0.96 ( 0.02) | 0.34 ( 0.02) |
5 Discussion and Future Work
We have taken the first steps towards a categorical framework for manifold learning. By defining an algorithm as a functor from a category of metric spaces, we can explicitly express the kind of dataset transformations under which it is equivariant. We have shown that for many popular manifold learning algorithms, including Metric Multidimensional Scaling and IsoMap, the optimization objective changes in a predictable way as we modify the metric space.
The functorial perspective also suggests a new strategy for exploratory data analysis with manifold learning. Since we can decompose manifold learning algorithms into two components (clustering and loss) we can examine how slight variations of the clustering algorithm affect the learned embeddings. We have shown in Section 2.1 that every manifold learning functor lies on a spectrum of interconnectedness between and , and we can form new algorithms by moving along this spectrum. For example, we saw in Section 4 that replacing the functor with in the Metric Multidimensional Scaling algorithm substantially changes the learned embeddings and improves performance on a DNA recombination task. There are also many algorithms that lie between these two options, including the -Path Scaling and -Vertex-Connected Scaling algorithms that we introduced in Section 2.3.
Another major benefit of expressing algorithms as functors is that functors preserve categorical properties like interleaving distance. This allows us to easily reason about the stability properties of both existing algorithms and new algorithms that we create by recombining functors. Other authors have used this strategy to prove stability properties of the homology of geometric filtered complexes [11]. In Section 3 we have used this strategy to define bounds on how dataset noise affects optimization quality. In future work we hope to use these techniques to derive more powerful theorems around the resilience of other kinds of unsupervised or supervised algorithms to noise. For example, we may also be able to tighten our bounds by switching our perspective from finite metric spaces to distributions [7] or even involving categorical probability [15] to replace interleaving distance with a probabilistic analog. Due to the simplicity and flexibility of this strategy, other researchers have begun to develop more flexible characterizations of interleaving distance that we can apply in even more situations [20].
References
- [1]
- [2] Hervรฉ Abdi (2007): Metric multidimensional scaling (MDS): analyzing distance matrices. Encyclopedia of Measurement and Statistics. SAGE Publications.
- [3] Hervรฉ Abdi & Lynneย J Williams (2010): Principal component analysis. Wiley interdisciplinary reviews: computational statistics 2(4), pp. 433โ459, 10.1002/wics.101.
- [4] Stephen Bailey (2012): Principal Component Analysis with Noisy and/or Missing Data. Publications of the Astronomical Society of the Pacific 124(919), pp. 1015โ1023, 10.1086/668105. Available at http://www.jstor.org/stable/10.1086/668105.
- [5] Mukund Balasubramanian (2002): The isomap algorithm and topological stability. Science 295(5552), 10.1126/science.295.5552.7a.
- [6] Andrewย J Blumberg & Michael Lesnick (2017): Universality of the homotopy interleaving distance. arXiv preprint arXiv:1705.01690.
- [7] Adam Brown, Omer Bobrowski, Elizabeth Munch & Bei Wang (2020): Probabilistic convergence and stability of random mapper graphs. Journal of Applied and Computational Topology, pp. 1โ42, 10.1007/s41468-020-00063-x.
- [8] Peter Bubenik & Jonathanย A Scott (2014): Categorification of persistent homology. Discrete & Computational Geometry 51(3), pp. 600โ627, 10.1007/s00454-014-9573-x.
- [9] Gunnar Carlsson & Facundo Mรฉmoli (2013): Classifying clustering schemes. Foundations of Computational Mathematics, 10.1007/s10208-012-9141-9.
- [10] Frรฉdรฉric Chazal, David Cohen-Steiner, Marc Glisse, Leonidasย J Guibas & Steveย Y Oudot (2009): Proximity of persistence modules and their diagrams. In: Proceedings of the twenty-fifth annual symposium on Computational geometry, pp. 237โ246, 10.1145/1542362.1542407.
- [11] Frรฉdรฉric Chazal, Vin Deย Silva & Steve Oudot (2014): Persistence stability for geometric complexes. Geometriae Dedicata 173(1), pp. 193โ214, 10.1007/s10711-013-9937-z.
- [12] Jared Culbertson, Danย P Guralnik, Jakob Hansen & Peterย F Stiller (2016): Consistency constraints for overlapping data clustering. arXiv preprint arXiv:1608.04331.
- [13] Jared Culbertson, Danย P Guralnik & Peterย F Stiller (2018): Functorial hierarchical clustering with overlaps. Discrete Applied Mathematics 236, pp. 108โ123, 10.1016/j.dam.2017.10.015.
- [14] Brendan Fong & Davidย I. Spivak (2019): An Invitation to Applied Category Theory: Seven Sketches in Compositionality. Cambridge University Press, 10.1017/9781108668804.
- [15] Tobias Fritz (2020): A synthetic approach to Markov kernels, conditional independence and theorems on sufficient statistics. Advances in Mathematics 370, p. 107239, 10.1016/j.aim.2020.107239.
- [16] Samuel Gerber, Tolga Tasdizen & Ross Whitaker (2007): Robust non-linear dimensionality reduction using successive 1-dimensional Laplacian eigenmaps. In: Proceedings of the 24th international conference on Machine learning, pp. 281โ288, 10.1145/1273496.1273532.
- [17] Jonย M Kleinberg (2003): An impossibility theorem for clustering. Advances in Neural Information Processing Systems, 10.5555/2968618.2968676.
- [18] Tom Leinster (2016): Basic Category Theory. Cambridge University Press, 10.1017/CBO9780511525858.
- [19] Leland McInnes, John Healy, Nathaniel Saul & Lukas Groรberger (2018): UMAP: Uniform Manifold Approximation and Projection. Journal of Open Source Software 3(29), p. 861, 10.21105/joss.00861.
- [20] Luisย N Scoccola (2020): Locally Persistent Categories And Metric Properties Of Interleaving Distances. PhD Thesis at Western University (Ontario). Available at https://ir.lib.uwo.ca/etd/7119/.
- [21] Dan Shiebler (2020): Functorial Clustering via Simplicial Complexes. Topological Data Analysis and Beyond Workshop at NeurIPS 2020. Available at https://openreview.net/pdf?id=ZkDLcXCP5sV.
- [22] Davidย I Spivak (2012): Metric realization of fuzzy simplicial sets. Self published notes. Available at http://math.mit.edu/~dspivak/files/metric_realization.pdf.
- [23] Joshuaย B Tenenbaum, Vin Deย Silva & Johnย C Langford (2000): A global geometric framework for nonlinear dimensionality reduction. Science 290(5500), pp. 2319โ2323, 10.1126/science.290.5500.2319.