Two Sample Test for Extrinsic Antimeans on Planar Kendall Shape Spaces with an Application to Medical Imaging
Abstract. In this paper one develops nonparametric inference procedures for comparing two extrinsic antimeans on compact manifolds. Based on recent Central limit theorems for extrinsic sample antimeans w.r.t. an arbitrary embedding of a compact manifold in a Euclidean space, one derives an asymptotic chi square test for the equality of two extrinsic antimeans. Applications are given to distributions on complex projective space w.r.t. the Veronese-Whitney embedding, that is a submanifold representation for the Kendall planar shape space. Two medical imaging analysis applications are also given.
1 Introduction
To date, the statistical analysis on object spaces is concerned with data analysis on sample spaces that have a complete metric space structure. In include, but not limited to, data extracted from DNA and RNA sequences, from medical images (such as classification and discrimination two different groups) and 3D computer vision outputs.
In those areas of research, one needs to extract certain geometric information, from images, or from DNA. In case of images, the features extracted are, often times, various sorts of shapes of labeled configurations of points. For similarity shape , we refer to Kendall (1984) [2], Kendall et al(1999) [3], Bookstein [5], Dryden and Mardia (2016)[22]. For affine shapes, projective shapes, or, in general, -shapes, see Patrangenaru and Ellingson(2015), , Section 3.5 [6].
The development of novel statistical principles methods for complex data types extracted from image data, relies upon key features of object spaces; namely non-linearity and topological Structure. Information extracted from an image is typically represented as point on non-linear spaces. Statistics and Geometry share an important commonality: they are both based on the concept of distance. However not just any distance is suitable for modeling a given object data: distances that give a homogeneous space structure are preferred, to allow for comparing location parameters of two distribution. In addition, in general one would like for the estimators of a random object’s location parameters to be consistent and furthermore allow for estimation of location parameters based on CLM results.
In classical nonparametric statistical theory on an object space with manifold structure, an estimator’s asymptotic behavior is described in terms of the asymptotic expansion around zero, of the tangential component of a consistent estimator aaround the true parameter (see Patrangenaru and Ellingson (2015, Cha.4)[6]), and our paper follows this direction. Fast computational algorithms are also key to the object data analysis, therefore we use an extrinsic analysis (see Bhattacharya et al.(2012)[1]), which, for such reasons, is preferred in Object Data Analysis, over so called “intrinsic” approach, if there is a choice for a distance on The extrinsic data analysis on is using a chord distance, which is the distance induced by the Euclidean distance via the embedding of , and is given by Here is the Euclidean norm.
If for a given probability measure on an embedded compact metric space the Fréchet function associated with the random object with is
(1.1) |
and its minimizers form the extrinsic mean set of (see Bhattacharya and Patrangenaru 2003 [10]). If there is a unique point in the extrinsic mean set, this point is the extrinsic mean, and is labeled . The maximizers in (1.1) form the extrinsic antimean set. In there is a unique point in the extrinsic antimean set, this point is the extrisic antimean, and is labeled , or simply when is known (See Patrangenaru and Guo 2016[12]). Also, given i.i.d.r.o.’s from , their extrinsic sample mean is the extrinsic mean of the empirical distribution and their extrinsic sample antimean is the extrinsic antimean of the empirical distribution assuming these sample extrinsic means or antimeans exist.
The main goal of this article is to establish some general methods for comparing extrinsic antimeans on a compact manifold embedded in an Euclidean space. In Section 2, some preliminary result, a central limit theorem for sample extrinsic antimeans on compact manifolds is brought up for future developments. Thereafter, a two sample test statistic for extrinsic antimeans on compact manifolds is given in Section 3. In Section 4, one focuses on the case of the planar Kendall shape space of k-ads, which is often represented in literature by the complex projective space ( see Patrangenaru and Ellingson (2015)[6], Ch.2, p.142 ). The embedding considered here is the Veronese-Whitney embedding of in the space of self adjoint complex matrices (see Patrangenaru and Ellingson (2015)[6], Ch.2,p. 154). In Section 5, we derive a nonparametric two sample test for the equality of the population Veronese-Whitey (VW) antimeans on the complex projective space . Finally, Section 6 illustrate a two example of Kendall shape data from Bookstein ((1997)[24]). The first example is on Apert syndrome vs clinically normal children, known as the University School data. The second example is on comparing brain scan shapes in schizophrenic vs clinically normal children.
2 Extrinsic mean and antimean of a random object
As discussed in the introduction, when is an embedding of a compact object space into a numerical space, the set of minimizers of the Fréchet function is the extrinsic mean set, and when this set has a unique point, this point is called the extrinsic mean. The maximizers of this function form the extrinsic antimean set, when this later set has one point only, this point is called the extrinsic antimean. We have following properties of the extrinsic antimean from Patrangenaru, Yao and Guo [12]
DEFINITION 2.1.
A point for which there is a unique point satisfying the equality,
(2.1) |
is called -nonfocal. A point which is not -nonfocal is said to be -focal. If y is an -nonfocal point, its farthest projection on is the unique point with .
DEFINITION 2.2.
A point for which there is a unique point satisfying the equality,
(2.2) |
is called -nonfocal. A point which is not -nonfocal is said to be -focal. If y is an -nonfocal point, its projection on is the unique point with .
DEFINITION 2.3.
A probability distribution on is said to be -nonfocal if the mean of is -nonfocal.
A probability distribution on is said to be -nonfocal if the mean of is -nonfocal.
Then we have the following theorem from Patrangenaru, Guo and Yao (2016)[21], which in particular is valid for a probability measure on an embedded compact object space with the chord distance
THEOREM 2.1.
Let be a probability measure associated with the random object on a compact metric space So we have is finite on (a) Then, given any , there exist a -null set and such that the Fréchet (sample) antimean set of is contained in the -neighborhood of the Fréchet antimean set of for all . (b) If the Fréchet antimean of exists then every measurable choice from the Fréchet (sample) antimean set of is a strongly consistent estimator of the Fréchet antimean of .
2.1 Previous Asymptotic Results for Extrinsic Sample Antimeans
In preparation, we are using the large sample distribution for extrinsic sample antimeans given in Patrangenaru et al (2016 [12]).
Assume is an embedding of a -dimensional manifold such that is closed in , and is a -nonfocal probability measure on such that has finite moments of order 2. Let and be the mean and covariance matrix of regarded as a probability measure on . Let be the set of -focal points of , and let be the farthest projection on . is differentiable at and has the differentiability class of around any nonfocal point. In order to evaluate the differential we consider a special orthonormal frame field that will ease the computations.
A local frame field , defined on an open neighborhood is adapted to the embedding if it is an orhonormal frame field and where is a local frame field on and is the value of the local vector field at
Let be the canonical basis of and assume is an adapted frame field around . Then is a linear combination of :
(2.3) |
where is the differential of at By the delta method, converges weakly to , where and
(2.4) | |||
Here is the covariance matrix of w.r.t the canonical basis
The asymptotic distribution is degenerate and the support of this distribution is on , since the range of is . Note that for .
The tangential component of , w.r.t the basis is given by
(2.5) |
Then, the random vector has the following anticovariance matrix w.r.t the basis :
(2.6) | |||
which is the anticovariance matrix of the random object . To simplify the notation, we set
(2.7) |
Similarly, given i.i.d.r.o.’s from , we define the sample anticovariance matrix as the anticovariance matrix associated with the empirical distribution
If, in addition, rank , then is invertible and if we define the -standardized antimean vector
(2.8) |
using basic large sample theory results, including a generalized Slutsky’s lemma ( see Patrangenaru and Ellingson(2015)[6], p.65), one has:
THEOREM 2.2.
Assume is a random sample from the -nonfocal distribution , and let . Let be an orthonormal frame field adapted to . Then (a) the tangential component at the extrinsic antimean of has asymptotically a multivariate normal distribution in the tangent space to at with mean and anticovariance matrix , (b) if is nonsingular, the j-standardized antimean vector converges weakly to a random vector with a distribution, and (c) under the assumptions of (b)
(2.9) |
3 A two sample test for extrinsic antimeans
We now turn to two-sample tests for extrinsic antimeans of distributions on an arbitrary dimensional compact manifold For a large sample test for equality of two extrinsic means see Bhattacharya and Bhattacharya (2012) [14], p.42. Let be two independent random samples drawn from distributions on and let be an embedding of into Denote by the mean of the induced probability and its anticovariance matrix Then the extrinsic antimean of is , assuming is -nonfocal. Write and let be the corresponding sample antimeans. By Theorem 2.2
(3.1) |
where is the extrinsic anticovariance matrix of and are the same as in Theorem 2.2, and (2.7), with replaced by (a=1,2). That is, is the matrix of an orthonormal basis (frame) of for in a neighborhood of and Similarly, The null hypothesis say, is equivalent to say. Then, under the null hypothesis, letting one has and
(3.2) |
For statistical inference one estimates by
(3.3) |
where is the sample covariance matrix of sample Also is replaced by where is a sample estimate of Under both and are consistent estimates of so we take a “pooled estimate”
(3.4) |
We, therefore, have the following result:
THEOREM 3.1.
Assume the extrinsic sample anticovariance matrix is nonsingular for Then, under one has:
(3.5) | |||
4 VW Means and VW Antimeans
4.1 VW Means and VW Antimeans on
We consider the case of a probability measure on the complex projective space . If we consider the action of the multiplicative group on , given by scalar multiplication
(4.1) |
the quotient space is the -dimensional complex projective space , set of all complex lines in going through . One can show that is a 2 dimensional real analytic manifold, using transition maps, similar to those in the case of (see Patrangenaru and Ellingson (2015) [20], chapter 3).
Here we are concerned with a landmark based non-parametric analysis of similarity shape data (see Bhattacharya and Patrangenaru (2014)[23]). Our analysis is for extrinsic antimeans. For landmark based shape data, one considers a -ad , which consists of labeled points in that represent coordinates of k-labeled landmarks.
In this subsection is the group direct similarities of . A similarity is a function , that uniformly scale the Euclidean distances, that is, for which there is , such that . Using the fundamental theorem of Euclidean geometry, one can show that a similarity is given by . A direct similarity is a similarity of this form, where has a positive determinant. Under composition direct similarities from a group. The object space considered here consist in orbits of the group action of on -ads, and is called direct similarities shape space of k-ads in .
For , or , a direct similarity (Kendall) shape is that the geometrical information that remains when location, scale and rotational effects are filtered out from a -ads. Two -ads and are said to have the same shape if there is a direct similarity in the plane, that is, a composition of a rotation, a translation and a homothety such that for . Having the same Kendall shape is an equivalence relationship in the space of planar -ads, and the set of all equivalence classes of nontrivial -ads is called the Kendall planar shape space of -ads, which is denoted (see Balan and Patrangenaru(2005) [15]).
4.2 VW Antimeans on
Kendall (1984) [2] introduced that planar direct similarity shapes of k-ads; which is a set of k labeled points at least two of which are distinct, can be represented as points on a complex projective space . A standard shape analysis method was also introduced by Kent(1992)[4] and is called Veronese-Whitney(VW) embedding of in the space of self adjoint complex matrices, to represent shape data in an Euclidean space. This Veronese–Whitney embedding , where is the space of Hermitian matrices, is given by
(4.2) |
This embedding is a equivariant embedding and is called the special unitary group matrices of determinant 1, since , . The corresponding extrinsic mean (set) of a random shape X on is called the VW mean (set) (see Patrangenaru and Ellingson (2015)[6], Ch.3), and the VW mean,when it exists,and is labeled or simply .The corresponding extrinsic antimean (set) of a random shape X,is called the VW antimean (set) and is labeled or .
For the VW-embedding of the complex projective space, we have the following theorem for sample VW-means from Bhattacharya and Patrangenaru(2003) [10]:
THEOREM 4.1.
Let be a probability distribution on and let be a i.i.d.r.o.’s from . is nonfocal iff the largest eigenvalue of is simple and in this case where is an eigenvector of corresponding to , with . The extrinsic sample mean , where is an eigenvector of norm 1 of , , corresponding to the largest eigenvalue of J.
We also have a similar result for sample VW-antimeans (see Wang, Patrangenaru and Guo [17]):
THEOREM 4.2.
Let be a probability distribution on and let be a i.i.d.r.o.’s from . is -nonfocal iff , the smallest eigenvalue of is simple and in this case where is an eigenvector of corresponding to , with . The extrinsic sample antimean , where is an eigenvector of norm 1 of , , corresponding to the smallest eigenvalue of J.
5 Two sample testing problem for VW antimeans on
5.1 Application on the planar shape space of -ads
We are concerned with a landmark based nonparametric analysis of similarity shape data. For landmark based shape data , we will denote a -ad by complex numbers: We center the -ad at via a translation; next we rotate it by an angle and scale it, operations that are achieved by multiplying by the complex number We can represent the shape of the -ad as the complex one dimensional vector subspace passing through of the linear subspace Thus, the space of -ads is the set of all complex lines on the hyperplane, Therefore the shape space of nontrivial planar -ads can be represented as the complex projective space the space of all complex lines through the origin in using an isomorphism of and As in the case of , it is convenient to represent the element of corresponding to a -ad by the curve on the unit sphere in
5.2 Test for VW antimeans on planar shape spaces
Let and be two probability measures on the shape space and let and denote the antimeans of and where is the VW-embedding that where thus . Suppose and are i.i.d. random objects from and respectively. Let be their images in which are random samples from and respectively. Suppose we are to test if the VW antimeans of and are equal, i.e.
It is known that , where and similarly, , where . We assume that both and have simple smallest eigenvalues. Then under their unit corresponding eigenvectors differ by a rotation.
Choose with the same farthest projection as and . Suppose and where are the eigenvalues of and are the corresponding eigenvectors. Also, we obtain an orthonormal basis for , which is given by and . Where has all entries zero except for those in the positions (a, b) and (b, a) that are equal to 1 and is a matrix with all entries zero except for those in the positions (a, b) and (b, a) that are equal to i, respectively -i.
It is defined as,
(5.1) |
(5.2) |
where is the standard canonical basis for . For any where SU is special unitary group of all complex matrices. and are an orthonormal frame for . Now, we will choose the orthonormal basis frame for . Then it can be shown that
(5.3) |
(5.4) |
Then, we write
(5.5) |
Then from equations (5.3) ,(5.4) and (5.5), it follows that
Define
(5.6) |
(5.7) |
Then we have,
(5.8) |
Under , and have mean zero, and as , suppose for some It follows that
(5.9) |
where and are the covariances of and respectively. Then
(5.10) |
Thus assuming and to be nonsingular,
(5.11) |
We may take Since and are unknown, we may estimate by the pooled sample mean . Note that and are consistent estimator of and . Thus, we may use and to obtain a two sample test statistic, where
(5.12) |
(5.13) |
Then the two sample test statistic can be estimated by
(5.14) |
Given the significance level , we reject if
(5.15) |
The expression for depends on the spectrum of through the orbit and the subspace spanned by If the population antimean exists, is a consistent estimator of , the projection on Span converges to that on Span . Thus from (5.11) and (5.14), has an asymptotic distribution. Hence the test in (5.15) has asympotic level
6 Application to medical imaging
6.1 Apert syndrome vs clinically normal children
Our data consists of shapes for a group of eighth midface labeled anatomical landmarks from X-rays of skulls of eight year old and fourteen year-old North American children(36 boys and 26 girls), known as the University School data. Each child’s skull was imaged twice, at age 8 and next at age 14. The data set, collected to study anatomical changes during children growth, represents coordinates of eight craniofacial landmarks, whose names and position on the skull are given in Bookstein ((1997)[24]), see also http://life.bio.sunysb.edu/morph/data/Book-UnivSch.txt. The data has two data set: the first one is the Apert data in Bookstein (pp. 405-406), which contains eight landmarks that describing the children who has Apert syndrome (a genetic craniosynostosis) and the second data set is clinically normal children, which contains about 40 anatomical landmarks on the skull. Out of these only 8 landmarks are registered in both groups. The two groups share only 5 registered landmarks: Posterior nasal spine,Anterior nasal spine, Sella, Sphenoethmoid registration, and Nasion. For operational definitions of these landmarks, see Bookstein (pp. 71). The shape variable of the 5 landmarks on the upper mid-face is valued in .


In our application, we take both two sample test for VW means (see Bhattacharya and Bhattacharya [14]) and for VW antimeans, to see if we can distinguish between the Apert group and clinically normal group, using such Kendall shape location parameters. Figure 1 shows the plots of an icon for the sample VW means and VW antimeans for the two groups along with the pooled sample VW mean and VW antimean.
For the VW mean test, the values of the two sample test statistic for testing equality of the population extrinsic mean shapes is , along with the asymptotic and for the VW antimean test, we get the result , along with the asymptotic .
From this application study, we reject the null hypothesis and conclude that both the VW mean test and the VW antimean test show that one may distinguish the Apert and clinically normal group on based on their VW antimeans, or on their VW means.
6.2 Brain scan shapes of schizophrenic vs clinically normal children
In this example from Bookstein (1991), 13 landmarks are recorded on a midsagittal two-dimensional slice from a Magnetic Resonance brain scan of each of 14 schizophrenic children and 14 normal children. It is of interest to study differences in shapes of brains between the two groups which can be used to detect schizophrenia. This is an application of disease detection. The shapes of the sample k-ads lie in To distinguish between the underlying distributions, we compare their VW mean and VW antimean shapes.
For testing the equality of the VW means we use the test in Bhattacharya and Bhattacharya [14]. In this application, in addition we consider the two sample test for the equality of VW antimeans developed in the previous section. Figure 2 shows the plots of the sample VW means and VW antimeans for the two groups along with the pooled sample VW mean and VW antimean for this data sets.


For the VW mean test, since the values of the two sample test statistic for testing equality of the population extrinsic mean shapes is , along with the asymptotic and for the VW antimean test, we get the result , along with the asymptotic which is significant at level .
From this application study, we therefore reject the null hypothesis and conclude that there is difference between the schizophrenic children and normal children for both the VW means or VW antimeans.
Acknowledgments. We would like to thank Yunfan Wang for helpful comments on an early version of our paper.
References
- [1] Bhattacharya, R. N.; Ellingson, L.; Liu, X.; Patrangenaru, V.; Crane, M. (2012). Extrinsic analysis on manifolds is computationally faster than intrinsic analysis with applications to quality control by machine vision. Appl. Stoch. Models Bus. Ind. 28, 222–235.
- [2] Kendall, D. G. (1984). Shape manifolds, Procrustean metrics, and complex projective spaces. Bull. London Math. Soc. 16 81-121.
- [3] Kendall, D. G., Barden, D., Carne, T. K. and Le, H. (1999). Shape and Shape Theory. Wiley, New York.
- [4] J.T. Kent (1992). New directions in shape analysis. The Art of Statistical Science, A Tribute to G.S. Watson, 115–127, Wiley Ser. Probab. Math. Statist., Wiley, Chichester.
- [5] Bookstein, F. L. (1991). Morphometric Tools for Landmark Data: Geometry and Biology. Cambridge Univ. Press.
- [6] Patrangenaru, V., and Ellingson, L. (2015). Nonparametric statistics on manifolds and their applications to object data analysis. CRC Press.
- [7] M. Fréchet (1948). Les élements aléatoires de nature quelconque dans un espace distancié, Ann. Inst. H. Poincaré 10, 215–310.
- [8] H.Ziezold(1977). On expected figures and a strong law of large numbers for random elements in quasi-metric spaces. Transactions of Seventh Prague Conference on Information Theory, Statistical Decision Functions, Random Processes A. 591–602.
- [9] Huckemann, Stephan(2011), Inference on 3d Procrustes means: Tree bole growth, rank deficient diffusion tensors and perturbation models, Scandinavian Journal of Statistics, 38, 424–446.
- [10] Bhattacharya, R.N. and Patrangenaru, V. (2003). Large sample theory of intrinsic and extrinsic sample means on manifolds-Part I,Ann. Statist. 31, no. 1, 1–29.
- [11] Bhattacharya, Rabi and Patrangenaru, Vic (2003). Large sample theory of intrinsic and extrinsic sample means on manifolds, The Annals of Statistics, 31, 1, 1–29.
- [12] Patrangenaru, Vic and Yao, K. David and Guo, Ruite(2016), Extrinsic means and antimeans, Nonparametric Statistics, 161–178.
- [13] Patrangenaru, V. (2001). New large sample and bootstrap methods on shape spaces in high level analysis of natural images. Communications in Statistics-Theory and Methods, 30(8-9), 1675–1693.
- [14] Bhattacharya, A. and Bhattacharya, R. (2012). Nonparametric statistics on manifolds with applications to shape spaces, Institute of Mathematical Statistics Monographs. Cambridge University Press.
- [15] Balan, Vladimir and Patrangenaru, Vic (2006) Geometry of shape spaces, Proc. of The 5-th Conference of Balkan Society of Geometers. Bucharest, 2005. Geometry Balkan Press. 28–33.
- [16] R. N. Bhattacharya and V. Patrangenaru (2002). Non parametric estimation of location and dispersion on Riemannian manifolds. J. Statist. Plan. Infer. Volume in honor of the 80th birthday of professor C.R.Rao. 108, 23–35.
- [17] Wang, Y., Patrangenaru, V. and Guo, R. (2020). A Central Limit Theorem for Extrinsic Antimeans and Estimation of Veronese-Whitney Means and Antimeans on Planar Kendall Shape Spaces. Journal of Multivariate Analysis, 178, 104600, 11 pp.
- [18] Kendall, David G. (1984) Shape manifolds, procrustean metrics, and complex projective spaces. Bulletin of the London Mathematical Society, Oxford University Press.
- [19] R. N. Bhattacharya and V. Patrangenaru(2005). Large sample theory of intrinsic and extrinsic sample means on manifolds, Part II, Ann. Statist. 33, 1211–1245.
- [20] V. Patrangenaru and L. Ellingson, Nonparametric Statistics on Manifolds and Their Applications to Object Data Analysis, CRC-Chapman & Hall. 2015.
- [21] V.Patrangenaru and Guo, Ruite and Yao, K. David. (2016), Nonparametric Inference for Location Parameters via Fréchet Functions,Second International Symposium on Stochastic Models in Reliability Engineering, Life Science & Operations Management ,254–262.
- [22] Ian L. Dryden and Kanti V. Mardia (2016). Statistical shape analysis with applications in R. Second edition. Wiley Series in Probability and Statistics. John Wiley Sons, Ltd., Chichester.
- [23] R. N. Bhattacharya and V. Patrangenaru (2014). Rejoinder of Discussion paper “Statistics on Manifolds and Landmarks Based Image Analysis: A Nonparametric Theory with Applications.” Journal of Statistical Planning and Inference. 145, 42–48.
- [24] F. L. Bookstein (1997). Morphometric tools for landmark data: geometry and biology, Cambridge University Press.
- [25] Rabi N. Bhattacharya, Marius Buibas, Ian L. Dryden, Leif A. Ellingson, David Groisser, Harrie Hendriks, Stephan Huckemann, Huiling Le, Xiuwen Liu, James S. Marron, Daniel E. Osborne, Vic Patrângenaru, Armin Schwartzman, Hilary W. Thompson, and Andrew T. A.Wood. (2013) Extrinsic data analysis on sample spaces with a manifold stratification. Advances in Mathematics, Invited Contributions at the Seventh Congress of Romanian Mathematicians, Brasov, 2011, Publishing House of the Romanian Academy (Editors: Lucian Beznea, Vasile Brîzanescu, Marius Iosifescu, Gabriela Marinoschi, Radu Purice and Dan Timotin), pp. 241–252.