This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Two Sample Test for Extrinsic Antimeans on Planar Kendall Shape Spaces with an Application to Medical Imaging

Aaid Algahtani 111Department of Statistics, Florida State University, Tallahassee, FL 32304, U.S.A.,King Saud University, Riyadh 11451, Saudi Arabia  Vic Patrangenaru 222Department of Statistics, Florida State University, Tallahassee, FL 32304, U.S.A.

Abstract. In this paper one develops nonparametric inference procedures for comparing two extrinsic antimeans on compact manifolds. Based on recent Central limit theorems for extrinsic sample antimeans w.r.t. an arbitrary embedding of a compact manifold in a Euclidean space, one derives an asymptotic chi square test for the equality of two extrinsic antimeans. Applications are given to distributions on complex projective space CPk2CP^{k-2} w.r.t. the Veronese-Whitney embedding, that is a submanifold representation for the Kendall planar shape space. Two medical imaging analysis applications are also given.

1 Introduction

To date, the statistical analysis on object spaces is concerned with data analysis on sample spaces that have a complete metric space structure. In include, but not limited to, data extracted from DNA and RNA sequences, from medical images (such as classification and discrimination two different groups) and 3D computer vision outputs.

In those areas of research, one needs to extract certain geometric information, from images, or from DNA. In case of images, the features extracted are, often times, various sorts of shapes of labeled configurations of points. For similarity shape , we refer to Kendall (1984) [2], Kendall et al(1999) [3], Bookstein [5], Dryden and Mardia (2016)[22]. For affine shapes, projective shapes, or, in general, 𝒢\mathcal{G}-shapes, see Patrangenaru and Ellingson(2015), , Section 3.5 [6].

The development of novel statistical principles methods for complex data types extracted from image data, relies upon key features of object spaces; namely non-linearity and topological Structure. Information extracted from an image is typically represented as point on non-linear spaces. Statistics and Geometry share an important commonality: they are both based on the concept of distance. However not just any distance is suitable for modeling a given object data: distances that give a homogeneous space structure are preferred, to allow for comparing location parameters of two distribution. In addition, in general one would like for the estimators of a random object’s location parameters to be consistent and furthermore allow for estimation of location parameters based on CLM results.

In classical nonparametric statistical theory on an object space \mathcal{M} with manifold structure, an estimator’s asymptotic behavior is described in terms of the asymptotic expansion around zero, of the tangential component of a consistent estimator aaround the true parameter (see Patrangenaru and Ellingson (2015, Cha.4)[6]), and our paper follows this direction. Fast computational algorithms are also key to the object data analysis, therefore we use an extrinsic analysis (see Bhattacharya et al.(2012)[1]), which, for such reasons, is preferred in Object Data Analysis, over so called “intrinsic” approach, if there is a choice for a distance on .\mathcal{M}. The extrinsic data analysis on \mathcal{M} is using a chord distance, which is the distance ρj\rho_{j} induced by the Euclidean distance via the embedding jj of \mathcal{M}, and is given by ρj(p,q)=j(q)j(p)02.\rho_{j}(p,q)={\|j(q)-j(p)\|}^{2}_{0}. Here 0\|\cdot\|_{0} is the Euclidean norm.

If for a given probability measure QQ on an embedded compact metric space (,ρj),(\mathcal{M},\rho_{j}), the Fréchet function associated with the random object X,X, with Q=PXQ=P_{X} is

(1.1) (p)=𝔼(ρj2(p,X))=j(x)j(p)02Q(dx),\mathcal{F}(p)=\mathbb{E}(\rho_{j}^{2}(p,X))=\int_{\mathcal{M}}\|j(x)-j(p)\|^{2}_{0}Q(dx),

and its minimizers form the extrinsic mean set of XX (see Bhattacharya and Patrangenaru 2003 [10]). If there is a unique point in the extrinsic mean set, this point is the extrinsic mean, and is labeled μj,E(Q)\mu_{j,E}(Q) . The maximizers in (1.1) form the extrinsic antimean set. In there is a unique point in the extrinsic antimean set, this point is the extrisic antimean, and is labeled αμj,E(Q)\alpha\mu_{j,E}(Q), or simply αμE,\alpha\mu_{E}, when jj is known (See Patrangenaru and Guo 2016[12]). Also, given X1,,XnX_{1},\dots,X_{n} i.i.d.r.o.’s from QQ, their extrinsic sample mean X¯E,\bar{X}_{E}, is the extrinsic mean of the empirical distribution Q^n=1ni=1nδXi\hat{Q}_{n}={\frac{1}{n}}\sum_{i=1}^{n}\delta_{X_{i}} and their extrinsic sample antimean aX¯Ea\bar{X}_{E} is the extrinsic antimean of the empirical distribution Q^n=1ni=1nδXi,\hat{Q}_{n}={\frac{1}{n}}\sum_{i=1}^{n}\delta_{X_{i}}, assuming these sample extrinsic means or antimeans exist.

The main goal of this article is to establish some general methods for comparing extrinsic antimeans on a compact manifold embedded in an Euclidean space. In Section 2, some preliminary result, a central limit theorem for sample extrinsic antimeans on compact manifolds is brought up for future developments. Thereafter, a two sample test statistic for extrinsic antimeans on compact manifolds is given in Section 3. In Section 4, one focuses on the case of the planar Kendall shape space of k-ads, Σ2k,\Sigma_{2}^{k}, which is often represented in literature by the complex projective space Pk2\mathbb{C}P^{k-2} ( see Patrangenaru and Ellingson (2015)[6], Ch.2, p.142 ). The embedding considered here is the Veronese-Whitney embedding of Pq\mathbb{C}P^{q} in the space of self adjoint (q+1)×(q+1)(q+1)\times(q+1) complex matrices (see Patrangenaru and Ellingson (2015)[6], Ch.2,p. 154). In Section 5, we derive a nonparametric two sample test for the equality of the population Veronese-Whitey (VW) antimeans on the complex projective space Pk2\mathbb{C}P^{k-2}. Finally, Section 6 illustrate a two example of Kendall shape data from Bookstein ((1997)[24]). The first example is on Apert syndrome vs clinically normal children, known as the University School data. The second example is on comparing brain scan shapes in schizophrenic vs clinically normal children.

2 Extrinsic mean and antimean of a random object

As discussed in the introduction, when jj is an embedding of a compact object space into a numerical space, the set of minimizers of the Fréchet function is the extrinsic mean set, and when this set has a unique point, this point is called the extrinsic mean. The maximizers of this function form the extrinsic antimean set, when this later set has one point only, this point is called the extrinsic antimean. We have following properties of the extrinsic antimean from Patrangenaru, Yao and Guo [12]

DEFINITION 2.1.

A point yRNy\in R^{N}for which there is a unique point pp\in\mathcal{M} satisfying the equality,

(2.1) supxyj(x)0=yj(p)0\sup_{x\in\mathcal{M}}||y-j(x)||_{0}={\|y-j(p)\|}_{0}

is called αj\alpha j-nonfocal. A point which is not αj\alpha j-nonfocal is said to be αj\alpha j-focal. If y is an αj\alpha j-nonfocal point, its farthest projection on j()j(\mathcal{M}) is the unique point z=PF,j(y)j()z=P_{F,j}(y)\in j(\mathcal{M}) with supxyj(x)0=d0(y,j(p))\sup_{x\in\mathcal{M}}||y-j(x)||_{0}=\,d_{0}(y,j(p)).

DEFINITION 2.2.

A point yRNy\in R^{N}for which there is a unique point pp\in\mathcal{M} satisfying the equality,

(2.2) infxyj(x)0=yj(p)0\inf_{x\in\mathcal{M}}||y-j(x)||_{0}={\|y-j(p)\|}_{0}

is called jj-nonfocal. A point which is not jj-nonfocal is said to be jj-focal. If y is an jj-nonfocal point, its projection on j()j(\mathcal{M}) is the unique point z=Pj(y)j()z=P_{j}(y)\in j(\mathcal{M}) with infxyj(x)0=d0(y,j(p))\inf_{x\in\mathcal{M}}||y-j(x)||_{0}=\,d_{0}(y,j(p)).

DEFINITION 2.3.

A probability distribution QQ on \mathcal{M} is said to be αj\alpha j-nonfocal if the mean μ\mu of j(Q)j(Q) is αj\alpha j-nonfocal.

A probability distribution QQ on \mathcal{M} is said to be jj-nonfocal if the mean μ\mu of j(Q)j(Q) is jj-nonfocal.

Then we have the following theorem from Patrangenaru, Guo and Yao (2016)[21], which in particular is valid for a probability measure on an embedded compact object space with the chord distance ρj:\rho_{j}:

THEOREM 2.1.

Let Q=PXQ=P_{X} be a probability measure associated with the random object XX on a compact metric space (M,ρ).(M,\rho). So we have F(p)=E(ρ2(p,X))F(p)=E(\rho^{2}(p,X)) is finite on M.M. (a) Then, given any ε>0\varepsilon>0, there exist a PP-null set NN and n(ω)<n(\omega)<\infty ωNc\forall\,\omega\in N^{c} such that the Fréchet (sample) antimean set of Q^n=Q^n,ω\hat{Q}_{n}=\hat{Q}_{n,\omega} is contained in the ε\varepsilon-neighborhood of the Fréchet antimean set of QQ for all nn(ω)n\geq n(\omega). (b) If the Fréchet antimean of QQ exists then every measurable choice from the Fréchet (sample) antimean set of Q^n\hat{Q}_{n} is a strongly consistent estimator of the Fréchet antimean of QQ.

2.1 Previous Asymptotic Results for Extrinsic Sample Antimeans

In preparation, we are using the large sample distribution for extrinsic sample antimeans given in Patrangenaru et al (2016 [12]).

Assume jj is an embedding of a dd-dimensional manifold \mathcal{M} such that j()j(\mathcal{M}) is closed in N\mathbb{R}^{N}, and Q=PXQ=P_{X} is a αj\alpha j-nonfocal probability measure on \mathcal{M} such that j(Q)j(Q) has finite moments of order 2. Let μ\mu and Σ\Sigma be the mean and covariance matrix of j(Q)j(Q) regarded as a probability measure on N\mathbb{R}^{N}. Let \mathcal{F} be the set of αj\alpha j-focal points of j()j(\mathcal{M}), and let PF,j:cj()P_{F,j}:{\mathcal{F}}^{c}\to j(\mathcal{M}) be the farthest projection on j()j(\mathcal{M}). PF,jP_{F,j} is differentiable at μ\mu and has the differentiability class of j()j(\mathcal{M}) around any αj\alpha j nonfocal point. In order to evaluate the differential dμPF,jd_{\mu}P_{F,j} we consider a special orthonormal frame field that will ease the computations.

A local frame field p(e1(p),,ek(p))p\to(e_{1}(p),\dots,e_{k}(p)), defined on an open neighborhood UNU\subseteq\mathbb{R}^{N} is adapted to the embedding jj if it is an orhonormal frame field and xj1(U),er(j(x))=dxj(fr(x)),r{1,,d},\forall x\in j^{-1}(U),e_{r}(j(x))=d_{x}j(f_{r}(x)),r\in\{1,\ldots,d\}, where (f1,,fd)(f_{1},\dots,f_{d}) is a local frame field on ,\mathcal{M}, and fr(x)f_{r}(x) is the value of the local vector field frf_{r} at x.x.

Let e1,,eNe_{1},\dots,e_{N} be the canonical basis of N\mathbb{R}^{N} and assume (e1(p),,eN(p))(e_{1}(p),\dots,e_{N}(p)) is an adapted frame field around PF,j(μ)=j(αμE)P_{F,j}(\mu)=j(\alpha\mu_{E}). Then dμPF,j(eb)TPF,j(μ)j()d_{\mu}P_{F,j}(e_{b})\in T_{P_{F,j}(\mu)}j(\mathcal{M}) is a linear combination of e1(PF,j(μ)),,ed(PF,j(μ))e_{1}(P_{F,j}(\mu)),\dots,e_{d}(P_{F,j}(\mu)):

(2.3) dμPF,j(eb)=a=1d(dμPF,j(eb))ea(PF,j(μ))ea(PF,j(μ))d_{\mu}P_{F,j}(e_{b})=\sum^{d}_{a=1}(d_{\mu}P_{F,j}(e_{b}))\cdot e_{a}(P_{F,j}(\mu))e_{a}(P_{F,j}(\mu))

where dμPF,jd_{\mu}P_{F,j} is the differential of PF,jP_{F,j} at μ.\mu. By the delta method, n1/2(PF,j(j(X)¯)PF,j(μ))n^{1/2}(P_{F,j}(\overline{j(X)})-P_{F,j}(\mu)) converges weakly to NN(0N,αΣμ)N_{N}(0_{N},\alpha\Sigma_{\mu}), where j(X)¯=1ni=1nj(Xi)\overline{j(X)}=\frac{1}{n}\sum^{n}_{i=1}j(X_{i}) and

(2.4) αΣμ=[a=1ddμPF,j(eb)ea(PF,j(μ))ea(PF,j(μ))]b{1,,N}\displaystyle\alpha\Sigma_{\mu}=[\sum^{d}_{a=1}d_{\mu}P_{F,j}(e_{b})\cdot e_{a}(P_{F,j}(\mu))e_{a}(P_{F,j}(\mu))]_{b\in\{1,\ldots,N\}}
×Σ[a=1ddμPF,j(eb)ea(PF,j(μ))ea(PF,j(μ))]b{1,,N}T.\displaystyle\times\Sigma[\sum^{d}_{a=1}d_{\mu}P_{F,j}(e_{b})\cdot e_{a}(P_{F,j}(\mu))e_{a}(P_{F,j}(\mu))]^{T}_{b\in\{1,\dots,N\}}{.}

Here Σ\Sigma is the covariance matrix of j(X1)j(X_{1}) w.r.t the canonical basis e1,,eN.e_{1},\dots,e_{N}.

The asymptotic distribution NN(0N,αΣμ)N_{N}(0_{N},\alpha\Sigma_{\mu}) is degenerate and the support of this distribution is on TPF,jj()T_{P_{F,j}}j(\mathcal{M}), since the range of dμPF,jd_{\mu}P_{F,j} is TPF,j(μ)j()T_{P_{F,j(\mu)}}j(\mathcal{M}). Note that dμPF,j(eb)ea(PF,j(μ))=0d_{\mu}P_{F,j}(e_{b})\cdot e_{a}(P_{F,j}(\mu))=0 for a{d+1,,N}a\in\{d+1,\dots,N\}.

The tangential component tan(v)tan(v) of vNv\in\mathbb{R}^{N}, w.r.t the basis ea(PF,j(μ))TPF,j(μ)j(),a{1,,d}e_{a}(P_{F,j}(\mu))\in T_{P_{F,j(\mu)}}j(\mathcal{M}),a\in\{1,\dots,d\} is given by

(2.5) tan(v)=[e1(PF,j(μ))Tv,,ed(PF,j(μ))Tv]T.tan(v)=[e_{1}{(P_{F,j}(\mu))}^{T}v,\dots,e_{d}{(P_{F,j}(\mu))}^{T}v]^{T}{.}

Then, the random vector (dαμEj)1(tan(PF,j((j(X))¯)PF,j(μ)))=a=1dX¯jafa{(d_{\alpha\mu_{E}}j)}^{-1}(tan(P_{F,j}(\overline{(j(X))})-P_{F,j}(\mu)))=\sum^{d}_{a=1}{\overline{X}}^{a}_{j}f_{a} has the following anticovariance matrix w.r.t the basis f1(αμE),,fd(αμE)f_{1}(\alpha\mu_{E}),\dots,f_{d}(\alpha\mu_{E}):

(2.6) αΣj,E=ea(PF,j(μ))TαΣμeb(PF,j(μ))1a,bd\displaystyle\alpha\Sigma_{j,E}=e_{a}{(P_{F,j}(\mu))}^{T}\alpha\Sigma_{\mu}e_{b}{(P_{F,j}(\mu))}_{1\leq a,b\leq d}
=[dμPF,j(eb)ea(PF,j(μ))]a{1,,d}Σ[dμPF,j(eb)ea(PF,j(μ))]a{1,,d}T,\displaystyle=[d_{\mu}P_{F,j}(e_{b})\cdot e_{a}(P_{F,j}(\mu))]_{a\in\{1,\dots,d\}}~{}\Sigma~{}[d_{\mu}P_{F,j}(e_{b})\cdot e_{a}(P_{F,j}(\mu))]^{T}_{a\in\{1,\dots,d\}}{,}

which is the anticovariance matrix of the random object XX. To simplify the notation, we set

(2.7) B=[dμPF,j(eb)ea(PF,j(μ))]a{1,,d}B=[d_{\mu}P_{F,j}(e_{b})\cdot e_{a}(P_{F,j}(\mu))]_{a\in\{1,\dots,d\}}

Similarly, given i.i.d.r.o.’s X1,,XnX_{1},\cdots,X_{n} from QQ, we define the sample anticovariance matrix aSj,E,naS_{j,E,n} as the anticovariance matrix associated with the empirical distribution Q^n.\hat{Q}_{n}.

If, in addition, rank αΣμ=d\alpha\Sigma_{\mu}=d, then αΣj,E\alpha\Sigma_{j,E} is invertible and if we define the jj-standardized antimean vector

(2.8) Z¯j,n=:n12αΣj,E12(X¯1j,,X¯dj)T,\overline{Z}_{j,n}=:n^{1\over 2}{\alpha\Sigma_{j,E}}^{-\frac{1}{2}}({\bar{X}^{1}}_{j},\cdots,{\bar{X}^{d}}_{j})^{T},

using basic large sample theory results, including a generalized Slutsky’s lemma ( see Patrangenaru and Ellingson(2015)[6], p.65), one has:

THEOREM 2.2.

Assume {Xr}r=1,,n\{X_{r}\}_{r=1,...,n} is a random sample from the αj\alpha j-nonfocal distribution QQ, and let μ=E(j(X1))\mu=E(j(X_{1})). Let (e1(p),e2(p),.,eN(p))(e_{1}(p),e_{2}(p),....,e_{N}(p)) be an orthonormal frame field adapted to jj. Then (a) the tangential component at the extrinsic antimean αμE\alpha\mu_{E} of dαμEj1tanPF,j(μ)(PF,j((j(X))¯)PF,j(μ)){d_{\alpha\mu_{E}}j^{-1}}tan_{P_{F,j}(\mu)}(P_{F,j}(\overline{(j(X))})-P_{F,j}(\mu)) has asymptotically a multivariate normal distribution in the tangent space to MM at αμE\alpha\mu_{E} with mean 0d0_{d} and anticovariance matrix n1αΣj,En^{-1}\alpha\Sigma_{j,E} , (b) if αΣj,E\alpha\Sigma_{j,E} is nonsingular, the j-standardized antimean vector Z¯j,n=αΣj,E12tanPF,j(μ)(PF,j((j(X))¯)PF,j(μ))\overline{Z}_{j,n}=\alpha\Sigma_{j,E}^{-\frac{1}{2}}tan_{P_{F,j}(\mu)}(P_{F,j}(\overline{(j(X))})-P_{F,j}(\mu)) converges weakly to a random vector with a Nd(0d,Id)N_{d}(0_{d},I_{d}) distribution, and (c) under the assumptions of (b)

(2.9) (aSj,E,n)12tanPF,j(μ)(PF,j((j(X))¯)PF,j(μ))2dχd2.\|(aS_{j,E,n})^{-\frac{1}{2}}tan_{P_{F,j}(\mu)}(P_{F,j}(\overline{(j(X))})-P_{F,j}(\mu))\|^{2}\to_{d}\chi^{2}_{d}.

3 A two sample test for extrinsic antimeans

We now turn to two-sample tests for extrinsic antimeans of distributions on an arbitrary mm dimensional compact manifold M.M. For a large sample test for equality of two extrinsic means see Bhattacharya and Bhattacharya (2012) [14], p.42. Let Xaka:ka=1,,na,a=1,2X_{ak_{a}}:k_{a}=1,\dots,n_{a},a=1,2 be two independent random samples drawn from distributions Qa,a=1,2Q_{a},a=1,2 on M,M, and let jj be an embedding of MM into N.\mathbb{R}^{N}. Denote by μa\mu_{a} the mean of the induced probability Qaj1Q_{a}\circ j^{-1} and αΣa,j\alpha\Sigma_{a,j} its anticovariance matrix (a=1,2).(a=1,2). Then the extrinsic antimean of QaQ_{a} is αμa,j=j1(PF,j(μa))\alpha\mu_{a,j}=j^{-1}(P_{F,j}(\mu_{a})), assuming QaQ_{a} is α\alpha-nonfocal. Write Yaka=j(Xaka)ka=1,,na,a=1,2Y_{ak_{a}}=j(X_{ak_{a}})k_{a}=1,\dots,n_{a},a=1,2 and let aY¯a,a=1,2a\bar{Y}_{a},a=1,2 be the corresponding sample antimeans. By Theorem 2.2

(3.1) naBa[PF,j(Y¯a)PF,j(μa)]d𝒩m(0,αΣa,j),a=1,2,\sqrt{n_{a}}B_{a}[P_{F,j}(\bar{Y}_{a})-P_{F,j}(\mu_{a})]\to_{d}\mathcal{N}_{m}(0,\alpha\Sigma_{a,j}),a=1,2,

where αΣa,j\alpha\Sigma_{a,j} is the extrinsic anticovariance matrix of Qa,Q_{a}, and BaB_{a} are the same as in Theorem 2.2, and (2.7), with QQ replaced by QaQ_{a} (a=1,2). That is, Ba(y)B_{a}(y) is the m×Nm\times N matrix of an orthonormal basis (frame) of Ty(j(M))Ty(N)=NT_{y}(j(M))\subset T_{y}(\mathbb{R}^{N})=\mathbb{R}^{N} for yy in a neighborhood of PF,j(μa),P_{F,j}(\mu_{a}), and Ba=Ba(PF,j(μa)).B_{a}=B_{a}(P_{F,j}(\mu_{a})). Similarly, CaΣμa,a=(DPF,j)μaΣa(DPF,j)μaT,(a=1,2).C_{a}\doteq\Sigma_{\mu_{a},a}=(DP_{F,j})_{\mu_{a}}\Sigma_{a}(DP_{F,j})_{\mu_{a}}^{T},(a=1,2). The null hypothesis H0:αμ1,j=αμ2,j,H_{0}:\alpha\mu_{1,j}=\alpha\mu_{2,j}, say, is equivalent to H0:PF,j(μ1)=PF,j(μ2)=π,H_{0}:P_{F,j}(\mu_{1})=P_{F,j}(\mu_{2})=\pi, say. Then, under the null hypothesis, letting B=B(π),B=B(\pi), one has B1=B2=B,B_{1}=B_{2}=B, and

[B(1n1C1+1n2C2)BT]1/2B[PF,j(Y¯1)PF,j(Y¯2)]d𝒩m(0m,Im),\displaystyle[B(\frac{1}{n_{1}}C_{1}+\frac{1}{n_{2}}C_{2})B^{T}]^{-1/2}B[P_{F,j}(\bar{Y}_{1})-P_{F,j}(\bar{Y}_{2})]\to_{d}\mathcal{N}_{m}(0_{m},I_{m}),
(3.2) asn=n1+n2,andn1nλ(0,1).\displaystyle\text{as}\ n=n_{1}+n_{2}\to\infty,\text{and}\frac{n_{1}}{n}\to\lambda\in(0,1).

For statistical inference one estimates CaC_{a} by

(3.3) C^a=(DPF,j)Y¯aΣ^a(DPF,j)Y¯aT\hat{C}_{a}=(DP_{F,j})_{\bar{Y}_{a}}\hat{\Sigma}_{a}(DP_{F,j})_{\bar{Y}_{a}}^{T}

where Σ^a\hat{\Sigma}_{a} is the sample covariance matrix of sample a(a=1,2).a(a=1,2). Also BB is replaced by B^=B(π^)\hat{B}=B(\hat{\pi}) where π^\hat{\pi} is a sample estimate of π.\pi. Under H0,H_{0}, both PF,j(Y¯1)P_{F,j}(\bar{Y}_{1}) and PF,j(Y¯2)P_{F,j}(\bar{Y}_{2}) are consistent estimates of π,\pi, so we take a “pooled estimate”

(3.4) π^=PF,j(1n1+n2(n1PF,j(Y¯1)+n2PF,j(Y¯2))).\hat{\pi}=P_{F,j}(\frac{1}{n_{1}+n_{2}}(n_{1}P_{F,j}(\bar{Y}_{1})+n_{2}P_{F,j}(\bar{Y}_{2}))).

We, therefore, have the following result:

THEOREM 3.1.

Assume the extrinsic sample anticovariance matrix α^Σa,j\hat{\alpha}\Sigma_{a,j} is nonsingular for a=1,2.a=1,2. Then, under H0:αμ1,j=αμ2,j,H_{0}:\alpha\mu_{1,j}=\alpha\mu_{2,j}, one has:

(B^[PF,j(Y¯1)PF,j(Y¯2)])T[B^(1n1C^1+1n2C^2)B^T]1(B^[PF,j(Y¯1)PF,j(Y¯2)])\displaystyle(\hat{B}[P_{F,j}(\bar{Y}_{1})-P_{F,j}(\bar{Y}_{2})])^{T}[\hat{B}(\frac{1}{n_{1}}\hat{C}_{1}+\frac{1}{n_{2}}\hat{C}_{2})\hat{B}^{T}]^{-1}(\hat{B}[P_{F,j}(\bar{Y}_{1})-P_{F,j}(\bar{Y}_{2})])
(3.5) dχm2,\displaystyle\to_{d}\chi_{m}^{2},
asn=n1+n2,andn1nλ(0,1).\displaystyle\text{as}\ n=n_{1}+n_{2}\to\infty,\text{and}\frac{n_{1}}{n}\to\lambda\in(0,1).

4 VW Means and VW Antimeans

4.1 VW Means and VW Antimeans on Pm\mathbb{C}P^{m}

We consider the case of a probability measure QQ on the complex projective space =Pq\mathcal{M}=\mathbb{C}P^{q}. If we consider the action of the multiplicative group =\{0}\mathbb{C}^{*}=\mathbb{C}\backslash\{0\} on q+1\{0}\mathbb{C}^{q+1}\backslash\{0\}, given by scalar multiplication

(4.1) α(λ,z)=λz\alpha(\lambda,z)=\lambda z

the quotient space is the qq-dimensional complex projective space Pq\mathbb{C}P^{q}, set of all complex lines in q+1\mathbb{C}^{q+1} going through 0q+10\in\mathbb{C}^{q+1}. One can show that Pq\mathbb{C}P^{q} is a 2qq dimensional real analytic manifold, using transition maps, similar to those in the case of Pq\mathbb{R}P^{q}  (see Patrangenaru and Ellingson (2015) [20], chapter 3).

Here we are concerned with a landmark based non-parametric analysis of similarity shape data (see Bhattacharya and Patrangenaru (2014)[23]). Our analysis is for extrinsic antimeans. For landmark based shape data, one considers a kk-ad x=(x1,,xk)(m)kx=(x^{1},\dots,x^{k})\in(\mathbb{R}^{m})^{k}, which consists of kk labeled points in m\mathbb{R}^{m} that represent coordinates of k-labeled landmarks.

In this subsection 𝒢\mathcal{G} is the group direct similarities of m\mathbb{R}^{m}. A similarity is a function f:mmf:\mathbb{R}^{m}\to\mathbb{R}^{m}, that uniformly scale the Euclidean distances, that is, for which there is K>0K>0, such that f(x)f(y)=Kxy,x,ym\lVert f(x)-f(y)\rVert=K\lVert x-y\rVert,\forall x,y\in\mathbb{R}^{m}. Using the fundamental theorem of Euclidean geometry, one can show that a similarity is given by f(x)=Ax+b,ATA=cIm,c>0f(x)=Ax+b,A^{T}A=cI_{m},c>0. A direct similarity is a similarity of this form, where AA has a positive determinant. Under composition direct similarities from a group. The object space considered here consist in orbits of the group action of 𝒢\mathcal{G} on kk-ads, and is called direct similarities shape space of k-ads in m\mathbb{R}^{m}.

For m=2m=2, or m=3m=3, a direct similarity (Kendall) shape is that the geometrical information that remains when location, scale and rotational effects are filtered out from a kk-ads. Two kk-ads (z1,z2,,zk)(z_{1},z_{2},\dots,z_{k}) and (z1,z2,,zk)(z^{{}^{\prime}}_{1},z^{{}^{\prime}}_{2},\dots,z^{{}^{\prime}}_{k}) are said to have the same shape if there is a direct similarity TT in the plane, that is, a composition of a rotation, a translation and a homothety such that T(zj)=zjT(z_{j})=z^{{}^{\prime}}_{j} for j=1,,kj=1,\dots,k. Having the same Kendall shape is an equivalence relationship in the space of planar kk-ads, and the set of all equivalence classes of nontrivial kk-ads is called the Kendall planar shape space of kk-ads, which is denoted Σ2k\Sigma_{2}^{k} (see Balan and Patrangenaru(2005) [15]).

4.2 VW Antimeans on Pm\mathbb{C}P^{m}

Kendall (1984) [2] introduced that planar direct similarity shapes of k-ads; which is a set of k labeled points at least two of which are distinct, can be represented as points on a complex projective space Pk2\mathbb{C}P^{k-2}. A standard shape analysis method was also introduced by Kent(1992)[4] and is called Veronese-Whitney(VW) embedding of Pk2\mathbb{C}P^{k-2} in the space of (k1)×(k1)(k-1)\times(k-1) self adjoint complex matrices, to represent shape data in an Euclidean space. This Veronese–Whitney embedding j:Pk2S(k1,)j:\mathbb{C}P^{k-2}\to S(k-1,\mathbb{C}), where S(k1,)S(k-1,\mathbb{C}) is the space of (k1)×(k1)(k-1)\times(k-1) Hermitian matrices, is given by

(4.2) j([𝐳])=𝐳𝐳,𝐳𝐳=1.j([{\bf z}])={\bf zz^{*},z^{*}z}=1.

This embedding is a SU(k1)SU(k-1) equivariant embedding and SU(k1)SU(k-1) is called the special unitary group (k1)×(k1)(k-1)\times(k-1) matrices of determinant 1, since j([Az])=Aj([z])Aj([Az])=Aj([z])A^{*}, ASU(k1)\forall A\in SU(k-1). The corresponding extrinsic mean (set) of a random shape X on Pk2\mathbb{C}P^{k-2} is called the VW mean (set) (see Patrangenaru and Ellingson (2015)[6], Ch.3), and the VW mean,when it exists,and is labeled μVW(X),μVW\mu_{VW}(X),\mu_{VW} or simply μE\mu_{E}.The corresponding extrinsic antimean (set) of a random shape X,is called the VW antimean (set) and is labeled αμVW(X),αμVW\alpha\mu_{VW}(X),\alpha\mu_{VW} or αμE\alpha\mu_{E}.

For the VW-embedding of the complex projective space, we have the following theorem for sample VW-means from Bhattacharya and Patrangenaru(2003) [10]:

THEOREM 4.1.

Let QQ be a probability distribution on Pk2\mathbb{C}P^{k-2} and let {[Zr],Zr=1,r=1,,n}\{[Z_{r}],\parallel Z_{r}\parallel=1,{r=1,\dots,n}\} be a i.i.d.r.o.’s from QQ. (a)(a) QQ is nonfocal iff λ\lambda the largest eigenvalue of E[Z1Z1]E[Z_{1}Z_{1}^{*}] is simple and in this case μEQ=[m],\mu_{E}{Q}=[m], where mm is an eigenvector of E[Z1Z1]E[Z_{1}Z_{1}^{*}] corresponding to λ\lambda, with m=1\parallel m\parallel=1. (b)(b) The extrinsic sample mean X¯E=[m]¯\overline{X}_{E}=\overline{[m]}, where mm is an eigenvector of norm 1 of J=1ni=1nZiZiJ=\frac{1}{n}\sum^{n}_{i=1}Z_{i}Z^{*}_{i}, Zi=1,i=1,,n\|Z_{i}\|=1,i=1,\dots,n, corresponding to the largest eigenvalue of J.

We also have a similar result for sample VW-antimeans (see Wang, Patrangenaru and Guo [17]):

THEOREM 4.2.

Let QQ be a probability distribution on Pk2\mathbb{C}P^{k-2} and let {[Zr],Zr=1r=1,,n}\{[Z_{r}],\parallel Z_{r}\parallel=1_{r=1,\dots,n}\} be a i.i.d.r.o.’s from QQ. (a)(a) QQ is α\alpha-nonfocal iff λ\lambda, the smallest eigenvalue of E[Z1Z1]E[Z_{1}Z_{1}^{*}] is simple and in this case αμj,EQ=[m],\alpha\mu_{j,E}{Q}=[m], where mm is an eigenvector of E[Z1Z1]E[Z_{1}Z_{1}^{*}] corresponding to λ\lambda, with m=1\parallel m\parallel=1. (b)(b) The extrinsic sample antimean αX¯E=[m]\alpha\overline{X}_{E}={[m]}, where mm is an eigenvector of norm 1 of J=1ni=1nZiZiJ=\frac{1}{n}\sum^{n}_{i=1}Z_{i}Z^{*}_{i}, Zi=1,i=1,,n\|Z_{i}\|=1,i=1,\dots,n, corresponding to the smallest eigenvalue of J.

5 Two sample testing problem for VW antimeans on Σ2k\Sigma_{2}^{k}

5.1 Application on the planar shape space of kk-ads

We are concerned with a landmark based nonparametric analysis of similarity shape data. For landmark based shape data , we will denote a kk-ad by kk complex numbers: zj=xj+iyj,1jk.z_{j}=x_{j}+iy_{j},~{}1\leq j\leq k. We center the kk-ad at z=1kj=1kzj\langle z\rangle={1\over k}\sum\limits_{j=1}^{k}z_{j} via a translation; next we rotate it by an angle θ\theta and scale it, operations that are achieved by multiplying zzz-\langle z\rangle by the complex number λ=reiθ.\lambda=re^{i\theta}. We can represent the shape of the kk-ad as the complex one dimensional vector subspace passing through 0 of the linear subspace Lk1={ζk,1kTζ=0}.L^{k-1}=\{\zeta\in\mathbb{C}^{k},1_{k}^{T}\zeta=0\}. Thus, the space of kk-ads is the set of all complex lines on the hyperplane, Lk1=wk\{0}:1kwj=0.L^{k-1}={w\in{\mathbb{C}}^{k}\backslash\{0\}}:~{}\sum\limits_{1}^{k}w_{j}=0. Therefore the shape space Σ2k\Sigma_{2}^{k} of nontrivial planar kk-ads can be represented as the complex projective space Pk2,\mathbb{C}P^{k-2}, the space of all complex lines through the origin in k1\mathbb{C}^{k-1} using an isomorphism of Lk1L^{k-1} and k1.\mathbb{C}^{k-1}. As in the case of Pk2\mathbb{C}P^{k-2}, it is convenient to represent the element of Σ2k\Sigma_{2}^{k} corresponding to a kk-ad zz by the curve γ(z)=[z]=eiθ((zz)/zz):0θ2π\gamma(z)=[z]={e^{i\theta}((z-\langle z\rangle)/\|z-\langle z\rangle\|):~{}0\leq\theta\leq 2\pi} on the unit sphere in Lk1k1.L^{k-1}\approx\mathbb{C}^{k-1}.

5.2 Test for VW antimeans on planar shape spaces

Let Q1Q_{1} and Q2Q_{2} be two probability measures on the shape space Σ2k,\Sigma_{2}^{k}, and let αμ1\alpha\mu_{1} and αμ2\alpha\mu_{2} denote the antimeans of Q1j1Q_{1}\circ j^{-1} and Q2j1,Q_{2}\circ j^{-1}, where jj is the VW-embedding that j([z])=uu,j([z])=uu^{*}, where u=1zz,u={1\over{\|z\|}}z, thus uu=1u^{*}u=1. Suppose [W1],,[Wn][W_{1}],\cdots,[W_{n}] and [Z1],,[Zm][Z_{1}],\cdots,[Z_{m}] are i.i.d. random objects from Q1Q_{1} and Q2Q_{2} respectively. Let Xi=j([Wi]),X_{i}=j([W_{i}]), Yj=j([Zj])Y_{j}=j([Z_{j}]) be their images in j(Pk2)j(\mathbb{C}P^{k-2}) which are random samples from Q1j1Q_{1}\circ j^{-1} and Q2j1,Q_{2}\circ j^{-1}, respectively. Suppose we are to test if the VW antimeans of Q1Q_{1} and Q2Q_{2} are equal, i.e.

H0:αμ1=αμ2H_{0}:\alpha\mu_{1}=\alpha\mu_{2}

It is known that αμ1=j1(PF(v1))\alpha\mu_{1}=j^{-1}(P_{F}(v_{1})), where v1=E(X1)v_{1}=E(X_{1}) and similarly, αμ2=j1(PF(v2))\alpha\mu_{2}=j^{-1}(P_{F}(v_{2})), where v2=E(Y1)v_{2}=E(Y_{1}). We assume that both v1v_{1} and v2v_{2} have simple smallest eigenvalues. Then under H0,H_{0}, their unit corresponding eigenvectors differ by a rotation.

Choose vS(k1,),v\in S(k-1,\mathbb{C}), with the same farthest projection as v1v_{1} and v2v_{2}. Suppose v=uΛu,v=u\Lambda u^{*}, and Λ=Diag(λ1<λ2λk1),\Lambda=Diag(\lambda_{1}<\lambda_{2}\leq\cdots\leq\lambda_{k-1}), where λa,a=1,,k1,\lambda_{a},a=1,\dots,k-1, are the eigenvalues of v,v, and u=[u1,u2,,uk1]u=[u_{1},u_{2},\cdots,u_{k-1}] are the corresponding eigenvectors. Also, we obtain an orthonormal basis for S(k1,)S(k-1,\mathbb{C}), which is given by {υba:1abk1}\{\upsilon_{b}^{a}:1\leq a\leq b\leq k-1\} and {ωba:1abk1}\{\omega_{b}^{a}:1\leq a\leq b\leq k-1\}. Where υba\upsilon_{b}^{a} has all entries zero except for those in the positions (a, b) and (b, a) that are equal to 1 and ωba\omega_{b}^{a} is a matrix with all entries zero except for those in the positions (a, b) and (b, a) that are equal to i, respectively -i.

It is defined as,

(5.1) υba={12(eaebt+ebeat),a<beaeat,a=b.\upsilon_{b}^{a}=\left\{\begin{array}[]{lr}\frac{1}{\sqrt{2}}(e_{a}e_{b}^{t}+e_{b}e_{a}^{t}),a<b\\ e_{a}e_{a}^{t},a=b.\end{array}\right.
(5.2) ωba=+i2(eaebtebeat),a<b\omega_{b}^{a}=+\frac{i}{\sqrt{2}}(e_{a}e_{b}^{t}-e_{b}e_{a}^{t}),a<b

where {ea:1ak1}\{e_{a}:1\leq a\leq k-1\} is the standard canonical basis for N\mathbb{C}^{N}. For any uSU(k1)(uu=uu=1,det(u)=+1)u\in SU(k-1)(uu^{*}=u^{*}u=1,det(u)=+1) where SU is special unitary group of all (k1)×(k1)(k-1)\times(k-1) complex matrices. {uυbau:1abk1}\{u\upsilon_{b}^{a}u^{*}:1\leq a\leq b\leq k-1\} and {uωbau:1abk1}\{u\omega_{b}^{a}u^{*}:1\leq a\leq b\leq k-1\} are an orthonormal frame for S(k1,)S(k-1,\mathbb{C}). Now, we will choose the orthonormal basis frame {uυbau,uωbau}\{u\upsilon_{b}^{a}u^{*},u\omega_{b}^{a}u^{*}\} for S(k1,)S(k-1,\mathbb{C}). Then it can be shown that

(5.3) dvPF(uυbau)={0,if1ab<k1,a=b=k1,(λk1λa)1uυbau,if1a<k,b=k1.d_{v}P_{F}(u\upsilon_{b}^{a}u^{*})=\left\{\begin{array}[]{lr}0,&\text{if}~{}1\leq a\leq b<k-1,a=b=k-1,\\ (\lambda_{k-1}-\lambda_{a})^{-1}u\upsilon_{b}^{a}u^{*},&\text{if}~{}1\leq a<k,b=k-1.\end{array}\right.
(5.4) dvPF(uωbau)={0,if1ab<k1,a=b=k1,(λk1λa)1uωbau,if1a<k,b=k1.d_{v}P_{F}(u\omega_{b}^{a}u^{*})=\left\{\begin{array}[]{lr}0,&\text{if}~{}1\leq a\leq b<k-1,a=b=k-1,\\ (\lambda_{k-1}-\lambda_{a})^{-1}u\omega_{b}^{a}u^{*},&\text{if}~{}1\leq a<k,b=k-1.\end{array}\right.

Then, we write

(5.5) X¯v=1ab<k1(X¯v),uυbauuυbau+1ab<k1(X¯v),uωbauuωbau\bar{X}-v=\mathop{\sum\sum}_{1\leq a\leq b<k-1}\langle\,(\bar{X}-v),u\upsilon_{b}^{a}u^{*}\rangle u\upsilon_{b}^{a}u^{*}+\mathop{\sum\sum}_{1\leq a\leq b<k-1}\langle\,(\bar{X}-v),u\omega_{b}^{a}u^{*}\rangle u\omega_{b}^{a}u^{*}

Then from equations (5.3) ,(5.4) and (5.5), it follows that

dvPF(X¯v)\displaystyle d_{v}P_{F}(\bar{X}-v) =a=2k12Re(uaX¯u1)(λaλ1)1uυa1u\displaystyle=\sum\limits_{a=2}^{k-1}\sqrt{2}Re(u_{a}^{*}\bar{X}u_{1}){(\lambda_{a}-\lambda_{1})}^{-1}u\upsilon_{a}^{1}u^{*}
+a=2k12Im(uaX¯u1)(λaλ1)1uωa1u\displaystyle\quad+\sum\limits_{a=2}^{k-1}\sqrt{2}Im(u_{a}^{*}\bar{X}u_{1}){(\lambda_{a}-\lambda_{1})}^{-1}u\omega_{a}^{1}u^{*}
=a=2k1(λaλ1)1(uaX¯u1)uau1\displaystyle=\sum\limits_{a=2}^{k-1}{(\lambda_{a}-\lambda_{1})}^{-1}(u_{a}^{*}\bar{X}u_{1})u_{a}u_{1}^{*}
+a=2k1(λaλ1)1(u1X¯ua)u1ua\displaystyle\quad+\sum\limits_{a=2}^{k-1}{(\lambda_{a}-\lambda_{1})}^{-1}(u_{1}^{*}\bar{X}u_{a})u_{1}u_{a}^{*}
dvPF(Y¯v)\displaystyle d_{v}P_{F}(\bar{Y}-v) =a=2k12Re(uaY¯u1)(λaλ1)1uυa1u\displaystyle=\sum\limits_{a=2}^{k-1}\sqrt{2}Re(u_{a}^{*}\bar{Y}u_{1}){(\lambda_{a}-\lambda_{1})}^{-1}u\upsilon_{a}^{1}u^{*}
+a=2k12Im(uaY¯u1)(λaλ1)1uωa1u\displaystyle\quad+\sum\limits_{a=2}^{k-1}\sqrt{2}Im(u_{a}^{*}\bar{Y}u_{1}){(\lambda_{a}-\lambda_{1})}^{-1}u\omega_{a}^{1}u^{*}
=a=2k1(λaλ1)1(uaY¯u1)uau1\displaystyle=\sum\limits_{a=2}^{k-1}{(\lambda_{a}-\lambda_{1})}^{-1}(u_{a}^{*}\bar{Y}u_{1})u_{a}u_{1}^{*}
+a=2k1(λaλ1)1(u1Y¯ua)u1ua\displaystyle\quad+\sum\limits_{a=2}^{k-1}{(\lambda_{a}-\lambda_{1})}^{-1}(u_{1}^{*}\bar{Y}u_{a})u_{1}u_{a}^{*}

Define

(5.6) T(αμ)ij={Re(ui+1Xju1),if1ik2,1jn,Im(uik+3Xju1),ifk1i2k4,1jn.{T(\alpha\mu)}_{ij}=\left\{\begin{array}[]{lr}Re(u_{i+1}^{*}X_{j}u_{1}),&\text{if}~{}1\leq i\leq k-2,~{}1\leq j\leq n,\\ Im(u_{i-k+3}^{*}X_{j}u_{1}),&\text{if}~{}k-1\leq i\leq 2k-4,~{}1\leq j\leq n.\end{array}\right.
(5.7) S(αμ)ij={Re(ui+1Yju1),if1ik2,1jm,Im(uik+3Yju1),ifk1i2k4,1jm.{S(\alpha\mu)}_{ij}=\left\{\begin{array}[]{lr}Re(u_{i+1}^{*}Y_{j}u_{1}),&\text{if}~{}1\leq i\leq k-2,~{}1\leq j\leq m,\\ Im(u_{i-k+3}^{*}Y_{j}u_{1}),&\text{if}~{}k-1\leq i\leq 2k-4,~{}1\leq j\leq m.\end{array}\right.

Then we have,

(5.8) T¯(αμ)=1nj=1nT(αμ),S¯(αμ)=1mj=1mS(αμ)\bar{T}(\alpha\mu)=\frac{1}{n}\sum\limits_{j=1}^{n}T(\alpha\mu),~{}\bar{S}(\alpha\mu)=\frac{1}{m}\sum\limits_{j=1}^{m}S(\alpha\mu)

Under H0H_{0}, T¯(αμ)\bar{T}(\alpha\mu) and S¯(αμ)\bar{S}(\alpha\mu) have mean zero, and as n,mn,m\to\infty, suppose (n/(m+n))p,(n/(m+n))\to p, (m/(m+n))q,(m/(m+n))\to q, for some p,q>0;p,q>0; p+q=1.p+q=1. It follows that

(5.9) nT¯(αμ)N(0,Σ1(αμ)),mS¯(αμ)N(0,Σ2(αμ))\sqrt{n}\bar{T}(\alpha\mu)\overset{\mathcal{L}}{\longrightarrow}N(0,\Sigma_{1}(\alpha\mu)),~{}\sqrt{m}\bar{S}(\alpha\mu)\overset{\mathcal{L}}{\longrightarrow}N(0,\Sigma_{2}(\alpha\mu))

where Σ1(αμ)\Sigma_{1}(\alpha\mu) and Σ2(αμ)\Sigma_{2}(\alpha\mu) are the covariances of T(αμ).1{T(\alpha\mu)}_{.1} and S(αμ).1,{S(\alpha\mu)}_{.1}, respectively. Then

(5.10) (n+m)(T¯(αμ)S¯(αμ))N2k4(0,1pΣ1(αμ)+1qΣ2(αμ))\sqrt{(n+m)}(\bar{T}(\alpha\mu)-\bar{S}(\alpha\mu))\overset{\mathcal{L}}{\longrightarrow}N_{2k-4}(0,\frac{1}{p}\Sigma_{1}(\alpha\mu)+\frac{1}{q}\Sigma_{2}(\alpha\mu))

Thus assuming Σ1(αμ),\Sigma_{1}(\alpha\mu), Σ2(αμ)\Sigma_{2}(\alpha\mu) and 1pΣ1(αμ)+1qΣ2(αμ)\frac{1}{p}\Sigma_{1}(\alpha\mu)+\frac{1}{q}\Sigma_{2}(\alpha\mu) to be nonsingular,

(5.11) (n+m)(T¯(αμ)S¯(αμ))(1pΣ1(αμ)+1qΣ2(αμ))1(T¯(αμ)S¯(αμ))χ2k42(n+m){(\bar{T}(\alpha\mu)-\bar{S}(\alpha\mu))}^{\prime}{(\frac{1}{p}\Sigma_{1}(\alpha\mu)+\frac{1}{q}\Sigma_{2}(\alpha\mu))}^{-1}(\bar{T}(\alpha\mu)-\bar{S}(\alpha\mu))\overset{\mathcal{L}}{\longrightarrow}\chi_{2k-4}^{2}

We may take j(αμ)=pj(αμ1)+qj(αμ2).j(\alpha\mu)=pj(\alpha\mu_{1})+qj(\alpha\mu_{2}). Since αμ1\alpha\mu_{1} and αμ2\alpha\mu_{2} are unknown, we may estimate αμ\alpha\mu by the pooled sample mean j(αμ^)=(nX¯+mY¯)/(m+n)j(\hat{\alpha\mu})=(n\bar{X}+m\bar{Y})/(m+n). Note that Σ^1(αμ^)\hat{\Sigma}_{1}(\hat{\alpha\mu}) and Σ^2(αμ^)\hat{\Sigma}_{2}(\hat{\alpha\mu}) are consistent estimator of Σ1(αμ)\Sigma_{1}(\alpha\mu) and Σ2(αμ)\Sigma_{2}(\alpha\mu). Thus, we may use Σ^1(αμ^)\hat{\Sigma}_{1}(\hat{\alpha\mu}) and Σ^2(αμ^)\hat{\Sigma}_{2}(\hat{\alpha\mu}) to obtain a two sample test statistic, where

(5.12) Σ^1(αμ^)=1nT(αμ)T(αμ)T¯(αμ)T¯(αμ)\hat{\Sigma}_{1}(\hat{\alpha\mu})=\frac{1}{n}T(\alpha\mu){T(\alpha\mu)}^{{}^{\prime}}-\bar{T}(\alpha\mu){\bar{T}(\alpha\mu)}^{{}^{\prime}}
(5.13) Σ^2(αμ^)=1mS(αμ)S(αμ)S¯(αμ)S¯(αμ)\hat{\Sigma}_{2}(\hat{\alpha\mu})=\frac{1}{m}S(\alpha\mu){S(\alpha\mu)}^{{}^{\prime}}-\bar{S}(\alpha\mu){\bar{S}(\alpha\mu)}^{{}^{\prime}}

Then the two sample test statistic can be estimated by

(5.14) aTnm=(T¯(αμ^)S¯(αμ^))(1nΣ^1(αμ^)+1mΣ^2(αμ^))1((T¯(αμ^)S¯(αμ^)))aT_{nm}={(\bar{T}(\hat{\alpha\mu})-\bar{S}(\hat{\alpha\mu}))}^{{}^{\prime}}{(\frac{1}{n}\hat{\Sigma}_{1}(\hat{\alpha\mu})+\frac{1}{m}\hat{\Sigma}_{2}(\hat{\alpha\mu}))}^{-1}((\bar{T}(\hat{\alpha\mu})-\bar{S}(\hat{\alpha\mu})))

Given the significance level β\beta, we reject H0H_{0} if

(5.15) aTnm>χ2k4,β2aT_{nm}>\chi_{2k-4,\beta}^{2}

The expression for aTnmaT_{nm} depends on the spectrum of j(αμ^)j(\hat{\alpha\mu}) through the orbit [Uk(αμ^)][U_{k}(\hat{\alpha\mu})] and the subspace spanned by {U2(αμ^),,Uk1(αμ^)}.\{U_{2}(\hat{\alpha\mu}),\dots,U_{k-1}(\hat{\alpha\mu})\}. If the population antimean exists, [Uk(αμ^)][U_{k}(\hat{\alpha\mu})] is a consistent estimator of [Uk(αμ)][U_{k}(\alpha\mu)], the projection on Span {U2(αμ^),,Uk1(αμ^)}\{U_{2}(\hat{\alpha\mu}),\dots,U_{k-1}(\hat{\alpha\mu})\} converges to that on Span {U2(αμ),,Uk1(αμ)}\{U_{2}(\alpha\mu),\dots,U_{k-1}(\alpha\mu)\} . Thus from (5.11) and (5.14), aTnmaT_{nm} has an asymptotic χ2k42\chi_{2k-4}^{2} distribution. Hence the test in (5.15) has asympotic level β.\beta.

6 Application to medical imaging

6.1 Apert syndrome vs clinically normal children

Our data consists of shapes for a group of eighth midface labeled anatomical landmarks from X-rays of skulls of eight year old and fourteen year-old North American children(36 boys and 26 girls), known as the University School data. Each child’s skull was imaged twice, at age 8 and next at age 14. The data set, collected to study anatomical changes during children growth, represents coordinates of eight craniofacial landmarks, whose names and position on the skull are given in Bookstein ((1997)[24]), see also http://life.bio.sunysb.edu/morph/data/Book-UnivSch.txt. The data has two data set: the first one is the Apert data in Bookstein (pp. 405-406), which contains eight landmarks that describing the children who has Apert syndrome (a genetic craniosynostosis) and the second data set is clinically normal children, which contains about 40 anatomical landmarks on the skull. Out of these only 8 landmarks are registered in both groups. The two groups share only 5 registered landmarks: Posterior nasal spine,Anterior nasal spine, Sella, Sphenoethmoid registration, and Nasion. For operational definitions of these landmarks, see Bookstein (pp. 71). The shape variable of the 5 landmarks on the upper mid-face is valued in 25\sum_{2}^{5}.

Refer to caption
Refer to caption
Figure 1: Landmarks from preshapes of extrinsic means-left,extrinsic antimeans-right

In our application, we take both two sample test for VW means (see Bhattacharya and Bhattacharya [14]) and for VW antimeans, to see if we can distinguish between the Apert group and clinically normal group, using such Kendall shape location parameters. Figure 1 shows the plots of an icon for the sample VW means and VW antimeans for the two groups along with the pooled sample VW mean and VW antimean.

For k=5,k=5, the VW mean test, the values of the two sample test statistic for testing equality of the population extrinsic mean shapes is Tnm=53.1140>χ2k4,0.052=12.5916T_{nm}=53.1140>\chi_{2k-4,0.05}^{2}=12.5916, along with the asymptotic pvalue=P(χ62>53.1140)=1.1129×109p-value=P(\chi^{2}_{6}>53.1140)=1.1129\times 10^{-9} and for the VW antimean test, we get the result aTnm=144.9891>χ2k42(1α)aT_{nm}=144.9891>\chi_{2k-4}^{2}(1-\alpha), along with the asymptotic pvalue=P(χ62>144.9891)=8.862936×1029p-value=P(\chi^{2}_{6}>144.9891)=8.862936\times 10^{-29}.

From this application study, we reject the null hypothesis and conclude that both the VW mean test and the VW antimean test show that one may distinguish the Apert and clinically normal group on based on their VW antimeans, or on their VW means.

6.2 Brain scan shapes of schizophrenic vs clinically normal children

In this example from Bookstein (1991), 13 landmarks are recorded on a midsagittal two-dimensional slice from a Magnetic Resonance brain scan of each of 14 schizophrenic children and 14 normal children. It is of interest to study differences in shapes of brains between the two groups which can be used to detect schizophrenia. This is an application of disease detection. The shapes of the sample k-ads lie in Σ2k,k=13.\Sigma^{k}_{2},k=13. To distinguish between the underlying distributions, we compare their VW mean and VW antimean shapes.

For testing the equality of the VW means we use the test in Bhattacharya and Bhattacharya [14]. In this application, in addition we consider the two sample test for the equality of VW antimeans developed in the previous section. Figure 2 shows the plots of the sample VW means and VW antimeans for the two groups along with the pooled sample VW mean and VW antimean for this data sets.

Refer to caption
Refer to caption
Figure 2: Landmarks from preshapes of extrinsic means-left,extrinsic antimeans-right

For the VW mean test, since k=13,k=13, the values of the two sample test statistic for testing equality of the population extrinsic mean shapes is Tnm=95.5476>χ2k4,0.052=33.9244T_{nm}=95.5476>\chi_{2k-4,0.05}^{2}=33.9244, along with the asymptotic pvalue=P(χ222>95.5476)=3.8316×1011p-value=P(\chi^{2}_{22}>95.5476)=3.8316\times 10^{-11} and for the VW antimean test, we get the result aTnm=139.1210>χ2k4,0.052aT_{nm}=139.1210>\chi_{2k-4,0.05}^{2}, along with the asymptotic pvalue=P(χ222>144.9891)<.00001,p-value=P(\chi^{2}_{22}>144.9891)<.00001, which is significant at level 0.050.05.

From this application study, we therefore reject the null hypothesis and conclude that there is difference between the schizophrenic children and normal children for both the VW means or VW antimeans.

Acknowledgments. We would like to thank Yunfan Wang for helpful comments on an early version of our paper.

References

  • [1] Bhattacharya, R. N.; Ellingson, L.; Liu, X.; Patrangenaru, V.; Crane, M. (2012). Extrinsic analysis on manifolds is computationally faster than intrinsic analysis with applications to quality control by machine vision. Appl. Stoch. Models Bus. Ind. 28, 222–235.
  • [2] Kendall, D. G. (1984). Shape manifolds, Procrustean metrics, and complex projective spaces. Bull. London Math. Soc. 16 81-121.
  • [3] Kendall, D. G., Barden, D., Carne, T. K. and Le, H. (1999). Shape and Shape Theory. Wiley, New York.
  • [4] J.T. Kent (1992). New directions in shape analysis. The Art of Statistical Science, A Tribute to G.S. Watson, 115–127, Wiley Ser. Probab. Math. Statist., Wiley, Chichester.
  • [5] Bookstein, F. L. (1991). Morphometric Tools for Landmark Data: Geometry and Biology. Cambridge Univ. Press.
  • [6] Patrangenaru, V., and Ellingson, L. (2015). Nonparametric statistics on manifolds and their applications to object data analysis. CRC Press.
  • [7] M. Fréchet (1948). Les élements aléatoires de nature quelconque dans un espace distancié, Ann. Inst. H. Poincaré 10, 215–310.
  • [8] H.Ziezold(1977). On expected figures and a strong law of large numbers for random elements in quasi-metric spaces. Transactions of Seventh Prague Conference on Information Theory, Statistical Decision Functions, Random Processes A. 591–602.
  • [9] Huckemann, Stephan(2011), Inference on 3d Procrustes means: Tree bole growth, rank deficient diffusion tensors and perturbation models, Scandinavian Journal of Statistics, 38, 424–446.
  • [10] Bhattacharya, R.N. and Patrangenaru, V. (2003). Large sample theory of intrinsic and extrinsic sample means on manifolds-Part I,Ann. Statist. 31, no. 1, 1–29.
  • [11] Bhattacharya, Rabi and Patrangenaru, Vic (2003). Large sample theory of intrinsic and extrinsic sample means on manifolds, The Annals of Statistics, 31, 1, 1–29.
  • [12] Patrangenaru, Vic and Yao, K. David and Guo, Ruite(2016), Extrinsic means and antimeans, Nonparametric Statistics, 161–178.
  • [13] Patrangenaru, V. (2001). New large sample and bootstrap methods on shape spaces in high level analysis of natural images. Communications in Statistics-Theory and Methods, 30(8-9), 1675–1693.
  • [14] Bhattacharya, A. and Bhattacharya, R. (2012). Nonparametric statistics on manifolds with applications to shape spaces, Institute of Mathematical Statistics Monographs. Cambridge University Press.
  • [15] Balan, Vladimir and Patrangenaru, Vic (2006) Geometry of shape spaces, Proc. of The 5-th Conference of Balkan Society of Geometers. Bucharest, 2005. Geometry Balkan Press. 28–33.
  • [16] R. N. Bhattacharya and V. Patrangenaru (2002). Non parametric estimation of location and dispersion on Riemannian manifolds. J. Statist. Plan. Infer. Volume in honor of the 80th birthday of professor C.R.Rao. 108, 23–35.
  • [17] Wang, Y., Patrangenaru, V. and Guo, R. (2020). A Central Limit Theorem for Extrinsic Antimeans and Estimation of Veronese-Whitney Means and Antimeans on Planar Kendall Shape Spaces. Journal of Multivariate Analysis, 178, 104600, 11 pp.
  • [18] Kendall, David G. (1984) Shape manifolds, procrustean metrics, and complex projective spaces. Bulletin of the London Mathematical Society, Oxford University Press.
  • [19] R. N. Bhattacharya and V. Patrangenaru(2005). Large sample theory of intrinsic and extrinsic sample means on manifolds, Part II, Ann. Statist. 33, 1211–1245.
  • [20] V. Patrangenaru and L. Ellingson, Nonparametric Statistics on Manifolds and Their Applications to Object Data Analysis, CRC-Chapman & Hall. 2015.
  • [21] V.Patrangenaru and Guo, Ruite and Yao, K. David. (2016), Nonparametric Inference for Location Parameters via Fréchet Functions,Second International Symposium on Stochastic Models in Reliability Engineering, Life Science & Operations Management ,254–262.
  • [22] Ian L. Dryden and Kanti V. Mardia (2016). Statistical shape analysis with applications in R. Second edition. Wiley Series in Probability and Statistics. John Wiley &\& Sons, Ltd., Chichester.
  • [23] R. N. Bhattacharya and V. Patrangenaru (2014). Rejoinder of Discussion paper “Statistics on Manifolds and Landmarks Based Image Analysis: A Nonparametric Theory with Applications.” Journal of Statistical Planning and Inference. 145, 42–48.
  • [24] F. L. Bookstein (1997). Morphometric tools for landmark data: geometry and biology, Cambridge University Press.
  • [25] Rabi N. Bhattacharya, Marius Buibas, Ian L. Dryden, Leif A. Ellingson, David Groisser, Harrie Hendriks, Stephan Huckemann, Huiling Le, Xiuwen Liu, James S. Marron, Daniel E. Osborne, Vic Patrângenaru, Armin Schwartzman, Hilary W. Thompson, and Andrew T. A.Wood. (2013) Extrinsic data analysis on sample spaces with a manifold stratification. Advances in Mathematics, Invited Contributions at the Seventh Congress of Romanian Mathematicians, Brasov, 2011, Publishing House of the Romanian Academy (Editors: Lucian Beznea, Vasile Brîzanescu, Marius Iosifescu, Gabriela Marinoschi, Radu Purice and Dan Timotin), pp. 241–252.