This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Longitudinal Canonical Correlation Analysis

Seonjoo Lee    Jongwoo Choi    Zhiqian Fang and F. DuBois Bowman
For the Alzheimer’s Disease Neuroimaging Initiative
Columbia University and New York State Psychiatric Institute, New York, U.S.A.
Department of Biostatistics, University of Michigan, New York, U.S.A.
[email protected]
Abstract

This paper considers canonical correlation analysis for two longitudinal variables that are possibly sampled at different time resolutions with irregular grids. We modeled trajectories of the multivariate variables using random effects and found the most correlated sets of linear combinations in the latent space. Our numerical simulations showed that the longitudinal canonical correlation analysis (LCCA) effectively recovers underlying correlation patterns between two high-dimensional longitudinal data sets. We applied the proposed LCCA to data from the Alzheimer’s Disease Neuroimaging Initiative and identified the longitudinal profiles of morphological brain changes and amyloid cumulation.

keywords:
Canonical correlation analysis, Longitudinal data analysis, Alzheimer’s disease
00footnotetext: Data used in preparation of this article were obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database (adni.loni.usc.edu). As such, the investigators within the ADNI contributed to the design and implementation of ADNI and/or provided data but did not participate in analysis or writing of this report. A complete listing of ADNI investigators can be found at: http://adni.loni.usc.edu/wp-content/uploads/how_to_apply/ADNI_Acknowledgement_List.pdf.

1 Introduction

Canonical correlation analysis (CCA) aims to find the correlation structures between two sets of multivariate variables. CCA seeks linear combinations within each set, such that the resulting linear combinations from variables are maximally correlated, but orthogonal with all other linear combinations in either set. Recently, CCA has been applied to high-dimensional multivariate variables via dimension reduction (Song et al., 2016), penalization (Avants et al., 2010; Bao et al., 2019; Fang et al., 2016; Gossmann et al., 2018; Lin et al., 2014; Wilms and Croux, 2016; Witten et al., 2009) and combining multiple datasets (Deleus and Van Hulle, 2011; Zhang et al., 2014; Kim et al., 2019).

CCA has been used frequently in medical applications to identify associations between clinical/behavior/imaging/genetic variables. For example, CCA has been applied to explore the associations between clinical symptoms and behavioral measures (Mihalik et al., 2019), functional connections in the brain and cognitive deficits (Adhikari et al., 2019) or clinical symptoms (Kang et al., 2016; Grosenick et al., 2019), two sets of imaging data (Avants et al., 2010), and gene-imaging associations (Lin et al., 2014; Kim et al., 2019).

When one of the multivariate variables is measured over time, CCA can be extended using a multi-set sparse canonical correlation approach via group lasso penalization (Hao et al., 2017) and temporal multi-task SCCA (T-MTSCCA) (Du et al., 2019). However, existing methods cannot handle missing values, irregular temporal sampling, and temporal misalignment between two variables.

To address this gap, we propose a new LCCA method that identifies the patterns of canonical variates that maximize the association between longitudinal trajectories. We model the longitudinal trajectory of each variable using random effects (e.g. random intercept and random slope). Then we find the linear combinations of the random effects that maximize the correlations. Since the dimensions of the two multivariate variables can be greater than the sample size, we employ dimension reduction via eigen decomposition. We implement estimation in the longitudinal principal component analysis framework. We conduct extensive simulation experiments to evaluate the performance of LCCA. We also apply LCCA to data from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) cohort (Mueller et al., 2005; Weiner et al., 2012). The ADNI data include longitudinal 18F-AV-45 (florbetapir) Positron emission tomography (PET) images to quantify brain amyloid loads over eight years and structural MRI data during the same follow-up. Our LCCA method shows stable performance in terms of dimension selection and yields accurate signal identification, and the method identifies distinct AD- related brain patterns in the ADNI data.

2 Method

2.1 Longitudinal Canonical Correlation Analysis

For each subject i=1,,ni=1,...,n and visits j=1,,Jij=1,\ldots,J_{i}, we observe the pp-dimensional vector 𝐗ij\mathbf{X}_{ij} at time tijt_{ij}. Similarly, for the visits k=1,,Kik=1,\ldots,K_{i}, we observe a qq-dimensional vector 𝐘ik\mathbf{Y}_{ik} at time siks_{ik}. For illustration, we consider the linear trajectories for each variable and later we generalize to the nonlinear trajectories via spline regression. We model each observation using a random intercept and slope model:

𝐗ij\displaystyle\mathbf{X}_{ij} =μijX+𝐙i,0X+tij𝐙i,1X+ϵijX,\displaystyle=\mathbf{\mu}_{ij}^{X}+\mathbf{Z}_{i,0}^{X}+t_{ij}\mathbf{Z}_{i,1}^{X}+{\boldsymbol{\epsilon}}_{ij}^{X}, (1)
𝐘ik\displaystyle\mathbf{Y}_{ik} =μikY+𝐙i,0Y+sik𝐙i,1Y+ϵikY,\displaystyle=\mathbf{\mu}_{ik}^{Y}+\mathbf{Z}_{i,0}^{Y}+s_{ik}\mathbf{Z}_{i,1}^{Y}+{\boldsymbol{\epsilon}}_{ik}^{Y}, (2)

where μijX\mathbf{\mu}_{ij}^{X} and μikY\mathbf{\mu}_{ik}^{Y} are fixed effects; 𝐙i,0X\mathbf{Z}_{i,0}^{X} and 𝐙i,0Y\mathbf{Z}_{i,0}^{Y} are random intercepts; 𝐙i,1X\mathbf{Z}_{i,1}^{X} and 𝐙i,1Y\mathbf{Z}_{i,1}^{Y} are random slopes; ϵijX{\boldsymbol{\epsilon}}_{ij}^{X} and ϵikY{\boldsymbol{\epsilon}}_{ik}^{Y} are errors. The random effects are stacked to a 2p2p-dimensional vector 𝐙iX:=(𝐙i,0X,𝐙i,1X)2p\mathbf{Z}_{i}^{X}:=\big{(}{{\mathbf{Z}}_{i,0}^{X}}^{\prime},{{\mathbf{Z}}_{i,1}^{X}}^{\prime}\big{)}^{\prime}\in\mathbb{R}^{2p} , and a 2q2q-dimensional vector 𝐙iY:=(𝐙i,0Y,𝐙i,1Y)2q{\mathbf{Z}}_{i}^{Y}:=\big{(}{{\mathbf{Z}}_{i,0}^{Y}}^{\prime},{{\mathbf{Z}}_{i,1}^{Y}}^{\prime}\big{)}^{\prime}\in\mathbb{R}^{2q}, and are assumed to be distributed with zero mean and a covariance 𝚺c=[𝚺c(0,0)𝚺c(0,1)𝚺c(1,0)𝚺c(1,1)]{\boldsymbol{\Sigma}}_{c}=\left[\begin{array}[]{rr}{\boldsymbol{\Sigma}}_{c}^{(0,0)}&{\boldsymbol{\Sigma}}_{c}^{(0,1)}\\ {\boldsymbol{\Sigma}}_{c}^{(1,0)}&{\boldsymbol{\Sigma}}_{c}^{(1,1)}\\ \end{array}\right], where 𝚺c(a,b)=𝔼(𝐙i,ac𝐙i,bc){\boldsymbol{\Sigma}}_{c}^{(a,b)}=\mathbb{E}\left({\mathbf{Z}}_{i,a}^{c}{{\mathbf{Z}}_{i,b}^{c}}^{\prime}\right), a,b{0,1},c{X,Y}a,b\in\{0,1\},c\in\{X,Y\} and uncorrelated with ϵijX{\boldsymbol{\epsilon}}_{ij}^{X} and ϵikY{\boldsymbol{\epsilon}}_{ik}^{Y}. The random intercepts, slopes and errors are assume to follow a Gaussian distribution.

Then, the objective function defining the canonical correlation is

rmXY=max𝐮m2p,𝐯m2q<𝐮m𝐙iX,𝐯m𝐙iY>.r_{m}^{XY}=max_{\mathbf{u}_{m}\in\mathcal{R}^{2p},\mathbf{v}_{m}\in\mathcal{R}^{2q}}<\mathbf{u}_{m}^{{}^{\prime}}{\mathbf{Z}}_{i}^{X},\mathbf{v}_{m}^{{}^{\prime}}{\mathbf{Z}}_{i}^{Y}>. (3)

When pp is greater than n/2n/2, we can employ principal component analysis to represent 𝐙iX𝚽X𝝃iX\mathbf{Z}_{i}^{X}\approx{\boldsymbol{\Phi}}_{X}{\boldsymbol{\xi}}_{i}^{X}, where 𝚽X=(𝚽0X,𝚽1X){\boldsymbol{\Phi}}_{X}=\left({{\boldsymbol{\Phi}}_{0}^{X}}^{\prime},{{\boldsymbol{\Phi}}_{1}^{X}}^{\prime}\right)^{\prime} is a 2p×NX2p\times N_{X} matrix of the leading NXN_{X} eigenvectors of 𝚺X=Var(𝐙iX){\boldsymbol{\Sigma}}_{X}=Var(\mathbf{Z}_{i}^{X}) and 𝝃iX=(ξi1X,,ξiNXX)NX{\boldsymbol{\xi}}_{i}^{X}=(\xi_{i1}^{X},\ldots,\xi_{iN_{X}}^{X})^{\prime}\in\mathbb{R}^{N_{X}} are the associated subject ii-specific eigenscores. Using longitudinal principal component analysis (LPCA) (Greven et al., 2011; Zipunnikov et al., 2014; Lee et al., 2015), each term is expressed as

𝐗ijμijX+l=1NXξilX(ΦlX,0+tijΦlX,1)+ϵijX,\mathbf{X}_{ij}\approx\mathbf{\mu}_{ij}^{X}+\sum_{l=1}^{N_{X}}\xi_{il}^{X}\left(\Phi_{l}^{X,0}+t_{ij}\Phi_{l}^{X,1}\right)+\epsilon_{ij}^{X}, (4)

where (ξil1X,ξil2X)(0,0;λXl1,λXl2;0)(\xi_{il_{1}}^{X},\xi_{il_{2}}^{X})\sim(0,0;\lambda_{X}^{l_{1}},\lambda_{X}^{l_{2}};0), in which “(μ1,μ2;σ12,σ22;ρ)\cdot\sim(\mu_{1},\mu_{2};\sigma_{1}^{2},\sigma_{2}^{2};\rho)” represents that a pair of gaussian variables has a distribution with mean (μ1,μ2)(\mu_{1},\mu_{2}), variance (σ12,σ22)(\sigma_{1}^{2},\sigma_{2}^{2}), and correlation ρ\rho. Similarly, 𝐘ik\mathbf{Y}_{ik} are expressed as

𝐘ikμikY+l=1NYξilY(ΦlY,0+sikΦlY,1)+ϵikY,\mathbf{Y}_{ik}\approx\mathbf{\mu}_{ik}^{Y}+\sum_{l=1}^{N_{Y}}\xi_{il}^{Y}\left(\Phi_{l}^{Y,0}+s_{ik}\Phi_{l}^{Y,1}\right)+\epsilon_{ik}^{Y}, (5)

where (ξil1Y,ξil2Y)(0,0;λYl1,λYl2;0)(\xi_{il_{1}}^{Y},\xi_{il_{2}}^{Y})\sim(0,0;\lambda_{Y}^{l_{1}},\lambda_{Y}^{l_{2}};0). The eigenvectors of (4) and (5) can be obtained by the least-squares estimation of the covariance matrices. Without loss of generality, we assume 𝐗ij\mathbf{X}_{ij} is demeaned to have mean zero. The p × p-covariance of 𝐗ij1\mathbf{X}_{ij_{1}} and 𝐗ij2\mathbf{X}_{ij_{2}} is given by

E𝐗ij1𝐗ij2=𝚺X(0,0)+tij2𝚺X(0,1)+tij1𝚺X(1,0)+tij1tij2𝚺X(1,1)+δj1,j2𝚺Xϵ,E\mathbf{X}_{ij_{1}}\mathbf{X}_{ij_{2}}^{\top}={\boldsymbol{\Sigma}}_{X}^{(0,0)}+t_{ij_{2}}{\boldsymbol{\Sigma}}_{X}^{(0,1)}+t_{ij_{1}}{\boldsymbol{\Sigma}}_{X}^{(1,0)}+t_{ij_{1}}t_{ij_{2}}{\boldsymbol{\Sigma}}_{X}^{(1,1)}+\delta_{j_{1},j_{2}}{\boldsymbol{\Sigma}}_{X}^{\epsilon}, (6)

where 𝚺Xϵ{\boldsymbol{\Sigma}}_{X}^{\epsilon} is the covariance matrix of ϵijX\epsilon_{ij}^{X} and δj1,j2=1\delta_{j_{1},j_{2}}=1 if j1=j2j_{1}=j_{2} and δj1,j2=0\delta_{j_{1},j_{2}}=0 otherwise. Denote vec()vec(\cdot) as the vectorization of a matrix by stacking columns of the matrix on top of one another. Then, we can form a matrix 𝐊X={vec(𝚺X(0,0)),vec(𝚺X(0,1)),vec(𝚺X(1,0)),vec(𝚺X(1,1)),\mathbf{K}_{X}=\{vec({\boldsymbol{\Sigma}}_{X}^{(0,0)}),vec({\boldsymbol{\Sigma}}_{X}^{(0,1)}),vec({\boldsymbol{\Sigma}}_{X}^{(1,0)}),vec({\boldsymbol{\Sigma}}_{X}^{(1,1)}), vec(𝚺Xϵ)}vec({\boldsymbol{\Sigma}}_{X}^{\epsilon})\} and 𝐟ij1j2=(1,tij2,tij1,tij1tij2,δj1,j2)\mathbf{f}_{ij_{1}j_{2}}=(1,t_{ij_{2}},t_{ij_{1}},t_{ij_{1}}t_{ij_{2}},\delta_{j_{1},j_{2}})^{\top} such that Evec(𝐗ij1𝐗ij2)=𝐊X𝐟ij1j2Evec(\mathbf{X}_{ij_{1}}\mathbf{X}_{ij_{2}}^{\top})=\mathbf{K}_{X}\mathbf{f}_{ij_{1}j_{2}}. By concatenating all vectors across all subjects and visits, we obtain a moment matrix identity for the p2×Jp^{2}\times J matrix 𝐗\mathbf{X}: 𝐗=𝐊X𝐅\mathbf{X}=\mathbf{K}_{X}\mathbf{F}, where J=i=1NJi2J=\sum_{i=1}^{N}J_{i}^{2}. Then covariance parameters 𝐊X\mathbf{K}_{X} can be unbiasedly estimated by using ordinary least squares (OLS): 𝐊^X=𝐗𝐅(𝐅𝐅)1\widehat{\mathbf{K}}_{X}=\mathbf{XF}^{\top}(\mathbf{FF}^{\top})^{-1}. Given the estimated covariance matrix, the eigenvectors are computed via eigen decomposition. Given the eigenvectors, the eigenscores are estimated by the best linear unbiased predictors (BLUP) from the equations (4) and (5) as discussed in Greven et al. (2011); Zipunnikov et al. (2014).

Denote the N×NXN\times N_{X} matrix 𝝃X={ξilX}i=1,,N;l=1,,NX{\boldsymbol{\xi}}^{X}=\{\xi_{il}^{X}\}_{i=1,\ldots,N;l=1,\ldots,N_{X}} and N×NYN\times N_{Y} matrix 𝝃Y={ξilY}i=1,,N;l=1,,NY{\boldsymbol{\xi}}^{Y}=\{\xi_{il}^{Y}\}_{i=1,\ldots,N;l=1,\ldots,N_{Y}}. Then, the objective function defining the canonical correlation is

rmXY=max𝐮mNX,𝐯mNY<𝝃X𝐮m,𝝃Y𝐯m>r_{m}^{XY}=max_{\mathbf{u}_{m}\in\mathcal{R}^{N_{X}},\mathbf{v}_{m}\in\mathcal{R}^{N_{Y}}}<{\boldsymbol{\xi}}^{X}\mathbf{u}_{m},{\boldsymbol{\xi}}^{Y}\mathbf{v}_{m}> (7)

The objective function in (7) is maximized under the restriction that each 𝐮m\mathbf{u}_{m} is orthogonal to the lower order 𝐮m\mathbf{u}_{m^{\prime}}, with 1m<mmin(NX,NY)1\leq m<m^{\prime}\leq min(N_{X},N_{Y}). Same restriction is applied for 𝐯m\mathbf{v}_{m} of the second set of variates. For identification purpose, we require a normalization condition for the canonical variates, 𝐮m𝝃iX,𝐯m𝝃iY,m=1,.,min(NX,NY)\mathbf{u}_{m}^{\top}{\boldsymbol{\xi}}_{i}^{X},\mathbf{v}_{m}^{\top}{\boldsymbol{\xi}}_{i}^{Y},m=1,.\dots,min(N_{X},N_{Y}), to have unit variance. The canonical vectors are estimated in the lower-dimensional space, and the mm-th longitudinal canonical vectors are computed as (𝚽0X+tij𝚽1X)𝐮m\left(\mathbf{\Phi}_{0}^{X}+t_{ij}\mathbf{\Phi}_{1}^{X}\right)\mathbf{u}_{m} and (𝚽0Y+sik𝚽1Y)𝐯m\left(\mathbf{\Phi}_{0}^{Y}+s_{ik}\mathbf{\Phi}_{1}^{Y}\right)\mathbf{v}_{m} for m=1,,min(NX,NY)m=1,...,min(N_{X},N_{Y}).

Depending on the availability of time points, this formation can be naturally extended to higher-order trends or modeled via B-spline basis expansion. Assume that the trajectory of the imaging measure at location vv is represented by a spline function with fixed knot sequence τ1<<τD\tau_{1}<\cdots<\tau_{D} and fixed degree dd.

𝐗ij(v)=μX(v)+r=1D+d+1Zi,r(v)br(tij)+ϵijX,\mathbf{X}_{ij}(v)=\mathbf{\mu}^{X}(v)+\sum_{r=1}^{D+d+1}Z_{i,r}(v)b_{r}(t_{ij})+\epsilon_{ij}^{X}, (8)

where the brb_{r} are a set of basis functions and Zi,r(v)Z_{i,r}(v) are the associated spline coefficients. The (D+d+1)V(D+d+1)V-dimensional stacked vector of 𝐙𝐢=(𝐙i,1,,𝐙i,D+d+1)\mathbf{Z_{i}}=(\mathbf{Z}_{i,1}^{\top},\ldots,\mathbf{Z}_{i,D+d+1}^{\top})^{\top} can be approximated by the principal component analysis: 𝐙𝐢l=1NXξilX𝚽lX\mathbf{Z_{i}}\approx\sum_{l=1}^{N_{X}}\xi_{il}^{X}\mathbf{\Phi}_{l}^{X} with each element expressed as Zi,r(v)l=1NXξilXΦlX,r(v)Z_{i,r}(v)\approx\sum_{l=1}^{N_{X}}\xi_{il}^{X}\Phi_{l}^{X,r}(v), for r=1,,D+d+1r=1,\ldots,D+d+1. Then, (8) is reorganized as

𝐗ij(v)=μX(v)+l=1NXξilXr=1D+d+1ΦlX,r(v)br(tij)+ϵijX.\mathbf{X}_{ij}(v)=\mathbf{\mu}^{X}(v)+\sum_{l=1}^{N_{X}}\xi_{il}^{X}\sum_{r=1}^{D+d+1}\Phi_{l}^{X,r}(v)b_{r}(t_{ij})+\epsilon_{ij}^{X}. (9)

Similarly, YY is modeled as

𝐘ik(v)=μY(v)+l=1NYξilYr=1D+d+1ΦlY,r(v)br(sik)+ϵikY.\mathbf{Y}_{ik}(v)=\mathbf{\mu}^{Y}(v)+\sum_{l=1}^{N_{Y}}\xi_{il}^{Y}\sum_{r=1}^{D+d+1}\Phi_{l}^{Y,r}(v)b_{r}(s_{ik})+\epsilon_{ik}^{Y}. (10)

This method does not require that XX’s and YY’s are observed at the same time. As long as they are observed within a time frame that the research question asks, the proposed method can extract association patterns.

2.2 The number of canonical covariates

To determine the dimension of the CCA, we employ a traditional likelihood ratio test-based approach in the latent space. Starting with m=0m=0, we test the null hypothesis H0H_{0}: d=md=m versus the alternative hypothesis H1H_{1}: d>md>m. If H0H_{0} is rejected, mm is incremented and a new test is conducted. This proceeds until H0H_{0} is not rejected or mm reaches M=min(NX,NY)M=min(N_{X},N_{Y}). For a given number of canonical variates, mm, the Wilk’s test statistics (Wilks, 1935; Bartlett, 1947; Friederichs and Hense, 2003) is given by

Λm=Πj=m+1min(NX,NY)(1r^j2).\Lambda_{m}=\Pi_{j=m+1}^{min(N_{X},N_{Y})}(1-\hat{r}_{j}^{2}). (11)

Based on Rao’s F-approximation (Rao et al., 1973), F=df2/df1(1Λm/(Λm))1/νF=df_{2}/df_{1}(1-\Lambda_{m}/(\Lambda_{m}))^{1/\nu} follows asymptotically Fdf1,df2F_{df_{1},df_{2}}, where ν=(df124)/((NXm)2+(NYm)25)\nu=\sqrt{(df_{1}^{2}-4)/((N_{X}-m)^{2}+(N_{Y}-m)^{2}-5)}, df1=(NXm)(NYm)df_{1}=(N_{X}-m)*(N_{Y}-m), and df2=(n1.5(NX+NY)/2)νdf1/2+1df_{2}=(n-1.5-(N_{X}+N_{Y})/2)\nu-df_{1}/2+1.

The performance of LCCA depends on the selection of NXN_{X} and NYN_{Y}. Previous works suggested using 80% threshold of the variance explained by the number of individual components (Greven et al., 2011; Lee et al., 2015) or pre-specified numbers (Zipunnikov et al., 2014). In practice, the performance depends on the signal-to-noise ratio, which is particularly relevant for imaging applications. Our numerical experiments show good performance using a threshold of 80-90% as a rule of thumb. The algorithms are implemented as an R package LCCA and available in github (https://seonjoo.github.io/lcca/).

3 Alzheimer’s Disease Neuroimaging Data Analysis

3.1 Participants

We apply our LCCA method to the ADNI data to identify longitudinal associations between brain morphometry and amyloid deposition. The data were downloaded from the ADNI database (http://adni.loni.usc.edu). The initial phase (ADNI-1) recruited 800 participants, including approximately 200 healthy controls, 400 patients with late mild cognitive impairment (MCI), and 200 patients clinically diagnosed with probable AD over 50 sites across the United States and Canada and followed up at 6- to 12-month intervals for 2–3 years. ADNI has been followed by ADNI-GO and ADNI-2 for existing participants and enrolled additional individuals, including early MCI. To be classified as MCI in ADNI, a subject needed an inclusive Mini-Mental State Examination score of between 24 and 30, subjective memory complaint, objective evidence of impaired memory calculated by scores of the Wechsler Memory Scale Logical Memory II adjusted for education, a score of 0.5 on the Global Clinical Dementia Rating, absence of significant confounding conditions such as current major depression, normal or near-normal daily activities, and absence of clinical dementia.

Table 1: Demographic characteristics at baseline and 5 year conversion rate.
AD (N=24) MCI (N=365) CN (N=291) Total (N=680) p value
Sex, Female, n (%) 10 (41.7) 162 (44.4) 147 (50.5) 319 (46.9) 0.257
Age, years, Mean (SD) 76.75 (7.15) 71.88 (7.74) 74.62 (6.47) 73.23 (7.35) <<0.001
Non-hispanic, n (%) 22 (91.7) 351 (96.2) 279 (95.9) 652 (95.9) 0.245
Race , White, n (%) 23 (95.8) 342 (93.7) 267 (91.8) 632 (92.9) 0.713
MMSE1, Mean (SD) 18.83 (4.90) 26.89 (3.34) 28.81 (1.63) 27.43 (3.39) << 0.001
ADAS2, Mean (SD) 37.35 (13.75) 16.64 (11.06) 9.05 (5.71) 14.10 (10.93) << 0.001
5 Yr AD Transition3, n (%) N/A 104 (28.5) 16 (5.5) 120 (18.3) << 0.001
1 Eight participants has missing values.
2 Nine participants had missing values.
3 Only non-demented participants at baseline were analyzed.

All studies were approved by their respective institutional review boards and all subjects or their surrogates provided informed consent compliant with HIPAA regulations. In total, the analysis included PET-MRI scan pairs of 680 subjects on average 2.7 (±\pm 0.79) visits over 3.7 (±\pm1.66) years on average. There are 291 cognitively normal (CN), 365 mild cognitive impairment (MCI), and 24 AD participants at baseline. Detailed characteristics of these individuals are given in Table 1.

3.2 Structural imaging processing

For internal consistency, 3.0 MPRAGE T1-weighted MR images were used. Cross-sectional image processing was performed using FreeSurfer Version 6. Region of interest (ROI)-specific cortical thickness and volume measures were extracted from the automated FreeSurfer v6 anatomical parcellation using the Desikan-Killiany Atlas (Desikan et al., 2006) for cortical regions; there were 68 ROIs (34 each on the left and right hemispheres), in which the longitudinal cortical thickness and volume measures were collected. The volume measures from 28 subcortical regions (Fischl et al., 2002) were computed, including Lateral Ventricle, Inferior Lateral Ventricle, Thalamus, Caudate, Putamen, Pallidum, Hippocampus, Amygdala, Accumbens area, Third Ventricle, Fourth Ventricle, and five corpus callosum subregions. In addition, the total intracranial volume was included.

We computed amyloid SUVR levels using the PetSurfer pipeline (Greve et al., 2014, 2016), which is available with Freesurfer version 6. The PetSurferpipeline first registers the PET image with the corresponding MRI scan, then applies Partial Volume Correction, and finally resamples the voxel-wise SUVR values onto the cortical surface. The 81 ROI-level summary was computed based on Desikan Atlas for the cortical and subcortical regions.

3.3 Naive approach

There are no existing CCA methods to handle the longitudinal data’s missing and irregular temporal sampling. Thus, we consider the following approach. First, we computed the intercept and slope of each variable for each subject. Then, within each modality, the vectors of intercepts and slopes were stacked, and we performed a canonical correlation analysis. Finally, the number of canonical variates was estimated in the same principles described in Section 2.2. We named this approach the naive approach.

3.4 Results

Refer to caption
Figure 1: Canonical correlation coefficients. Seven canonical correlation coefficients were significant (p<0.05p<0.05) using F-test based on Rao’s F-approximation of Wilk’s Lambda.

The longitudinal CCA identified 7 canonical variates. The significant canonical correlation coefficients are displayed in Figure 1. The longitudinal canonical vectors were reorganized as a function of time (𝚽0X+t𝚽1X)𝐮m\left(\mathbf{\Phi}_{0}^{X}+t\mathbf{\Phi}_{1}^{X}\right)\mathbf{u}_{m} and (𝚽0Y+s𝚽1Y)𝐯m\left(\mathbf{\Phi}_{0}^{Y}+s\mathbf{\Phi}_{1}^{Y}\right)\mathbf{v}_{m} for m=1,,min(NX,NY)m=1,...,min(N_{X},N_{Y}) are displayed in Figures 2 and 3 at time=0,1,2,3,4. For visualization, the vectors are standardized with the total variance of the vectors, and the ROIs with at least one element of the vector over 1.64 are included in the heatmaps. The rows are clustered using hierarchical cluster analysis with complete linkage to visualize which variables behave similarly over time. Figures 2 and 3 include the longitudinal canonical weights of the six longitudinal canonical variates that are associated with either baseline diagnosis status or AD transition. We also performed naive approach and compared its performance.

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 2: LCCA: Pairs of the longitudinal canonical vectors are displayed at time=0,1,2,3,4.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 3: LCCA: Pairs of the longitudinal canonical vectors are displayed at time=0,1,2,3,4.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 4: Naive: Pairs of the longitudinal canonical vectors are displayed at time=0,1,2,3,4.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 5: Naive: Pairs of the longitudinal canonical vectors are displayed at time=0,1,2,3,4.
Table 2: ANOVA of diagnosis group comparison and AD conversion with age, sex and education adjusted for LCCA.
T1 PET
LCV F p ηp2\eta^{2}_{p} outliers F p ηp2\eta^{2}_{p} outliers
Group 1 3.47 0.032 0.010 0 0.57 0.566 0.002 1
Comparison 2 19.83 <<0.001 0.056 0 38.73 <<0.001 0.103 0
3 2.13 0.120 0.006 1 1.08 0.342 0.003 2
4 4.86 0.008 0.014 1 3.34 0.036 0.010 2
5 0.95 0.386 0.003 0 11.60 <<0.001 0.033 2
6 2.42 0.090 0.007 0 2.46 0.086 0.007 1
7 8.26 <<0.001 0.024 0 3.86 0.022 0.011 2
AD 1 11.74 0.001 0.018 0 5.76 0.017 0.009 1
Conversion 2 102.21 <<0.001 0.136 0 164.90 <<0.001 0.202 0
3 1.51 0.220 0.002 0 0.08 0.773 0.000 1
4 6.11 0.014 0.009 0 7.70 0.006 0.012 1
5 0.00 0.972 0.000 0 0.75 0.386 0.001 1
6 3.19 0.074 0.005 0 7.05 0.008 0.011 0
7 2.80 0.095 0.004 0 3.44 0.064 0.005 2
Table 3: ANOVA of diagnosis group comparison and AD conversion with age, sex and education adjusted for the Naive approach. : adjusted p<<0.05.
T1 PET
CV F p ηp2\eta^{2}_{p} outliers F p ηp2\eta^{2}_{p} outliers
Group 1 0.02 0.984 0.000 3 3.23 0.040 0.010 4
Comparison 2 4.24 0.015 0.012 3 1.18 0.308 0.004 4
3 47.83 <<0.001 0.124 1 51.68 <<0.001 0.133 0
4 0.75 0.472 0.002 0 1.92 0.148 0.006 1
5 8.74 <<0.001 0.025 1 2.17 0.115 0.006 1
6 2.03 0.132 0.006 2 8.47 <<0.001 0.025 1
7 0.24 0.787 0.001 1 1.19 0.304 0.004 3
8 0.78 0.458 0.002 0 1.60 0.202 0.005 1
9 0.53 0.589 0.002 0 1.71 0.183 0.005 2
10 0.25 0.780 0.001 0 1.47 0.230 0.004 0
11 1.00 0.367 0.003 0 1.71 0.182 0.005 2
12 1.17 0.310 0.003 0 4.96 0.007 0.015 0
AD 1 0.64 0.424 0.001 2 1.89 0.170 0.003 3
Conversion 2 0.80 0.372 0.001 2 0.94 0.334 0.001 3
3 248.28 <<0.001 0.276 1 226.29 <<0.001 0.258 0
4 0.47 0.494 0.001 0 1.00 0.317 0.002 1
5 0.23 0.635 0.000 1 0.03 0.874 0.000 1
6 8.72 0.003 0.013 2 13.11 <<0.001 0.020 1
7 0.38 0.541 0.001 1 0.81 0.368 0.001 3
8 0.02 0.886 0.000 0 3.43 0.064 0.005 1
9 0.09 0.769 0.000 0 1.16 0.282 0.002 2
10 5.84 0.016 0.009 0 11.10 0.001 0.017 0
11 0.01 0.910 0.000 0 0.14 0.713 0.000 2
12 0.31 0.578 0.001 0 1.21 0.271 0.002 0

Further, we investigated the longitudinal canonical variates (CV). First, we compared CVs by each subject’s baseline clinical status. For group comparison, linear regression was conducted with age, sex and years of education as covariates, followed by multiple comparison correction controlling for false discovery rate (Benjamini and Hochberg, 1995). Table 2 reports the F-statistics, unadjusted p-values and multiple comparison corrected p-values. For all analyses, we excluded extreme outliers beyond 3 IQR from the first and third quartiles. For post-hoc pair-wise group comparison, least squared mean differences were computed. Figure 6 shows jittered boxplots and pair-wise comparison results with unadjusted p-values for the CVs with significant group difference after multiple comparison correction.

The CV 2, 4, and 7 of T1 and CVs 2 and 5 of PET showed group differences after multiple comparison correction. The CV2 indicates that AD patients at baseline showed atrophy in bilateral hippocampus and amygdala. The amyloid deposition showed a decreased pattern in bilateral Pallidum, and increasing amyloid deposition in the cortical areas (right lingual, right pericalcarine, right fusiform, and inferior temporal cortices). The CV4 of T1 showed the enlarged volume in right caudal middle frontal gyus and longitudinal atrophy in right pars orbitalis in AD and MCI comparing to CN, while PET did not showed any differences. The CV5 of T1 did not show differences across baseline clinical status. The AD patients showed brain atrophy in the left Accumbens area volume, and right paracentral thickness, smaller volume in the left temporal transverse volume and right caudal middle frontal volume, while thicker thickness in the left isthmus cingulate and right lateral orbitofrontal gyrus. The CV5 of av45 showed a longitudinal increase in the left hippocampus and right Pallidum and left paracentral area in AD participants, while MCI and CN participants did not show that pattern on average. CV7 showed enlarged volumes in right lingual gyrus, but atrophy in left caudal ACC and pars orbitals, while no significant difference in amyloid deposition. The naive method identified CV 3 and 5 with significant T1 differences and CVs 9 and 12 for PET. Unlike LCCA, naive method did not identify CVs showed group differences between diagnosis status in both T1 and PET. For AD transition, only CV3 T1 and CV9 PET showed significant group difference.

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 6: LCCA: Baseline diagnosis status comparison. Significant CVs are reported. Age, sex and education were adjusted for statistical testing. For post-hoc pair-wise group comparison, least squared mean differences were computed, and unadjusted p-values are reported.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 7: Naive Approach: Baseline diagnosis status comparison. Significant CVs are reported. Age, sex and education were adjusted for statistical testing. For post-hoc pair-wise group comparison, least squared mean differences were computed, and unadjusted p-values are reported.

Secondly, among the non-demented participants at baseline, we compared CVs between the participants who translated to AD within 5 years and those who remained non-demented. Among 656 non-demented participants, 120 participants (18.3%) converted to AD in 5 years. Table 2 reports the F-statistics, unadjusted p-values and multiple comparison corrected p-values. For post-hoc pair-wise group comparison, least squared mean differences were computed. Figure 8 shows jittered boxplots and pair-wise comparison results with unadjusted p-values for the CVs with significant group difference after multiple comparison correction. The CVs 1, 2 and 4 for T1 and CVs 1, 2, 4 and 6 of PET showed a significant difference between those who converted to AD and those who did not.

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 8: LCCA: Group comparison of the canonical variates between participants who transition to dementia within 5 years compared to those who did not. Age, sex and education were adjusted.
Refer to caption
Refer to caption
Refer to caption
Figure 9: Naive Approach: Group comparison of the canonical variates between participants who transition to dementia within 5 years compared to those who did not. Age, sex and education were adjusted.

4 Simulations

We conducted extensive numerical experiments to evaluate the performance of the LCCA comparing to the naive approach. The first simulation setting was designed to evaluate general performance of LCCA when the data were generated from the LCCA model. The second simulation was designed when the data were generated with a low signal-to-noise ratio. We also conducted a third simulation to determine the ability of LCCA to recover subgroups across a range of sample sizes and a varied degree of subgroup imbalance.

For all simulation settings, we generated 100 independent datasets and compared the performance of LCCA according to the following criteria: (1) number of estimated canonical variates;(2) bias of the estimated canonical correlation coefficients; (3) accuracy of the estimated canonical loading using cosine similarity. For the third simulation, we report the correlation between the true and estimated canonical variates in the place of (2). Since the true number of canonical variates is one, we compared the first canonical variate’s performance even if the LCCA or the naive approach selected different numbers of canonical variates.

4.1 Simulation Setting 1.

Data were generated from the model 4 setting NXN_{X} and NYN_{Y} to 3. The LPC loadings were generated as depicted in Figure 10; we set p=144p=144 and q=81q=81. The LPC scores ξikX{\xi}_{ik}^{X} and ξikY\xi_{ik}^{Y} were generated from the N(0,λk)N(0,\lambda_{k}), λ1=8,λ2=4,λ3=2\lambda_{1}=8,\lambda_{2}=4,\lambda_{3}=2, where Cor(ξi2X,ξi2Y)=rCor({\xi}_{i2}^{X},{\xi}_{i2}^{Y})=r. We note that we correlated the second pairs of the LPCs to evaluate LCCA’s performance when the correlated signals explain more minor variances. For each subject, the number of visits were generated from Poisson(1) followed by adding 3 such that each subject has on average 4 time points. The time intervals between visits were generated from U[1,2]U[1,2]. Figure 11 shows examples of the time variables tijt_{ij} and siks_{ik} from the first 10 subjects from a simulation setting. The time points between XX and YY variables are not aligned, and the numbers of visits differ. We conducted simulations for r=0.1,0.3,0.5r=0.1,0.3,0.5, at n=100,200,400n=100,200,400, and different threshold for LPC dimension selection (80% or 90%). For each scenario, 100 independent datasets were generated.

Refer to caption
Figure 10: Simulation 1. LFPCA models for simulation.
Refer to caption
Figure 11: Example of the time variables tijt_{ij} and siks_{ik} from the first 10 subjects from a simulation setting. The time points between XX and YY variables are not aligned, and the numbers of visits differ.
Refer to caption
Figure 12: Simulation 1. Performance evaluation.

Overall, LCCA outperforms the naive approach in the three criteria. Figure 12 shows that both LCCA and the naive approach tend to overestimate canonical correlation when the sample size is small and the true canonical correlation is smaller, while LCCA performs better than the naive approach. The results show that the approximation-based CV dimension estimation performs better as the sample size increases, and the underlying true canonical correlation is larger. LCCA identifies the number of canonical variates more accurately than the naive approach. Similar patterns were found in the cosine similarity measures. The cosine similarity between estimated LCV loadings performs better as the sample size increases, and the underlying true canonical correlation is larger.

4.2 Simulation Setting 2.

Refer to caption
Figure 13: Simulation 2. LFPCA models for simulation.

Data were generated similarly to the first simulation, except the first five LPC loadings were taken from ADNI data from Section 3 (Figure 13). The LPC scores ξikX{\xi}_{ik}^{X} and ξikY\xi_{ik}^{Y} were generated from the N(0,λk)N(0,\lambda_{k}), λ1=45.4,λ2=17.2,λ3=7.0,λ4=4,1,λ5=3.8\lambda_{1}=45.4,\lambda_{2}=17.2,\lambda_{3}=7.0,\lambda_{4}=4,1,\lambda_{5}=3.8, where Cor(ξi2X,ξi2Y)=rCor({\xi}_{i2}^{X},{\xi}_{i2}^{Y})=r. For each subject, the number of visits were generated from Poisson(1) followed by adding 3 such that each subject has on average 4 time points. The time intervals between visits were generated from U[1,2]U[1,2]. We conducted simulations for r=0.1,0.3,0.5r=0.1,0.3,0.5, at n=100,200,400n=100,200,400, and different threshold for LPC dimension selection (80% or 90%). For each scenario, 100 independent datasets were generated. We found very similar results as Simulation 1 (Figure 14, assuring our data analysis results in Section 3 have less bias in estimating canonical weights and correlations. The both methods tend to underestimate canonical correlation when the threshold (=0.8) is smaller, while this issue does not persist when the threshold is more lenient.

Refer to caption
Figure 14: Simulation 2. Performance evaluation.

4.3 Simulation Setting 3.

We also examined whether the LCCA can recover subgroups. We employed the same simulation settings from Simulation 2, except longitudinal canonical variates (LCVs). Instead of generating the LCVs from a multivariate normal distribution, we imposed two subgroups. One group was generated from MVN(0,λ𝐈2)MVN(0,\lambda\mathbf{I}_{2}) and the other group was generated from N(μ𝟏2,𝐈2)N(\mu\mathbf{1}_{2},\mathbf{I}_{2}) as shown in Figure 15. We varied the proportion of subsamples from 0.1 to 0.5 and evaluated whether the LCCA can identify subgroups.

To evaluate whether LCCA recovers the true LCVs with subgroups, the correlations between the true and estimated LCVs were calculated. Figure 16 shows that the LCCA identified LCVs, LC vectors, and the number of LCVs well, even with smaller sample sizes and a highly unbalanced subgroups (subsample proportion=0.1).

Refer to caption
Figure 15: Simulation 3. An example of the longitudinal canonical variates (LCVs) at the different ratios of subgroups.
Refer to caption
Figure 16: Simulation 3. Performance evaluation.

5 Conclusion

In this manuscript, we proposed a new longitudinal CCA method that can handle different temporal samplings and missing values that often occur in longitudinal data. The proposed method is very flexible in its ability to handle linear and nonlinear trajectories. The application to ADNI data, the longitudinal CCA revealed the most relevant patterns between Amyloid deposition associated neuronal loss measured as cortical thickness and subcortical volume. The canonical variates are associated with the baseline clinical status and predicted AD transition. The numerical experiments showed that LCCA outperformed the naive approach, also showing the performance of LCCA is not very sensitive to the selection of the threshold but requires adequate sample size or an effect size of the correlation.

Our two-stage approach is computationally fast, and our numerical experiments showed good performance using a threshold of 80-90% as a rule of thumb. However, the dimension reduction step using the LPCA possibly removes small but important biological signals in other neuroimaging modalities such as functional magnetic resonance imaging (fMRI). Furthermore, the two-step approach does not optimize both dimension reduction and dimension selection for CCA simultaneously, unlike the unified estimation approach for cross-sectional high-dimensional data (Song et al., 2016). Nonetheless, our novel LCCA approach offers a computationally efficient and well performing tool that could have important applications in neuroimaging and other settings producing high-dimensional, longitudinal, multivariate data.

Acknowledgement

This work was supported by NIH R01AG062578 (PI: Lee).

Data collection and sharing for this project was funded by the Alzheimer’s Disease Neuroimaging Initiative (ADNI) (National Institutes of Health Grant U01 AG024904) and DOD ADNI (Department of Defense award number W81XWH-12-2-0012). ADNI is funded by the National Institute on Aging, the National Institute of Biomedical Imaging and Bioengineering, and through generous contributions from the following: AbbVie, Alzheimer’s Association; Alzheimer’s Drug Discovery Foundation; Araclon Biotech; BioClinica, Inc.; Biogen; Bristol-Myers Squibb Company; CereSpir, Inc.; Cogstate; Eisai Inc.; Elan Pharmaceuticals, Inc.; Eli Lilly and Company; EuroImmun; F. Hoffmann-La Roche Ltd and its affiliated company Genentech, Inc.; Fujirebio; GE Healthcare; IXICO Ltd.; Janssen Alzheimer Immunotherapy Research & Development, LLC.; Johnson & Johnson Pharmaceutical Research & Development LLC.; Lumosity; Lundbeck; Merck & Co., Inc.; Meso Scale Diagnostics, LLC.; NeuroRx Research; Neurotrack Technologies; Novartis Pharmaceuticals Corporation; Pfizer Inc.; Piramal Imaging; Servier; Takeda Pharmaceutical Company; and Transition Therapeutics. The Canadian Institutes of Health Research is providing funds to support ADNI clinical sites in Canada. Private sector contributions are facilitated by the Foundation for the National Institutes of Health (www.fnih.org). The grantee organization is the Northern California Institute for Research and Education, and the study is coordinated by the Alzheimer’s Therapeutic Research Institute at the University of Southern California. ADNI data are disseminated by the Laboratory for Neuro Imaging at the University of Southern California.

References

  • Adhikari et al. (2019) Adhikari, B. M., Hong, L. E., Sampath, H., Chiappelli, J., Jahanshad, N., Thompson, P. M., Rowland, L. M., Calhoun, V. D., Du, X., Chen, S. and Kochunov, P. (2019) Functional network connectivity impairments and core cognitive deficits in schizophrenia. Human Brain Mapping.
  • Avants et al. (2010) Avants, B. B., Cook, P. A., Ungar, L., Gee, J. C. and Grossman, M. (2010) Dementia induces correlated reductions in white matter integrity and cortical thickness: A multivariate neuroimaging study with sparse canonical correlation analysis. NeuroImage, 50, 1004–1016.
  • Bao et al. (2019) Bao, Z., Hu, J., Pan, G. and Zhou, W. (2019) Canonical correlation coefficients of high-dimensional Gaussian vectors: Finite rank case. The Annals of Statistics, 47, 612–640.
  • Bartlett (1947) Bartlett, M. (1947) The general canonical correlation distribution. The Annals of Mathematical Statistics, 1–17.
  • Benjamini and Hochberg (1995) Benjamini, Y. and Hochberg, Y. (1995) Controlling the false discovery rate: a practical and powerful approach to multiple testing. Journal of the Royal statistical society: series B (Methodological), 57, 289–300.
  • Deleus and Van Hulle (2011) Deleus, F. and Van Hulle, M. M. (2011) Functional connectivity analysis of fMRI data based on regularized multiset canonical correlation analysis. Journal of Neuroscience Methods, 197, 143–157.
  • Desikan et al. (2006) Desikan, R. S., Ségonne, F., Fischl, B., Quinn, B. T., Dickerson, B. C., Blacker, D., Buckner, R. L., Dale, A. M., Maguire, R. P., Hyman, B. T. et al. (2006) An automated labeling system for subdividing the human cerebral cortex on mri scans into gyral based regions of interest. Neuroimage, 31, 968–980.
  • Du et al. (2019) Du, L., Liu, K., Zhu, L., Yao, X., Risacher, S. L., Guo, L., Saykin, A. J. and Shen, L. (2019) Identifying progressive imaging genetic patterns via multi-task sparse canonical correlation analysis: A longitudinal study of the ADNI cohort. Bioinformatics, 35, i474–i483.
  • Fang et al. (2016) Fang, J., Lin, D., Schulz, S. C., Xu, Z., Calhoun, V. D. and Wang, Y.-P. (2016) Joint sparse canonical correlation analysis for detecting differential imaging genetics modules. Bioinformatics (Oxford, England), 32, 3480–3488.
  • Fischl et al. (2002) Fischl, B., Salat, D. H., Busa, E., Albert, M., Dieterich, M., Haselgrove, C., Van Der Kouwe, A., Killiany, R., Kennedy, D., Klaveness, S. and others (2002) Whole brain segmentation: Automated labeling of neuroanatomical structures in the human brain. Neuron, 33, 341–355.
  • Friederichs and Hense (2003) Friederichs, P. and Hense, A. (2003) Statistical inference in canonical correlation analyses exemplified by the influence of north atlantic sst on european climate. Journal of Climate, 16, 522–534.
  • Gossmann et al. (2018) Gossmann, A., Zille, P., Calhoun, V. and Wang, Y.-P. (2018) FDR-Corrected Sparse Canonical Correlation Analysis With Applications to Imaging Genomics. IEEE transactions on medical imaging, 37, 1761–1774.
  • Greve et al. (2016) Greve, D. N., Salat, D. H., Bowen, S. L., Izquierdo-Garcia, D., Schultz, A. P., Catana, C., Becker, J. A., Svarer, C., Knudsen, G. M., Sperling, R. A. and Johnson, K. A. (2016) Different partial volume correction methods lead to different conclusions: An 18F-FDG-PET study of aging. NeuroImage, 132, 334–343.
  • Greve et al. (2014) Greve, D. N., Svarer, C., Fisher, P. M., Feng, L., Hansen, A. E., Baare, W., Rosen, B., Fischl, B. and Knudsen, G. M. (2014) Cortical surface-based analysis reduces bias and variance in kinetic modeling of brain PET data. NeuroImage, 92, 225–236.
  • Greven et al. (2011) Greven, S., Crainiceanu, C., Caffo, B. and Reich, D. (2011) Longitudinal functional principal component analysis. In Recent Advances in Functional Data Analysis and Related Topics, 149–154. Springer.
  • Grosenick et al. (2019) Grosenick, L., Shi, T. C., Gunning, F. M., Dubin, M. J., Downar, J. and Liston, C. (2019) Functional and Optogenetic Approaches to Discovering Stable Subtype-Specific Circuit Mechanisms in Depression. Biological Psychiatry. Cognitive Neuroscience and Neuroimaging, 4, 554–566.
  • Hao et al. (2017) Hao, X., Li, C., Yan, J., Yao, X., Risacher, S. L., Saykin, A. J., Shen, L., Zhang, D. and Alzheimer’s Disease Neuroimaging Initiative (2017) Identification of associations between genotypes and longitudinal phenotypes via temporally-constrained group sparse canonical correlation analysis. Bioinformatics (Oxford, England), 33, i341–i349.
  • Kang et al. (2016) Kang, J., Bowman, F. D., Mayberg, H. and Liu, H. (2016) A depression network of functionally connected regions discovered via multi-attribute canonical correlation graphs. NeuroImage, 141, 431–441.
  • Kim et al. (2019) Kim, M., Won, J. H., Youn, J. and Park, H. (2019) Joint-connectivity-based sparse canonical correlation analysis of imaging genetics for detecting biomarkers of Parkinson’s disease. IEEE transactions on medical imaging.
  • Lee et al. (2015) Lee, S., Zipunnikov, V., Reich, D. S. and Pham, D. L. (2015) Statistical image analysis of longitudinal RAVENS images. Frontiers in neuroscience, 9, 368.
  • Lin et al. (2014) Lin, D., Calhoun, V. D. and Wang, Y.-P. (2014) Correspondence between fMRI and SNP data by group sparse canonical correlation analysis. Medical Image Analysis, 18, 891–902.
  • Mihalik et al. (2019) Mihalik, A., Ferreira, F. S., Rosa, M. J., Moutoussis, M., Ziegler, G., Monteiro, J. M., Portugal, L., Adams, R. A., Romero-Garcia, R., V\a’ertes, P. E., Kitzbichler, M. G., V\a’aša, F., Vaghi, M. M., Bullmore, E. T., Fonagy, P., Goodyer, I. M., Jones, P. B., Dolan, R. and Mourão-Miranda, J. (2019) Brain-behaviour modes of covariation in healthy and clinically depressed young people. Scientific Reports, 9.
  • Mueller et al. (2005) Mueller, S. G., Weiner, M. W., Thal, L. J., Petersen, R. C., Jack, C., Jagust, W., Trojanowski, J. Q., Toga, A. W. and Beckett, L. (2005) The Alzheimer’s Disease Neuroimaging Initiative. Neuroimaging Clinics of North America, 15, 869–877.
  • Rao et al. (1973) Rao, C. R., Rao, C. R., Statistiker, M., Rao, C. R. and Rao, C. R. (1973) Linear statistical inference and its applications, vol. 2. Wiley New York.
  • Song et al. (2016) Song, Y., Schreier, P. J., Ramírez, D. and Hasija, T. (2016) Canonical correlation analysis of high-dimensional data with very small sample support. Signal Processing, 128, 449–458.
  • Weiner et al. (2012) Weiner, M. W., Veitch, D. P., Aisen, P. S., Beckett, L. A., Cairns, N. J., Green, R. C., Harvey, D., Jack, C. R., Jagust, W., Liu, E., Morris, J. C., Petersen, R. C., Saykin, A. J., Schmidt, M. E., Shaw, L., Siuciak, J. A., Soares, H., Toga, A. W., Trojanowski, J. Q. and Alzheimer’s Disease Neuroimaging Initiative (2012) The Alzheimer’s Disease Neuroimaging Initiative: A review of papers published since its inception. Alzheimer’s & Dementia, 8, S1–S68.
  • Wilks (1935) Wilks, S. (1935) On the independence of k sets of normally distributed statistical variables. Econometrica, Journal of the Econometric Society, 309–326.
  • Wilms and Croux (2016) Wilms, I. and Croux, C. (2016) Robust sparse canonical correlation analysis. BMC Systems Biology, 10, 72.
  • Witten et al. (2009) Witten, D. M., Tibshirani, R. and Hastie, T. (2009) A penalized matrix decomposition, with applications to sparse principal components and canonical correlation analysis. Biostatistics, 10, 515–534.
  • Zhang et al. (2014) Zhang, Y., Zhou, G., Jin, J., Wang, X. and Cichocki, A. (2014) Frequency recognition in SSVEP-based BCI using multiset canonical correlation analysis. International Journal of Neural Systems, 24, 1450013.
  • Zipunnikov et al. (2014) Zipunnikov, V., Greven, S., Shou, H., Caffo, B., Reich, D. S. and Crainiceanu, C. (2014) Longitudinal High-Dimensional Principal Components Analysis with Application to Diffusion Tensor Imaging of Multiple Sclerosis. The annals of applied statistics, 8, 2175.