Jackstraw Inference for AJIVE Data Integration
Abstract
In the age of big data, data integration is a critical step especially in the understanding of how diverse data types work together and work separately. Among data integration methods, the Angle-Based Joint and Individual Variation Explained (AJIVE) approach is particularly attractive because it not only studies joint behavior but also individual behavior. Typically AJIVE scores indicate important relationships between data objects, such as clusters. An important challenge is understanding which features, i.e. variables, are associated with those relationships. This challenge is addressed by the proposal of a hypothesis test for assessing statistical significance of features. The new test is inspired by the related jackstraw method developed for Principal Component Analysis. We use a high-dimensional muti-genomic cancer data set as our strong motivation and deep illustration of the methodology.
keywords:
Data integration; AJIVE; jackstraw; statistically significant.1 Introduction
As modern data sets have grown large they have also grown more complex. A canonical example is cancer genomics research. That field made a major transformation in the direction of Big Data over 20 years ago with the advent of micro-arrays for high throughput measurement of gene expression, resulting in simultaneous quantification of the activity of tens of thousands of genes in a single tissue sample. Since then gene expression measurements have been refined by several orders of magnitude using next generation sequencing also called RNA seq. These technologies led to many new biological discoveries, and also drove the development of many novel statistical methodologies.
An important method for obtaining deep biological insights from such high dimensional data has been Principal Component Analysis (PCA) [1]. As noted in Section 3.1.4 of [2], PCA provides many data insights through modes of variation. These are a sequence of orthogonal rank one approximations of the mean centered data matrix, representing maximal variation explained. They are the products of loading vectors, representing directions of maximal variation in the data space, with scores vectors, that are the projection coefficients of each data case onto the loading vectors. Alternatively, modes of variation can be defined using Singular Value Decomposition (SVD) of the centered data , where and are matrices comprised of orthonormal columns and is a diagonal matrix with positive entries [3]. Each rank 1 matrix is a mode of variation.
PCA scores are very interpretable as they give a clear impression of how the high dimensional data vectors relate to each other (often called dimension reduction). Scatter plot matrix views of scores often give clear views of data set structure, e.g. finding scientifically important clusters or groups of samples with shared features (see Section 4.1 of [2]). The feature loadings are also key to interpretability as they provide insight about the features driving the population structure discovered in the score plots. As an example, genes with major involvement in cancer subtypes (clusters) tend to have large magnitude loadings. An important question is which of these loadings are statistically significant which is not evident from PCA analysis alone. To overcome this, the jackstraw method [4] was developed and can viewed as a permutation version of jacknife [5] designed for high-dimensional problems.
Despite the many useful biological discoveries made from measuring high dimensional gene expression, cancer still remains a major medical challenge, due to incredible diversity and complexity. In addition to gene expression, there are many other high-dimensional assays used in cancer that include other measurements of biological functions including mutations, copy number variation, DNA methylation, miRNAs, and protein expression. It is important to understand how these different measurements work together (e.g. a copy number variation that amplifies a gene region leading to more copies of the gene and increased gene expression) or where a data type has independent information. This has motivated the statistical area of multi-block analysis (also referred to multi-view in the machine learning literature or multi-omics in bioinformatics). These analyses consist of multiple data matrices (one for each data type) called blocks, which consist of data vectors coming from a common set of experimental units (e.g. tissue samples).
An important example of multi-omics data is The Cancer Genome Atlas (TCGA) [6]. The present paper focuses on just two of the data types available in TCGA. The first is Gene Expression (GE), widely known to be fundamental to cancer biology. The second is Copy Number Variation (CNV), which quantifies repeated replication and also deletion of chromosomal parts (that play an important role in many cancer types). In this paper, we work with Copy Number Region (CNR), medians of CNV over chromosomal regions called cytobands. There is strong biological interest in the association between GE and CNV, which motivates our AJIVE - Jackstraw analyses.
A useful method for analyzing multi-block data is the Angle-based Joint and Individual Variation Explained (AJIVE), proposed by [7]. In a similar spirit to PCA, AJIVE reveals population structure through a decomposition into modes of variation. However, AJIVE takes multi-block analysis to the next level by providing two types of such modes of variation. The first type is joint between data blocks, in the sense of having common scores vectors for each block. Here, each data block represents a different type of data measurement made on the same set of experimental samples and therefore it is possible to have common scores, but not possible to have common loadings. The second type are modes of variation that are individual in the sense of having scores vectors that are orthogonal to the joint scores. Both types of scores are usefully visualized to find interesting population structure. The loadings indicate which features in each data block work together (joint) or separately (individual).
The main contribution of the present paper is the adaptation of jackstraw for inference on loadings in AJIVE analysis. In particular, it assigns statistical significance to both joint and individual loadings, which indicates the significance of the ranked feature loadings. Basic ideas are illustrated using a Toy Example in Section 1.1. Section 2 gives the algorithmic development of our novel jackstraw for AJIVE. Application of our AJIVE based jackstraw, in the context of cancer genomics research appears in Section 3.
1.1 Introduction to AJIVE
Following the standard in bioinformatics, the columns of these matrices represent data vectors or different experimental units and the rows represents features sometimes called variables or traits. In particular, all data matrices have the same number of columns (experimental units) but potentially different number of rows (features). This is the transpose of what is usual in classical statistics. Using this convention, AJIVE [7] assumes the following model for each of the mean centered data blocks:
(1) |
Here the low rank matrix of joint scores is common for all the data blocks and is composed of orthonormal columns. The low rank matrices of individual scores are composed of orthonormal columns that are orthogonal to . Notice that opposite to the usual PCA convention, to give the AJIVE analysis scale free properties, it is the scores that are normalized to have norm 1 and the loadings (columns of and ) reflect the variation quantified by singular values. The matrix is determined only up to a left multiplication by an orthogonal matrix. The matrices are chosen so that, just as in SVD, the columns of are orthogonal and decreasing in norm. The are full rank residual noise matrices.
The AJIVE modes of variation are obtained as outer products of the corresponding columns of the loading and score matrices, i.e., if and , , are the th column of and respectively, the th joint mode of variation for block is the rank 1 matrix . Analogously, the th individual mode of variation is the rank 1 matrix , .
The AJIVE estimator of is called the Common Normalized Scores (CNS) matrix, and the estimators of are called the Block Specific Scores (BSS). Similar to PCA scores, the CNS and BSS indicate the relationships between the data vectors. Again an important difference from traditional PCA is that the columns of CNS and BSS are normalized to be orthonormal vectors, hence they are also direction vectors in . This is because AJIVE inference is based on studying uncertainty in the angle distribution of the scores. In particular, the joint space, i.e. subspace of spanned by the columns of CNS, defines the estimated joint variation of all data blocks among the experimental subjects. Similarly the subspace of spanned by columns of each of the BSS defines the estimated individual variation.
Estimated modes of variation are obtained from the estimated scores and loadings matrices as outer products of the corresponding columns. The joint and individual estimated modes of variation reflect how the data blocks work together and separately respectively. An important measure of the effectiveness of AJIVE in recovering these modes of variation is the angle between the estimated loadings vector and the true underlying direction.
Using a simple Gaussian simulation, we illustrate how AJIVE identifies modes of variation from multiple data blocks that is the basis of the inference developed here (Figure 1). This toy example is constructed as in (1) across two data blocks to visibly demonstrate the difference between joint and individual modes of variation. In real data, this is not as clear as in this toy example. Figure 1 shows a number of panels. Each is a heat map view of a data matrix whose entries are coded by colors. The color red is coding values that are greater than 1 and blue is coding smaller than -1. White codes zero and in-between values are lighter shades of red or blue. As discussed above, the columns of these matrices represent data vectors or different experimental units. The rows of the matrices represents features sometimes called variables or traits. This toy example has two data blocks: Datablock 1 (shown on the top) and Datablock 2 (bottom). Both have 120 features (rows) and 160 cases (columns). The input data matrices of the toy sample in the far-left panels are the sum of the other three matrices representing joint variation, individual variation, and noise respectively shown in the other panels. The second panels on the left are rank-1 joint matrices. The first 40 rows in the top (Datablock 1) have a similar pattern to the first 40 rows in the bottom, representing important joint features which should be flagged as significant by jackstraw and the remaining rows should be insignificant regarding the joint modes of variation. The third panels on the left are rank-1 individual matrices and the top and bottom panels are very different. They have a relatively weak signal characterized by a minimum value of -0.8 and a maximum of 0.8, as indicated by the paler red and blue colors. The individual jackstraw should label the bottom 80 rows as significant. The far-right panels show simulated independent Gaussian noise with mean 0 and variance 2.
The different joint/individual signal strengths in the toy example in Figure 1 are designed to show how AJIVE recovers signals that are poorly recovered by PCA. This example will also highlight the fact that this richer information improves the statistical power of AJIVE jackstraw tests relative to PCA.
As noted above the jackstraw method [4] for statistical inference on PCA loadings has inspired our approach to AJIVE inference. Technical details are given in Section 2. To demonstrate the improvement of jackstraw with AJIVE over PCA, we applied jackstraw to three different sets of data. 1) AJIVE extracted rank-1 Joint and Individual datasets (Figure 2, top row), 2) PCA modes of variation on Datablocks 1 and 2 separately (Figure 2, middle row) ordered by best correspondence to the row above, and 3) PCA on the concatenated datablocks 1 and 2 (Figure 2, bottom row). Jackstraw significance is indicated by horizontal black lines on the right of each heatmap matrix (using the same color code as Figure 1).
The first major take way from this analysis, was that stronger signal, indicated through deeper red and blue colors, are found in AJIVE compared to PCA (Figure 2, row 1 versus row 2). Concatenation of the datablocks did improve ability of PCA to find joint signal but was not able to resolve individual modes of variation similar to AJIVE (Figure 2, row 3 versus row 1). The second major take away is that the ability for jackstraw to identify the significant joint and individual components. This is observed by the higher number of joint (top 40 rows) and individual (bottom 80 rows) identified as significant by jackstraw in AJIVE compared to PCA.
2 Jackstraw Inference
Recall the definitions of CNS and BSS and corresponding notation at the beginning of Section 1.1. Let be one column of either , i.e. scores for one joint mode of variation, or , i.e., scores for one individual mode of variation.
For one datablock , jackstraw inference is based on either joint or individual loading vectors. The corresponding loading vector (one column in the estimated loading matrix) can be computed as:
This can also be viewed as the least squares solution of the following linear regression problem:
(2) |
Note that Equation 2 contains separate simple linear regressions without an intercept in the following sense:
-
1.
Each column of is the response in the linear regression. It is centered before AJIVE so no intercept needs to be considered.
-
2.
represents the predictor of the linear regression.
-
3.
The corresponding entry of is the unknown simple linear regression coefficient, i.e. in simple linear regression.
-
4.
Rows of are the residual error terms.
The aim of the proposed jackstraw test is to find which entries of the loadings vector are statistically significant. The following hypothesis test structure forms the basis of our jackstraw inference. Let . For each , we consider a hypothesis test:
For each of these tests, we define the test statistic as
where and are the residual sums of squares under and .
Unlike in a classical regression model, we do not expect to follow an F distribution due to the data not necessarily being normal. For each entry , jackstraw could be viewed as a regular permutation test, where entries of the corresponding row of , i.e. response in the corresponding simple linear regression, are permuted. Therefore we proposed to estimate the distribution of under the null hypothesis using jackstraw. The proposed algorithm generates the simulated null distribution by randomly choosing rows and within each row randomly permuting the observations. To do this correctly, one needs to rerun an AJIVE to re-estimate the matrices. This is computationally expensive and therefore we recommend selecting rows, permuting those labels then rerunning AJIVE. After that, we do the simple linear regressions corresponding to the rows of . To generate the null distributions this is repeated times. In high dimensional settings we expect some of the rows of the data matrix to be very correlated. Therefore, we only want to permute a small number of rows perhaps even .
Thus we have test statistics simulated under the null hypothesis. The parameters and are chosen as follows:
-
1.
is the number of rows selected in one step, which should be much less than , such as 1, 10, or 100.
-
2.
is the number of times the permutation step is repeated, which should be at least 10 times .
The key algorithmic steps are:
-
1.
The test statistics, , are obtained by fitting a linear regression with the rows of mean-centered AJIVE input data as the response and with columns of the selected CNS or BSS of interest as the predictors.
-
2.
Randomly select rows of the original data matrix and permute the observations within each row and recalculate the AJIVE CNS and BSS. Then recalculate the test statistics as in Step 1 using the permuted data matrix and the same columns as in Step 1 of the new CNSs or BSSs to get K samples from the simulated null distribution. Then Repeat times to generate samples from the null distribution .
-
3.
Calculate an empirical p-value: for each observed test statistic , where denotes the indicator function.
-
4.
Adjust p-values for multiple testing. We recommend using the Bonferroni adjustment [8] for the level of the test, i.e. the regular p-value times the number of tests (in this case the number of features ) as the adjusted p-value. A reasonable alternative would be to consider the false discovery rate [9].
As stated in [4], smaller (in Step 2) gives more precise estimation, and larger gives faster calculation. The analyses in this paper use and parallel computation for maximal precision per unit time.
2.1 Approximate Algorithm for Faster Computations
Each of the resamples requires one round of the AJIVE algorithm. When the data matrices are large recomputing AJIVE times could be very computationally intensive. In many situation, the rows are highly correlated and individual features tend not to be particularly influential such as the data in Section 3. Thus permuting a small number out of a large number of rows will lead to minor changes in the AJIVE results. This is especially true in the case of a joint space estimated from multiple data blocks. In that case, one can save a large amount of computational effort by just using the original CNS rather than recomputing AJIVE times. Note this results in a classical permutation test.
We next demonstrate the effects of not recomputing AJIVE for each of the replications of jackstraw using TCGA breast cancer data, which will be analyzed in detail in Section 3. As noted above, we have 2 data blocks (data matrices): CNV that includes as features (rows) CNR, and GE that includes as features normalized log of gene expression.
After AJIVE inference, we identify 3 joint components and we proceed to apply jackstraw to find the significant genes/CNRs corresponding to each of the 3 joint components. As an investigation of the Monte Carlo variation, we run the 2 algorithms (full and approximate) 2 times using 2 different seeds, denoted as seed1 and seed2. For convenience of notation, let ‘App’ denote the approximate algorithm and ‘Full’ denote full algorithm. In Table 1, we will report the number of significant CNRs/genes. After a careful comparison, more than 95% of the significant CNRs/genes stay significant if we use different algorithms or random permutations. In each row, the subset corresponding to a smaller number of significant features is always included in the subset corresponding to the larger number. Thus, jackstraw analysis of TCGA breast cancer data is not sensitive to random permutations or algorithms.
Datablock | Joint Component | App-1 | Full-1 | App-2 | Full-2 | Total number |
---|---|---|---|---|---|---|
GE | 1 | 11288 | 11450 | 11450 | 11051 | 20249 |
GE | 2 | 7058 | 7765 | 7859 | 6918 | 20249 |
GE | 3 | 6274 | 6510 | 6455 | 6288 | 20249 |
CNV | 1 | 526 | 557 | 526 | 537 | 806 |
CNV | 2 | 247 | 197 | 276 | 276 | 806 |
CNV | 3 | 302 | 235 | 240 | 247 | 806 |
While the algorithms gave similar results, the approximate algorithm took about 30 mins and the full algorithm takes about mins for GE without parallel computation. With access to parallel computation, the full algorithm can be much faster. In practice, we recommend making a careful choice between the 2 algorithms based on whether any features are expected to be dominant.
2.2 Jackstraw Diagnostic Graph
Jackstraw typically involves a very large number of individual hypothesis tests. We recommend three diagnostic graphs (kernal density estimates, jackstraw p-values, and Kolmogorov-Smirnov tests) as useful methods for quickly summarizing the results of the numerous individual hypothesis tests such as those shown in Figure 3 of the toy example in Section 1.1. In the left panel, the black curve is a kernel density estimate visualization of the distribution (on the scale) calculated in Step 2 of the algorithm in Section 2. The red (significant) and blue (not significant) dots are the observed F statistics: calculated in Step 1. Significance is at the level of 0.05 using the Bonferroni correction. The middle panel shows the sorted jackstraw p-values, computed in Step 3, for all the genes. Out of 120 features, we find 40 significant features in the joint component. The right panel shows the results of a Kolmogorov-Smirnov test (K-S test)[10] for whether the first component overall has useful information. This tests the uniformity of the distribution of these p-values. In this panel, the colored curve is relatively far from the 45-degree line and the K-S p-value is .
3 TCGA Breast Cancer Data
As noted in the introduction, The Cancer Genome Atlas (TCGA) [6] is a multicenter effort to generate multiple different data types for a large cohort of patients to comprehensively characterize cancer. TCGA contains many cancer types and several different data platforms. We are particularly interested in the breast cancer [11] cohort, where prior work has identified molecular subtypes of breast cancer, based on gene expression [12], [13]. Therefore, this data set is an excellent case to study the application of jackstraw on the integration of multiple data platforms. The 1038 patients are classified into 4 subtypes (185 Basal cases, 81 Her2, 556 LumA, and 216 LumB). Of particular interest are the relationships between GE and CNV and how those relationships vary by subtype. Hence we do a three block AJIVE analysis here: GE (), CNV (), and subtype () and . Inclusion of the third block provides a supervised version of AJIVE, which enables careful focusing on the subtypes. The indicator (subtype) matrix has a column corresponding to each of the 1038 patients. Each column of the matrix has four entries, which indicate the four subtypes using a one in the appropriate row and zeroes for the other entries.
We use the count GE values. To construct the CNV data, we use the ratio per gene values from GISTIC downloaded from TCGA FireBrowse website. All data blocks are feature-centered as the input of AJIVE. Hence, the sign of all features is relative to the average.
A low-rank approximation is used in the first step of AJIVE for the first two input matrices. Using AJIVE diagnostic plots we established low ranks of 77 for CNV, 70 for GE and 3 for the subtype indicator matrix. This resulted in a rank 3 joint space i.e. 3 joint components overall, 74 and 67 individual components for CNV and GE respectively. In the following sections, we investigate the statistical significance of the loadings of the AJIVE joint and individual spaces using jackstraw.
3.1 Joint Space
The following discussion demonstrates the kind of information that can be learned from the Common Normalized Scores (CNS). Figure 4 shows a scatter plot matrix view of the CNS, which indicates how the cancer patients relate to each other in terms of GE and CNV. This shows a strong visual separation of the subtypes, that is a consequence of subtypes playing an important role together in GE and CNV, as well as supervision using the third data block. The CNS ( in (2)) is a 1038 by 3 matrix because of 3 joint components and 1038 cases/patients. The plots on the diagonal are one-dimensional views of the scores, columns of the CNS matrix. The subtypes are illustrated with subplot kernel density estimates (KDE). Scores are shown using jitter plots (i.e. random heights) to visually separate them in the one-dimensional plots. The overall KDEs are shown in black and the subtype KDEs are shown in subtype colors. The area under the curve of each colored KDE is proportional to its abundance so that the sum of the areas under the colored curves equals 1, the area under the black curve. The off-diagonal plots are scatter plots, i.e. two-dimensional projections of the data onto the planes generated by the corresponding CNS vectors. The first joint component strongly separates the Basal subtypes (red). Joint components 2 and 3 also seem to be strongly related to subtypes even though the separation is less than component 1. The second joint component separates the three other subtypes in the order of LumA (blue), LumB (cyan), and Her2 (magenta). This is sensible because these subtypes are known to be closer together and Her2 is known to be more distinct from LumA. The third joint component separates Her2 (magenta), LumA (blue), and LumB (cyan).
For a deeper understanding of this relationship between subtypes, Figure 5 shows the subtype loading for each joint mode of variation from Figure 4. These loadings indicate the impact of each subtype on the mode of variation. This gives another interpretation of the 3 joint components:
-
1.
contrast of Basal vs the rest; contains little Her2 information;
-
2.
contrast of (Her2 & LumB) vs LumA; contains little Basal information;
-
3.
contrast of LumB vs (Her2 & LumA); contains little Basal information.
While AJIVE analysis indicates distinct joint spaces, we implement the jackstraw method to focus on the driving set of significant features for each joint component. The column App-1 in Table 1 shows the number of jackstraw significant genes and CNRs for each joint component, using the approximate algorithm. For joint component 1, the strong differences between Basal-like subtype and non-Basals in Figures 4 and 5 appear in column App-1 in Table 1 as a large number of significant genes. This may be surprising to statisticians who have approached gene selection via sparsity methods, which has a goal of very few significant genes. However, it clearly underlines the biological complexity of this cancer. The differences within the non-Basal subtypes are weaker, which reflects itself in a smaller number of significant genes.
Our AJIVE version of jackstraw allows us to focus on a set of statistically significant features, giving insight into how GE and CNV work together. In cancer, one important mechanism is copy number variation that can influence expression by increasing or decreasing the numbers of copies per gene. Hence significant copy number variation is expected to lead to significant gene variation. We want to see if genes that are significant in joint space are joint because the copy number is directly regulating expression, however, it’s important to note that there can be many genes located in one CNR. We are interested in the cis effects where the copy number variation directly regulates the gene expression. As shown in the second row of Table 2, almost all significant CNRs contain some significant genes. This is consistent with GE being influenced by its local copy number. However, as shown in the third row of Table 2, not all significant genes in joint space have a dependency on significant CNRs. There are other mechanisms for GE regulation that are still important and show up in the joint space without having a direct dependency on the CNRs. These may be related to a trans effect (non-local regulation) or to presence of strongly correlated genes in the same pathway.
Overlapped Joint Significant % | comp 1 | comp 2 | comp 3 |
---|---|---|---|
Significant CNRs containing significant genes (%) | 95.27 | 94.57 | 91.67 |
Significant genes located in significant CNRs (%) | 69.17 | 39.68 | 36.27 |
We consider how the jackstraw significance relates to the gene/CNR loadings in Figure 6.Each panel contains colored curves which show the sorted loadings for each gene/CNR. The x-axis represents the full set of the sorted features. The y-axis shows the corresponding loadings. Recall loading vectors have length one, so the sum of the squared entries must be one. In each panel, the curves show joint components 1 (black), 2 (green), and 3 (yellow). Each variable is plotted using jitter plots (i.e. random heights) for joint components 1,2,3. Jackstraw significant genes or CNRs are identified with a red dot and non-significant features with a blue dot. The GE loadings are very similar across the three curves in the left panel, yet the CNV loadings are substantially different in the right panel. In particular, CNV has very few CNRs with very negative loadings. Relatively speaking, the signal and gene expression is spread over more genes, so there are fewer loadings that are much different from zero. There are 20249 genes and only 806 CNRs, thus the jitter plots on the left are much denser than those on the right. In both panels, joint component 1 has more significant features than joint components 2 and 3. This is consistent with the fact that joint component 1 has the most shared variation and is driven mostly by Basal subtype information, which is known to be associated with large genomic changes [14]. Generally, features with large loadings tend to be more significant. Comparing these significant genes with prior work [15] shows the jackstraw analysis has picked up biologically important regions previously identified in these types of cancers. Detailed analysis is shown in Section 4.3 of Xi Yang’s dissertation [16].
Next we investigate feature selection using jackstraw. In particular, we study classification by comparing the results based on each of jackstraw significant features, jackstraw non-significant features and all features. Significance is assessed using the Direction-Projection-Permutation high dimensional hypothesis test (DiProPerm) [17], [18] to quantify subtype separations based on different feature sets. Test results are on the scale of Population Difference Criterion (PDC) as defined in [18]. PDC works on the scale of Gaussian Z-score, so larger than 2 gives statistical significance with larger values reflecting stronger significance. Error bars are included to reflect the simulated variation of the PDC using balanced permutations as explained in Section 3.2 in [18]. Each error bar in Figure 7 shows the confidence interval of the PDC of each test, where a large PDC indicates more significance.
In each panel of Figure 7, we quantify the amount of separation between subtypes and corresponding components indicated in the titles using each of the three feature subsets as the input GE data:
-
1
All 20249 features (shown as ‘All’ in the left)
-
2
Jackstraw significant features for each joint component (shown as ‘Sig.’ in the middle)
-
3
Non-significant features for each joint component (shown as ‘Non.sig’ in the right)
The comparator groups in Figure 7 were taken from the relationships observed in Figure 5. Comp1 corresponds to Basal vs Rest, comp2 corresponds to (Her2&LumB) vs LumA, and comp 3 corresponds to (Her2&LumA) vs LumB. For all 3 panels, the tests using all features have very strong PDCs, indicating strong separations of the subtypes. When we focus on significant features only, the separations become even stronger as indicated by larger PDCs demonstrating the value of focusing on the jackstraw significance. However, the non-significant features have relatively small PDCs, but still retain some signal (PDCs greater than 2). The confidence intervals show that all of these differences are statistically significant relative to the natural permutation variation. In summary, the jackstraw significant feature sets provide the strongest subtype distinction.
3.2 Individual space
As we have seen AJIVE is very effective at finding the joint structure between all three data blocks: GE, CNV, and subtype. Next we discuss the individual information left for each data block. The DiProPerm investigation of the individual GE did not find significant subtype information, PDC of 0.59 (note less than 2). The individual CNV did reveal a significant difference between LumA and LumB, i.e. PDC of 12, which is not joint with GE. This was surprising because class labels were determined by gene expression, which is expected to have information not contained in CNV. In particular, further investigation of the individual space show that there remains informative GE-independent CNRs that can differentiate LumA and LumB. Detailed analysis can be found in Section 4.3 of Xi Yang’s dissertation [16].
4 Discussion
In this paper we propose a novel combination of jackstraw and AJIVE to bring needed inference. The new methodology gives more precise statistical inference about the relationship between gene expression and copy number. In particular, we found:
-
1.
The toy example demonstrated the value of AJIVE jackstraw relative to PCA analyses.
-
2.
Jackstraw analysis has picked up biologically important regions previously identified for these types of cancers.
-
3.
A large number of genes/copy number regions are related to the very substantial differences between Basal and non-Basals.
-
4.
The smaller number of significant genes/copy number regions reflect the weaker difference within the non-Basal subtypes.
-
5.
Focusing on significant features (i.e., gene/copy number regions) provides a stronger subtype distinction.
5 Acknowledgements
This research has been partially supported by the National Science Foundation under Grant No. IIS-1633074, DMS-1916115 and DMS-211340.
References
- [1] H. Hotelling, Analysis of a complex of statistical variables into principal components., Journal of educational psychology 24 (6) (1933) 417.
- [2] J. S. Marron, I. L. Dryden, Object Oriented Data Analysis, Chapman and Hall/CRC, 2021.
- [3] G. Golub, W. Kahan, Calculating the singular values and pseudo-inverse of a matrix, Journal of the Society for Industrial and Applied Mathematics Series B Numerical Analysis 2 (2) (1965) 205–224. doi:10.1137/0702016.
- [4] N. C. Chung, J. D. Storey, Statistical significance of variables driving systematic variation in high-dimensional data, Bioinformatics 31 (4) (2014) 545–554.
-
[5]
M. H. Quenouille, Problems in
Plane Sampling, The Annals of Mathematical Statistics 20 (3) (1949) 355 –
375.
doi:10.1214/aoms/1177729989.
URL https://doi.org/10.1214/aoms/1177729989 - [6] C. Hutter, J. C. Zenklusen, The cancer genome atlas: creating lasting value beyond its data, Cell 173 (2) (2018) 283–285.
- [7] Q. Feng, M. Jiang, J. Hannig, J. Marron, Angle-based joint and individual variation explained, Journal of multivariate analysis 166 (2018) 241–265.
- [8] C. Bonferroni, Teoria statistica delle classi e calcolo delle probabilita, Pubblicazioni del R Istituto Superiore di Scienze Economiche e Commericiali di Firenze 8 (1936) 3–62.
- [9] Y. Benjamini, Y. Hochberg, Controlling the false discovery rate: a practical and powerful approach to multiple testing, Journal of the Royal statistical society: series B (Methodological) 57 (1) (1995) 289–300.
- [10] J. E. Walsh, Bounded probability properties of kolmogorov-smirnov and similar statistics for discrete data, Annals of the Institute of Statistical Mathematics 15 (1) (1963) 153–158.
- [11] G. Ciriello, M. L. Gatza, A. H. Beck, M. D. Wilkerson, S. K. Rhie, A. Pastore, H. Zhang, M. McLellan, C. Yau, C. Kandoth, et al., Comprehensive molecular portraits of invasive lobular breast cancer, Cell 163 (2) (2015) 506–519.
- [12] C. M. Perou, T. Sørlie, M. B. Eisen, M. Van De Rijn, S. S. Jeffrey, C. A. Rees, J. R. Pollack, D. T. Ross, H. Johnsen, L. A. Akslen, et al., Molecular portraits of human breast tumours, nature 406 (6797) (2000) 747–752.
- [13] T. Sørlie, C. M. Perou, R. Tibshirani, T. Aas, S. Geisler, H. Johnsen, T. Hastie, M. B. Eisen, M. Van De Rijn, S. S. Jeffrey, et al., Gene expression patterns of breast carcinomas distinguish tumor subclasses with clinical implications, Proceedings of the National Academy of Sciences 98 (19) (2001) 10869–10874.
- [14] K. A. Hoadley, C. Yau, T. Hinoue, D. M. Wolf, A. J. Lazar, E. Drill, R. Shen, A. M. Taylor, A. D. Cherniack, V. Thorsson, et al., Cell-of-origin patterns dominate the molecular classification of 10,000 tumors from 33 types of cancer, Cell 173 (2) (2018) 291–304.
- [15] V. J. Weigman, H.-H. Chao, A. A. Shabalin, X. He, J. S. Parker, S. H. Nordgard, T. Grushko, D. Huo, C. Nwachukwu, A. Nobel, et al., Basal-like breast cancer dna copy number losses identify genes involved in genomic instability, response to therapy, and patient survival, Breast cancer research and treatment 133 (3) (2012) 865–880.
- [16] X. Yang, Machine learning methods in hdlss settings, Ph.D. thesis, The University of North Carolina at Chapel Hill (2021).
- [17] S. Wei, C. Lee, L. Wichers, J. Marron, Direction-projection-permutation for high-dimensional hypothesis tests, Journal of Computational and Graphical Statistics 25 (2) (2016) 549–569.
- [18] X. Yang, J. Hannig, J. Marron, Visual high dimensional hypothesis testing, arXiv preprint arXiv:2101.00362.