Covariance test and universal bootstrap by operator norm
Abstract
Testing covariance matrices is crucial in high-dimensional statistics. Traditional methods based on the Frobenius and supremum norms are widely used but often treat the covariance matrix as a vector and neglect its inherent matrix structure. This paper introduces a new testing framework based on the operator norm, designed to capture the spectral properties of the covariance matrix more accurately. The commonly used empirical bootstrap and multiplier bootstrap methods are shown to fail for operator norm-based statistics. To derive the critical values of this type of statistics, we propose a universal bootstrap procedure, utilizing the concept of universality from random matrix theory. Our method demonstrates consistency across both high- and ultra-high-dimensional regimes, accommodating scenarios where the dimension-to-sample-size ratio converges to a nonzero constant or diverges to infinity. As a byproduct, we provide the first proof of the Tracy-Widom law for the largest eigenvalue of sample covariances with non-Gaussian entries as . We also show such universality does not hold for the Frobenius norm and supremum norm statistics. Extensive simulations and a real-world data study support our findings, highlighting the favorable finite sample performance of the proposed operator norm-based statistics.
keywords:
[class=MSC]keywords:
, and
1 Introduction
Testing large covariance has received considerable attention due to its critical role in high-dimensional statistics. Generally, statistics for testing one-sample high-dimensional covariance matrices are based on the distance between the sample covariance and the hypothesized covariance matrices. As highlighted by Chen, Qiu and Zhang (2023), this distance is typically measured using two types of norms: the Frobenius norm (e.g., Chen, Zhang and Zhong (2010); Cai and Ma (2013)) and the supremum norm (e.g., Jiang (2004); Cai and Jiang (2011)), similar to their application in high-dimensional mean tests Chen and Qin (2010); Cai, Liu and Xia (2014).
In contrast to the Frobenius norm and the supremum norm which treat the covariance matrices as vectors, the operator norm captures the spectral structure of the covariance matrices and has recently gained significant attention. For independent, not necessarily identically distributed random vectors with zero mean and a common covariance matrix , consider the statistics
(1) |
where is the sample covariance matrix, is the hypothesized covariance matrix, and the operator norm of a matrix is defined as . The statistic is fundamental in principal component analysis, as it bounds the error of sample eigenvalues and eigenvectors. As a result, numerous studies have developed non-asymptotic upper bounds for the tail of , e.g., Adamczak et al. (2011); Bunea and Xiao (2015); Koltchinskii and Lounici (2017). However, these upper bounds are often conservative when used to construct confidence intervals for . Only a few studies have explored the asymptotic distribution of due to the complex nature of the spectral analysis, including Han, Xu and Zhou (2018); Lopes (2022a); Lopes, Erichson and Mahoney (2023); Giessing (2023). Nonetheless, these works assume that either the dimension grows slower than , or that the eigenvalues of and decay fast enough, making its effective rank smaller than . Under such conditions, the empirical covariance matrix is a consistent estimator of with respect to the operator norm. Essentially, these assumptions reduce the intrinsic dimension of the problem, leaving the true high-dimensional cases unresolved. This article studies the behavior of under both high- and ultra-high-dimensional regimes, characterized by constants and , such that
(2) |
Furthermore, no eigen-decay assumptions are made for or . In this context, the dimension-to-sample size ratio may converge to a nonzero constant or diverge to infinity, and no consistent estimators exist for under the operator norm, as discussed in Ding and Wang (2023); Ding, Hu and Wang (2024). This setting, while challenging, is common in practice. For example, the simple case of with growing proportionally to falls into this inconsistent regime. Consequently, existing results are not applicable to such cases. The theoretical difficulties in determining the distribution of restrict the use of the operator norm in high-dimensional covariance testing within the inconsistent regime.
1.1 Random matrix theory
Building upon prior work, various approaches have been developed for analyzing and its variants in specific scenarios using random matrix theory. When , the distribution of can be derived by examining the limiting behavior of the extreme eigenvalues of the sample covariance matrix . The study of extreme spectral properties of high-dimensional covariance matrices has attracted substantial interest, resulting in a number of influential contributions, including Johnstone (2001); Bianchi et al. (2011); Onatski, Moreira and Hallin (2013); Johnstone and Paul (2018). Building on the progress of these studies, El Karoui (2007); Bao, Pan and Zhou (2015); Lee and Schnelli (2016); Knowles and Yin (2017) established that under conditions where the dimension and sample size grow proportionally and other certain regularity conditions, the limiting distribution of the largest eigenvalue of follows the Tracy–Widom law. In cases where is invertible, Bao, Pan and Zhou (2015) proposed transforming the data into , allowing for testing whether the covariance of is the identity matrix . This transformation results in the statistic taking the form of Roy’s largest root statistic
(3) |
and the Tracy–Widom distribution results hold for when and grow proportionally, i.e., . However, similar distributional results for have not yet been obtained for general , even when is bounded. To address this gap, we leverage random matrix theory to characterize the extreme singular values for the matrix of the form with a general matrix . When taking , this gives the limiting properties of . Additionally, we introduce a novel technique, the universal bootstrap, designed to directly approximate the distribution of in the context of general .
1.2 Bootstrap
The bootstrap method, a generic resampling technique to approximate the distribution of a statistic, was first introduced by Efron (1979). Originally designed for fixed-dimensional problems, the bootstrap method has recently been adapted for high-dimensional settings through substantial work. Chernozhukov, Chetverikov and Kato (2013, 2017); Lopes, Lin and Müller (2020); Lopes (2022b); Chernozhukov, Chetverikov and Koike (2023) developed Gaussian approximation methods and established the approximation rates for empirical bootstrap and multiplier bootstrap for the supremum norm of the sum of high-dimensional vectors. Moreover, Han, Xu and Zhou (2018); Lopes, Blandino and Aue (2019); Yao and Lopes (2021); Lopes (2022a); Lopes, Erichson and Mahoney (2023) explored bootstrap methods for spectral statistics of covariance matrices, while these works often assume a low intrinsic dimension or effective rank relative to the sample size . Despite such advances in high-dimensional bootstrap methods, El Karoui and Purdom (2019) and Yu, Zhao and Zhou (2024) demonstrated that empirical and multiplier bootstraps can be inconsistent for , even when if and grow proportionally and the eigenvalues of and do not decay.
To address the inconsistency issues associated with high-dimensional bootstraps based on Gaussian approximation, motivated by the concept of universality, a topic of recent interest in random matrix theory (Hu and Lu, 2022; Montanari and Saeed, 2022), we develop a novel resampling method informed by universality principles. Specifically, traditional bootstrap methods (Chernozhukov, Chetverikov and Kato, 2013, 2017; Chernozhukov, Chetverikov and Koike, 2023), depending on the high-dimensional central limit theorem, treat the large sample covariance matrix as a sum of random matrices and approximate its distribution with a Gaussian matrix sharing the same covariance as . Nonetheless, as the dimension increases, the accuracy of this approximation diminishes, eventually leading to inconsistency, as shown by El Karoui and Purdom (2019). This inconsistency indicates that in high-dimensional contexts, the large matrix structures dominate the central limit theorem’s applicability, causing the sample covariance matrix to diverge from Gaussian behavior. To circumvent this limitation, the universal property leverages high-dimensional structures. Broadly, the universality suggests that the asymptotic distribution of only depends on the distribution of through their first two moments. This allows us to construct the universal bootstrap statistic by substituting independent Gaussian samples in place of in the definition of (1),
(4) |
where . The key insight here is that while the universal bootstrap matrix need not approximate a Gaussian matrix, it effectively replicates the structure of , which is disrupted by the empirical and multiplier bootstraps. Although this method relies on sophisticated random matrix theory, the implementation remains straightforward.
1.3 Our contributions
We summarize our contribution as three-fold. First, we establish the anisotropic local law for matrices of the form within both high- and ultra-high-dimensional settings (2), where may have positive, negative, or zero eigenvalues. Erdős, Yau and Yin (2012); Knowles and Yin (2017) demonstrated that the anisotropic local law is essential for proving eigenvalue rigidity and universality. The anisotropic local law for has been previously established by Knowles and Yin (2017) for and by Ding and Wang (2023); Ding, Hu and Wang (2024) for . However, the existence of alters the structure of the sample covariance matrix , rendering existing techniques invalid. Additionally, due to the potential ultra-high dimension, each block of the Green function matrix (see (21) for definition) may converge at different rates, making the bound derived in Ding and Wang (2023) suboptimal. To address these challenges, we introduce a new ”double Schur inversion” technique to represent the Green function matrix using local blocks. We also define an auxiliary parameter-dependent covariance matrix and establish the corresponding parameter-dependent Marčenko-Pastur law. This parameter-dependent structure enables us to prove the entry-wise local law, which represents a weaker version of the anisotropic local law. Furthermore, we present an improved representation of the anisotropic local law, enhancing the results of Ding and Wang (2023) and Ding, Hu and Wang (2024). This unified structure provides a valuable tool for establishing the anisotropic local law.
Second, we establish the universality results for the statistic , based on which the universal bootstrap procedure is introduced. Leveraging this universality, we demonstrate that the empirical distribution generated by the universal bootstrap in (4) effectively approximates the distribution of . Additionally, the asymptotic distribution of is shown to be independent of the third and fourth moments of , allowing us to focus primarily on the covariance. In contrast, we find that two commonly used norms, the Frobenius norm and supremum norm, lack this universality, as their asymptotic distributions explicitly depend on all fourth moments of the data. Therefore, standardized statistics based on these norms requires fourth-moment estimation, which is generally more complex than estimating the covariance. This difference highlights a key advantage of employing the operator norm for covariance testing.
Third, we perform size and power analyses for the statistics and and propose a combined approach to enhance testing power. We also develop a generalized universality theorem to show the consistency of the universal bootstrap for statistics based on extreme eigenvalues, including this combined statistic. For size analysis, we demonstrate that the universality results apply to . As a byproduct, we extend the Tracy–Widom law for the largest eigenvalue of sample covariance with general entries under the ultra-high dimension regime, addressing a long-standing gap since it was only proven for Gaussian entries by Karoui (2003). For power performance, we analyze both and within the generalized spiked model framework from Bai and Yao (2012); Jiang and Bai (2021). We show that performs better in worst-case scenarios while excels on average performance. To further enhance the power, we propose a new combined statistic, , supported by a generalized universality theorem that confirms the validity of the universal bootstrap for . This also serves as a theoretical guarantee for the universal bootstrap applicable to a broader range of other statistics based on extreme eigenvalues. Extensive simulations validate these findings, highlighting the superior performance of our combined statistics across a range of scenarios.
1.4 Notations and paper organization
Throughout the paper, we reserve boldfaced symbols for vectors and matrices. For a complex number , we use and for its real part and imaginary part, respectively. For a vector , we use for its Euclidean norm. For a matrix , we use , , and to represent its operator norm, Frobenius norm and the supremum norm, respectively. Denote the singular values of by , where . When , we use for the eigenvalues of . The trace of a square matrix is denoted by . For two sequences , , we use or to show there exists a constant not depending on such that for all . Denote if , and if both and .
The paper is organized as follows. Section 2 provides an overview of the proposed method and presents our main results on the consistency of the proposed universal bootstrap procedure. Section 3 describes the key tools from random matrix theory and applies the universality theorem to covariance testing problems. Section 4 presents the power analysis of the universal bootstrap procedure. Section 5 provides simulation results and a real data example, demonstrating the numerical performance of our operator norm-based statistics. Detailed technical proofs are provided in the supplementary material. The data and codes are publicly available in a GitHub repository (https://github.com/zhang-guoyu/universal_bootstrap).
2 Proposed method and theoretical outlines
In this section, we introduce the covariance testing problem and present an informal summary of our main universal results. The formal results are provided in Section 3.3. We also discuss statistical applications of our proposed universal bootstrap method, which utilizes the concept of universality. Consider independent random vectors with zero mean and covariance matrix , which are not necessarily identically distributed. For a given by non-negative matrix , we aim to test the hypothesis
(5) |
As we focus on testing the covariance structure, we allow the third and fourth moments of to differ across .
To simplify notation, we arrange the data into an matrix , where denotes the transpose of , and the sample covariance matrix is expressed by . Recall the definition of statistic in (1), we aim to control the size of our procedure by charaterizing the asymptotic distribution of under the null hypothesis . Consider the Gaussian data matrix , with . Its sample covariance is defined accordingly as . Throughout this article, we work under both the proportionally growing high-dimensional regime and the ultra-high dimensional regime (2). Under these regimes, the key quantity of interest to bound is as follows
(6) |
where represents operator norm ball with center and the radius .
A simplified form is given to illustrate our main results.
Result 1 (Informal).
We have
(7) |
for some constants and , where is the covariance matrix of .
This result enables us to approximate the asymptotic distribution of using , as defined in (4). Under , and the distribution of does not contain any unknown quantities. This inspires the procedure of universal bootstrap. Given the analytical complexity of ’s distribution, we generate independent samples, , from the same distribution as . We define for each , allowing us to compute
The empirical distribution of for serves as an approximation for the distribution of . In particular, we use the empirical upper- quantile, , of , , as the threshold for the test (5) based on . Defining , we reject if . Results like (7) ensure the consistency of to the upper- quantile of under as and , validating the universal bootstrap procedure.
To contextualize our findings, we briefly compare (7) with existing high-dimensional bootstrap results. Generally, these results are presented as
where is a statistic, is its Gaussian counterpart, and represents a specified family of sets. For example, Chernozhukov, Chetverikov and Kato (2013, 2017); Chernozhukov, Chetverikov and Koike (2023) considered as mean estimators and as all rectangular in . A more related choice of is in Zhai (2018); Xu, Zhang and Wu (2019); Fang and Koike (2024), who also considered mean estimators but take to be the sets of Euclidean balls and convex sets in . Their results demonstrated that under mild conditions and for sets of Euclidean balls , converge to if and only if , meaning the Gaussian approximation holds when . For comparison, we observe that the operator norm ball for vector in coincides with the Euclidean balls, and our results show that the universality approximation holds when converges to a nonzero constant or even diverges to infinity. For the covariance test, Han, Xu and Zhou (2018) took as the sample covariance matrix and as all sets of -sparse operator norm balls (defined in their work). Especially, with as all operator norm balls, i.e. , they required , limiting to be considerably smaller than . Similarly, Lopes (2022a); Lopes, Erichson and Mahoney (2023) considered as sample covariance with as operator norm balls, but imposed a decay rate for the -th largest eigenvalue of with , implying a low intrinsic test dimension. Likewise, Giessing (2023) required the effective rank to satisfy . In contrast, we impose no such assumptions, allowing each eigenvalue of to be of comparable scale. To summarize, previous work has typically assumed either or fast-decaying eigenvalues, yielding consistent estimates for . In our setting, however, no consistent estimator of exists (see Cai and Ma (2013)). These comparisons underscore the advantages of our proposal, even in regimes lacking consistency.
Given these improvements on existing results, we establish a universal result that extends beyond (7).
Result 2 (Informal).
For that commutes with , we have
(8) |
for some constants and . See Theorem 3.5 for a formal description.
This result generalizes the universal bootstrap consistency of under to alternative covariance distinct from . The commutativity requirement between and assumes the two matrices share eigenvectors. Similar assumptions appear in Zhou, Bai and Hu (2023); Zhou et al. (2024). The result in (8) further guarantees universal bootstrap consistency for statistics beyond . For instance, we can estimate with shrinkage estimators such as for as proposed by Schäfer and Strimmer (2005), or for some as in Tsukuma (2016). Using these estimators, covariance can be tested with statistics or . Under , universal bootstrap consistency for and can be shown by and , as and commute with .
Despite the broad applications of (8), certain statistics using extreme eigenvalues remain outside its scope. For example, as outlined in Section 1, we define to combine and for enhanced statistical power, where is given by
(9) |
While (8) demonstrates universal bootstrap consistency for and separately, it does not apply to . This limitation arises from ’s dependence on the joint law of and , which introduces a complex dependence structure. To address these limitations, we develop a generalized universality theorem that accounts for dependencies among various extreme eigenvalues. We begin by introducing the generalized operator norm ball
(10) |
where is positive-definite, and with for each . Since the operator norm equals the largest singular value , we obtain . Next, we define the following quantity for symmetric matrices positive-definite matrices , ,
(11) |
where we use to represent for each .
Result 3 (Informal).
For that commutes with for , we have
(12) |
Take , , , , (12) recovers (8). But (12) provides a more general result, showing that the joint law of
(13) |
can be approximated by the universal bootstrap. This result allows for constructing statistics by combining extreme eigenvalues in (13) in various ways, with (12) confirming universal bootstrap consistency for these statistics.
3 Universal bootstrap
3.1 Preliminaries
Our results rely on the universality from the random matrix theory. To proceed, we first introduce some preliminary results relevant to our analysis. Denote , where has zero mean and identity covariance matrix for . We accordingly define . The primary matrix of interest is . This inspires us to consider the matrix of a general form
(14) |
where and . Here, is a symmetric matrix with eigenvalues that may be positive, negative, or zero, and is normalized to address cases where , as per Ding and Wang (2023). We impose some assumptions to determine the limit of the Stieltjes transform of .
Assumption 1.
Suppose that and are independent with , . There exists a positive sequence such that for and .
Assumption 2.
The matrix and are bounded in spectral norm, i.e. there exists some positive such that . Furthermore, there are constants such that the empirical spectral distribution of satisfies .
Assumption 3.
The matrix and are commutative, i.e. .
Assumptions 1 and 2 are standard and often appear in the random matrix literature. Notably, Assumption 1 relies only on independence, without assuming identical distribution as in Qiu, Li and Yao (2023). While all moments exist under Assumption 1, this condition could be relaxed as discussed in Ding and Yang (2018). We do not pursue this here. We also permit to be singular, a less restrictive condition than the invertibility assumed in Ding and Wang (2023); Ding, Hu and Wang (2024). This generalization enables testing singular , where is invalid but remains applicable. Lastly, Assumption 3, imposed for technical requirements, necessitates that and share the same eigenvectors. The same Assumption is required in the signal-plus-noise model as in Zhou, Bai and Hu (2023); Zhou et al. (2024).
We introduce the following deterministic equivalence matrix
(15) |
where and is the fixed point of the equation . Using , we define the associated Stieltjes transform as . According to Couillet, Debbah and Silverstein (2011), and are well-defined for every , and serves as the Stieltjes transform of a measure on . We further denote and as the endpoints of . Formally, if we define , we have and . Combining results of Knowles and Yin (2017) for the case and Ding and Wang (2023) for the case, it follows that and . Intuitively, the largest eigenvalue approaches , while approaches . The next subsection characterizes the fluctuations of and .
3.2 Universality
In this subsection, we establish the anisotropic local law, which forms the foundation for our universality results. Specifically, we aim to describe the fluctuations of and . This requires us to analyze the convergence of the Stieltjes transform near the endpoints and . We define the local domains and as
(16) |
for a fixed parameter . For , let , and for , define . We will focus on the behavior of the Green function on . This definition ensures that for , both the largest and smallest eigenvalues of are controlled, so that can also be controlled. While for the case , we always have , thus only the largest eigenvalue requires attention.
To facilitate the presentation of the anisotropic local law, we assume in this subsection that
(17) |
Under (17), we can rewrite as where . Since is quadratic in , we follow Knowles and Yin (2017) to define the linearization matrix and the corresponding linearized Green function
(21) |
Notice that is a large matrix. We also define the deterministic equivalent Green function as
(25) |
where . The anisotropic local law aims to control the difference for . A crucial observation is that defining the parameter -dependent covariance with the corresponding measure , where is the Dirac point measure at , we have the following -dependent deformed Marčenko-Pastur law,
(26) |
This result mirrors the form of the deformed Marčenko-Pastur law in Ding and Wang (2023), with as the covariance matrix. This demonstrates that the effect of can be represented by turning into the -dependent covariance . This insight simplifies our proof and presentation. For the denominator in (26), we impose the following technical assumption.
Assumption 4.
When we require that there exists such that
(27) |
We provide several remarks regarding Assumption 4. Informally, condition (27) ensures that the extreme eigenvalues of do not spread near the endpoints , thereby preventing spikes outside the support of . Similar assumptions have been made in the literature for the universality of , using the Stieltjes transform in place of and in place of , as in Bao, Pan and Zhou (2015); Knowles and Yin (2017). This aligns with the intuition that the effect of can be expressed through the transformation from to . Moreover, (27) is automatically satisfied when , i.e. . We thus impose Assumption 4 only in the case .
To state our main results, we define the concept of stochastic domination, introduced by Erdős, Knowles and Yau (2013) and widely applied in random matrix theory (Knowles and Yin, 2017). For two families of non-negative random variables and , where is -dependent parameter set, we say is stochastically dominated by uniformly in if for any and , we have for large enough . We use the notation to represent this relationship. When is a family of negative or complex random variables, we also write or to indicate . With these definitions, we state the anisotropic local law as follows.
Theorem 3.1 (Anisotropic local law).
(i) We have
(28) |
uniformly in vectors and , where the weight matrix is defined as and the augemented covariance matrix is .
(ii) Moreover, we have the average local law
(29) |
uniformly in .
The anisotropic local law provides a delicate characterization of the discrepancy between the random Green function and its deterministic counterpart, yielding a precise bound to analyze the behavior of the extreme eigenvalues of . When , similar results have been established for in Knowles and Yin (2017) and for in Ding and Wang (2023); Ding, Hu and Wang (2024). As demonstrated in Ding and Wang (2023), the results for general hold without requiring the dimension and sample size to grow proportionally, resulting in different convergence rates for the blocks of . Ding and Wang (2023) separately provides convergence rates for each block and offers a coarse bound for the anisotropic local law
In this study, we enhance these findings by introducing the weight matrix , allowing us to express the anisotropic local law (28) in a more compact form. This reformulation refines the convergence rate and simplifies our proof.
Building on the anisotropic local law theorem, we proceed to establish our universality result. Our first step involves deriving key implications from Theorem 3.1. Recall that is the measure on whose Stieltjes transform is the deterministic equivalence . We define the quantile sequence of as such that for . We aim to demonstrate that the eigenvalues of remain close to for .
Theorem 3.2 (Rigidity of eigenvalues).
(i) When , we have for any fixed integer number ,
(30) |
uniformly for .
(ii)When , we have
(31) |
uniformly for .
Theorem 3.2 establishes the rigidity of the eigenvalues of , with two main implications. First, when , equation (30) shows that the largest and smallest -th eigenvalues lie within an -neighborhood of , respectively, for any . Furthermore, we have , leading to uniformly for . Second, with Theorem 3.2, we demonstrate that for , revealing a fast approximation rate for the diverging quantities and .
To establish the universality result, we provide bounds for the discrepancy between the distribution of and its Gaussian counterpart. We denote by and the probability and expectation under the additional assumption that are independent standard Gaussian variables. With this notation, we present the following result.
Theorem 3.3 (Universality of the largest eigenvalue).
The universality Theorem 3.3 shows that the asymptotic distributions of , , and consequently , rely solely on the first two moments of . In contrast, as discussed in Section 4.3, other widely-used norms of , such as and ,are influenced by the first four moments of . This universality characteristic of the operator norm offers a straightforward yet effective framework for constructing rejection regions in hypothesis testing. A detailed exploration of this application is presented in Section 3.3. Before applying universality to covariance matrix testing, we outline several corollaries of Theorem 3.3, which are of notable independent interest.
Corollary 1.
Consider the case , , . Define , . Under Assumption 1, we have as ,
(33) |
where “” represents convergence in law, and is the Tracy-Widom distribution of type .
One of the central problems in random matrix theory is establishing the asymptotic distribution of the largest eigenvalue of a sample covariance matrix. In this context, (33) was derived for Gaussian in the setting where (Karoui, 2003). However, for general distributions of where , no further results have been established. Our universality Theorem 3.3 addresses this significant gap in random matrix theory.
Corollary 2.
As demonstrated in Section 5 in the Supplement, when follows a Gaussian distribution, (34) can be derived using the findings from Ji and Park (2021). For non-Gaussian distributions of , our universality Theorem 3.3 extends this result to the same form. These results expand the traditional analysis of extreme eigenvalues in the sample covariance matrix to a more general setting , where , with broader applications in covariance testing within high-dimensional statistics.
3.3 Universal bootstrap for covariance test
Theorem 3.3 is commonly termed universality in random matrix literature concerning extreme eigenvalue, see Bao, Pan and Zhou (2015); Knowles and Yin (2017). However, this universality does not ensure the validity of our universal bootstrap procedure, which depends on bounds like (7) and (8). Specifically, (3.3) provides bounds on and with , while we aim to bound and . Establishing this bound requires demonstrating the closeness between and . This step relies on the anti-concentration inequality (Chernozhukov, Chetverikov and Kato, 2013) that will be established below.
We first define the Lévy anti-concentration function for random variable as
(35) |
for . We shall provide bounds for . We note that the theorems derived thus far apply specifically to the extreme eigenvalues and . As we will show, extending these results to extreme singular values requires only weaker assumptions.
Assumption 4′.
When we require that there exists such that
(i) if , we assume for .
(ii) if , we assume for .
(iii) if , we assume for .
Assumption 4′ is evidently weaker than Assumption 4. The reasoning behind Assumption 4′ is straightforward: given that the singular value , if , then with probability , for sufficiently large , meaning only the spectral behavior near the right edge is necessary to consider. A similar argument applies when . While this relaxation may appear minor, the following example demonstrates its merit in extending the applicability of our theorem, even in the simplest case.
Example 1.
This demonstrates the relevance of introducing Assumption 4′ when focusing on . Specifically, in the case where , the smallest eigenvalue, , becomes isolated at the left edge and does not satisfy Assumption 4. However, we can show that when , , allowing Assumption 4′ to hold consistently. Accordingly, we adopt Assumption 4′ in our analysis of extreme singular values. Recalling that is the largest singular value of , we are led to define the concept of an admissible pair.
Definition 1 (Admissible pair).
Theorem 3.4 (Anti-concentration inequality).
Under Assumption 1, for any admissible pair and , we have
(36) |
Equipped with these results, we present the universal consistency theorem for . Recall the test in (5), where we reject if . Here and denotes the empirical upper -th quantile of , . To establish the theoretical validity of this universal bootstrap procedure, we provide a bound on the uniform Gaussian approximation error
(37) |
where is the operator norm ball centered at with radius . The probability in (37) is defined with respect to and , whose rows have covariance matrix , which may differ from . We summarize the results for universal bootstrap consistency as follows.
Theorem 3.5 (Universal bootstrap consistency).
Under Assumption 1, for any admissible pair , we have the uniform Gaussian approximation bound
(38) |
for some constants , .
Moreover, we have with probability approaching ,
(39) |
Theorem 3.5 establishes the uniform Gaussian approximation bound and Type I error bound of the universal bootstrap. Expression (39) ensures that we can uniformly control the test size with an error of at most , which vanishes as and . This result, therefore, guarantees the uniform consistency of the universal bootstrap procedure. To appreciate these results, we compare our universal bootstrap with the widely used multiplier bootstrap that approximates the distribution of using the distribution of conditioned on , where are random variables with zero mean and unit variance. To ensure consistency of multiplier bootstrap, Han, Xu and Zhou (2018) required a low dimension condition, while Lopes (2022a); Lopes, Erichson and Mahoney (2023); Giessing (2023) imposed a fast eigen-decay condition, as discussed in Section 2. These methods rely on the consistency of sample covariance, i.e. , which does not hold even when and . This highlights the key advantage of our proposal for covariance testing in a high-dimensional setting where estimation consistency is unattainable, yet test consistency is assured.
3.4 Sharp uniform simultaneous confidence intervals
The approximation of the distribution of the largest singular value has numerous applications, including determining the number of spikes, as discussed in Ding and Yang (2022). Next, we present an additional important application of Theorem 3.5. Define the inner product of two matrices and as , and the Schatten norm of matrix as . We shall construct sharp uniform simultaneous confidence intervals for and for all and . To achieve this, we note that where the spectral matrix is defined as . Although the complete matrix cannot be consistently estimated in our setting, certain estimators of the spectral matrix have been shown to be consistent under a suitable normalized distance. For instance, given the distance , Ledoit and Wolf (2015); Kong and Valiant (2017) proposed estimators satisfying as when . Therefore, the estimated threshold can replace the unknown .
Theorem 3.6.
Under Assumption 1, for any admissible pair and estimators of , we have the following sharp uniform simultaneous confidence intervals for
(40) | ||||
As a special case, we can also construct sharp uniform simultaneous confidence intervals for
(41) | ||||
Theorem 3.6 establishes sharp uniform simultaneous confidence intervals for both and . Specifically, intervals (40) and (41) are sharp, as their probability converges uniformly to , rather than no less than . When and are chosen as the canonical basis vectors in , we obtain a uniform entry-wise confidence interval for all . Notably, Corollary 3.6 extends this result by providing confidence intervals for all linear combinations of entries of . Constructing confidence intervals for all such combinations is considerably more challenging than for individual entries. For a mean vector in the simpler vector case, achieving confidence intervals for each entry requires only for some , see Chernozhukov, Chetverikov and Kato (2013, 2017); Chernozhukov, Chetverikov and Koike (2023). In contrast, confidence intervals for all linear combinations of the mean vector require , see Zhai (2018); Xu, Zhang and Wu (2019); Fang and Koike (2024). Remarkably, we are able to construct sharp simultaneous confidence intervals for all even as converges or diverges to infinity.
4 Power analysis and the generalized universal bootstrap approximation
4.1 Power analysis for operator norm-based statistics
As noted in Section 1, several popular statistics utilize the operator norm in covariance testing, such as Roy’s largest root statistic in (3). Our Theorem 3.5 also applies to Roy’s largest root with universal bootstrap, prompting the question of which test to use. In this subsection, we perform a power analysis across a family of statistics, with and examined as specific cases.
We conduct a power analysis under the generalized spike model setting, which allows variability in the bulk eigenvalues and does not assume block independence between the spiked and bulk parts, as in Bai and Yao (2012) and further developed by Jiang and Bai (2021). We assume the following generalized spike structure
(48) |
where , are diagonal matrices, is a fixed number, and is orthogonal matrix. Thus
(53) |
where , , . We assume that , , are bounded, without requiring the eigenvalues of to be the same, unlike the classical spike model. We consider the family of statistics
(54) |
and reject if . Notably, , are both included in this family. Define , and as in Section 3 with in in (14). Intuitively, , and are limits of the upper and lower bound and the Stieltjes transform of bulk eigenvalues of . For simplicity, we restrict our analysis to and .
Theorem 4.1.
(i) If for some we have , then
i.e. the power goes to .
(ii) If for all , we have and , then
i.e. there is no power under this setting.
Theorem 4.1 provides the power analysis of the test based on the statistic (54). Specifically, it identifies the phase transition point
which closely resembles the well-known Baik-Ben Arous-Péché (BBP) phase transition for the largest eigenvalues of the sample covariance matrix, see Baik, Ben Arous and Péché (2005); Baik and Silverstein (2006); Paul (2007). When the spike eigenvalues exceed this phase transition point, they lie outside the bulk eigenvalue support, yielding full test power. Conversely, if the spike eigenvalues fall below this point, distinguishing them from the null case using becomes infeasible, as the spike and bulk eigenvalues are indistinguishably close. It is noted that Theorem 4.1 focuses solely on the largest eigenvalue spike, as analyzing the smallest eigenvalue spike is more complex. Since this power analysis serves as a preliminary illustration, we do not extend our investigation in that direction.
With Theorem 4.1, we can now deduce the power of and as special cases.
Corollary 3.
Under the same assumption in theorem 4.1, define , .
(i) If for some , we have , then
i.e. the power goes to . Here satisfies .
(ii) If for some , we have , then
i.e. the power goes to .
We compare the phase transition points (i) and (ii) for and , respectively. These points reveal distinct behaviors: as , the phase transition point of (i) decreases with , whereas the phase transition point of (ii) increases with . This indicates that when spikes affect the larger eigenvalues, i.e. for large , then the statistics achieves full power more effectively than the normalized due to a lower phase transition point. Conversely, the normalized statistic performs better for smaller . Thus, Corollary 3 quantifies this intuitive observation.
4.2 The generalized universal bootstrap
Given that neither and dominates across all settings, a natural approach is to combine these two statistics to enhance testing power, see in (9) as an example. In Section 5, simulations demonstrate that can outperform both and in certain scenarios. However, conducting tests with requires calculating an appropriate threshold. While Theorem 3.5 provides thresholds for and individually, it cannot be directly applied to due to the complex dependence between and . To address this, we develop a generalized universality theorem that applies to a broad class of statistics involving extreme singular values.
We consider the statistic as a general function of extreme singular values
(55) |
where is a measurable function, are matrices, and denotes integers for . The universal bootstrapped version, , is defined by replacing with in (55). We then define the empirical universal bootstrap threshold as the upper -th quantile of of the i.i.d. samples . This setup enables the following generalized universal bootstrap consistency result.
Theorem 4.2 (Generalized universal bootstrap consistency).
Under Assumption 1, given admissible pairs and fixed integer numbers for , we have the uniform Gaussian approximation bound
(56) |
for some constants , , where is defined in (11).
Moreover, we have
(57) |
Theorem 4.2 extends Theorem 3.5 by demonstrating that universal bootstrap consistency applies not only to and , but also to combinations like , the sum of the largest -th singular values of proposed by Ke (2016), and to all statistics based on extreme singular values with arbitrary dependencies in the form of (55). This result reinforces our universal bootstrap principle: to approximate the distribution of statistics involving extreme eigenvalues, it suffices to substitute all random variables with their Gaussian counterparts, showcasing the practical advantage of universality.
4.3 Comparison with other norms
In this subsection, we compare statistics based on the operator norm with those based on the two widely used norms: the Frobenius norm and the supremum norm. Define the statistics
(58) |
While the universality Theorem 3.3 holds for operator norm-based statistics and , we demonstrate that it does not extend to or . Theorem 3.3 specifies that the asymptotic distribution of and depends solely on the first two moments of , allowing for the construction of a universal bootstrap procedure. In contrast, the asymptotic distributions of and rely on the first four moments of . We present the following modified assumptions to establish the asymptotic distribution of .
Assumption 1′.
Suppose that are i.i.d. random variables with , , . Assume that for some .
Proposition 1.
Under Assumption 1′, if or , we have
To facilitate the Gaussian approximation of , we introduce the following sub-Gaussian assumption. This assumption is employed here for simplicity and can be extended to a more general, though tedious bound if needed.
Assumption 1′′.
Suppose that for ,
for some positive sequence .
Assumption 2′′.
Proposition 2.
Propositions 1 and 2 follow as direct corollaries from Qiu, Li and Yao (2023) and Chernozhuokov et al. (2022), and we omit the proofs here. These propositions demonstrate that statistics based on the Frobenius and supremum norms lack the second-moment universality as the operator norm. Consequently, tests relying on the Frobenius and supremum norms require normalization through fourth-moment estimates of , a process more complex than estimating the covariance structure itself. This complexity arises because the Frobenius and supremum norms reduce the covariance matrices to vector forms, calculating the and norms, respectively. As a result, deriving the asymptotic distribution necessitates the second moment of the sample covariance, equivalent to the fourth moment of . Thus, the operator norm emerges as a more easy-to-implement option for covariance testing.
5 Numerical results
5.1 Simulation
In this subsection, we evaluate the empirical size and power of the proposed universal bootstrap for the operator norm-based statistics in (1) (Opn) and the combined statistics in (9) (Com). These are compared with the operator norm-based statistic in (3) (Roy), the Frobenius norm-based linear spectral statistic from Qiu, Li and Yao (2023) (Lfn), the debiased Frobenius norm-based U statistic from Cai and Ma (2013) (Ufn), and the supremum norm-based statistic inspired by Cai and Jiang (2011); Cai, Liu and Xia (2013) (Supn). The expressions for these statistics are provided in Section 12 of the Supplement.
To evaluate the robustness of our method, we generate independent, non-identically distributed samples of -dimensional random vectors with zero means and covariance matrix . We set sample sizes and and vary the dimension over , , , , and to cover both proportional and ultra-high-dimensional cases. For each configuration, we calculate the average number of rejections based on replications, with the universal bootstrap procedure using resamplings. The significance level for all tests is fixed at . We consider three different structures for the null covariance matrix :
(a). Exponential decay. The elements .
(b). Block diagonal. Let be the number of blocks, with each diagonal block being .
(c). Signed sub-exponential decay. The elements .
These structures have been previously analyzed in Cai, Liu and Xia (2013); Zhu et al. (2017); Yu, Li and Xue (2024) for models (a) and (b), and in Cai, Liu and Xia (2013); Yu, Li and Xue (2024) for model (c), highlighting the broad applicability of our approach. To assess the empirical size, we examine the following distributions of :
(1). Gaussian distribution: are i.i.d. standard normal random variables.
(2). Uniform and t-distributions: are i.i.d. normalized uniform random variables on , while are i.i.d. normalized t random variables with degree of freedom .
(3). Gaussian and uniform distributions: are i.i.d. standard normal random variables, while are i.i.d. normalized uniform random variables on .
(4). Gaussian and t-distributions: are i.i.d. standard normal random variables, while are i.i.d. normalized t random variables with degree of freedom .
Common covariance test cases often consider the Gaussian and uniform distributions Chen, Qiu and Zhang (2023), while the t-distribution with degrees of freedom is examined in Cai, Liu and Xia (2013); Zhu et al. (2017). All distributions are standardized to zero mean and unit variance. Except for the Gaussian baseline, the other cases share identical covariance matrices but differ in distribution, creating challenging yet crucial scenarios for testing covariance structures without imposing additional distributional assumptions. These cases reflect the broad applicability of our universality result.
Exp. decay(0.6) | Block diagonal | Signed subExp. decay(0.4) | ||||||||||||||
100 | 200 | 500 | 1000 | 2000 | 100 | 200 | 500 | 1000 | 2000 | 100 | 200 | 500 | 1000 | 2000 | ||
Gaussian, n=100 | ||||||||||||||||
Supn | 0.203 | 0.240 | 0.316 | 0.406 | 0.487 | 0.203 | 0.240 | 0.316 | 0.406 | 0.487 | 0.203 | 0.240 | 0.316 | 0.406 | 0.487 | |
Lfn | 0.049 | 0.050 | 0.049 | 0.050 | 0.055 | 0.049 | 0.050 | 0.049 | 0.050 | 0.055 | 0.049 | 0.050 | 0.049 | 0.050 | 0.055 | |
Ufn | 0.048 | 0.056 | 0.046 | 0.049 | 0.056 | 0.048 | 0.056 | 0.046 | 0.049 | 0.056 | 0.048 | 0.056 | 0.046 | 0.049 | 0.056 | |
Roy | 0.050 | 0.050 | 0.045 | 0.053 | 0.049 | 0.050 | 0.050 | 0.045 | 0.053 | 0.049 | 0.050 | 0.050 | 0.045 | 0.053 | 0.049 | |
Opn | 0.051 | 0.046 | 0.048 | 0.044 | 0.049 | 0.048 | 0.047 | 0.048 | 0.047 | 0.052 | 0.053 | 0.055 | 0.049 | 0.049 | 0.052 | |
Com | 0.050 | 0.046 | 0.046 | 0.047 | 0.049 | 0.050 | 0.046 | 0.047 | 0.045 | 0.050 | 0.055 | 0.054 | 0.048 | 0.052 | 0.051 | |
Gaussian, n=300 | ||||||||||||||||
Supn | 0.073 | 0.066 | 0.071 | 0.070 | 0.076 | 0.073 | 0.066 | 0.071 | 0.070 | 0.076 | 0.073 | 0.066 | 0.071 | 0.070 | 0.076 | |
Lfn | 0.056 | 0.048 | 0.045 | 0.045 | 0.053 | 0.056 | 0.048 | 0.045 | 0.045 | 0.053 | 0.056 | 0.048 | 0.045 | 0.045 | 0.053 | |
Ufn | 0.053 | 0.048 | 0.044 | 0.044 | 0.057 | 0.053 | 0.048 | 0.044 | 0.044 | 0.057 | 0.053 | 0.048 | 0.044 | 0.044 | 0.057 | |
Roy | 0.048 | 0.049 | 0.050 | 0.052 | 0.051 | 0.048 | 0.049 | 0.050 | 0.052 | 0.051 | 0.048 | 0.049 | 0.050 | 0.052 | 0.051 | |
Opn | 0.057 | 0.052 | 0.047 | 0.051 | 0.052 | 0.052 | 0.050 | 0.052 | 0.052 | 0.049 | 0.042 | 0.046 | 0.054 | 0.049 | 0.054 | |
Com | 0.057 | 0.050 | 0.046 | 0.052 | 0.052 | 0.053 | 0.050 | 0.050 | 0.051 | 0.048 | 0.042 | 0.048 | 0.052 | 0.049 | 0.055 | |
uniform and t(df=12), n=100 | ||||||||||||||||
Supn | 0.251 | 0.293 | 0.391 | 0.492 | 0.613 | 0.251 | 0.293 | 0.391 | 0.492 | 0.613 | 0.251 | 0.293 | 0.391 | 0.492 | 0.613 | |
Lfn | 0.057 | 0.050 | 0.053 | 0.052 | 0.062 | 0.057 | 0.050 | 0.053 | 0.052 | 0.062 | 0.057 | 0.050 | 0.053 | 0.052 | 0.062 | |
Ufn | 0.056 | 0.051 | 0.048 | 0.051 | 0.064 | 0.056 | 0.051 | 0.048 | 0.051 | 0.064 | 0.056 | 0.051 | 0.048 | 0.051 | 0.064 | |
Roy | 0.050 | 0.045 | 0.051 | 0.050 | 0.050 | 0.050 | 0.045 | 0.051 | 0.050 | 0.050 | 0.050 | 0.051 | 0.051 | 0.050 | 0.050 | |
Opn | 0.045 | 0.033 | 0.043 | 0.041 | 0.047 | 0.052 | 0.043 | 0.042 | 0.051 | 0.043 | 0.047 | 0.047 | 0.051 | 0.060 | 0.049 | |
Com | 0.045 | 0.033 | 0.044 | 0.041 | 0.050 | 0.051 | 0.044 | 0.041 | 0.055 | 0.046 | 0.049 | 0.047 | 0.046 | 0.063 | 0.055 | |
uniform and t(df=12), n=300 | ||||||||||||||||
Supn | 0.093 | 0.089 | 0.098 | 0.109 | 0.125 | 0.093 | 0.089 | 0.098 | 0.109 | 0.125 | 0.093 | 0.089 | 0.098 | 0.109 | 0.125 | |
Lfn | 0.055 | 0.047 | 0.050 | 0.039 | 0.053 | 0.055 | 0.047 | 0.050 | 0.039 | 0.053 | 0.055 | 0.047 | 0.050 | 0.039 | 0.053 | |
Ufn | 0.056 | 0.044 | 0.049 | 0.040 | 0.056 | 0.056 | 0.044 | 0.049 | 0.040 | 0.056 | 0.056 | 0.044 | 0.049 | 0.040 | 0.056 | |
Roy | 0.050 | 0.046 | 0.038 | 0.053 | 0.054 | 0.050 | 0.046 | 0.038 | 0.053 | 0.054 | 0.050 | 0.046 | 0.038 | 0.053 | 0.054 | |
Opn | 0.040 | 0.051 | 0.049 | 0.051 | 0.050 | 0.042 | 0.053 | 0.051 | 0.047 | 0.051 | 0.046 | 0.042 | 0.036 | 0.042 | 0.061 | |
Com | 0.046 | 0.051 | 0.041 | 0.048 | 0.047 | 0.043 | 0.052 | 0.050 | 0.046 | 0.052 | 0.049 | 0.045 | 0.038 | 0.045 | 0.058 |
Table 1 reports the empirical sizes of the supremum, the Frobenius, and the operator norm tests at the significance level across various sample size , dimension , covariance and data generation distributions of Gaussian and the uniform and t distributions, (distributions (1) and (2)). For results on Gaussian-uniform and Gaussian- distributions (distributions (3) and (4)), see Table 2 in the Supplement, which demonstrates similar patterns. The three operator norm tests are performed using the proposed universal bootstrap procedure, while the supremum and Frobenius norm tests rely on their respective asymptotic distribution formulas. Both Table 1 and Table 2 in the Supplement show that the universal bootstrap effectively controls the size of all operator norm-based statistics at the nominal significance level across all tested scenarios. This approach performs well under both Gaussian and non-Gaussian distributions, for both i.i.d. and non-i.i.d. data, in proportional and ultra-high dimensional settings, and across various covariance structures. These findings confirm the universality and consistency of the universal bootstrap procedure, providing empirical support for our theoretical results in Theorem 3.5 and Theorem 4.2. Additionally, the Frobenius norm-based test maintains appropriate size control, while the supremum norm-based test exhibits substantial size inflation at and moderate inflation even at a larger sample size of .
To evaluate the empirical power of these statistics, we consider a setting where , with having a low rank. This setup corresponds to the power analysis presented in Theorem 4.1. As shown in Theorem 4.1, the statistic outperforms when the eigenspaces of align with the eigenvectors of corresponding to larger eigenvalues. In contrast, performs better when aligns with the smaller eigenvectors of . For a fair comparison, we define , where is the fifth-largest eigenvector of , is uniformly sampled on the unit sphere in , and is the signal strength. We term this configuration the spike setting. For completeness, we also conduct a power analysis where has full rank by setting , where represents the signal level. This configuration, termed the white noise setting, represents the covariance structure after adding white noise to the null covariance. Together, the spike and white noise settings cover both low- and full-rank deviations from the null covariance. For illustration, we consider sample sizes and , with dimension . Based on the universality results in Theorem 3.5 and empirical size performance, we conduct the power analysis using a Gaussian distribution for simplicity. Further empirical power comparisons under varying sample sizes and dimensions are provided in Table 1 of the Supplement table.












Figure 1 illustrates the empirical power performance of six test statistics across different signal levels and null covariance structures under the spike setting, where the dashed line represents the significance level of . Under this setting, the proposed operator norm-based statistics, and exhibit similar power, while the two Frobenius norm-based statistics and the normalized operator norm statistic perform comparably. When , the operator norm-based statistics and maintain appropriate test size and achieve the highest power, approaching quickly across all three covariance configurations. The normalized operator norm statistic and the two Frobenius norm-based statistics and also control the test size but exhibit lower power, among which slightly outperforms. The supremum norm-based statistic, however, exhibits significant size inflation and fails to detect signals, as its power curve does not increase with signal level. The pattern is similar for , except that the power curves for the supremum norm-based statistic remain near the dashed line, indicating limited power performance. This provides empirical evidence for the advantage of operator norm-based statistics over the other norms in the spike setting, particularly supporting the proposed statistic over . We further make comparisons under the white noise setting across various signal levels in Figure 2. In this setting, the combined statistic and the normalized statistic outperform other statistics. Because the white noise setting applies a uniform signal across eigenvector directions of with both large and small eigenvalues, the normalized statistic achieves higher power than , which is consistent with our power analysis Theorem 4.1. Notably, the combined statistic performs robustly in both the spike and white noise settings, demonstrating enhanced power without size inflation, as supported by the generalized universal bootstrap consistency Theorem 4.2. As observed in the spike setting, Frobenius norm-based statistics display lower power, while the supremum norm-based statistic fails to control size. In summary, when the covariance structure is known, we recommend using the statistic . When the covariance structure is unknown, we prefer the more robust combined statistic, .
5.2 Data application
We apply our test procedure to annual mean near-surface air temperature data from the global scale for the period , using the HadCRUT4 dataset (Morice et al. (2012)) and the Coupled Model Intercomparison Project Phase 5 (CMIP5, Taylor, Stouffer and Meehl (2012)), as detailed in Li et al. (2021). The global dataset includes monthly anomalies of near-surface air temperature across grid boxes. To reduce dimensionality, these grid boxes are aggregated into larger boxes, resulting in spatial grid boxes. To mitigate distribution shifts due to long-term trends, we divide the period into five decades and analyze the data for each decade separately. The monthly data for each decade are averaged over five-year intervals, yielding a temporal dimension for each ten-year period. To illustrate the rationale of the covariance test in this context, we first outline the commonly used modeling procedure in climatology. Following Li et al. (2021), one considers a high-dimensional linear regression with errors-in-variables
where is the vectorized observed mean temperature across spatial and temporal dimensions for the -th decades after , with dimension . The vectors , represent the expected unobserved climate response under the ANT and NAT forcings, and the unobserved noise, respectively. The coefficients , are unknown scaling factors of interest for statistical inference. To estimate , , in addition to the observed data , we have noisy observation fingerprints of the climate system
In our dataset, , . For privacy, the data are preprocessed by adding white noise, with the according covariance of the noise added to the hypothesized matrix defined below. The climate models also provide simulations , with provided in this dataset, which are assumed to follow the same distribution as the unobserved noise . To efficiently estimate , a typical assumption is that the natural variability simulated by the climate models matches the observed variability, i.e., the covariance of , , equals the covariance matrix of . Since is unobserved, a common choice for is the sample covariance of the simulations . Therefore, it is important to test the hypothesis
(59) |
As noted by Olonscheck and Notz (2017), this equivalence (59) is crucial for optimal fingerprint estimation in climate models. However, few studies have validated (59) with statistical evidence, which calls for the need to test this hypothesis.
We construct the data matrix by combining quantities for and for , where and . Here . We apply several statistics to the data matrix with the hypothesized matrix : the proposed operator norm-based statistics (Opn) and (Com), the operator norm-based statistics (Roy), the Frobenius norm-based statistics (Lfn) and (Ufn), and the supremum norm-based statistic (Supn). For comparison, we generate i.i.d. Gaussian samples for as a control group. The observed data forms the observation group. The p-value results for these statistics are summarized in Table 2. The supremum norm-based statistic fails to reject the observation group for all periods but rejects the control group in the periods , , and . The Frobenius norm-based statistics and fail to reject the observation group in the periods and , and rejects the control group during . Only the operator norm-based statistics , , and reject the null hypothesis in the observation group while not rejecting it in the control group for all years. This suggests that the commonly assumed hypothesis (59) should be rejected, and a more suitable assumption should be used to estimate and for all .
p-value | observation group | control group | ||||||||||
year | Supn | Lfn | Ufn | Roy | Opn | Com | Supn | Lfn | Ufn | Roy | Opn | Com |
1960-1970 | .783 | .006 | .001 | .000 | .000 | .000 | .002 | .566 | .521 | .805 | .646 | .734 |
1970-1980 | .406 | .146 | .115 | .000 | .000 | .000 | .004 | .230 | .243 | .401 | .452 | .427 |
1980-1990 | .249 | .001 | .000 | .000 | .000 | .000 | .000 | .419 | .451 | .638 | .713 | .677 |
1990-2000 | .905 | .634 | .410 | .000 | .000 | .000 | .223 | .072 | .042 | .445 | .383 | .414 |
2000-2010 | .984 | .002 | .001 | .000 | .000 | .000 | .719 | .368 | .376 | .244 | .288 | .268 |
In summary, the operator norm-based statistics demonstrate strong performance across various covariance structures and signal configurations, with the universal bootstrap ensuring the consistency of tests constructed from diverse combinations of operator norms. Numerical results align with the theoretical framework, indicating that the proposed tests exhibit robust power properties and that the universal bootstrap procedures maintain appropriate size control.
Fang Yao is the corresponding author. This research is partially supported by the National Key Research and Development Program of China (No. 2022YFA1003801), the National Natural Science Foundation of China (No. 12292981, 11931001), the LMAM, the Fundamental Research Funds for the Central Universities, Peking University, and LMEQF.
References
- Adamczak et al. (2011) {barticle}[author] \bauthor\bsnmAdamczak, \bfnmRadosław\binitsR., \bauthor\bsnmLitvak, \bfnmAlexander E\binitsA. E., \bauthor\bsnmPajor, \bfnmAlain\binitsA. and \bauthor\bsnmTomczak-Jaegermann, \bfnmNicole\binitsN. (\byear2011). \btitleSharp bounds on the rate of convergence of the empirical covariance matrix. \bjournalComptes Rendus. Mathématique \bvolume349 \bpages195–200. \endbibitem
- Bai and Yao (2012) {barticle}[author] \bauthor\bsnmBai, \bfnmZhidong\binitsZ. and \bauthor\bsnmYao, \bfnmJianfeng\binitsJ. (\byear2012). \btitleOn sample eigenvalues in a generalized spiked population model. \bjournalJournal of Multivariate Analysis \bvolume106 \bpages167–177. \endbibitem
- Baik, Ben Arous and Péché (2005) {barticle}[author] \bauthor\bsnmBaik, \bfnmJinho\binitsJ., \bauthor\bsnmBen Arous, \bfnmGérard\binitsG. and \bauthor\bsnmPéché, \bfnmSandrine\binitsS. (\byear2005). \btitlePhase transition of the largest eigenvalue for nonnull complex sample covariance matrices. \endbibitem
- Baik and Silverstein (2006) {barticle}[author] \bauthor\bsnmBaik, \bfnmJinho\binitsJ. and \bauthor\bsnmSilverstein, \bfnmJack W\binitsJ. W. (\byear2006). \btitleEigenvalues of large sample covariance matrices of spiked population models. \bjournalJournal of multivariate analysis \bvolume97 \bpages1382–1408. \endbibitem
- Bao, Pan and Zhou (2015) {barticle}[author] \bauthor\bsnmBao, \bfnmZhigang\binitsZ., \bauthor\bsnmPan, \bfnmGuangming\binitsG. and \bauthor\bsnmZhou, \bfnmWang\binitsW. (\byear2015). \btitleUniversality for the largest eigenvalue of sample covariance matrices with general population. \endbibitem
- Bianchi et al. (2011) {barticle}[author] \bauthor\bsnmBianchi, \bfnmPascal\binitsP., \bauthor\bsnmDebbah, \bfnmMerouane\binitsM., \bauthor\bsnmMa\̈text{i}da, \bfnmMylène\binitsM. and \bauthor\bsnmNajim, \bfnmJamal\binitsJ. (\byear2011). \btitlePerformance of statistical tests for single-source detection using random matrix theory. \bjournalIEEE Transactions on Information theory \bvolume57 \bpages2400–2419. \endbibitem
- Bunea and Xiao (2015) {barticle}[author] \bauthor\bsnmBunea, \bfnmFlorentina\binitsF. and \bauthor\bsnmXiao, \bfnmLuo\binitsL. (\byear2015). \btitleOn the sample covariance matrix estimator of reduced effective rank population matrices, with applications to fPCA. \endbibitem
- Cai and Jiang (2011) {barticle}[author] \bauthor\bsnmCai, \bfnmT Tony\binitsT. T. and \bauthor\bsnmJiang, \bfnmTiefeng\binitsT. (\byear2011). \btitleLimiting laws of coherence of random matrices with applications to testing covariance structure and construction of compressed sensing matrices. \endbibitem
- Cai, Liu and Xia (2013) {barticle}[author] \bauthor\bsnmCai, \bfnmTony\binitsT., \bauthor\bsnmLiu, \bfnmWeidong\binitsW. and \bauthor\bsnmXia, \bfnmYin\binitsY. (\byear2013). \btitleTwo-sample covariance matrix testing and support recovery in high-dimensional and sparse settings. \bjournalJournal of the American Statistical Association \bvolume108 \bpages265–277. \endbibitem
- Cai, Liu and Xia (2014) {barticle}[author] \bauthor\bsnmCai, \bfnmT Tony\binitsT. T., \bauthor\bsnmLiu, \bfnmWeidong\binitsW. and \bauthor\bsnmXia, \bfnmYin\binitsY. (\byear2014). \btitleTwo-sample test of high dimensional means under dependence. \bjournalJournal of the Royal Statistical Society Series B: Statistical Methodology \bvolume76 \bpages349–372. \endbibitem
- Cai and Ma (2013) {barticle}[author] \bauthor\bsnmCai, \bfnmT Tony\binitsT. T. and \bauthor\bsnmMa, \bfnmZongming\binitsZ. (\byear2013). \btitleOptimal hypothesis testing for high dimensional covariance matrices. \endbibitem
- Chen and Qin (2010) {barticle}[author] \bauthor\bsnmChen, \bfnmSong Xi\binitsS. X. and \bauthor\bsnmQin, \bfnmYing-Li\binitsY.-L. (\byear2010). \btitleA two-sample test for high-dimensional data with applications to gene-set testing. \endbibitem
- Chen, Qiu and Zhang (2023) {barticle}[author] \bauthor\bsnmChen, \bfnmSong Xi\binitsS. X., \bauthor\bsnmQiu, \bfnmYumou\binitsY. and \bauthor\bsnmZhang, \bfnmShuyi\binitsS. (\byear2023). \btitleSharp optimality for high-dimensional covariance testing under sparse signals. \bjournalThe Annals of Statistics \bvolume51 \bpages1921–1945. \endbibitem
- Chen, Zhang and Zhong (2010) {barticle}[author] \bauthor\bsnmChen, \bfnmSong Xi\binitsS. X., \bauthor\bsnmZhang, \bfnmLi-Xin\binitsL.-X. and \bauthor\bsnmZhong, \bfnmPing-Shou\binitsP.-S. (\byear2010). \btitleTests for high-dimensional covariance matrices. \bjournalJournal of the American Statistical Association \bvolume105 \bpages810–819. \endbibitem
- Chernozhukov, Chetverikov and Kato (2013) {barticle}[author] \bauthor\bsnmChernozhukov, \bfnmVictor\binitsV., \bauthor\bsnmChetverikov, \bfnmDenis\binitsD. and \bauthor\bsnmKato, \bfnmKengo\binitsK. (\byear2013). \btitleGaussian approximations and multiplier bootstrap for maxima of sums of high-dimensional random vectors. \endbibitem
- Chernozhukov, Chetverikov and Kato (2017) {barticle}[author] \bauthor\bsnmChernozhukov, \bfnmVictor\binitsV., \bauthor\bsnmChetverikov, \bfnmDenis\binitsD. and \bauthor\bsnmKato, \bfnmKengo\binitsK. (\byear2017). \btitleCentral limit theorems and bootstrap in high dimensions. \endbibitem
- Chernozhukov, Chetverikov and Koike (2023) {barticle}[author] \bauthor\bsnmChernozhukov, \bfnmVictor\binitsV., \bauthor\bsnmChetverikov, \bfnmDenis\binitsD. and \bauthor\bsnmKoike, \bfnmYuta\binitsY. (\byear2023). \btitleNearly optimal central limit theorem and bootstrap approximations in high dimensions. \bjournalThe Annals of Applied Probability \bvolume33 \bpages2374–2425. \endbibitem
- Chernozhuokov et al. (2022) {barticle}[author] \bauthor\bsnmChernozhuokov, \bfnmVictor\binitsV., \bauthor\bsnmChetverikov, \bfnmDenis\binitsD., \bauthor\bsnmKato, \bfnmKengo\binitsK. and \bauthor\bsnmKoike, \bfnmYuta\binitsY. (\byear2022). \btitleImproved central limit theorem and bootstrap approximations in high dimensions. \bjournalThe Annals of Statistics \bvolume50 \bpages2562–2586. \endbibitem
- Couillet, Debbah and Silverstein (2011) {barticle}[author] \bauthor\bsnmCouillet, \bfnmRomain\binitsR., \bauthor\bsnmDebbah, \bfnmMérouane\binitsM. and \bauthor\bsnmSilverstein, \bfnmJack W\binitsJ. W. (\byear2011). \btitleA deterministic equivalent for the analysis of correlated MIMO multiple access channels. \bjournalIEEE Transactions on Information Theory \bvolume57 \bpages3493–3514. \endbibitem
- Ding, Hu and Wang (2024) {barticle}[author] \bauthor\bsnmDing, \bfnmXiucai\binitsX., \bauthor\bsnmHu, \bfnmYichen\binitsY. and \bauthor\bsnmWang, \bfnmZhenggang\binitsZ. (\byear2024). \btitleTwo sample test for covariance matrices in ultra-high dimension. \bjournalJournal of the American Statistical Association \bpages1–12. \endbibitem
- Ding and Wang (2023) {barticle}[author] \bauthor\bsnmDing, \bfnmXiucai\binitsX. and \bauthor\bsnmWang, \bfnmZhenggang\binitsZ. (\byear2023). \btitleGlobal and local CLTs for linear spectral statistics of general sample covariance matrices when the dimension is much larger than the sample size with applications. \bjournalarXiv preprint arXiv:2308.08646. \endbibitem
- Ding and Yang (2018) {barticle}[author] \bauthor\bsnmDing, \bfnmXiucai\binitsX. and \bauthor\bsnmYang, \bfnmFan\binitsF. (\byear2018). \btitleA necessary and sufficient condition for edge universality at the largest singular values of covariance matrices. \bjournalThe Annals of Applied Probability \bvolume28 \bpages1679–1738. \endbibitem
- Ding and Yang (2022) {barticle}[author] \bauthor\bsnmDing, \bfnmXiucai\binitsX. and \bauthor\bsnmYang, \bfnmFan\binitsF. (\byear2022). \btitleTracy-Widom distribution for heterogeneous Gram matrices with applications in signal detection. \bjournalIEEE Transactions on Information Theory \bvolume68 \bpages6682–6715. \endbibitem
- Efron (1979) {barticle}[author] \bauthor\bsnmEfron, \bfnmBradley\binitsB. (\byear1979). \btitleBootstrap Methods: Another Look at the Jackknife. \bjournalThe Annals of Statistics \bvolume7 \bpages1–26. \endbibitem
- El Karoui (2007) {barticle}[author] \bauthor\bsnmEl Karoui, \bfnmNoureddine\binitsN. (\byear2007). \btitleTracy–Widom limit for the largest eigenvalue of a large class of complex sample covariance matrices. \endbibitem
- El Karoui and Purdom (2019) {binproceedings}[author] \bauthor\bsnmEl Karoui, \bfnmNoureddine\binitsN. and \bauthor\bsnmPurdom, \bfnmElizabeth\binitsE. (\byear2019). \btitleThe non-parametric bootstrap and spectral analysis in moderate and high-dimension. In \bbooktitleThe 22nd International Conference on Artificial Intelligence and Statistics \bpages2115–2124. \bpublisherPMLR. \endbibitem
- Erdős, Knowles and Yau (2013) {binproceedings}[author] \bauthor\bsnmErdős, \bfnmLászló\binitsL., \bauthor\bsnmKnowles, \bfnmAntti\binitsA. and \bauthor\bsnmYau, \bfnmHorng-Tzer\binitsH.-T. (\byear2013). \btitleAveraging fluctuations in resolvents of random band matrices. In \bbooktitleAnnales Henri Poincaré \bvolume14 \bpages1837–1926. \bpublisherSpringer. \endbibitem
- Erdős, Yau and Yin (2012) {barticle}[author] \bauthor\bsnmErdős, \bfnmLászló\binitsL., \bauthor\bsnmYau, \bfnmHorng-Tzer\binitsH.-T. and \bauthor\bsnmYin, \bfnmJun\binitsJ. (\byear2012). \btitleRigidity of eigenvalues of generalized Wigner matrices. \bjournalAdvances in Mathematics \bvolume229 \bpages1435–1515. \endbibitem
- Fang and Koike (2024) {barticle}[author] \bauthor\bsnmFang, \bfnmXiao\binitsX. and \bauthor\bsnmKoike, \bfnmYuta\binitsY. (\byear2024). \btitleLarge-dimensional central limit theorem with fourth-moment error bounds on convex sets and balls. \bjournalThe Annals of Applied Probability \bvolume34 \bpages2065–2106. \endbibitem
- Giessing (2023) {barticle}[author] \bauthor\bsnmGiessing, \bfnmAlexander\binitsA. (\byear2023). \btitleGaussian and Bootstrap Approximations for Suprema of Empirical Processes. \bjournalarXiv preprint arXiv:2309.01307. \endbibitem
- Han, Xu and Zhou (2018) {barticle}[author] \bauthor\bsnmHan, \bfnmFang\binitsF., \bauthor\bsnmXu, \bfnmSheng\binitsS. and \bauthor\bsnmZhou, \bfnmWen-Xin\binitsW.-X. (\byear2018). \btitleOn Gaussian comparison inequality and its application to spectral analysis of large random matrices. \endbibitem
- Hu and Lu (2022) {barticle}[author] \bauthor\bsnmHu, \bfnmHong\binitsH. and \bauthor\bsnmLu, \bfnmYue M\binitsY. M. (\byear2022). \btitleUniversality laws for high-dimensional learning with random features. \bjournalIEEE Transactions on Information Theory \bvolume69 \bpages1932–1964. \endbibitem
- Ji and Park (2021) {barticle}[author] \bauthor\bsnmJi, \bfnmHong Chang\binitsH. C. and \bauthor\bsnmPark, \bfnmJaewhi\binitsJ. (\byear2021). \btitleTracy-Widom limit for free sum of random matrices. \bjournalarXiv preprint arXiv:2110.05147. \endbibitem
- Jiang (2004) {barticle}[author] \bauthor\bsnmJiang, \bfnmTiefeng\binitsT. (\byear2004). \btitleThe asymptotic distributions of the largest entries of sample correlation matrices. \endbibitem
- Jiang and Bai (2021) {barticle}[author] \bauthor\bsnmJiang, \bfnmDandan\binitsD. and \bauthor\bsnmBai, \bfnmZhidong\binitsZ. (\byear2021). \btitleGeneralized four moment theorem and an application to CLT for spiked eigenvalues of high-dimensional covariance matrices. \endbibitem
- Johnstone (2001) {barticle}[author] \bauthor\bsnmJohnstone, \bfnmIain M\binitsI. M. (\byear2001). \btitleOn the distribution of the largest eigenvalue in principal components analysis. \bjournalThe Annals of statistics \bvolume29 \bpages295–327. \endbibitem
- Johnstone and Paul (2018) {barticle}[author] \bauthor\bsnmJohnstone, \bfnmIain M\binitsI. M. and \bauthor\bsnmPaul, \bfnmDebashis\binitsD. (\byear2018). \btitlePCA in high dimensions: An orientation. \bjournalProceedings of the IEEE \bvolume106 \bpages1277–1292. \endbibitem
- Karoui (2003) {barticle}[author] \bauthor\bsnmKaroui, \bfnmNoureddine El\binitsN. E. (\byear2003). \btitleOn the largest eigenvalue of Wishart matrices with identity covariance when n, p and p/n tend to infinity. \bjournalarXiv preprint math/0309355. \endbibitem
- Ke (2016) {barticle}[author] \bauthor\bsnmKe, \bfnmZheng Tracy\binitsZ. T. (\byear2016). \btitleDetecting rare and weak spikes in large covariance matrices. \bjournalarXiv preprint arXiv:1609.00883. \endbibitem
- Knowles and Yin (2017) {barticle}[author] \bauthor\bsnmKnowles, \bfnmAntti\binitsA. and \bauthor\bsnmYin, \bfnmJun\binitsJ. (\byear2017). \btitleAnisotropic local laws for random matrices. \bjournalProbability Theory and Related Fields \bvolume169 \bpages257–352. \endbibitem
- Koltchinskii and Lounici (2017) {barticle}[author] \bauthor\bsnmKoltchinskii, \bfnmVladimir\binitsV. and \bauthor\bsnmLounici, \bfnmKarim\binitsK. (\byear2017). \btitleConcentration inequalities and moment bounds for sample covariance operators. \bjournalBernoulli \bpages110–133. \endbibitem
- Kong and Valiant (2017) {barticle}[author] \bauthor\bsnmKong, \bfnmWeihao\binitsW. and \bauthor\bsnmValiant, \bfnmGregory\binitsG. (\byear2017). \btitleSpectrum estimation from samples. \endbibitem
- Ledoit and Wolf (2015) {barticle}[author] \bauthor\bsnmLedoit, \bfnmOlivier\binitsO. and \bauthor\bsnmWolf, \bfnmMichael\binitsM. (\byear2015). \btitleSpectrum estimation: A unified framework for covariance matrix estimation and PCA in large dimensions. \bjournalJournal of Multivariate Analysis \bvolume139 \bpages360–384. \endbibitem
- Lee and Schnelli (2016) {barticle}[author] \bauthor\bsnmLee, \bfnmJi Oon\binitsJ. O. and \bauthor\bsnmSchnelli, \bfnmKevin\binitsK. (\byear2016). \btitleTracy–Widom distribution for the largest eigenvalue of real sample covariance matrices with general population. \endbibitem
- Li et al. (2021) {barticle}[author] \bauthor\bsnmLi, \bfnmYan\binitsY., \bauthor\bsnmChen, \bfnmKun\binitsK., \bauthor\bsnmYan, \bfnmJun\binitsJ. and \bauthor\bsnmZhang, \bfnmXuebin\binitsX. (\byear2021). \btitleUncertainty in optimal fingerprinting is underestimated. \bjournalEnvironmental Research Letters \bvolume16 \bpages084043. \endbibitem
- Lopes (2022a) {barticle}[author] \bauthor\bsnmLopes, \bfnmMiles E\binitsM. E. (\byear2022a). \btitleImproved rates of bootstrap approximation for the operator norm: A coordinate-free approach. \bjournalarXiv preprint arXiv:2208.03050. \endbibitem
- Lopes (2022b) {barticle}[author] \bauthor\bsnmLopes, \bfnmMiles E\binitsM. E. (\byear2022b). \btitleCentral limit theorem and bootstrap approximation in high dimensions: Near rates via implicit smoothing. \bjournalThe Annals of Statistics \bvolume50 \bpages2492–2513. \endbibitem
- Lopes, Blandino and Aue (2019) {barticle}[author] \bauthor\bsnmLopes, \bfnmMiles E\binitsM. E., \bauthor\bsnmBlandino, \bfnmAndrew\binitsA. and \bauthor\bsnmAue, \bfnmAlexander\binitsA. (\byear2019). \btitleBootstrapping spectral statistics in high dimensions. \bjournalBiometrika \bvolume106 \bpages781–801. \endbibitem
- Lopes, Erichson and Mahoney (2023) {barticle}[author] \bauthor\bsnmLopes, \bfnmMiles E\binitsM. E., \bauthor\bsnmErichson, \bfnmN Benjamin\binitsN. B. and \bauthor\bsnmMahoney, \bfnmMichael W\binitsM. W. (\byear2023). \btitleBootstrapping the operator norm in high dimensions: Error estimation for covariance matrices and sketching. \bjournalBernoulli \bvolume29 \bpages428–450. \endbibitem
- Lopes, Lin and Müller (2020) {barticle}[author] \bauthor\bsnmLopes, \bfnmMiles E\binitsM. E., \bauthor\bsnmLin, \bfnmZhenhua\binitsZ. and \bauthor\bsnmMüller, \bfnmHans-Georg\binitsH.-G. (\byear2020). \btitleBootstrapping max statistics in high dimensions: Near-parametric rates under weak variance decay and application to functional and multinomial data. \endbibitem
- Montanari and Saeed (2022) {binproceedings}[author] \bauthor\bsnmMontanari, \bfnmAndrea\binitsA. and \bauthor\bsnmSaeed, \bfnmBasil N\binitsB. N. (\byear2022). \btitleUniversality of empirical risk minimization. In \bbooktitleConference on Learning Theory \bpages4310–4312. \bpublisherPMLR. \endbibitem
- Morice et al. (2012) {barticle}[author] \bauthor\bsnmMorice, \bfnmColin P\binitsC. P., \bauthor\bsnmKennedy, \bfnmJohn J\binitsJ. J., \bauthor\bsnmRayner, \bfnmNick A\binitsN. A. and \bauthor\bsnmJones, \bfnmPhil D\binitsP. D. (\byear2012). \btitleQuantifying uncertainties in global and regional temperature change using an ensemble of observational estimates: The HadCRUT4 data set. \bjournalJournal of Geophysical Research: Atmospheres \bvolume117. \endbibitem
- Olonscheck and Notz (2017) {barticle}[author] \bauthor\bsnmOlonscheck, \bfnmDirk\binitsD. and \bauthor\bsnmNotz, \bfnmDirk\binitsD. (\byear2017). \btitleConsistently estimating internal climate variability from climate model simulations. \bjournalJournal of Climate \bvolume30 \bpages9555–9573. \endbibitem
- Onatski, Moreira and Hallin (2013) {barticle}[author] \bauthor\bsnmOnatski, \bfnmAlexei\binitsA., \bauthor\bsnmMoreira, \bfnmMarcelo J\binitsM. J. and \bauthor\bsnmHallin, \bfnmMarc\binitsM. (\byear2013). \btitleAsymptotic power of sphericity tests for high-dimensional data. \endbibitem
- Paul (2007) {barticle}[author] \bauthor\bsnmPaul, \bfnmDebashis\binitsD. (\byear2007). \btitleAsymptotics of sample eigenstructure for a large dimensional spiked covariance model. \bjournalStatistica Sinica \bpages1617–1642. \endbibitem
- Qiu, Li and Yao (2023) {barticle}[author] \bauthor\bsnmQiu, \bfnmJiaxin\binitsJ., \bauthor\bsnmLi, \bfnmZeng\binitsZ. and \bauthor\bsnmYao, \bfnmJianfeng\binitsJ. (\byear2023). \btitleAsymptotic normality for eigenvalue statistics of a general sample covariance matrix when p/n and applications. \bjournalThe Annals of Statistics \bvolume51 \bpages1427–1451. \endbibitem
- Schäfer and Strimmer (2005) {barticle}[author] \bauthor\bsnmSchäfer, \bfnmJuliane\binitsJ. and \bauthor\bsnmStrimmer, \bfnmKorbinian\binitsK. (\byear2005). \btitleA shrinkage approach to large-scale covariance matrix estimation and implications for functional genomics. \bjournalStatistical applications in genetics and molecular biology \bvolume4. \endbibitem
- Taylor, Stouffer and Meehl (2012) {barticle}[author] \bauthor\bsnmTaylor, \bfnmKarl E\binitsK. E., \bauthor\bsnmStouffer, \bfnmRonald J\binitsR. J. and \bauthor\bsnmMeehl, \bfnmGerald A\binitsG. A. (\byear2012). \btitleAn overview of CMIP5 and the experiment design. \bjournalBulletin of the American meteorological Society \bvolume93 \bpages485–498. \endbibitem
- Tsukuma (2016) {barticle}[author] \bauthor\bsnmTsukuma, \bfnmHisayuki\binitsH. (\byear2016). \btitleEstimation of a high-dimensional covariance matrix with the Stein loss. \bjournalJournal of Multivariate Analysis \bvolume148 \bpages1–17. \endbibitem
- Xu, Zhang and Wu (2019) {barticle}[author] \bauthor\bsnmXu, \bfnmMengyu\binitsM., \bauthor\bsnmZhang, \bfnmDanna\binitsD. and \bauthor\bsnmWu, \bfnmWei Biao\binitsW. B. (\byear2019). \btitlePearson’s chi-squared statistics: approximation theory and beyond. \bjournalBiometrika \bvolume106 \bpages716–723. \endbibitem
- Yao and Lopes (2021) {barticle}[author] \bauthor\bsnmYao, \bfnmJunwen\binitsJ. and \bauthor\bsnmLopes, \bfnmMiles E\binitsM. E. (\byear2021). \btitleRates of bootstrap approximation for eigenvalues in high-dimensional PCA. \bjournalarXiv preprint arXiv:2104.07328. \endbibitem
- Yu, Li and Xue (2024) {barticle}[author] \bauthor\bsnmYu, \bfnmXiufan\binitsX., \bauthor\bsnmLi, \bfnmDanning\binitsD. and \bauthor\bsnmXue, \bfnmLingzhou\binitsL. (\byear2024). \btitleFisher’s combined probability test for high-dimensional covariance matrices. \bjournalJournal of the American Statistical Association \bvolume119 \bpages511–524. \endbibitem
- Yu, Zhao and Zhou (2024) {barticle}[author] \bauthor\bsnmYu, \bfnmLong\binitsL., \bauthor\bsnmZhao, \bfnmPeng\binitsP. and \bauthor\bsnmZhou, \bfnmWang\binitsW. (\byear2024). \btitleTesting the number of common factors by bootstrapped sample covariance matrix in high-dimensional factor models. \bjournalJournal of the American Statistical Association \bpages1–12. \endbibitem
- Zhai (2018) {barticle}[author] \bauthor\bsnmZhai, \bfnmAlex\binitsA. (\byear2018). \btitleA high-dimensional CLT in distance with near optimal convergence rate. \bjournalProbability Theory and Related Fields \bvolume170 \bpages821–845. \endbibitem
- Zhou, Bai and Hu (2023) {barticle}[author] \bauthor\bsnmZhou, \bfnmHuanchao\binitsH., \bauthor\bsnmBai, \bfnmZhidong\binitsZ. and \bauthor\bsnmHu, \bfnmJiang\binitsJ. (\byear2023). \btitleThe Limiting Spectral Distribution of Large-Dimensional General Information-Plus-Noise-Type Matrices. \bjournalJournal of Theoretical Probability \bvolume36 \bpages1203–1226. \endbibitem
- Zhou et al. (2024) {barticle}[author] \bauthor\bsnmZhou, \bfnmHuanchao\binitsH., \bauthor\bsnmHu, \bfnmJiang\binitsJ., \bauthor\bsnmBai, \bfnmZhidong\binitsZ. and \bauthor\bsnmSilverstein, \bfnmJack W\binitsJ. W. (\byear2024). \btitleAnalysis of the Limiting Spectral Distribution of Large-dimensional General Information-Plus-Noise-Type Matrices. \bjournalJournal of Theoretical Probability \bvolume37 \bpages1199–1229. \endbibitem
- Zhu et al. (2017) {barticle}[author] \bauthor\bsnmZhu, \bfnmLingxue\binitsL., \bauthor\bsnmLei, \bfnmJing\binitsJ., \bauthor\bsnmDevlin, \bfnmBernie\binitsB. and \bauthor\bsnmRoeder, \bfnmKathryn\binitsK. (\byear2017). \btitleTesting high-dimensional covariance matrices, with application to detecting schizophrenia risk genes. \bjournalThe annals of applied statistics \bvolume11 \bpages1810. \endbibitem