A survey of some recent developments in measures of association
Abstract
This paper surveys some recent developments in measures of association related to a new coefficient of correlation introduced by the author. A straightforward extension of this coefficient to standard Borel spaces (which includes all Polish spaces), overlooked in the literature so far, is proposed at the end of the survey.
Key words and phrases. Correlation, dependence, measures of association, standard Borel space
2020 Mathematics Subject Classification. 62H20, 62H15.
In honor of friend and teacher Prof. Rajeeva L. Karandikar on the occasion of his 65th birthday.
1 Introduction
Measuring associations between variables is one of the central goals of data analysis. Arguably, the three most popular classical measures of association are Pearson’s correlation coefficient, Spearman’s , and Kendall’s . Although these coefficients are powerful for detecting monotonic associations, a practical problem is that they are not effective for detecting associations that are not monotonic. There have been many proposals to address this deficiency of the classical coefficients [66], such as the maximal correlation coefficient [59, 45, 92, 17], various coefficients based on joint cumulative distribution functions and ranks [43, 34, 54, 8, 83, 123, 124, 125, 29, 89, 61, 13, 94, 95, 30, 122], kernel-based methods [86, 48, 50, 99, 130], information theoretic coefficients [71, 74, 93], coefficients based on copulas [32, 76, 108, 98, 127], and coefficients based on pairwise distances [117, 115, 58, 41, 77].
This survey is about some recent developments in this area, beginning with a new coefficient of correlation proposed by the author in the paper [22]. This coefficient has the following desirable features: (a) It has a simple expression, like the classical coefficients. (b) It is a consistent estimator of a measure of dependence which is if and only if the variables are independent and if and only if one is a measurable function of the other. (c) It has a simple asymptotic theory under the hypothesis of independence.
The new coefficient is defined as follows. Let be a pair of random variables defined on the same probability space, where is not a constant. Let be i.i.d. pairs of random variables with the same law as , where . Rearrange the data as such that . (Note that is just the -value ‘paired with’ in the original data, and not the order statistic of the ’s.) If there are ties among the ’s, then choose an increasing rearrangement as above by breaking ties uniformly at random. Let be the number of such that , and let to be the number of such that . Then define
(1.1) |
This is the correlation coefficient proposed in [22]. When there are no ties among the ’s, is just a permutation of , and so the denominator in the above expression is just . The following theorem is the main consistency result for .
Theorem 1.1 ([22]).
If is not almost surely a constant, then as , converges almost surely to the deterministic limit
(1.2) |
where is the law of . This limit belongs to the interval . It is if and only if and are independent, and it is if and only if there is a measurable function such that almost surely.
The limiting value appeared in the literature prior to [22], in a paper of Dette, Siburg, and Stoimenov [32] (see also [43, 70]). The paper [32] gave a copula-based estimator for when and are continuous, that is consistent under smoothness assumptions on the copula and is computable in time for an optimal choice of tuning parameters.
Note that neither nor the limiting value is symmetric in and . A symmetrized version of can be constructed by taking the maximum of and .
2 Why does it work?
The complete proof of Theorem 1.1 is available in the supplementary materials of [22], and also in the arXiv version of the paper. It is not too hard to see why has the properties listed in Theorem 1.1, although filling in the details takes some work. The proof of convergence of to is less obvious. The following is a very rough sketch of the proof, reproduced from a similar discussion in [22].
For simplicity, consider only the case of continuous and , where the denominator in (1.2) is simply . First, note that by the Glivenko–Cantelli theorem, , where is the cumulative distribution function of . Thus,
(2.1) |
where is the unique index such that is immediately to the right of if we arrange the ’s in increasing order. If is the rightmost value, define arbitrarily; it does not matter since the contribution of a single term in the above sum is of order . Next, observe that for any ,
(2.2) |
where is the law of . This is true because the integrand is between and and outside.
Suppose that we condition on . Since is likely to be very close to , the random variables and are likely to be approximately i.i.d. after this conditioning. This leads to the approximation
This gives
Combining this with (2.2), we get
But note that , and . Thus,
Therefore by (2.1),
where the last identity holds because , as shown above. This establishes the convergence of to . Concentration inequalities are then used to show that almost surely.
3 Asymptotic distribution
Let , and be as in the previous section. For each , let and . Let . Define
(3.1) |
where are independent copies of . The following theorem gives the limiting distribution of under the null hypothesis that and are independent.
Theorem 3.1 ([22]).
Suppose that and are independent. Then converges to in distribution as , where is given by the formula (3.1) stated above. The number is strictly positive if is not a constant, and equals if is continuous.
The reason why does not depend on the law of if is continuous is that in this case and are Uniform random variables, which implies that the expectations in (3.1) do not depend on the law of . If is not continuous, then may depend on the law of . For example, it is not hard to show that if is a Bernoulli random variable, then . Fortunately, if is not continuous, there is a simple way to estimate from the data using the estimator
where , , and are defined as follows. For each , let
(3.2) |
(Note that and are different than and defined earlier.) Let be an increasing rearrangement of . Let for . Define
Then the following holds.
Theorem 3.2 ([22]).
The estimator can be computed in time , and converges to almost surely as .
The question of proving a central limit theorem for in the absence of independence is much more difficult than the independent case. This was left as an open question in [22] and recently resolved in complete generality by Lin and Han [73], following earlier proofs of Deb, Ghosal, and Sen [31] and Shi, Drton, and Han [104] under additional assumptions. Lin and Han [73] also give a consistent estimator of the asymptotic variance of in the absence of independence, solving another question that was left open in [22]. A central limit theorem for the symmetrized version of (defined as the maximum of and ) under the hypothesis of independence was proved by Zhang [129].
4 Power for testing independence
A deficiency of , as already pointed out in [22] through simulated examples, is that it has low power for testing independence against ‘standard’ alternatives, such as linear or monotone associations. This was theoretically confirmed by Shi, Drton, and Han [105], where it was shown that the test of independence using is rate-suboptimal against a family of local alternatives, whereas three other nonparametric tests of independence proposed in [61, 13, 8, 125] are rate-optimal. Like , the three competing test statistics considered in [105] are also computable in time. Similar results were obtained for a different type of competing test statistic by Cao and Bickel [21].
A more detailed analysis of the power properties of was carried out by Auddy, Deb, and Nandy [2], where the asymptotic distribution of under any changing sequence of alternatives converging to the null hypothesis of independence was computed. This analysis yielded exact detection thresholds and limiting power under natural alternatives converging to the null, such as mixture models, rotation models and noisy nonparametric regression. The detection boundary lies at distance from the null, instead of the more standard . This is similar to the power properties of other ‘graph-based’ statistics for testing independence, such as the Friedman–Rafsky statistic [41, 11].
A proposal for ‘boosting’ the power of for testing independence, by incorporating multiple nearby ranks instead of only the nearest ones, was recently proposed by Lin and Han [72]. This modified estimator was shown to attain near-optimal rates of power against certain classes of alternative hypotheses.
The conceptual reason behind the absence of local power of statistics such as was explained by Bickel [12]. An interesting question that remains unexplained is the following. It is seen in simulations that although has low power for testing independence against standard alternatives such as linear and monotone, it becomes more powerful as the signal starts to get more and more oscillatory [22]. This gives an advantage over other coefficients in applications where oscillatory signals arise naturally [25, 97], and suggests that may be efficient for certain kinds of local alternatives. No result of this sort has yet been proven.
5 Multivariate extensions
Many methods have been proposed for testing independence nonparametrically in the multivariate setting. This includes classical tests [88, 49, 117] as well as a flurry of recent ones proposed in the last ten years [57, 56, 58, 131, 124, 68, 30, 106, 10, 107, 31].
Most of these papers are concerned only with testing independence, and not with measuring the strength of dependence as measured by a correlation coefficient such as . Unfortunately, the coefficient and many other popular univariate coefficients do not readily generalize to the multivariate setting because they are based on ranks. Statisticians have started taking a new look at this old problem in recent years by considering a multivariate notion of rank defined using optimal transport. Roughly speaking, the idea is as follows. Let be a ‘reference measure’ in , akin to the uniform distribution on in . Given any probability measure in , let be the map that ‘optimally transports’ to — that is, if then , and minimizes among all such . By a theorem of McCann [80], such a map exists and is unique if and are both absolutely continuous with respect to Lebesgue measure. For example, when , is just the cumulative distribution function of , which transforms into a Uniform random variable. For properties of this map, see, e.g., Figalli [38] and Hallin et al. [51].
The above idea suggests a natural definition of multivariate rank. If are i.i.d. samples from , one can try to estimate using this data. Let be an estimate. Then can act as a ‘multivariate rank’ of among , divided by . Since , we can then assume that is approximately distributed according to , and then try to test for independence of random vectors using a test for independence that works when the marginal distributions are both . This idea has been made precise in a number of recent works in the statistics literature, such as Chernozhukov et al. [27], Deb, Ghosal, and Sen [31], Deb and Sen [30], Hallin et al. [51], Manole et al. [78], Shi, Drton, and Han [106], Ghosal and Sen [47], Mordant and Segers [82] and Shi et al. [107, 103]. For a survey and further discussions, see Han [52].
A direct generalization of to higher dimensional spaces has not been proposed so far, although the variant proposed in Deb et al. [31] satisfies the same properties as provided that the space on which takes values admits a nonnegative definite kernel and the space on which takes values has a metric. This covers most spaces that arise in practice. There are a couple of other generalizations, proposed by Azadkia and Chatterjee [3] and Gamboa et al. [44], on measuring the dependence between a univariate random variable and a random vector . The coefficient proposed in [3] (discussed in detail in Section 6) is based on a generalization of the ideas behind the construction of . The coefficient proposed in [44] combines the construction of with ideas from the theory of Sobol indices. A new contribution of the present paper is a simple generalization of to standard Borel spaces, which has been overlooked in the literature until now. This is presented in Section 8.
6 Measuring conditional dependence
The problem of measuring conditional dependence has received less attention than the problem of measuring unconditional dependence, partly because it is a more difficult task. Non-parametric conditional independence can be tested for discrete data using the classical Cochran–Mantel–Haenszel test [28, 79], which can be adapted for continuous random variables by binning the data [62] or using kernels [42, 128, 111, 33, 100]. Besides these, there are methods based on estimating conditional cumulative distribution functions [75, 85], conditional characteristic functions [112, 67], conditional probability density functions [113], empirical likelihood [114], mutual information and entropy [96, 65, 87], copulas [7, 109, 119], distance correlation [121, 37, 116], and other approaches [101]. A number of interesting ideas based on resampling and permutation tests have been proposed in recent years [20, 100, 9].
In Azadkia and Chatterjee [3], a new coefficient of conditional dependence was proposed, based on the ideas behind the -coefficient defined in [22]. Like the -coefficient, this one also has a long list of desirable features, such as being fully nonparametric and working under minimal assumptions. The coefficient is defined as follows.
Let be a random variable and and be random vectors, all defined on the same probability space. Here and . The value means that has no components at all. Let be the law of . The following quantity was proposed in [3] as a measure of the degree of conditional dependence of and given :
(6.1) |
If the denominator equals zero, is undefined. If , then has no components, and the conditional expectations and variances given should be interpreted as unconditional expectations and variances. In this case we write instead of . Note that is a generalization of the statistic appearing in Theorem 1.1. The following theorem summarizes the main properties of .
Theorem 6.1 ([3]).
Suppose that is not almost surely equal to a measurable function of (when , this means that is not almost surely a constant). Then is well-defined and . Moreover, if and only if and are conditionally independent given , and if and only if is almost surely equal to a measurable function of given . When , conditional independence given simply means unconditional independence.
Now suppose we have data consisting of i.i.d. copies of the triple , where . For each , let be the index such that is the nearest neighbor of with respect to the Euclidean metric on , where ties are broken uniformly at random. Let be the index such that is the nearest neighbor of in , again with ties broken uniformly at random. Let be the rank of , that is, the number of such that . If , define
If , let be the number of such that , let denote the such that is the nearest neighbor of (ties broken uniformly at random), and let
In both cases, is undefined if the denominator is zero. The following theorem shows that is a consistent estimator of .
Theorem 6.2 ([3]).
Suppose that is not almost surely equal to a measurable function of . Then as , almost surely.
For various other properties of , such as rate of convergence, performance in simulations and real data, etc., see [3]. One problem that was left unsolved in [3] was the question of proving a central limit theorem for under the null hypothesis, which is crucial for carrying out tests for conditional independence. This question was partially resolved by Shi, Drton, and Han [104], who proved a central limit theorem for under the assumption that is independent of . An improved version of this result was proved recently by Lin and Han [73]. A version for data supported on manifolds was proved by Han and Huang [53].
The paper of Shi et al. [104] also develops the ‘conditional randomization test’ (CRT) framework of Candès et al. [20] to test conditional independence using , and find that , like is an inefficient test statistic. To address this concern, an improved generalization of , called ‘kernel partial correlation’ (KPC), was proposed by Huang, Deb, and Sen [63]. Unlike , KPC has the flexibility to use more than one nearest neighbor, which gives it better power properties.
Note that by the above theorems, a test of conditional independence based on is consistent against all alternatives. The problem is that in the absence of an asymptotic theory for , it is difficult to control the significance level of such a test. This is in fact an impossible problem, by a recent result of Shah and Peters [102] that proves hardness of conditional independence testing in the absence of smoothness assumptions. Assuming some degree of smoothness, minimax optimal conditional independence tests were recently constructed by Neykov, Balakrishnan, and Wasserman [84] and Kim et al. [69].
7 Application to nonparametric variable selection
The commonly used variable selection methods in the statistics literature rely on linear or additive models. This includes classical methods [14, 46, 26, 118, 35, 40, 55, 81] as well as modern ones [19, 132, 133, 126, 36, 91]. These methods are powerful and widely used in practice. However, they sometimes run into problems when significant interaction effects or nonlinearities are present. Such problems can be overcome by model-free methods [20, 60, 1, 15, 39, 55, 18, 6, 120, 16]. On the flip side, the theoretical foundations of model-free methods are usually weaker than those of model-based methods.
In an attempt to combine the best of both worlds, a new method of variable selection called Feature Ordering by Conditional Independence (FOCI), was proposed in Azadkia and Chatterjee [3]. This method uses the conditional dependence coefficient described in the previous section in a stepwise fashion, as follows. Let be the response variable and let be the set of predictors. The data consists of i.i.d. copies of . First, choose to be the index that maximizes . Having obtained , choose to be the index that maximizes . Continue like this until arriving at the first such that , and then declare the chosen subset to be . If there is no such , define to be the whole set of variables. It may also happen that . In that case declare to be empty. Note that this variable selection procedure involves no choice of tuning parameters, which may be an advantage in practice.
It was shown in [3] that under mild conditions, the method selects a ‘correct’ set of variables with high probability. More precisely, it was shown that with high probability, the set selected by FOCI has the property that and are conditionally independent given . In other words, all the information about that one can get from is contained in . For further properties of FOCI and its performance in simulations and real data sets, see [3].
An improved version of FOCI called KFOCI (‘Kernel FOCI’) was proposed by Huang et al. [63]. An application of FOCI to causal inference, via an algorithm named DAG-FOCI, was introduced in Azadkia, Taeb, and Bühlmann [5]. For another application to causal inference, see Chatterjee and Vidyasagar [24].
8 A new proposal: Generalization to standard Borel spaces
In this section, a simple but wide ranging generalization of is proposed. In hindsight, this generalization seems obvious, but somehow this was overlooked both in the original paper [22] as well as in the subsequent developments listed in Section 5.
Recall that two measurable spaces are said to be isomorphic to each other if there is a bijection between the two spaces which is measurable and whose inverse is also measurable. Recall that a standard Borel space is a measurable space that is isomorphic to a Borel subset of a Polish space [110, Chapter 3]. In particular, every Borel subset of every Polish space is a standard Borel space. The Borel isomorphism theorem says that any uncountable standard Borel space is isomorphic to the real line (see Rao and Srivastava [90] for an elementary proof). In particular, if is any standard Borel space, there is a measurable map such that is injective, is Borel, and is measurable on . We will say that is an isomorphism between and a Borel subset of the real line, or simply, a ‘Borel isomorphism’.
Now let and be two standard Borel spaces. Let be a Borel isomorphism of and be a Borel isomorphism of . Let be an -valued pair of random variables, and let be i.i.d. copies of . Let and , so that is a pair of real-valued random variables. Let and for each . Finally, define
where is defined using as in equation (1.1). Note that the definition of depends on our choices of and . Different choices of isomorphisms would lead different definitions of . The following theorem generalizes Theorem 1.1.
Theorem 8.1.
If is not almost surely a constant, then as , converges almost surely to the deterministic limit , which equals , defined as in (1.2) with and in place of and . This limit belongs to the interval . It is if and only if and are independent, and it is if and only if there is a measurable function such that almost surely. Moreover, the asymptotic distribution of under the hypothesis of independence, as given by Theorems 3.1 and 3.2, also holds, provided that and are computed using and instead of and .
Proof.
The convergence is clear from Theorem 1.1. Also, by Theorem 1.1, if and only if and are independent, and if and only if is a measurable function of . Since and , it follows that and are independent if and only if and are independent. For the same reason, is a measurable function of almost surely if and only if is a measurable function of almost surely. Note that is not almost surely a constant if and only if is not almost surely a constant. Lastly, since is just , any result about the asymptotic distribution of , including Theorems 3.1 and 3.2 of this draft, can be transferred to . ∎
Just like the univariate case, the generalized has the advantage of working under zero assumptions and having a simple asymptotic theory, as shown by Theorem 8.1. On the other hand, just like the univariate coefficient, one can expect the generalized coefficient to also suffer from low power for testing independence.
Theorem 8.1 is a nice, clean result, but to implement the idea in practice, one needs to work with actual Borel isomorphisms. Here is an example of a Borel isomorphism between and a Borel subset of . Take any . Let
be the binary expansion of . Filling in extra ’s at the beginning if necessary, let us assume that . Then, let us ‘interlace’ the digits to get the number
This is an encoding of the -tuple . But we also want to encode the signs of the ’s. Let if and if . Sticking in front of the above list, we get an encoding of the vector as a real number. (The in front ensures there is no ambiguity arising from some of the leading ’s being .) It is easy to verify that this mapping is measurable, injective, and its inverse (defined on its range) is also measurable.
Numerical simulations with computed using the above scheme produced satisfactory results. Some examples are as follows. The examples show one potential problem with using statistics such as in a high dimensional setting: The bias may be quite large (even though the standard deviation is small), resulting in slow convergence of to .
Example 8.2 (Points on a sphere).
Non-uniform random points were generated on the unit sphere in by drawing uniformly from , drawing uniformly from , and defining
Taking and , was computed for and . One thousand simulations were done for each . The histograms of the values of are displayed in Figure 1. For , had a mean value of with a standard deviation of . For , the mean value was and the standard deviation was . Simulations were also done for , where the mean value of turned out to be and the standard deviation was . The slow convergence due to large bias is clearly apparent in this example.


Example 8.3 (Points on a sphere plus noise).
This is the same example as the previous one, except that independent random noises were added to , and . The noise variables were taken to be normal with mean zero and standard deviation , for various values of . Table 1 displays the means and standard deviations of in simulations, for and , and , and .
100 | () | () | () |
---|---|---|---|
1000 | () | () | () |
Example 8.4 (Marginal independence versus joint dependence).
In this example, we have a pair of random variables which is a function of a -tuple , but , , and are marginally independent of . The variables are constructed as follows. Let , and be independent Uniform random variables. Let
Here denotes the ‘remainder modulo ’ of a real number . It is easy to see that individually, , , and are independent of , since the operation of adding the uniform random variable and taking the remainder modulo erases all information about the quantity to which is added. However, can be recovered from the -tuple because and , where the second identity holds in each case because and . With the above definitions, i.i.d. copies of were sampled, and was computed, along with the asymptotic P-value for testing independence. This was repeated one thousand times. The average values of the coefficients and the P-values for and are reported in Table 2. The table shows that even for as small as , the hypothesis that and are independent is rejected with average P-value , whereas the average P-value for the hypothesis and and are independent is . On the other hand, even for as large as , the average P-value for the hypothesis that and are independent is , whereas the average P-value for the hypothesis that and are independent is .
Average value | Average P-value | ||||
---|---|---|---|---|---|
The above examples indicate that it may not be ‘crazy’ to use the generalized version of as a measure of association in the multivariate setting and beyond. Further investigations, through applications in simulated and real data, would be necessary to arrive at a concrete verdict.
The generalized version of can also be used to define a coefficient of conditional dependence for random variables taking values in standard Borel spaces, as follows. Let , and be standard Borel spaces, and , and be random variables taking values in , and , respectively. Let denote the -valued random variable . Using Borel isomorphisms, let us define the coefficients and as before, based on i.i.d. data . One can then define a coefficient of conditional dependence between and given as
leaving it undefined if the denominator is zero. The following theorem justifies its use as a measure of conditional dependence.
Theorem 8.5.
Suppose that is not almost surely equal to a measurable function of . Then as , converges almost surely to a deterministic limit , which is if and only if and are independent given , and if and only if is almost surely equal to a measurable function of given .
Proof.
Let , , and be the Borel isomorphisms of , , and that we are using in our definitions of . Let , , and . By Theorem 1.1 and the assumption that is not almost surely a measurable function of , we get that converges almost surely to
Let denote the law of . Then
Similarly,
From the above expressions, we see that is nothing but the quantity displayed in equation (6.1). The claims of the theorem now follow from Theorem 6.1. ∎
9 R packages
An R package for calculating , as well as the generalized coefficient proposed in this survey, and P-values for testing independence — named XICOR — is available on CRAN [23]. An R package for calculating the conditional dependence coefficient and implementing the FOCI algorithm, called FOCI, is also available [4]. The KFOCI algorithm of Huang et al. [63] is implemented in an R package by the same name [64].
References
- Amit and Geman [1997] Y. Amit and D. Geman. Shape quantization and recognition with randomized trees. Neural Computation, 9(7):1545–1588, 1997.
- Auddy et al. [2021] A. Auddy, N. Deb, and S. Nandy. Exact Detection Thresholds for Chatterjee’s Correlation. arXiv preprint arXiv:2104.15140, 2021.
- Azadkia and Chatterjee [2021] M. Azadkia and S. Chatterjee. A simple measure of conditional dependence. Annals of Statistics, 49(6):3070–3102, 2021.
- Azadkia et al. [2020] M. Azadkia, S. Chatterjee, and N. S. Matloff. FOCI: Feature Ordering by Conditional Independence, 2020. URL https://CRAN.R-project.org/package=FOCI.
- Azadkia et al. [2021] M. Azadkia, A. Taeb, and P. Bühlmann. A fast non-parametric approach for causal structure learning in polytrees. arXiv preprint arXiv:2111.14969, 2021.
- Battiti [1994] R. Battiti. Using mutual information for selecting features in supervised neural net learning. IEEE Transactions on Neural Networks, 5(4):537–550, 1994.
- Bergsma [2004] W. Bergsma. Testing conditional independence for continuous random variables. Report Eurandom, 2004048, 2004.
- Bergsma and Dassios [2014] W. Bergsma and A. Dassios. A consistent test of independence based on a sign covariance related to Kendall’s tau. Bernoulli, 20(2):1006–1028, 2014.
- Berrett et al. [2020] T. B. Berrett, Y. Wang, R. F. Barber, and R. J. Samworth. The conditional permutation test for independence while controlling for confounders. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 82(1):175–197, 2020.
- Berrett et al. [2021] T. B. Berrett, I. Kontoyiannis, and R. J. Samworth. Optimal rates for independence testing via -statistic permutation tests. Annals of Statistics, 49(5):2457–2490, 2021.
- Bhattacharya [2019] B. B. Bhattacharya. A general asymptotic framework for distribution-free graph-based two-sample tests. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 81(3):575–602, 2019.
- Bickel [2022] P. J. Bickel. Measures of independence and functional dependence. arXiv preprint arXiv:2206.13663, 2022.
- Blum et al. [1961] J. Blum, J. Kiefer, and M. Rosenblatt. Distribution free tests of independence based on the sample distribution function. Annals of Mathematical Statistics, 32(2):485–498, 1961.
- Breiman [1995] L. Breiman. Better subset regression using the nonnegative garrote. Technometrics, 37(4):373–384, 1995.
- Breiman [1996] L. Breiman. Bagging predictors. Machine Learning, 24(2):123–140, 1996.
- Breiman [2001] L. Breiman. Random forests. Machine Learning, 45(1):5–32, 2001.
- Breiman and Friedman [1985] L. Breiman and J. H. Friedman. Estimating optimal transformations for multiple regression and correlation. Journal of the American statistical Association, 80(391):580–598, 1985.
- Breiman et al. [1984] L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stone. Classification and Regression Trees. Wadsworth Press, 1984.
- Candès and Tao [2007] E. Candès and T. Tao. The Dantzig Selector: Statistical estimation when is much larger than . Annals of Statistics, 35(6):2313–2351, 2007.
- Candès et al. [2018] E. Candès, Y. Fan, L. Janson, and J. Lv. Panning for gold: ‘model-X’ knockoffs for high dimensional controlled variable selection. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 80(3):551–577, 2018.
- Cao and Bickel [2020] S. Cao and P. J. Bickel. Correlations with tailored extremal properties. arXiv preprint arXiv:2008.10177, 2020.
- Chatterjee [2021] S. Chatterjee. A new coefficient of correlation. Journal of the American Statistical Association, 116(536):2009–2022, 2021.
- Chatterjee and Holmes [2020] S. Chatterjee and S. Holmes. XICOR: Association measurement through cross rank increments, 2020. URL https://CRAN.R-project.org/package=XICOR.
- Chatterjee and Vidyasagar [2022] S. Chatterjee and M. Vidyasagar. Estimating large causal polytree skeletons from small samples. arXiv preprint arXiv:2209.07028, 2022.
- Chen [2020] L.-P. Chen. A note of feature screening via rank-based coefficient of correlation. arXiv preprint arXiv:2008.04456, 2020.
- Chen and Donoho [1994] S. Chen and D. Donoho. Basis pursuit. In Proceedings of 1994 28th Asilomar Conference on Signals, Systems and Computers, volume 1, pages 41–44. IEEE, 1994.
- Chernozhukov et al. [2017] V. Chernozhukov, A. Galichon, M. Hallin, and M. Henry. Monge–Kantorovich depth, quantiles, ranks and signs. Annals of Statistics, 45(1):223–256, 2017.
- Cochran [1954] W. G. Cochran. Some methods for strengthening the common tests. Biometrics, 10(4):417–451, 1954.
- Csörgő [1985] S. Csörgő. Testing for independence by the empirical characteristic function. Journal of Multivariate Analysis, 16(3):290–299, 1985.
- Deb and Sen [2021] N. Deb and B. Sen. Multivariate rank-based distribution-free nonparametric testing using measure transportation. Journal of the American Statistical Association, pages 1–16, 2021.
- Deb et al. [2020] N. Deb, P. Ghosal, and B. Sen. Measuring association on topological spaces using kernels and geometric graphs. arXiv preprint arXiv:2010.01768, 2020.
- Dette et al. [2013] H. Dette, K. F. Siburg, and P. A. Stoimenov. A copula-based non-parametric measure of regression dependence. Scandinavian Journal of Statistics, 40(1):21–41, 2013.
- Doran et al. [2014] G. Doran, K. Muandet, K. Zhang, and B. Schölkopf. A permutation-based kernel conditional independence test. In Uncertainty in Artificial Intelligence, pages 132–141. AUAI, 2014.
- Drton et al. [2020] M. Drton, F. Han, and H. Shi. High-dimensional consistent independence testing with maxima of rank correlations. Annals of Statistics, 48(6):3206–3227, 2020.
- Efron et al. [2004] B. Efron, T. Hastie, I. Johnstone, and R. Tibshirani. Least angle regression. Annals of statistics, 32(2):407–499, 2004.
- Fan and Li [2001] J. Fan and R. Li. Variable selection via nonconcave penalized likelihood and its oracle properties. Journal of the American Statistical Association, 96(456):1348–1360, 2001.
- Fan et al. [2020] J. Fan, Y. Feng, and L. Xia. A projection-based conditional dependence measure with applications to high-dimensional undirected graphical models. Journal of Econometrics, 218(1):119–139, 2020.
- Figalli [2018] A. Figalli. On the continuity of center-outward distribution and quantile functions. Nonlinear Analysis, 177:413–421, 2018.
- Freund and Schapire [1996] Y. Freund and R. E. Schapire. Experiments with a new boosting algorithm. In Machine Learning: Proceedings of the Thirteenth International Conference, pages 148–156, 1996.
- Friedman [1991] J. H. Friedman. Multivariate adaptive regression splines. Annals of Statistics, 19(1):1–67, 1991.
- Friedman and Rafsky [1983] J. H. Friedman and L. C. Rafsky. Graph-theoretic measures of multivariate association and prediction. Annals of Statistics, pages 377–391, 1983.
- Fukumizu et al. [2007] K. Fukumizu, A. Gretton, X. Sun, and B. Schölkopf. Kernel measures of conditional dependence. In Advances in Neural Information Processing Systems, volume 20. Curran Associates, Inc., 2007.
- Gamboa et al. [2018] F. Gamboa, T. Klein, and A. Lagnoux. Sensitivity analysis based on Cramér–von Mises distance. SIAM/ASA Journal on Uncertainty Quantification, 6(2):522–548, 2018.
- Gamboa et al. [2022] F. Gamboa, P. Gremaud, T. Klein, and A. Lagnoux. Global sensitivity analysis: A novel generation of mighty estimators based on rank statistics. Bernoulli, 28(4):2345–2374, 2022.
- Gebelein [1941] H. Gebelein. Das statistische problem der korrelation als variations-und eigenwertproblem und sein zusammenhang mit der ausgleichsrechnung. Zeitschrift für Angewandte Mathematik und Mechanik, 21(6):364–379, 1941.
- George and McCulloch [1993] E. I. George and R. E. McCulloch. Variable selection via Gibbs sampling. Journal of the American Statistical Association, 88(423):881–889, 1993.
- Ghosal and Sen [2022] P. Ghosal and B. Sen. Multivariate ranks and quantiles using optimal transport: Consistency, rates and nonparametric testing. Annals of Statistics, 50(2):1012–1037, 2022.
- Gretton et al. [2005a] A. Gretton, O. Bousquet, A. Smola, and B. Schölkopf. Measuring statistical dependence with Hilbert–Schmidt norms. In Proceedings of the 16th International Conference on Algorithmic Learning Theory, pages 63–77. Springer, Berlin, 2005a.
- Gretton et al. [2005b] A. Gretton, A. Smola, O. Bousquet, R. Herbrich, A. Belitski, M. Augath, Y. Murayama, J. Pauls, B. Schölkopf, and N. Logothetis. Kernel constrained covariance for dependence measurement. In International Workshop on Artificial Intelligence and Statistics, pages 112–119. PMLR, 2005b.
- Gretton et al. [2007] A. Gretton, K. Fukumizu, C. Teo, L. Song, B. Schölkopf, and A. Smola. A kernel statistical test of independence. In Advances in Neural Information Processing Systems, volume 20. Curran Associates, Inc., 2007.
- Hallin et al. [2021] M. Hallin, E. Del Barrio, J. Cuesta-Albertos, and C. Matrán. Distribution and quantile functions, ranks and signs in dimension : A measure transportation approach. Annals of Statistics, 49(2):1139–1165, 2021.
- Han [2021] F. Han. On extensions of rank correlation coefficients to multivariate spaces. Bernoulli News, 28(2):7–11, 2021.
- Han and Huang [2022] F. Han and Z. Huang. Azadkia–Chatterjee’s correlation coefficient adapts to manifold data. arXiv preprint arXiv:2209.11156, 2022.
- Han et al. [2017] F. Han, S. Chen, and H. Liu. Distribution-free tests of independence in high dimensions. Biometrika, 104(4):813–828, 2017.
- Hastie et al. [2009] T. Hastie, R. Tibshirani, and J. H. Friedman. The Elements of Statistical Learning: Data Mining, Inference, and Prediction. Springer, Berlin, 2nd edition, 2009.
- Heller and Heller [2016] R. Heller and Y. Heller. Multivariate tests of association based on univariate tests. In Advances in Neural Information Processing Systems, volume 29. Curran Associates, Inc., 2016.
- Heller et al. [2012] R. Heller, M. Gorfine, and Y. Heller. A class of multivariate distribution-free tests of independence based on graphs. Journal of Statistical Planning and Inference, 142(12):3097–3106, 2012.
- Heller et al. [2013] R. Heller, Y. Heller, and M. Gorfine. A consistent multivariate test of association based on ranks of distances. Biometrika, 100(2):503–510, 2013.
- Hirschfeld [1935] H. O. Hirschfeld. A connection between correlation and contingency. Mathematical Proceedings of the Cambridge Philosophical Society, 31(4):520–524, 1935.
- Ho [1998] T. K. Ho. The random subspace method for constructing decision forests. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20(8):832–844, 1998.
- Hoeffding [1948] W. Hoeffding. A non-parametric test of independence. Annals of Mathematical Statististics, 19(4):546–557, 1948.
- Huang [2010] T.-M. Huang. Testing conditional independence using maximal nonlinear conditional correlation. Annals of Statistics, 38(4):2047–2091, 2010.
- Huang et al. [2020] Z. Huang, N. Deb, and B. Sen. Kernel partial correlation coefficient — a measure of conditional dependence. arXiv preprint arXiv:2012.14804, 2020.
- Huang et al. [2022] Z. Huang, N. Deb, and B. Sen. KPC: Kernel Partial Correlation Coefficient, 2022. URL https://cran.r-project.org/web/packages/KPC.
- Joe [1989] H. Joe. Relative entropy measures of multivariate dependence. Journal of the American Statistical Association, 84(405):157–164, 1989.
- Josse and Holmes [2016] J. Josse and S. Holmes. Measuring multivariate association and beyond. Statistics Surveys, 10:132, 2016.
- Ke and Yin [2019] C. Ke and X. Yin. Expected conditional characteristic function-based measures for testing independence. Journal of the American Statistical Association, 115(530):985–996, 2019.
- Kim et al. [2020] I. Kim, S. Balakrishnan, and L. Wasserman. Robust multivariate nonparametric tests via projection averaging. Annals of Statistics, 48(6):3417–3441, 2020.
- Kim et al. [2021] I. Kim, M. Neykov, S. Balakrishnan, and L. Wasserman. Local permutation tests for conditional independence. arXiv preprint arXiv:2112.11666, 2021.
- Kong et al. [2019] E. Kong, Y. Xia, and W. Zhong. Composite coefficient of determination and its application in ultrahigh dimensional variable screening. Journal of the American Statistical Association, 114(528):1740–1751, 2019.
- Kraskov et al. [2004] A. Kraskov, H. Stögbauer, and P. Grassberger. Estimating mutual information. Physical Review E, 69(6):066138, 2004.
- Lin and Han [2022+] Z. Lin and F. Han. On boosting the power of Chatterjee’s rank correlation. Biometrika, 2022+. Forthcoming.
- Lin and Han [2022] Z. Lin and F. Han. Limit theorems of Chatterjee’s rank correlation. arXiv preprint arXiv:2204.08031, 2022.
- Linfoot [1957] E. H. Linfoot. An informational measure of correlation. Information and Control, 1(1):85–89, 1957.
- Linton and Gozalo [1997] O. Linton and P. Gozalo. Conditional independence restrictions: testing and estimation. Cowles Foundation Discussion Paper, 1140, 1997.
- Lopez-Paz et al. [2013] D. Lopez-Paz, P. Hennig, and B. Schölkopf. The randomized dependence coefficient. In Advances in Neural Information Processing Systems, volume 26. Curran Associates, Inc., 2013.
- Lyons [2013] R. Lyons. Distance covariance in metric spaces. Annals of Probability, 41(5):3284–3305, 2013.
- Manole et al. [2021] T. Manole, S. Balakrishnan, J. Niles-Weed, and L. Wasserman. Plugin estimation of smooth optimal transport maps. arXiv preprint arXiv:2107.12364, 2021.
- Mantel and Haenszel [1959] N. Mantel and W. Haenszel. Statistical aspects of the analysis of data from retrospective studies of disease. Journal of the National Cancer Institute, 22(4):719–748, 1959.
- McCann [1995] R. J. McCann. Existence and uniqueness of monotone measure-preserving maps. Duke Mathematical Journal, 80(2):309–323, 1995.
- Miller [2002] A. Miller. Subset Selection in Regression. Chapman and Hall, 2002.
- Mordant and Segers [2022] G. Mordant and J. Segers. Measuring dependence between random vectors via optimal transport. Journal of Multivariate Analysis, 189:104912, 2022.
- Nandy et al. [2016] P. Nandy, L. Weihs, and M. Drton. Large-sample theory for the Bergsma–Dassios sign covariance. Electronic Journal of Statistics, 10(2):2287–2311, 2016.
- Neykov et al. [2021] M. Neykov, S. Balakrishnan, and L. Wasserman. Minimax optimal conditional independence testing. Annals of Statistics, 49(4):2151–2177, 2021.
- Patra et al. [2016] R. K. Patra, B. Sen, and G. J. Székely. On a nonparametric notion of residual and its applications. Statistics & Probability Letters, 109:208–213, 2016.
- Pfister et al. [2018] N. Pfister, P. Bühlmann, B. Schölkopf, and J. Peters. Kernel-based tests for joint independence. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 80(1):5–31, 2018.
- Póczos and Schneider [2012] B. Póczos and J. Schneider. Nonparametric estimation of conditional information and divergences. In Artificial Intelligence and Statistics, pages 914–923. PMLR, 2012.
- Puri et al. [1970] M. Puri, P. Sen, and D. Gokhale. On a class of rank order tests for independence in multivariate distributions. Sankhyā, Series A, 32(3):271–298, 1970.
- Puri and Sen [1971] M. L. Puri and P. K. Sen. Nonparametric methods in multivariate analysis. Wiley, New York, 1971.
- Rao and Srivastava [1994] B. Rao and S. Srivastava. An elementary proof of the Borel isomorphism theorem. Real Analysis Exchange, 20(1):347–349, 1994.
- Ravikumar et al. [2009] P. Ravikumar, J. Lafferty, H. Liu, and L. Wasserman. Sparse additive models. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 71(5):1009–1030, 2009.
- Rényi [1959] A. Rényi. On measures of dependence. Acta Mathematica Hungarica, 10(3-4):441–451, 1959.
- Reshef et al. [2011] D. N. Reshef, Y. A. Reshef, H. K. Finucane, S. R. Grossman, G. McVean, P. J. Turnbaugh, E. S. Lander, M. Mitzenmacher, and P. C. Sabeti. Detecting novel associations in large data sets. Science, 334(6062):1518–1524, 2011.
- Romano [1988] J. P. Romano. A bootstrap revival of some nonparametric distance tests. Journal of the American Statistical Association, 83(403):698–708, 1988.
- Rosenblatt [1975] M. Rosenblatt. A quadratic measure of deviation of two-dimensional density estimates and a test of independence. Annals of Statistics, pages 1–14, 1975.
- Runge [2018] J. Runge. Conditional independence testing based on a nearest-neighbor estimator of conditional mutual information. In International Conference on Artificial Intelligence and Statistics, pages 938–947. PMLR, 2018.
- Sadeghi [2022] B. Sadeghi. Chatterjee Correlation Coefficient: a robust alternative for classic correlation methods in geochemical studies-(including “TripleCpy” Python package). Ore Geology Reviews, page 104954, 2022.
- Schweizer and Wolff [1981] B. Schweizer and E. F. Wolff. On nonparametric measures of dependence for random variables. Annals of Statistics, 9(4):879–885, 1981.
- Sen and Sen [2014] A. Sen and B. Sen. Testing independence and goodness-of-fit in linear models. Biometrika, 101(4):927–942, 2014.
- Sen et al. [2017] R. Sen, A. T. Suresh, K. Shanmugam, A. G. Dimakis, and S. Shakkottai. Model-powered conditional independence test. In Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017.
- Seth and Príncipe [2012] S. Seth and J. C. Príncipe. Conditional association. Neural Computation, 24(7):1882–1905, 2012.
- Shah and Peters [2020] R. D. Shah and J. Peters. The hardness of conditional independence testing and the generalised covariance measure. Annals of Statistics, 48(3):1514–1538, 2020.
- Shi et al. [2021a] H. Shi, M. Drton, M. Hallin, and F. Han. Center-outward sign-and rank-based quadrant, Spearman, and Kendall tests for multivariate independence. arXiv preprint arXiv:2111.15567, 2021a.
- Shi et al. [2021b] H. Shi, M. Drton, and F. Han. On Azadkia–Chatterjee’s conditional dependence coefficient. arXiv preprint arXiv:2108.06827, 2021b.
- Shi et al. [2022a] H. Shi, M. Drton, and F. Han. On the power of Chatterjee’s rank correlation. Biometrika, 109(2):317–333, 2022a.
- Shi et al. [2022b] H. Shi, M. Drton, and F. Han. Distribution-free consistent independence tests via center-outward ranks and signs. Journal of the American Statistical Association, 117(537):395–410, 2022b.
- Shi et al. [2022c] H. Shi, M. Hallin, M. Drton, and F. Han. On universally consistent and fully distribution-free rank tests of vector independence. Annals of Statistics, 50(4):1933–1959, 2022c.
- Sklar [1959] M. Sklar. Fonctions de répartition à dimensions et leurs marges. Publ. Inst. Statist. Univ. Paris, 8:229–231, 1959.
- Song [2009] K. Song. Testing conditional independence via Rosenblatt transforms. Annals of Statistics, 37(6B):4011–4045, 2009.
- Srivastava [1998] S. M. Srivastava. A Course on Borel Sets. Springer-Verlag, New York, 1998.
- Strobl et al. [2019] E. V. Strobl, K. Zhang, and S. Visweswaran. Approximate kernel-based conditional independence tests for fast non-parametric causal discovery. Journal of Causal Inference, 7(1), 2019.
- Su and White [2007] L. Su and H. White. A consistent characteristic function-based test for conditional independence. Journal of Econometrics, 141(2):807–834, 2007.
- Su and White [2008] L. Su and H. White. A nonparametric Hellinger metric test for conditional independence. Econometric Theory, 24(4):829–864, 2008.
- Su and White [2014] L. Su and H. White. Testing conditional independence via empirical likelihood. Journal of Econometrics, 182(1):27–44, 2014.
- Székely and Rizzo [2009] G. J. Székely and M. L. Rizzo. Brownian distance covariance. Annals of Applied Statistics, 3(4):1236–1265, 2009.
- Székely and Rizzo [2014] G. J. Székely and M. L. Rizzo. Partial distance correlation with methods for dissimilarities. Annals of Statistics, 42(6):2382–2412, 2014.
- Székely et al. [2007] G. J. Székely, M. L. Rizzo, and N. K. Bakirov. Measuring and testing dependence by correlation of distances. Annals of Statistics, 35(6):2769–2794, 2007.
- Tibshirani [1996] R. Tibshirani. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society: Series B (Methodological), 58(1):267–288, 1996.
- Veraverbeke et al. [2011] N. Veraverbeke, M. Omelka, and I. Gijbels. Estimation of a conditional copula and association measures. Scandinavian Journal of Statistics, 38(4):766–780, 2011.
- Vergara and Estévez [2014] J. R. Vergara and P. A. Estévez. A review of feature selection methods based on mutual information. Neural Computing and Applications, 24(1):175–186, 2014.
- Wang et al. [2015] X. Wang, W. Pan, W. Hu, Y. Tian, and H. Zhang. Conditional distance correlation. Journal of the American Statistical Association, 110(512):1726–1734, 2015.
- Wang et al. [2017] X. Wang, B. Jiang, and J. S. Liu. Generalized R-squared for detecting dependence. Biometrika, 104(1):129–139, 2017.
- Weihs et al. [2016] L. Weihs, M. Drton, and D. Leung. Efficient computation of the Bergsma–Dassios sign covariance. Computational Statistics, 31(1):315–328, 2016.
- Weihs et al. [2018] L. Weihs, M. Drton, and N. Meinshausen. Symmetric rank covariances: a generalized framework for nonparametric measures of dependence. Biometrika, 105(3):547–562, 2018.
- Yanagimoto [1970] T. Yanagimoto. On measures of association and a related problem. Annals of the Institute of Statistical Mathematics, 22(1):57–63, 1970.
- Yuan and Lin [2006] M. Yuan and Y. Lin. Model selection and estimation in regression with grouped variables. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 68(1):49–67, 2006.
- Zhang [2019] K. Zhang. BET on independence. Journal of the American Statistical Association, 114(528):1620–1637, 2019.
- Zhang et al. [2012] K. Zhang, J. Peters, D. Janzing, and B. Schölkopf. Kernel-based conditional independence test and application in causal discovery. arXiv preprint arXiv:1202.3775, 2012.
- Zhang [2022] Q. Zhang. On the asymptotic distribution of the symmetrized Chatterjee’s correlation coefficient. arXiv preprint arXiv:2205.01769, 2022.
- Zhang et al. [2018] Q. Zhang, S. Filippi, A. Gretton, and D. Sejdinovic. Large-scale kernel methods for independence testing. Statistics and Computing, 28(1):113–130, 2018.
- Zhu et al. [2017] L. Zhu, K. Xu, R. Li, and W. Zhong. Projection correlation between two random vectors. Biometrika, 104(4):829–843, 2017.
- Zou [2006] H. Zou. The adaptive lasso and its oracle properties. Journal of the American Statistical Association, 101(476):1418–1429, 2006.
- Zou and Hastie [2005] H. Zou and T. Hastie. Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 67(2):301–320, 2005.