A Transparent and Nonlinear Method for Variable Selection 11footnotemark: 1
Abstract
Variable selection is a procedure to attain the truly important predictors from inputs. Complex nonlinear dependencies and strong coupling pose great challenges for variable selection in high-dimensional data. In addition, real-world applications have increased demands for interpretability of the selection process. A pragmatic approach should not only attain the most predictive covariates, but also provide ample and easy-to-understand grounds for removing certain covariates. In view of these requirements, this paper puts forward an approach for transparent and nonlinear variable selection. In order to transparently decouple information within the input predictors, a three-step heuristic search is designed, via which the input predictors are grouped into four subsets: the relevant to be selected, and the uninformative, redundant, and conditionally independent to be removed. A nonlinear partial correlation coefficient is introduced to better identify the predictors which have nonlinear functional dependence with the response. The proposed method is model-free and the selected subset can be competent input for commonly used predictive models. Experiments demonstrate the superior performance of the proposed method against the state-of-the-art baselines in terms of prediction accuracy and model interpretability.
keywords:
Variable selection , High-dimensional , Interpretation , Nonlinear relevanceabstract
1 Introduction
Predictive modeling often encounters high-dimensional data (Yin et al., 2022; Hossny et al., 2020; Chaudhari & Thakkar, 2023; Lu et al., 2023), and selecting the truly important predictors is the key to achieving accurate and reliable predictions (Guyon & Elisseeff, 2003). However, variable selection often encounters great challenges from complex nonlinearity and strong coupling that are widespread in high-dimensional data (Hastie et al., 2001). Variable selection methods, e.g., feature screening (Fan & Lv, 2008; Li et al., 2012), and stepwise (Efroymson, 1960; Buhlmann et al., 2010), have been devoted to selecting the subset of predictors which the response is most related to, with the expectation of improving prediction accuracy, and reducing computational cost. Although improving the transparency of variable selection has important implications in enhancing model interpretability (Murdoch et al., 2019; Rudin et al., 2022), it has seldomly mentioned in high-dimensional predictions.
How to transparently decouple the information contained in the inputs is an important, yet easily overlooked area of concern when designing variable selection methods. Collected based on limited experience, the input set of predictors of high-dimensional data are usually “dirty” (Cai et al., 2018). Only several predictors may be truly relevant to predict the response, and the rest may be lack of information, convey the same information as other predictors or carry irrelevant information to the response (Wan et al., 2022). To achieve transparent information decoupling, two issues need further discussion. (i) how to effectively select the truly important predictors; (ii) how to transparently delete predictors which are of no use for the prediction.
When selecting the predictors which are useful for predicting the response, most prevailing approaches for variable selection usually assume that the response and predictors follow some simple forms of functional dependencies, e.g., linear (Tibshirani, 1996), monotonic (Zhu et al., 2011), or additive (Marra & Wood, 2011). In reality, however, there are more diverse and complicated nonlinear forms of dependencies, e.g., nonmonotonic or even oscillatory functional dependencies between predictors and response (Chatterjee, 2021), and the interactions among predictors (Wan et al., 2021). Traditional approaches have difficulty in effectively identifying and selecting such complex nonlinear dependencies. The omission of some key nonlinear relevant predictors can greatly damage the accuracy of predictive modeling (Azadkia & Chatterjee, 2021).
When deleting the predictors which are of no use to predict the response, prevailing approaches mainly rank the predictors based on their correlations to the response, and divide the predictors into two categories, i.e., those relevant to the response, and those independent of the response (Song et al., 2017; Dessì & Pes, 2015). Most approaches make assumptions that the inputs exclude predictors without information, or there is no multicollinearity between predictors (Fan et al., 2020). However, the uninformative predictors and collinearity widely exist in real-world applications (Li et al., 2017a). It is difficult for classical approaches to categorize and remove multiple types of predictors separately, which leads to intransparent selection process, and diminishes their efficiency and interpretability.
To alleviate these challenges in selecting and deleting predictors, this article constructs a Transparent and Nonlinear Variable Selection (TNVS) method for high-dimensional data. The input predictors are divided into four nonoverlapping subsets to achieve Transparent Information Decoupling (TID), i.e., the relevant predictors to be selected, and the uninformative, redundant and conditionally independent predictors to be deleted. The transparent selection and deletion improve predictive accuracy and model interpretability. The main contributions are as follows.
-
1.
Equipped with the recently proposed nonlinear partial correlation, TNVS is able to select predictors with a diversity of complex nonlinear relevance to the response, including nonmonotonic or oscillatory functional dependence between the response and predictors, and interactions among predictors.
-
2.
Information entropy is adopted to filter out uninformative predictor, Gram-Schmidt orthogonalization is adopted to remove redundant predictors which are collinear with the relevant predictors, and the nonlinear partial correlation coefficient is adopted to remove the conditionally independent predictors. In this way, the predictors are classified into different types, and the reasons for removing certain predictors are clearly indicated.
-
3.
The effectiveness and interpretability of the proposed method are demonstrated against the state-of-the-art baselines. The proposed method categorizes the predictors with high accuracy on high-dimensional nonlinear simulations, and the selected subset of predictors improve out-of-sample predictions on real datasets.
The remainder of the paper is organized as follows. Section 2 reviews some prevailing methods in related realms. Section 3 describes the concepts involved in information decoupling and the proxy measures. Section 4 presents the search framework of the proposed TNVS. Section 5 demonstrates the effectiveness and interpretability of the proposed method on simulation problems. The performance of the proposed method on real data applications is discussed in Section 6, including the predictive effectiveness of the selected subset, model interpretability, post hoc interpretability, and tuning parameter stability. Conclusions and future work are summarized in Section 7.
2 Related work
Since our purpose is to derive a transparent variable selection method for high-dimensional data with nonlinear functional dependencies, we reviewed the most relevant methods in this section, including feature screening approaches based on correlations, feature screening approaches based on partial correlations, and stepwise selection methods based on partial correlations.
Correlation coefficients, and other statistical measures for dependencies between two variables, have been widely introduced into feature screening to identify the nonlinear relevance between the response and the predictor in high-dimensional data. Some feature screening methods tried to adopt nonlinear correlations, e.g., generalized correlation (Hall & Miller, 2009), distance correlation (Li et al., 2012), ball correlation (Pan et al., 2019), and projection correlation (Liu et al., 2022). Others attempted to enhance the Pearson’s correlation Zhu et al. (2011) and the leverage score of singular value decompositions (Zhong et al., 2021) through the slicing scheme and inverse regression idea (Li, 1991). However, it is difficult for these methods to recruit those predictors that have weak marginal utilities but are jointly strongly correlated with the response. In this study, we turn to adopt the partial correlation rather than the correlations to better detect the interacted covariates.
Feature screening methods based on Partial correlations (PC) were able to avoid the false omission of some important predictors (Barut et al., 2016; Wang et al., 2018). Given a set of known key variables, these methods could detect and select the important predictors on which the response was strongly dependent through a one-pass screening and ordering algorithm. Meanwhile, the predictors which the response was conditionally independent of were screened out. However, given the newly enlarged selected predictor set, some removed predictors were likely to become jointly correlated with the response. This indicates that if complex interactions exist among the predictors, the selection may still be insufficient to capture all the important predictors. In addition, such methods require prior knowledge of key variables, but such knowledge is usually unobtainable in practical problems. This stimulates the design of stepwise methods for more sufficient selection and more generalized initialization.
Stepwise variable selection based on PCs can be a substitute when the predictors have complex interactions or there is little prior knowledge of the key variables, e.g., the PC-simple algorithm (Buhlmann et al., 2010) and the thresholded partial correlation (TPC) (Li et al., 2017b). Nevertheless, a limitation of PC-simple and TPC is that they are both designed with the linear partial correlation coefficient, and thus they are solid only in linear regression models. This motivates us to introduce nonlinear PCs into stepwise methods, so that these methods can better identify the nonlinear conditional relevance between the response and predictors, and be extended to more general settings.
A nonlinear partial correlation coefficient, namely Conditional Dependence Coefficient (CODEC) (Azadkia & Chatterjee, 2021), was a recent significant development in this field. CODEC was a fully nonparametric measure of the nonlinear conditional dependence between two variables given a set of other variables. Based on the CODEC, a forward selection method was further presented, namely Feature Ordering by Conditional Independence (FOCI). FOCI could identify the conditionally dependent predictors from a diversity of complicated associations. However, the unimportant predictors were evaluated in each iteration and not removed until the end of the search, which is time-consuming. Besides, the removed subset was considered as a whole, though it may include various types of predictors, which is hard to interpret. Our rationale is that by introducing multiple time-saving measures to remove the unimportant predictors during the search, the subset to be evaluated will shrink much faster than FOCI, which increases efficiency. Moreover, the removed subset can be divided in a quasi-chunked fashion, which increases interpretability.
The related methods are summarized in Table 1. Prevailing feature screening methods based on correlations are listed, including Sure Independent Screening (SIS), Sure Independent Ranking and Screening (SIRS), Distance Correlation based Sure Independence Screening (DC-SIS), and Weighted Leverage Score (WLS). Stepwise methods based on partial correlations are also considered, including PC-simple, TPC, and FOCI. From Table 1, it is observed that a nonlinear method which can both select nonmonotonic nonlinear relevance, and remove the uninformative and redundant predictors, does not exist. A fully interpretable method, which can transparently select and delete certain types of predictors, is highly needed. These motivate us to present a transparent manner to decouple complex information for high-dimensional data, and design an effective and interpretable scheme for variable selection. To our knowledge, ours is the first report to design variable selection method directed by transparent information decoupling.
Category | Method | Nonlinear relevant | Redundant | Uninformative | Interpretable |
Correlation, feature screening | SIS (Fan & Lv, 2008) | × | × | Partially | |
SIRS (Zhu et al., 2011) | Monotonic | × | Partially | ||
DC-SIS (Li et al., 2012) | Monotonic | × | Partially | ||
WLS (Zhong et al., 2021) | × | Partially | |||
Partial correlation, stepwise | PC-simple (Buhlmann et al., 2010) | × | × | Partially | |
TPC (Li et al., 2017b) | × | × | Partially | ||
FOCI (Azadkia & Chatterjee, 2021) | (inefficient) | × | Partially | ||
proposed TNVS | Fully |
3 Information decoupling and the proxy measures
The proposed variable selection can be regarded as an information decoupling process of the input set of predictors. The input set is transparently divided into four disjoint subsets, i.e., the subset which is relevant to the response, the uninformative subset , the redundant subset , and the conditionally independent subset , as shown in Fig. 1. In this section, the four subsets in information decoupling are defined, and their corresponding measures in nonlinear supervised learning are described. An example of the four types of predictors are given in Appendix A1.
3.1 The relevant subset and its measure
Let be the response, and be the -dimensional input predictors, where , , and is the sample size. Let be the index set of the input predictors. For any subset of indices , let be the data matrix composed of all the that satisfies . For an , if there is a function that makes , where is a stochastic error, then stands for the index set of the true model, which is named the relevant subset in this paper. In high-dimensional data, is close to or even larger than . In such cases, sparsity assumption usually holds that the dimension of is much less than .
In real problems, and can be linearly correlated, or they can also be associated in the form of a variety of complex nonlinear functions. To date, handling complex nonlinear correlations remains a great challenge (Fan et al., 2020). The primary goal of the proposed variable selection method is to obtain the correlated subset from the input set . For a given group of predictors , , and an unknown predictor , , the Conditional Dependence Coefficient (CODEC) (Azadkia & Chatterjee, 2021) can measure the nonlinear conditional correlation between and given , as well as the interaction between and in explaining . One of the most important features of the CODEC is that it converges to a limit in . Given , the limit is if and only if and are conditionally independent, and is if and only if is almost surely equal to a measurable function of . If , let , where . Let be the -th observation of , where . CODEC is calculated as
(1) |
where denotes the rank of observation , i.e., the number of such that . denotes the index of the nearest neighbor of with respect to the Euclidean metric on , denotes index such that is the nearest neighbor of in , and the ties are broken uniformly at random for both and . denotes the rank of , i.e., the number of such that , and denotes the rank of .
The CODEC can also measure the unconditional correlation between and a predictor , where , in the absence of any given predictors, i.e., when . In this case, the CODEC is interpreted as an unconditional dependence coefficient, and is calculated as
(2) |
where denotes the rank of . denotes the index of the nearest neighbor of , and the ties are broken uniformly at random for . denotes the rank of . denotes the number of such that .
The calculations above require continuous , and , but the CODEC can also be applied to measure the correlations between discrete predictors if the ties are unbonded randomly. If the denominator of is , the CODEC is undefined. At this point, is almost surely equal to a measurable function of , and can be regarded as sufficient for predicting .
3.2 The uninformative subset and its measure
High-dimensional data usually contains uninformative predictors because of some restrictions in data collection. Removing these predictors has limited influence on explaining the response (Li et al., 2017a). For example, in the study of handwriting digits, pixels in the marginal areas may have little explanatory power and can be ignored in the construction of deep neural networks (Chen et al., 2021). Real datasets do not always contain uninformative predictors, since the predictors are usually carefully chosen based on expert experience before being collected to save storage space. However, if we have little prior knowledge, and obtain a dataset with a large number of uninformative predictors, removing them beforehand can greatly improve the efficiency of variable selection.
In this paper, Shannon entropy is used as the measure to distinguish the uninformative predictors from the remaining predictors. Shannon entropy is the probability of all possible values of a predictor, which represents the expectation of the amount of information contained in the predictor (Gray, 2011). Consider a discrete predictor that takes a finite number of possible values with corresponding probabilities . Its entropy is defined as
(3) |
In general, there is . If the distribution of is highly biased toward one of the possible value , is the lowest. At this point, if , is defined as an uninformative predictor, i.e., the quantity of information contained in is . Shannon entropy can be applied only to discrete predictors, and data discretization is required beforehand for continuous predictors (Brown et al., 2012).
3.3 The redundant subset and its measure
If a candidate predictor and a subset of relevant predictors are collinear, keeping both of them may affect the robustness of model estimation (Yu & Liu, 2004). Although there is a consensus on the adverse effects of multicollinearity (Fan & Lv, 2010), most variable selection methods manage to avoid discussing the issue. In this paper, such predictors are named redundant predictors, and are measured separately from other predictors to be deleted.
Gram‒Schmidt Orthogonalization (GSO) is adopted in this paper to decompose and identify the information contained in the predictor set, and further to measure multicollinearity among predictors (Wang et al., 2020; Lyu et al., 2017). The Gram‒Schmidt theorem in Euclidean space indicates that for an index subset of predictors and the corresponding matrix , if is linearly independent, one can always construct an orthogonal basis via GSO, where is the linear combination of , and it spans the same space as .
For each , , the orthogonalized variable is
(4) |
where is the inner product of and , and is the norm of .
Let be the rank of . Suppose , and is linearly independent, where the index subset ; then, an orthogonal basis of can be obtained using GSO. For any , one can obtain the orthogonal variable of with GSO based on . If the variance of satisfies , all the information of is already contained in , and accordingly, is a redundant predictor, which is collinear with .
3.4 The conditionally independent subset and its measure
In addition to uninformative and redundant predictors, there is another group of predictors that needs to be deleted. One of the most obvious features is that is conditionally independent of these predictors given the relevant subset . In this paper, these predictors are called conditionally independent predictors. Modeling with the conditionally independent predictors does more harm than good. Not only will it damage interpretability, but it may also reduce prediction accuracy. The CODEC can be used to measure conditionally independent predictors. In particular, given , if there is close to for a predictor , where , can be considered a conditionally independent predictor.
4 The Proposed Transparent Variable Selection for Nonlinear and High-dimensional Data
In this section, we propose a Transparent and Nonlinear Variable Selection (TNVS) approach for high-dimensional data. Denote as an index set of the candidate predictors, which contains the indices of all predictors that are promising to explain the response. A three-step heuristic search is established to transparently separate the candidate set into the subset to be selected and those to be deleted. The indices of the selected predictors are reserved in the relevant subset , and those of the removed predictors are respectively categorized into in the uninformative subset , the redundant subset , and the conditionally independent subset . A flowchart of the proposed TNVS is illustrated in Fig. 2.
If we have no prior knowledge of the key variables, the initial is set as the index set of input predictors , and the initial , , , and are set as empty sets. The heuristic search of TNVS contains three steps, i.e., prefiltering, forward selection, and batch deletion. In the prefiltering step, uninformative predictors containing little information are identified and removed. Forward selection and batch deletion are iterated alternately. Every time a relevant predictor is selected, a deletion step is performed to remove redundant predictors that are collinear with all the selected predictors.
In the prefiltering step, uninformative predictors are distinguished from other predictors using an Uninformative Score (UinS), and the indices of these uninformative predictors are added to and excluded from . In TNVS, the Shannon entropy of a predictor serves as the UinS. For an index , if , where is the uninformative threshold, will be accordingly regarded as an uninformative predictor.
The forward selection step identifies the most relevant predictor to the response given , and places its index in to form a new relevant subset. The degree of relevance between any and is determined by the Relevance Score (RelS). In TNVS, the CODEC is adopted as the measure of RelS. Given , the RelS of any index is
(5) |
The index with maximum RelS is selected from , and the corresponding predictor is most relevant to considering its interaction with .
The batch deletion step identifies multicollinearity in given the relevant subset . Every time forward selection is performed, all redundant predictors that are collinear with the new are detected from the candidates. Their indices are added in and removed from . In TNVS, GSO is adopted to identify the redundant predictors given . First, the orthogonal basis of is obtained. Then, GSO is performed on all candidate predictors based on to determine which candidates are collinear with . The Redundancy Score (RedS) of given is the variance of the orthogonalized variable . If , where is the redundant threshold, is regarded as a redundant predictor which is collinear with .
The termination criterion of TNVS is the cardinality of the relevant subset reaching the predetermined upper bound , or or undefined for all , where is the relevant threshold. If the termination criterion is satisfied, all remaining elements in belong to the conditionally independent subset . When TNVS is finished, the original index set of predictors is divided into a relevant subset , an uninformative subset , a redundant subset , and a conditionally independent subset , resulting in and . The pseudocode of TNVS is shown in Algorithm 1, and an example to illustrate the procedure of TNVS is provided in Appendix A2.
The computational time of the proposed TNVS is , where is the number of selected predictors. First, the computational cost of the prefiltering step is . The complexity of the forward selection step is since it takes to calculate the CODEC of all the predictors and sort the most relevant predictor (Azadkia & Chatterjee, 2021). For the batch deletion step, GSO requires to screen all predictors. Forward selection and batch deletion are iterated times on average. The total cost of TNVS is . In the worst case when is proportional to , the time complexity is . Suppose there are no less than processors, predictors can be measured in parallel in each step. In summary, the computational time of TNVS is on average. As is shown in simulations, the proposed TNVS is efficient compared with baselines even with limited processors.
The main advantages of the presented TNVS are summarized as follows. First, a transparent framework of variable selection is achieved using a three-step heuristic search, which groups predictors into four mutually disjoint subsets. In every step of the framework, TNVS can provide reasonable interpretations to select or delete certain predictors. Second, the recently proposed CODEC is introduced in TNVS to identify the complex nonlinear associations between the response and the predictors. TNVS can handle not only high-dimensional data in which the response is monotonic and additive functionally dependent on the predictors, but also those data with oscillatory functional dependence and interactions. Moreover, TNVS is model-free. The selected subset of predictors has strong generality and can be used to construct many kinds of learning models.
5 Simulation studies
Substantial experiments on simulated datasets are implemented to illustrate the effectiveness and interpretability of TNVS. First, the variable selection results of TNVS are compared with several competitive baselines to show its validity and conciseness. Then, the subsets of predictors are further evaluated to demonstrate model interpretability.
5.1 The simulated datasets
We generate a regression simulation problem with both monotonic and nonmonotonic functional dependence, interactions, uninformative predictors, multicollinearity, and conditionally independent predictors. In the problem, of the -dimensional input predictors have information, and the remaining are uninformative. The first predictors are consisted of 9 signals , where , and the rest are redundant predictors that are collinear with these signals. The signals are independently sampled from the standard normal distribution . The redundant predictors are highly correlated with one of the signals, and there is , where is the stochastic error that follows , and is set as . The last predictors are uninformative with only a few nonzero observations. The proportion of nonzero observations is set as , and nonzero observations are randomly generated from . The response is a nonlinear measurable function of some predictors
(6) |
where is the stochastic error that follows . is nonlinearly correlated with these four predictors or any of their collinear predictors. is oscillatory functionally dependent on and , and there are interactions between and , and and . The problem has the following three settings by specifying and . For each setting, datasets are generated.
-
Setting 1: ;
-
Setting 2: ;
-
Setting 3: .
5.2 Baselines and parameter settings
In our experiments, TNVS is evaluated against six representative baselines, including two up-to-date model-free and nonlinear methods, namely FOCI (Azadkia & Chatterjee, 2021) and WLS (Zhong et al., 2021), two classic model-free and nonlinear feature screening methods, namely DC-SIS (Li et al., 2012) and SIRS (Zhu et al., 2011), and two stepwise methods based on linear partial correlation, namely PC-simple (Buhlmann et al., 2010) and TPC (Li et al., 2017b).
The parameter settings of the variable selection methods are listed as follows. For TNVS, the uninformative threshold , relevant threshold , and redundant threshold are empirically set as , as the default setting in line with previous research (Liu et al., 2018). is set as , which represents the smallest positive integer no less than . The termination criterion of FOCI is that the CODECs of all candidates are no larger than 0, or the number of selected predictors reaches . The significance levels of PC-simple and TPC are set as (Buhlmann et al., 2010; Li et al., 2017b). The number of selected predictors of WLS, DC-SIS and SIRS are empirically set as in line with the literature. Considering the computing capability, we set the upper bound of the CPU time as seconds for all stepwise methods, including TNVS, FOCI, PC-Simple and TPC. In addition, predictors with zero variance are removed beforehand for FOCI. If a selected subset is empty, i.e., no predictor is considered to be correlated with the response, the mean of the response is used as the predictive value for regression problems, and a random category of the response is chosen as the predictive value for classification problems.
10-fold cross-validation is performed on each dataset, where the samples of the dataset are randomly divided into equal folds. Each unique fold of samples is used for testing and the remaining samples are used for training. The performance in each setting is the average result of all folds on all the corresponding datasets, i.e., each variable selection method is tested for times in each setting to evaluate their variable selection capability.
All the variable selection methods are implemented in R 4.0.3 for a fair comparison222Our code is available at https://github.com/kywang95/TNVS.. The experiments are performed using an Intel Core i7 3.4 GHz processor with 16 GB of RAM.
5.3 Effectiveness and efficiency of the proposed method on nonlinear simulations
The following four criteria are adopted to evaluate the effectiveness and efficiency of TNVS and baselines on high-dimensional simulations with complex nonlinear functional dependencies. The results of each simulation in repetitions are summarized in Table 2.
-
1.
: probability of the discovery of all truly important predictors over all repetitions. The closer is to , the more robust the variable selection method is to retain all true predictors. For simulations, true predictors are any of the relevant predictors and the redundant predictors that are collinear with the relvant predictors. For example, for Setting , the indices of true predictors are any combinations of four elements from the subsets , , , and , and is the frequency that contains at least one element from each of the four sets.
-
2.
: the minimum model size required to include all true predictors in all repetitions in which none of the true predictors are omitted. The closer is to the true model size, the more concise the selected subset is. The average and the standard deviation of are obtained.
-
3.
Coverage: The number of true predictors covered by the selected subset. Coverage is less than or equal to the number of true predictors. The average and standard deviation of the coverage in all repetitions are obtained. In the simulated datasets, the closer the coverage is to , the fewer true predictors are omitted.
-
4.
Time: the average and standard deviation of running time (in seconds) of variable selection in repetitions.
Setting | Index | TNVS | FOCI | WLS | SIRS | DC-SIS | PC-simple | TPC |
1 | 0.98 | 0.00 | 0.02 | 0.00 | 0.00 | |||
Coverage | ||||||||
Time(s) | ||||||||
2 | 0.93 | 0.00 | 0.00 | 0.00 | 0.00 | |||
Coverage | ||||||||
Time(s) | ||||||||
3 | 0.97 | 0.00 | 0.00 | 0.00 | 0.00 | |||
Coverage | ||||||||
Time(s) |
Table 2 shows that our TNVS, and the recently proposed FOCI and WLS, outperform the other traditional methods, and TNVS is the most effective and efficient one among these three methods. First, the-close to-1 in all settings indicates that these methods can reserve all true predictors in most cases. Among these three methods, TNVS has the smallest minimum model size, which means that the subset selected by TNVS is the most concise. In addition, computational time of TNVS is the least among these three methods, which shows that it is the most efficient. The framework of TNVS is most similar to that of the FOCI. Comparing these two methods, TNVS has a smaller minimum model size and a larger coverage, indicating that TNVS is more accurate and reliable than FOCI.
Table 2 also shows that these nonlinear simulation problems are extremely difficult to solve with the other four prevailing methods, i.e., SIRS, DC-SIS, PC-simple and TPC. Their are close to 0, indicating that these methods omit some true predictors and are invalid. PC-simple and TPC are designed for linear regression models, and close-to-0 coverages indicate their weak abilities to identify nonlinearly relevant predictors. SIRS and DC-SIS can only identify the monotonic nonlinear relevance, and having difficulty in removing the multicollinearity, and close-to-0 coverages indicate they cannot retain all important predictors.
Experiments on simulations demonstrate that every module of the proposed TNVS is effective and indispensable. First, compared with PC-simple and TPC, these two stepwise methods with linear partial correlation, TNVS adopts the nonlinear CODEC. This leads to the unsatisfying of these two baselines and the close-to-1 of TNVS, which indicates that introducing CODEC ensures that TNVS can effectively identify the complex nonlinear associations, such as interacted predictors, and oscillatory functional dependence with interactions. Second, FOCI is a stepwise method with CODEC, but ignoring to sorted disposal the uninformative and redundant predictors. Compared with FOCI, TNVS avoids retaining the large percentages of these unnecessary predictors, and achieves larger , smaller model size, larger coverage, and less computational time. This implies that prefiltering and batch deletion inherently improve effectiveness and efficiency of TNVS.
5.4 Interpretability of the proposed method
We further demonstrate the model interpretability of the proposed variable selection method. Since in simulations, the truly relevant, uninformative, redundant, and conditionally independent predictors are known in advance, we can obtain the proportions of four types of predictors in the selected subset for all methods. Taking a closer look at the three competitive methods mentioned in Table 2, Table 3 shows the proportions of the ground truth relevant (), uninformative (), redundant (), and conditionally independent () predictors in the selected subset of TNVS, FOCI, and WLS. The average of each quantity in all repetitions are summarized. The truly relevant predictors in the selected subset are true positive, whereas the rest three types are false positive. Thus, is actually the precision of each methods. The larger is, and the smaller the proportions of , , and are, the more concise the selected subset is.
Setting | Index | TNVS | FOCI | WLS |
1 | 0.02 | |||
0.83 | ||||
0.42 | ||||
0.53 | ||||
2 | 0.02 | |||
0.92 | ||||
0.42 | ||||
0.51 | ||||
3 | 0.02 | |||
0.87 | ||||
0.41 | ||||
0.50 |
Table 3 shows that TNVS has the largest proportion of in its selected subsets, or equivalently, the largest precision. Nearly all the predictors selected by TNVS are relevant predictors. Uninformative, redundant, and conditionally independent predictors are rarely included in the selected subset of TNVS. The subset selected by FOCI contains a large percentage of uninformative predictors and some redundant predictors. Although WLS can select all relevant predictors, they only take a small part of its selected subset, and many unimportant predictors are falsely picked before the truly important predictors. These findings demonstrate that TNVS enhances the capability of FOCI to handle the lack of information and multicollinearity, and it also achieves a better balance between accuracy and conciseness than WLS.
The most significant difference of TNVS against other methods is that TNVS can divide the predictors into four disjoint subsets, instead of only the selected and the deleted ones. Table 4 shows the proportions of the ground truth types of predictors in the four subsets obtained by TNVS (the relevant subset , the uninformative subset , the redundant subset , and the conditionally independent subset ). The average of each quantity in all repetitions are summarized. The truly relevant predictors grouped into the selected relevant subset are true positive, whereas those grouped into the other three types are false negative. Thus, elements in the diagonal are actually the recall of TNVS. The larger elements in the diagonal are, and the smaller the other elements are, the more accurate the variable selection is.
Setting | Index | ||||
1 | 0.995 | 0 | 0 | 0.005 | |
0 | 1 | 0 | 0 | ||
0 | 0 | 0.995 | 0.005 | ||
0 | 0 | 0 | 1 | ||
2 | 0.980 | 0 | 0 | 0.020 | |
0 | 1 | 0 | 0 | ||
0 | 0 | 0.980 | 0.020 | ||
2e-5 | 0 | 3.98e-3 | 0.996 | ||
3 | 0.995 | 0 | 0 | 0.005 | |
0 | 1 | 0 | 0 | ||
0 | 0 | 0.995 | 0.005 | ||
0 | 0 | 0 | 1 |
Table 4 demonstrates the unique strength of TNVS. Besides the selected relevant subset, it further divides the predictors to be removed into the uninformative, redundant, and conditionally independent subsets. The diagonal elements show that the recall of TNVS is high on all four subsets. All uninformative predictors are correctly deleted. A very few relevant predictors are falsely identified as conditionally independent in rare cases, and the misclassification of relevant predictors hinders TNVS from distinguishing the redundancy and conditionally independence. Sufficient evidence shows that when tackling simulations with complex nonlinear functional dependencies, uninformative predictors and collinearity, TNVS provides credible explanations on why certain predictors are removed. To sum up, TNVS outperforms the baselines on model interpretability.
6 Real data applications
6.1 Datasets
Four prevailing real datasets from various domains are chosen to demonstrate the performance of the proposed TNVS (Li et al., 2017a). The descriptive statistics of the datasets are listed in Table 5. In this paper, the original probes, features, or pixels are directly used as input predictors without any preprocessing such as feature extraction, since the aim of these experiments is to demonstrate the predictive capability and interpretability of the proposed variable selection method in high-dimensional data, not to design a distinct method for specific domain, such as face images recognition.
Name | Description | |||
arcene | 2 | 200 | 10 000 | mass spectrometry |
isolet | 6 | 300 | 617 | voices of spoken letters |
warpAR10P | 10 | 240 | 2 400 | face images |
ORL | 40 | 400 | 4 096 | face images |
The reasons for choosing these datasets are as follows. First, these biological or image datasets are high-dimensional data with a large number of predictors. The response and predictors are often nonlinearly correlated in such datasets, and the inputs usually contain a large number of uninformative, redundant, or conditionally independent predictors. Moreover, the face images are considered as the preference since we can easily visualize the variable selection results on the original images, and literature on facial landmark detection (Köstinger et al., 2011) has provided some key reference points. By checking the relevance between our selected predictors and these key reference points, we can easily observe whether the selection results have evident post hoc interpretability. Two ORL datasets with different resolutions are selected to examine whether there are more redundant predictors in images with higher resolution, i.e., whether the images with lower resolution is distinct enough for classification. The datasets are described in detail.
-
1.
arcene is a binary classification dataset to discriminate between patients with and without cancer. The data are collected from mass spectrometry analysis of blood serum, and the predictors are continuous. The sample contains observations with ovarian or prostate cancer and observations of healthy patients. Each observation has real features and random probes. These random probes are permuted least-informative feats, and some of them can probably be identified as uninformative predictors.
-
2.
isolet is an audio dataset for predicting what letter name is spoken. The dataset contains observations, which are audio recordings of speakers. Each subject speaks names of letters of the alphabet twice. Each observation contains features. In this experiment, we generate a high-dimensional dataset with observations from the raw isolet: the first categories are considered, and observations are randomly sampled from each category.
-
3.
warpAR10P is a facial recognition dataset with 10 distinct subjects. For each subject, images are captured under various lighting conditions, with different facial expressions and occlusions such as scarfs and sunglasses. The size of each image is pixels, with gray levels per pixel.
-
4.
ORL is another facial recognition dataset containing face images collected from individuals, and each image contains pixels with gray levels per pixel. These images exhibit multiple changes, such as various picture-taking times, illumination changes, different expressions, and facial detail changes.
Note that Z-score normalization is adopted for input predictors in all datasets, i.e., , where and are the mean and standard deviation of predictor , respectively.
6.2 Learning models and parameter settings
The proposed TNVS is inherently model-free and the selection process relies on no assumptions of the learning model. The effectiveness and generality of the selected subsets are consequently evaluated in terms of their performance on multiple learning models. Specifically, five prevailing predictive models commonly adopted in literature are considered (Salesi et al., 2021; Wan et al., 2022), including Support Vector Regression (SVR) / Support Vector Machine (SVM), Random Forest (RF), Decision Tree (DT), Light Gradient Boosting Machine (lightGBM) and Multi-layer Perceptron (MLP). To examine the predictive capabilities of these variable selection methods, each learning model is trained with the selected predictors, and the results are then compared. These experiments aim to demonstrate the superior generality of variable selection methods rather than the enhanced prediction performance of well-tuned models. Thus, the configurations of the learning models are set to be identical for all variable selection methods to avoid favoring any individual method by combining it with more competitive models. Most parameters are kept as the default settings in scikit-learn (Buitinck et al., 2013). The only adjustment is that the number of iterations is set as for MLP. All the learning models are implemented in Python 3.9 using the application programming interface (API) of scikit-learn.
For real datasets, repeated -fold cross-validation is adopted to evaluate each combination of variable selection methods and learning models. The -fold cross-validation is repeated times on each dataset. All variable selection methods and learning models share the same cross-validation splits so that unbiased estimates are obtained. For all splits, the ratio of classes in the training set is kept the same as that in the full sample. The predictive results are evaluated with out-of-sample accuracy (ACC), recall, , and the Cohen Kappa Score (Kappa). The ranges of ACC, recall and are within ; the larger these indicators are, the better the prediction is. The range of Kappa is within ; the closer the score is to 1, the better the prediction is.
6.3 Effectiveness on real datasets
Since the true models of real datasets are unknown, the predictive results of selected subsets are used to indirectly demonstrate the effectiveness of variable selection. The predictive results obtained using multiple combinations of variable selections and learning models are demonstrated as follows. First, Fig. 3 shows the boxplot of the four predictive indicators obtained with lightGBM on warpAR10P, which compares the predictive capability of the proposed TNVS with the baselines in this case. The predictive results of TNVS are better than those of the baselines, especially PC-Simple and TPC. Other statistical results of the average ACC, recall, , and Kappa on the real datasets are listed in Table LABEL:tab:appendix_A1 of Appendix B. The results of other datasets and other learning models substantially agree with those in Fig. 3, which demonstrates that TNVS is robustly superior to the baselines.
A statistical test is performed to demonstrate the predictive capability of the proposed TNVS. For every dataset, the Friedman test is performed on statistical results of the four indicators obtained by all learning models to determine the overall performance of a certain variable selection method. The Friedman test provides a mean rank of these methods, and indicates whether there are significant differences among these methods. A higher mean rank means that the method has a better-ranked performance over the competitors. Table 6 shows the result of Friedman test, including the mean rank of these methods, the order of the mean rank, and the p-value of the Friedman test. As seen, the predictors selected by TNVS attain the top two among the seven methods on real datasets under the condition of retaining less predictors than feature screening methods, and there are significant differences among these methods, which further confirms the strength and robustness of TNVS in prediction enhancement.
Dataset | Indicator | TNVS | FOCI | WLS | SIRS | DC-SIS | PC-simple | TPC | -value |
arcene | Mean rank | 2.60 | 4.20 | 4.55 | 2.90 | 1.75 | 0.000 | ||
Order | 6 | 4 | 3 | 5 | 7 | ||||
isolet | Mean rank | 5 | 3.00 | 4.00 | 2.00 | 1.00 | 0.000 | ||
Order | 3 | 5 | 4 | 6 | 7 | ||||
warpAR10P | Mean rank | 4.20 | 3.65 | 4.15 | 1.00 | 2.00 | 0.000 | ||
Order | 3 | 5 | 4 | 7 | 6 | ||||
ORL | Mean rank | 5.20 | 3.40 | 3.60 | 1.65 | 1.35 | 0.000 | ||
Order | 3 | 5 | 4 | 6 | 7 |
6.4 Interpretability on real datasets
TNVS transparently groups the predictors into four subsets, i.e., relevant, uninformative, redundant, and conditionally independent, which improves the model interpretability of TNVS. The average cardinalities of the four subsets are counted to check whether these types of predictors exist in the real datasets. Besides, the average compression rate is calculated, which is defined as the proportion of relevant predictors in the inputs. The results of repetitions on each dataset are shown in Table 7.
Dataset | Relevant | Compression rate | Uninformative | Redundant | Conditionally independent |
arcene | 0.24% | ||||
isolet | 5.80% | ||||
warpAR10P | 0.76% | ||||
ORL | 0.91% |
As shown in Table 7, the percentage of relevant predictors is lower than in all datasets, and is even lower than in arcene, warpAR10P, and ORL, which indicates that TNVS compresses the input set of predictors to a large degree. Redundant predictors are identified in all datasets, which shows the usefulness of GSO. TNVS identifies the uninformative predictors in arcene as expected.
In addition to model interpretability, we also visualize variable selection results of the ORL dataset as examples to demonstrate post hoc interpretability of TNVS. Under the premise of the transparent search, the results of TNVS are intuitionistic. The literature on facial landmark detection has highlighted some key reference points on the face that could help identity an individual, e.g., the Annotated Facial Landmarks in the Wild (AFLW) marked facial feature points (Köstinger et al., 2011), including points on the eyes and eyebrows, on the mouth, on the nose, on each ear, and on the chin. Checking whether our selected pixels cover most if not all of these key reference points, we can observe qualitatively whether the selection results provide insight to capture the key predictors of these observations.
One of the variable selection results is marked on the original images of ORL. For TNVS, we color the relevant pixels in yellow, the redundant pixels in red, and the unmarked pixels are conditionally independent pixels, as shown in Figure 4. For other methods, we color the selected pixels in yellow, and the unselected pixels are left unmarked. For ORL, TNVS selects relevant pixels, and removes redundant pixels, and conditionally independent pixels. FOCI selects pixels, WLS, DC-SIS, and SIRS select pixels, and PC-simple and TPC select pixels. Compared with the predictive results of baselines in Table LABEL:tab:appendix_A1 of Appendix B, TNVS achieves comparable or even better performance with more concise subset of predictors.
As is shown in Table LABEL:tab:appendix_A1 of Appendix B, TNVS and WLS are the best two methods on ORL dataset. Their selected pixels both cover the facial points of AFLW, which mainly lie in eye corners, eyebrow corners, tips of nose, corners of mouth, and chins. The accurate detection of key reference points intuitively explains the superior facial discrimination of TNVS and WLS. TNVS further reveals that the redundant pixels of ORL are concentrated mainly in the foreheads, indicating that pixels in these areas are linearly correlated with the relevant pixels. Most regions belong to the conditionally independent subset, indicating that only a very small number of pixels is adequate to explain the response to a certain extent. Some pixels around the eye corners are usually symmetric, and TNVS only select pixels on one side, since given these areas, the pixels in symmetric areas will be conditionally independent. FOCI selects less pixels than TNVS, and these two methods select similar regions of the faces, but FOCI cannot identify the redundant pixels. SIRS and DC-SIS select a large number of redundant pixels in the forehead, and they ignore some feature points in the eye corners, eyebrow corners, and tips of the nose. PC-simple and TPC omit many important facial feature points. Together with the results in Table LABEL:tab:appendix_A1 of Appendix B, this example shows that the predictors selected by TNVS can precisely predict the response, and the reserved or removed predictors can be explicitly interpreted.
6.5 Robustness in tuning the parameters
We further demonstrate stability of TNVS as the three thresholds vary. These ranges should be tuned within certain ranges. If these thresholds are too large, the predictors in each subset will violate their definitions, i.e., the uninformative subset may include informative predictors, and the redundant subset may include predictors which are not linearly independent. If the method is only robust within small ranges, it will be easily influnced by noise in real data. Here, we take a large enough range for each threshold to show that the outperformance is persistently significant, and the proposed method is robust against hyperparameter variations.
First, the uninformative threshold is tuned within range , and nonlinear correlations are evaluated between the response and the uninformative predictors , and the nonlinear partial correlations between the response and uninformative predictors given the selected predictors for the four real datasets. If is undefined, it can be regarded that is almost surely a measurable function of , and we set . For warpAR10P and ORL, the numbers of uninformative predictors are . For arcene and isolet, the numbers of identified uninformative predictors, their correlations and partial correlations are shown in Table 8. The number of uninformative predictors in arcene is the same when , and that in isolet is the same when . The correlations and partial correlations between the response and the uninformative predictors are close to in both datasets, which illustrates that the uninformative predictors have no positive effect on predicting the response. Identifying the uninformative subsets does not heavily depend on the settings of threshold , which shows the robustness of TNVS at the prefiltering step.
Dataset | Indicator | ||||||
arcene | |||||||
isolet | |||||||
We next tune the remaining two parameters by taking the value of the relevant threshold and the value of the redundant threshold . For each combination of and , a 10-fold cross-validation is performed on the four real datasets to demonstrate the stability of TNVS as these two parameters vary. Fig. 5 shows the average cardinality of the selected subsets obtained by TNVS under different values of and . The number of selected predictors fluctuates slightly in response to the change in and , and there is no obvious and unified trend in different datasets. Fig. C1 of Appendix C further describes the prediction accuracy using the above selected predictors. As expected, the changes in both thresholds have little influence on the predictive results. In summary, the proposed TNVS is insensitive to changes in the thresholds within certain ranges.
7 Conclusions and future work
In this paper, a Transparent and Nonlinear Variable Selection (TNVS) method was proposed for high-dimensional data. Transparent information decoupling was achieved with a three-step heuristic search, where the predictors which were relevant to the response were selected, and uninformative, collinear, and conditionally independent predictors were deleted. Introducing a recently proposed nonlinear partial correlation, TNVS was able to identify very complex nonlinear functional dependencies, including not only monotonic and additive functional dependencies between the response and predictors, but also nonmonotonic or oscillatory functional dependence and interactions among predictors. According to scores based on information entropy, Gram-Schmidt orthogonalization, and nonlinear partial correlation, the unimportant predictors were classified into three unimportant subsets. The clear selecting and deleting process enhanced the effectiveness and model interpretability of the proposed method for variable selection.
We should note that limitations exist in the proposed method, and extensions could be developed. First, the three thresholds were designed empirically, although the selection kept stable within certain ranges of these thresholds. In the future, more advanced measures can be introduced to improve the empirical thresholds by using techniques such as nonparametric statistical tests (Shi et al., 2021). In addition, though the proposed method made progress in identifying the interacted predictors, it had difficulty in identifying the marginal independent but jointly correlated predictors. In the future, domain knowledge can be considered in the initial stage to better support the detection of such hard-to-identified interactions (Wu et al., 2022). Last but not least, although the three-step heuristic search adopted in the proposed method was more efficient than other stepwise methods, it could be time-consuming when a large number of predictors were relevant to the response. In the future, more advanced algorithms such as genetic algorithm (Saibene & Gasparini, 2023) and other metaheuristic algorithms (Alcaraz et al., 2022; Pramanik et al., 2023) can be designed to improve the efficiency of TNVS in a less sparse high-dimensional scenario.
Acknowledgments
The work was supported by grants from the National Natural Science Foundation of China (Grant Nos. 72021001 and 71871006).
CRediT authorship contribution statement
Keyao Wang: Conceptualization, Methodology, Software, Writing - Original draft preparation. Huiwen Wang: Conceptualization, Resources, Supervision, Writing - Reviewing and Editing. Jichang Zhao: Validation, Writing - Reviewing and Editing. Lihong Wang: Supervision, Resources, Writing - Reviewing and Editing.
Declaration of Competing Interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Appendix A
A simple example is provided to further demonstrate the measures and the three-step heuristic search of the proposed method for variable selection.
A1 Relevant, uninformative, multicollinear, and conditionally independent predictors
We generate a regression example with nonlinear relevance, uninformative predictors, multicollinearity, and conditionally independent predictors. The predictors and response are defined as follows:
-
1.
, , and are mutually independent predictors all sampled from the standard normal distribution .
-
2.
, and .
-
3.
only observations of are nonzero, and these nonzero observations are randomly generated from .
-
4.
.
In this example, is an uninformative predictor, and its histogram is shown in Fig. A1a. is nonlinearly correlated with and , or any combination of their collinear predictors, i.e., and , and . There are interactions between and , and given , is functionally dependent on . Thus, if and are identified as relevant predictors, is the redundant predictor which is collinear with and . Given the relevant predictors, and are conditionally independent predictors. The dependency between and given is shown in Fig. A1b, and that between and given is shown in Fig. A1c.
A2 The procedure of the proposed variable selection method
The heuristic search of TNVS includes prefiltering, forward selection, and batch deletion steps. According to the flowchart in Fig. 2, the procedure of the proposed method is as follows. In the prefiltering step, uninformative predictors, i.e., in the example is identified and removed according to UniS. In the first iteration of forward selection, the predictor with largest RelS will be selected. Assume is the selected relevant predictor. Then, a deletion step is performed to remove redundant predictors that are collinear with . Here, none of the predictors are identified. In the second iteration, assume is selected as the relevant predictor, and then will be identified as redundant predictor since it is the linear combination of and . After two iterations, RelSs of the rest predictors and are both less than . The search will stop, and the rest two predictors are conditionally independent. The inputs are separated into the uninformative subset , the relevant subset , the redundant subset , and the conditionally independent subset .
Appendix B
Dataset | Model | Indicator | TNVS | FOCI | WLS | SIRS | DC-SIS | PC-simple | TPC |
SVM | ACC | ||||||||
recall | |||||||||
Kappa | |||||||||
RF | ACC | ||||||||
recall | |||||||||
Kappa | |||||||||
arcene | DT | ACC | |||||||
recall | |||||||||
Kappa | |||||||||
lightGBM | ACC | ||||||||
recall | |||||||||
Kappa | |||||||||
MLP | ACC | ||||||||
recall | |||||||||
Kappa | |||||||||
SVM | ACC | ||||||||
recall | |||||||||
Kappa | |||||||||
RF | ACC | ||||||||
recall | |||||||||
Kappa | |||||||||
isolet | DT | ACC | |||||||
recall | |||||||||
Kappa | |||||||||
lightGBM | ACC | ||||||||
recall | |||||||||
Kappa | |||||||||
MLP | ACC | ||||||||
recall | |||||||||
Kappa | |||||||||
SVM | ACC | ||||||||
recall | |||||||||
Kappa | |||||||||
RF | ACC | ||||||||
recall | |||||||||
Kappa | |||||||||
warpAR10P | DT | ACC | |||||||
recall | |||||||||
Kappa | |||||||||
lightGBM | ACC | ||||||||
recall | |||||||||
Kappa | |||||||||
MLP | ACC | ||||||||
recall | |||||||||
Kappa | |||||||||
SVM | ACC | ||||||||
recall | |||||||||
Kappa | |||||||||
RF | ACC | ||||||||
recall | |||||||||
Kappa | |||||||||
ORL | DT | ACC | |||||||
recall | |||||||||
Kappa | |||||||||
lightGBM | ACC | ||||||||
recall | |||||||||
Kappa | |||||||||
MLP | ACC | ||||||||
recall | |||||||||
Kappa |
Appendix C
References
- Alcaraz et al. (2022) Alcaraz, J., Labbé, M., & Landete, M. (2022). Support vector machine with feature selection: A multiobjective approach. Expert Syst. Appl., 204, 117485.
- Azadkia & Chatterjee (2021) Azadkia, M., & Chatterjee, S. (2021). A simple measure of conditional dependence. Ann. Stat., 49, 3070–3102.
- Barut et al. (2016) Barut, E., Fan, J., & Verhasselt, A. (2016). Conditional sure independence screening. J. Am. Stat. Assoc., 111, 1266–1277.
- Brown et al. (2012) Brown, G., Pocock, A., Zhao, M.-J., & Luján, M. (2012). Conditional likelihood maximisation: A unifying framework for information theoretic feature selection. J. Mach. Learn. Res., 13, 27–66.
- Buhlmann et al. (2010) Buhlmann, P., Kalisch, M., & Maathuis, M. H. (2010). Variable selection in high-dimensional linear models: Partially faithful distributions and the pc-simple algorithm. Biometrika, 97, 261–278.
- Buitinck et al. (2013) Buitinck, L., Louppe, G., Blondel, M., Pedregosa, F., Mueller, A., Grisel, O., Niculae, V., Prettenhofer, P., Gramfort, A., Grobler, J., Layton, R., VanderPlas, J., Joly, A., Holt, B., & Varoquaux, G. (2013). Api design for machine learning software: experiences from the scikit-learn project. In ECML PKDD Workshop: Languages for Data Mining and Machine Learning (pp. 108–122).
- Cai et al. (2018) Cai, J., Luo, J., Wang, S., & Yang, S. (2018). Feature selection in machine learning: A new perspective. Neurocomputing, 300, 70–79.
- Chatterjee (2021) Chatterjee, S. (2021). A new coefficient of correlation. J. Am. Stat. Assoc., 116, 2009–2022.
- Chaudhari & Thakkar (2023) Chaudhari, K., & Thakkar, A. (2023). Neural network systems with an integrated coefficient of variation-based feature selection for stock price and trend prediction. Expert Syst. Appl., 219, 119527.
- Chen et al. (2021) Chen, Y., Gao, Q., Liang, F., & Wang, X. (2021). Nonlinear variable selection via deep neural networks. J. Comput. Graph. Stat., 30, 484–492.
- Dessì & Pes (2015) Dessì, N., & Pes, B. (2015). Similarity of feature selection methods: An empirical study across data intensive classification tasks. Expert Syst. Appl., 42, 4632–4642.
- Efroymson (1960) Efroymson, M. A. (1960). Multiple regression analysis. Math. Methods Digit. Comput., 1, 191–203.
- Fan et al. (2020) Fan, J., Li, R., Zhang, C.-H., & Zou, H. (2020). Statistical Foundations of Data Science. (1st ed.). Chapman and Hall/CRC.
- Fan & Lv (2008) Fan, J., & Lv, J. (2008). Sure independence screening for ultrahigh dimensional feature space. J. R. Stat. Soc. Ser. B-Stat. Methodol., 70, 849–911.
- Fan & Lv (2010) Fan, J., & Lv, J. (2010). A selective overview of variable selection in high dimensional feature space. Stat. Sin., 20, 101–148.
- Gray (2011) Gray, R. M. (2011). Entropy. In Entropy and Information Theory (pp. 61–95). Boston, MA: Springer US.
- Guyon & Elisseeff (2003) Guyon, I., & Elisseeff, A. (2003). An introduction to variable and feature selection. J. Mach. Learn. Res., 3, 1157–1182.
- Hall & Miller (2009) Hall, P., & Miller, H. (2009). Using generalized correlation to effect variable selection in very high dimensional problems. J. Comput. Graph. Stat., 18, 533–550.
- Hastie et al. (2001) Hastie, T., Friedman, J., & Tibshirani, R. (2001). Basis expansions and regularization. In The Elements of Statistical Learning: Data Mining, Inference, and Prediction (pp. 115–163). New York, NY: Springer New York.
- Hossny et al. (2020) Hossny, A. H., Mitchell, L., Lothian, N., & Osborne, G. (2020). Feature selection methods for event detection in twitter: A text mining approach. Soc. Netw. Anal. Min., 10, 61.
- Köstinger et al. (2011) Köstinger, M., Wohlhart, P., Roth, P. M., & Bischof, H. (2011). Annotated facial landmarks in the wild: A large-scale, real-world database for facial landmark localization. In 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops) (pp. 2144–2151).
- Li et al. (2017a) Li, J., Cheng, K., Wang, S., Morstatter, F., Trevino, R. P., Tang, J., & Liu, H. (2017a). Feature selection: A data perspective. ACM Comput. Surv., 50.
- Li (1991) Li, K.-C. (1991). Sliced inverse regression for dimension reduction. J. Am. Stat. Assoc., 86, 316–327.
- Li et al. (2017b) Li, R., Liu, J., & Lou, L. (2017b). Variable selection via partial correlation. Stat. Sin., 27, 983–996.
- Li et al. (2012) Li, R., Zhong, W., & Zhu, L. (2012). Feature screening via distance correlation learning. J. Am. Stat. Assoc., 107, 1129–1139.
- Liu et al. (2018) Liu, R., Wang, H., & Wang, S. (2018). Functional variable selection via gram–Schmidt orthogonalization for multiple functional linear regression. J. Stat. Comput. Simul., 88, 3664–3680.
- Liu et al. (2022) Liu, W., Ke, Y., Liu, J., & Li, R. (2022). Model-free feature screening and fdr control with knockoff features. J. Am. Stat. Assoc., 117, 428–443.
- Lu et al. (2023) Lu, S., Yu, M., & Wang, H. (2023). What matters for short videos’ user engagement: A multiblock model with variable screening. Expert Syst. Appl., 218, 119542.
- Lyu et al. (2017) Lyu, H., Wan, M., Han, J., Liu, R., & Wang, C. (2017). A filter feature selection method based on the maximal information coefficient and gram-schmidt orthogonalization for biomedical data mining. Comput. Biol. Med., 89, 264–274.
- Marra & Wood (2011) Marra, G., & Wood, S. N. (2011). Practical variable selection for generalized additive models. Comput. Stat. Data Anal., 55, 2372–2387.
- Murdoch et al. (2019) Murdoch, W. J., Singh, C., Kumbier, K., Abbasi-Asl, R., & Yu, B. (2019). Definitions, methods, and applications in interpretable machine learning. Proc. Natl. Acad. Sci. U. S. A., 116, 22071–22080.
- Pan et al. (2019) Pan, W., Wang, X., Xiao, W., & Zhu, H. (2019). A generic sure independence screening procedure. J. Am. Stat. Assoc., 114, 928–937.
- Pramanik et al. (2023) Pramanik, R., Pramanik, P., & Sarkar, R. (2023). Breast cancer detection in thermograms using a hybrid of ga and gwo based deep feature selection method. Expert Syst. Appl., 219, 119643.
- Rudin et al. (2022) Rudin, C., Chen, C., Chen, Z., Huang, H., Semenova, L., & Zhong, C. (2022). Interpretable machine learning: Fundamental principles and 10 grand challenges. Statist. Surv., 16.
- Saibene & Gasparini (2023) Saibene, A., & Gasparini, F. (2023). Genetic algorithm for feature selection of eeg heterogeneous data. Expert Syst. Appl., 217, 119488.
- Salesi et al. (2021) Salesi, S., Cosma, G., & Mavrovouniotis, M. (2021). Taga: Tabu asexual genetic algorithm embedded in a filter/filter feature selection approach for high-dimensional data. Inf. Sci., 565, 105–127.
- Shi et al. (2021) Shi, H., Drton, M., & Han, F. (2021). On azadkia-chatterjee’s conditional dependence coefficient. arXiv preprint arXiv:2108.06827, .
- Song et al. (2017) Song, Q., Jiang, H., & Liu, J. (2017). Feature selection based on fda and f-score for multi-class classification. Expert Syst. Appl., 81, 22–27.
- Tibshirani (1996) Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B-Stat. Methodol., 58, 267–288.
- Wan et al. (2022) Wan, J., Chen, H., Li, T., Huang, W., Li, M., & Luo, C. (2022). R2ci: Information theoretic-guided feature selection with multiple correlations. Pattern Recognit., 127, 108603.
- Wan et al. (2021) Wan, J., Chen, H., Li, T., Yang, X., & Sang, B. (2021). Dynamic interaction feature selection based on fuzzy rough set. Inf. Sci., 581, 891–911.
- Wang et al. (2020) Wang, H., Liu, R., Wang, S., Wang, Z., & Saporta, G. (2020). Ultra-high dimensional variable screening via gram–Schmidt orthogonalization. Comput. Stat., 35, 1153–1170.
- Wang et al. (2018) Wang, X., Wen, C., Pan, W., & Huang, M. (2018). Sure independence screening adjusted for confounding covariates with ultrahigh-dimensional data. Stat. Sin., 28, 293–317.
- Wu et al. (2022) Wu, X., Tao, Z., Jiang, B., Wu, T., Wang, X., & Chen, H. (2022). Domain knowledge-enhanced variable selection for biomedical data analysis. Inf. Sci., 606, 469–488.
- Yin et al. (2022) Yin, D., Chen, D., Tang, Y., Dong, H., & Li, X. (2022). Adaptive feature selection with shapley and hypothetical testing: Case study of eeg feature engineering. Inf. Sci., 586, 374–390.
- Yu & Liu (2004) Yu, L., & Liu, H. (2004). Efficient feature selection via analysis of relevance and redundancy. J. Mach. Learn. Res., 5, 1205–1224.
- Zhong et al. (2021) Zhong, W., Liu, Y., & Zeng, P. (2021). A model-free variable screening method based on leverage score. J. Am. Stat. Assoc., 1, 1–12.
- Zhu et al. (2011) Zhu, L.-P., Li, L., Li, R., & Zhu, L.-X. (2011). Model-free feature screening for ultrahigh-dimensional data. J. Am. Stat. Assoc., 106, 1464–1475.