Decision tree heuristics can fail, even in the smoothed setting
Abstract
Greedy decision tree learning heuristics are mainstays of machine learning practice, but theoretical justification for their empirical success remains elusive. In fact, it has long been known that there are simple target functions for which they fail badly (Kearns and Mansour, STOC 1996).
Recent work of Brutzkus, Daniely, and Malach (COLT 2020) considered the smoothed analysis model as a possible avenue towards resolving this disconnect. Within the smoothed setting and for targets that are -juntas, they showed that these heuristics successfully learn with depth- decision tree hypotheses. They conjectured that the same guarantee holds more generally for targets that are depth- decision trees.
We provide a counterexample to this conjecture: we construct targets that are depth- decision trees and show that even in the smoothed setting, these heuristics build trees of depth before achieving high accuracy. We also show that the guarantees of Brutzkus et al. cannot extend to the agnostic setting: there are targets that are very close to -juntas, for which these heuristics build trees of depth before achieving high accuracy.
1 Introduction
Greedy decision tree learning heuristics are among the earliest and most basic algorithms in machine learning. Well-known examples include ID3 [Qui86], its successor C4.5 [Qui93], and CART [BFSO84], all of which continue to be widely employed in everyday ML applications. These simple heuristics build a decision tree for labeled dataset in a greedy, top-down fashion. They first identify a “good” attribute to query as the root of the tree. This induces a partition of into and , and the left and right subtrees are built recursively using and respectively.
In more detail, each heuristic is associated with an impurity function that is concave, symmetric around , and satisfies and . Examples include the binary entropy function that is used by ID3 and C4.5, and the Gini impurity function that is used by CART; Kearns and Mansour [KM96] proposed and analyzed the function . For a target function and a distribution over , these heuristics build a decision tree hypothesis for as follows:
-
1.
Split: Query as the root of the tree, where and are chosen to (approximately) maximize the purity gain with respect to :
where the expectations and probabilities above are with respect to randomly drawn labeled examples where , and denotes the restriction of to inputs satisfying (and similarly for ).
-
2.
Recurse: Build the left and right subtrees by recursing on and respectively.
-
3.
Terminate: The recursion terminates when the depth of the tree reaches a user-specified depth parameter. Each leaf of the tree is labeled by , where we associate with the restriction corresponding to the root-to- path within the tree and .
Given the popularity and empirical success of these heuristics111CART and C4.5 were named as two of the “Top 10 algorithms in data mining” by the International Conference on Data Mining (ICDM) community [WKQ+08]; other algorithms on this list include -means, -nearest neighbors, Adaboost, and PageRank, all of whose theoretical properties are the subjects of intensive study. C4.5 has also been described as “probably the machine learning workhorse most widely used in practice to date” [WFHP16]., it is natural to seek theoretical guarantees on their performance:
Let be a target function and be a distribution over . Can we obtain a high-accuracy hypothesis for by growing a depth- tree using these heuristics, where is not too much larger than , the optimal decision tree depth for ? ()
1.1 Background and prior work
A simple and well-known impossibility result.
Unfortunately, it has long been known [KM96, Kea96] that no such guarantee is possible even under favorable feature and distributional assumptions. Consider the setting of binary features (i.e. ) and the uniform distribution over , and suppose is the parity of two unknown features for . It can be easily verified that for all impurity functions , all features have the same purity gain: for all , regardless of whether . Therefore, these heuristics may build a tree of depth , querying irrelevant variables where , before achieving any nontrivial accuracy. This is therefore an example where the target is computable by a decision tree of depth , and yet these heuristics may build a tree of depth before achieving any nontrivial accuracy.
Smoothed analysis.
In light of such impossibility results, a line of work has focused on establishing provable guarantees for restricted classes of target functions [FP04, Lee09, BDM19, BLT20b, BLT20a]; we give an overview of these results in Section 1.3.
The focus of our work is instead on smoothed analysis as an alternative route towards evading these impossibility results, an approach that was recently considered by Brutzkus, Daniely, and Malach [BDM20]. Smoothed analysis is by now a standard paradigm for going beyond worst-case analysis. Roughly speaking, positive results in this model show that “hard instances are pathological.” Smoothed analysis has been especially influential in accounting for the empirical effectiveness of algorithms widely used in practice, a notable example being the simplex algorithm for linear programming [ST04]. The idea of analyzing greedy decision tree learning heuristics through the lens of smoothed analysis is therefore very natural.
A smoothed product distribution over , a notion introduced by Kalai, Samrodnitsky, and Teng [KST09], is obtained by randomly and independently perturbing the bias of each marginal of a product distribution. For smoothed product distributions, Brutzkus et al. proved strong guarantees on the performance of greedy decision tree heuristics when run on targets that are juntas, functions that depend only on a small number of its features. For a given impurity function , let us write to denote the corresponding decision tree learning heuristic.
Theorem 1 (Performance guarantee for targets that are -juntas [BDM20]).
For all impurity functions and for all target functions that are -juntas, if is trained on examples drawn from a smoothed product distribution, it learns a decision tree hypothesis of depth that achieves perfect accuracy.
(Therefore Theorem 1 shows that the smoothed setting enables one to circumvent the impossibility result discussed above, which was based on targets that are -juntas.)
Every -junta is computable by a depth- decision tree, but a depth- decision tree can depend on as many as variables. Brutzkus et al. left as an open problem of their paper a conjecture that the guarantees of Theorem 1 hold more generally for targets that are depth- decision trees:
Conjecture 1 (Performance guarantee for targets that are depth- decision trees).
For all impurity functions and for all target functions that are depth- decision trees, if is trained on examples drawn from a smoothed product distribution, it learns a decision tree hypothesis of depth that achieves high accuracy.
In other words, Conjecture 1 states that for all targets , the sought-for guarantee () holds if the heuristics are trained on examples drawn from a smoothed product distribution.
1.2 This work: Lower bounds in the smoothed setting
Our main result is a counterexample to Conjecture 1. We construct targets that are depth- decision trees for which all greedy impurity-based heuristics, even in the smoothed setting, may grow a tree of depth before achieving high accuracy. This lower bound is close to being maximally large since Theorem 1 implies an upper bound of . Our result is actually stronger than just a lower bound in the smoothed setting: our lower bound holds with respect to any product distribution that is balanced in the sense that its marginals are not too skewed.
Theorem 2 (Our main result: a counterexample to Conjecture 1; informal).
Conjecture 1 is false: For all , there are target functions that are depth- decision trees such that for all impurity functions , if is trained on examples drawn from any balanced product distribution, its decision tree hypothesis does not achieve high accuracy unless it has depth .
By building on our proof of Theorem 2, we also show that the guarantees of Brutzkus et al. for -juntas cannot extend to the agnostic setting:
Theorem 3 (Theorem 1 does not extend to the agnostic setting; informal).
For all and , there are target functions that are -close to a -junta such that for all impurity functions , if is trained on examples drawn from any balanced product distribution, its decision tree hypothesis does not achieve high accuracy unless it has depth .
In particular, there are targets that are -close to -juntas, for which these heuristics have to construct a decision tree hypothesis of depth before achieving high accuracy. Taken together with the positive result of Brutzkus et al., Theorems 2 and 3 add to our understanding of the strength and limitations of greedy decision tree learning heuristics.
Our lower bounds are based on new generalizations of the addressing function. Since the addressing function is often a useful extremal example in a variety of settings, we are hopeful that these generalizations and our analysis of them will see further utility beyond the applications of this paper.
1.3 Related Work
As mentioned above, there has been a substantial line of work on establishing provable guarantees for greedy decision tree heuristics when run in restricted classes of target functions. Fiat and Pechyony [FP04] considered the class of read-once DNF formulas and halfspaces; the Ph.D. thesis of Lee [Lee09] considered the class of monotone functions; Brutzkus, Daniely, and Malach [BDM19] considered conjunctions and read-once DNF formulas; recent works of [BLT20b, BLT20a] build on the work of Lee and further studied monotone target functions. (All these works focus on the case of binary features and product distributions over examples.)
Kearns and Mansour [KM96], in one of the first papers to study these heuristics from a theoretical perspective, showed that they can be viewed as boosting algorithms, with internal nodes of the decision tree hypothesis playing the role of weak learners. Their subsequent work with Dietterich [DKM96] provide experimental results that complement the theoretical results of [KM96]; see also the survey of Kearns [Kea96].
Finally, we mention that decision trees are one of the most intensively studied concept classes in learning theory. The literature on this problem is rich and vast (see e.g. [EH89, Riv87, Blu92, Han93, Bsh93, KM93, BFJ+94, HJLT96, KM99, MR02, JS06, OS07, GKK08, KS06, KST09, KST09, HKY18, CM19, BGLT20]), studying it from a variety of perspectives and providing both positive and negative results. However, the algorithms developed in these works do not resemble the greedy heuristics used in practice, and indeed, most of them are not proper (in the sense of returning a hypothesis that is itself a decision tree).222Quoting [KM96], “it seems fair to say that despite their other successes, the models of computational learning theory have not yet provided significant insight into the apparent empirical success of programs like C4.5 and CART.”
2 Preliminaries
Recall that an impurity function is concave, symmetric with respect to , and satisfies and . We further quantify the concavity and smoothness of as follows:
Definition 1 (Impurity functions).
is an -impurity function if is -strongly concave and -smooth, i.e., is twice-differentiable and for every .
For a boolean function and index , we write and to denote the restricted functions obtained by fixing the -th input bit of to either or . Formally, each is a function over defined as , where denotes the string obtained by setting the -th bit of to . More generally, a restriction is a list of constraints of form “” in which every index appears at most once. For restriction , the restricted function is similarly defined as .
Definition 2 (Purity gain).
Let be a distribution over and . The -purity gain of querying variable on boolean function is defined as
In a decision tree, each node naturally corresponds to a restriction formed by the variables queried by the ancestors of (excluding itself). We use as a shorthand for . We say that a decision tree learning algorithm is impurity-based if, in the tree returned by the algorithm, every internal node queries a variable that maximizes the purity gain with respect to .
Definition 3 (Impurity-based algorithms).
A decision tree learning algorithm is -impurity-based if the following holds for every and distribution over : When learning on , the algorithm outputs a decision tree such that for every internal node , the variable that is queried at satisfies for every .
The above definition assumes that the algorithm exactly maximizes the -purity gain at every split, while in reality, the purity gains can only be estimated from a finite dataset. We therefore consider an idealized setting that grants the learning algorithm with infinitely many training examples, which, intuitively, strengthens our lower bounds. (Our lower bounds show that in order for an algorithm to recover a good tree—a high-accuracy hypothesis whose depth is close to that of the target—it would need to query a variable that has exponentially smaller purity gain than that of the variable with the largest purity gain. Hence, if purity gains are estimated using finitely many random samples as is done in reality, the strength of our lower bounds imply that with extremely high probability, impurity-based heuristics will fail to build a good tree; see Remark 5 for a detailed discussion.)
When a decision tree queries variable on function , it naturally induces two restricted functions and . The following lemma states that the purity gain of querying is roughly the squared difference between the averages of the two functions, up to a factor that depends on the impurity function and the data distribution . We say that a product distribution over is -balanced if the expectation of each of the coordinates is in .
Lemma 1.
For any , -balanced product distribution over and -impurity function , it holds for and every that
Proof of Lemma 1.
Let and respectively. Then, we have , and the purity gain can be written as
Since is -strongly concave and -smooth, the above is bounded between and . Since is -balanced, we have . It follows that
Thus, the ratio is bounded between and . ∎
Our lower bounds hold with respect to all -balanced product distributions. We compare this to the definition of a -smoothened -balanced product distribution from [BDM20].
Definition 4 (Smooth distributions).
A -smoothened -balanced product distribution is a random product distribution over where the marginal for the bit is with probability for fixed and drawn i.i.d. from .
Since our lower bounds hold against all -balanced product distributions, it also holds against all -smoothened -balanced product distributions.
3 Proof overview and formal statements of our results
Our goal is to construct a target function that can be computed by a depth- decision tree, but on which impurity-based algorithms must build to depth or have large error. To do so, we construct a decision tree target where the variables with largest purity gain are at the bottom layer of (adjacent to its leaves). Intuitively, impurity-based algorithms will build their decision tree hypothesis for by querying all the variables in the bottom layer of before querying any of the variables higher up in . Our construction will be such that until the higher up variables are queried, it is impossible to approximate the target with any nontrivial error. Summarizing informally, we show that impurity-based algorithms build its decision tree hypothesis for our target by querying variables in exactly the “wrong order”.
The starting point of our construction is the well known addressing function. For , the addressing function is defined as follows: Given “addressing bits” and “memory bits” , the output is the bit of , where “ bit” is computed by interpreting as a base-2 integer. Note that the addressing function is computable by a decision tree of depth that first queries the addressing bits, followed by the appropriate memory bit.
For our lower bound, we would like the variables with the highest purity gain to be the memory bits. However, for smoothed product distributions, the addressing bits might have higher purity gain than the memory bits, and impurity-based algorithms might succeed in learning the addressing function. We therefore modify the addressing function by making each addressing bit the parity of multiple new bits. We show that by making each addressing bit the parity of sufficiently many new bits, we can drive the purity gain of these new bits down to the point where the memory bits have the highest purity gain as desired—in fact, larger than the addressing bits by a multiplicative factor of . (Making each addressing bit the parity of multiple new bits increases the depth of the target, so this introduces technical challenges we have to overcome in order to achieve the strongest parameters.)
Our main theorem is formally restated as follows.
Theorem 4 (Formal version of Theorem 2).
Fix and . There are boolean functions such that: (1) is computable by a decision tree of depth ; (2) For every -balanced product distribution over the domain of and every -impurity function , any -impurity based decision tree heuristic, when learning on , returns a tree that has either depth or an error.
An extension of our construction and its analysis shows that the guarantees of Brutzkus et al. for targets that are -juntas cannot extend to the agnostic setting. Roughly speaking, while our variant of the addressing function from Theorem 4 is far from all -juntas, it can be made close to one by fixing most of the memory bits. We obtain our result by showing that our analysis continues to hold under such a restriction.
Theorem 5 (Formal version of Theorem 3).
Fix , and . There are boolean functions such that for every -balanced product distribution over the domain of : (1) is -close to an -junta with respect to ; (2) For every -impurity function , any -impurity based decision tree heuristic, when learning on , returns a tree that has either a depth of or an error.
4 Warm-Up: A Weaker Lower Bound
We start by giving a simplified construction that proves a weaker version of Theorem 4, in which the depth in condition (1) is relaxed to . For integers , we define a boolean function as follows. The input of is viewed as two parts: addressing bits indexed by and , and memory bits indexed by . The function value is defined by first computing for every , and then assigning .
In other words, is a disjoint composition of the -bit addressing function and the parity function over bits. Given addressing bits and memory bits , the function first computes a -bit address by taking the XOR of the addressing bits in each group of size , and then retrieves the memory bit with the corresponding address. Clearly, can be computed by a decision tree of depth that first queries all the addressing bits and then queries the relevant memory bit in the last layer.
4.1 Address is Almost Uniform
Drawing input from a distribution naturally defines a distribution over of the -bit address . The following lemma states that when is a -balanced product distribution, the distribution of is almost uniform in the sense. Furthermore, this almost uniformity holds even if one of the addressing bits is fixed.
Lemma 2.
Suppose that and is a -balanced product distribution over the domain of . Then,
Furthermore, for every , and ,
The proof of Lemma 2 uses the following simple fact, which states that the XOR of independent biased random bits is exponentially close to an unbiased coin flip.
Fact 3.
Suppose that are independent Bernoulli random variables, each with an expectation between and . Then, .
Proof of Lemma 2.
Since and is -balanced, Fact 3 gives
Note that the bits of are independent, so is given by
where the third step applies for and integers . Similarly,
where the last two steps apply and . This proves the first part.
The proof of the “furthermore” part is essentially the same, except that conditioning on , becomes the XOR of independent bits and . By Fact 3, we have
and the rest of the proof is the same. ∎
4.2 Memory Bits are Queried First
The following technical lemma states that the purity gain of is maximized by a memory bit, regardless of the impurity function and the data distribution. Therefore, when an impurity-based algorithm (in the sense of Definition 3) learns , the root of the decision tree will always query a memory bit. Furthermore, this property also holds for restrictions of as long as the restriction only involves the memory bits.
Lemma 4.
Fix and . Let and , where is chosen as in Lemma 1. The following holds for every function with and : For any -impurity function , -balanced product distribution and restriction of size that only contains the memory bits of , the purity gain is maximized by a memory bit.
Proof of Lemma 4.
Fix and and shorthand for . We will prove a stronger claim: with respect to , every memory bit (that is not in ) gives a much higher purity gain than every addressing bit does.
Purity gain of the memory bits.
Fix a memory bit () that does not appear in restriction . Let for . By the law of total expectation,
Here the second step holds since evaluates to when the address equals , and agrees with when . Since only the first term above depends on , we have
where the second step follows from and Lemma 2. Finally, by Lemma 1, .
Purity gain of the addressing bits.
Similarly, we fix an addressing bit and define the average . Since is a product distribution, is equal to the conditional expectation . Then, by the law of total expectation, we can write as
Here the second step holds since and are independent conditioning on the address ; in other words, once we know the value of , it doesn’t matter how is set in determining the output of .
Let denote , and let be the distribution of conditioning on . Then, is exactly given by . Since each is in , is upper bounded by the total variation distance between and :
(Lemma 2) |
Finally, applying Lemma 1 shows that .
Recall that , so we have . Therefore, the purity gain of every memory bit outside the restriction is strictly larger than that of any addressing bit, and the lemma follows immediately. ∎
Remark 5.
The proof above bounds the purity gain of each memory bit and each addressing bit by and respectively. For Lemma 4 to hold when the purity gains are estimated from a finite dataset, it suffices to argue that each estimate is accurate up to an additive error. By a standard concentration argument, to estimate the purity gains for all restriction of size , training examples are sufficient. When applied later in the proof of Theorem 4, this finite-sample version of Lemma 4 would imply that impurity-based algorithms need to build a tree of depth as soon as the sample size reaches .
4.3 Proof of the Weaker Version
Now we are ready to prove the weaker version of Theorem 4. We will apply Lemma 4 to argue that the tree returned by an impurity-based algorithm never queries an addressing bit (unless all the memory bits have been queried), and then show that every such decision tree must have an error of .
Proof of Theorem 4 (weaker version).
Fix integer and consider the functions . Since each is represented by a decision tree of depth , it remains to show that impurity-based algorithms fail to learn . Fix integer (where is chosen as in Lemma 4) and -balanced product distribution over the domain of . In the following, we use shorthand for .
Small trees never query addressing bits.
Let be the decision tree returned by a -impurity-based algorithm when learning on . If has depth , we are done, so we assume that has depth at most . We claim that never queries the addressing bits of . Suppose otherwise, that an addressing bit is queried at node in , and no addressing bits are queried by the ancestors of . Then, the restriction associated with node only contains the memory bits of . Since has depth , the size of is strictly less than . Then, by Lemma 4, the -purity gain with respect to is maximized by a memory bit. This contradicts the assumption that the algorithm is -impurity-based.
Trivial accuracy if no addressing bits are queried.
We have shown that only queries the memory bits of . We may further assume that queries all the memory bits before reaching any of its leaves, i.e., is a full binary tree of depth . This assumption is without loss of generality because we can add dummy queries on the memory bits to the leaves of depth , and label all the resulting leaves with the same bit. This change does not modify the function represented by .
Assuming that is full, every leaf of is labeled by bits , meaning that each memory bit is fixed to on the root-to- path. The expectation of the restricted function is then given by . Clearly, the error of is minimized when each leaf is labeled with , and the conditional error when reaching leaf is .
It remains to show that for a large fraction of leaves , is bounded away from and , so that is large. When leaf is randomly chosen according to distribution , the corresponding is given by
(1) |
where are independent Bernoulli random variables with means in .
By Lemma 2 and our choice of , holds for every . Thus, each term in (1) is bounded between and . Furthermore, since each has expectation at least , . Then, Hoeffding’s inequality guarantees that over the random choice of , holds with probability at least , which is lower bounded by for all sufficiently large . By a symmetric argument, also holds with probability . Therefore, with probability over the choice of leaf , holds and thus the conditional error on leaf is at least . This shows that the error of over distribution is lower bounded by , which completes the proof. ∎
5 Proof of Theorem 4
When proving the weaker version of Theorem 4, each hard instance has addressing bits grouped into disjoint subsets, and the -bit address is defined by the XOR of bits in each subset. We will prove Theorem 4 using a slightly different construction that computes address from overlapping subsets of only addressing bits.
For integers and a list of sets where each , we define a boolean function as follows. The input of is again divided into two parts: addressing bits and memory bits indexed by a -bit address . The function value is computed by taking and then . Clearly, can be computed by a decision tree of depth that first queries all the addressing bits , and then queries the relevant memory bit .
Let denote the -ary symmetric difference of sets through , i.e., the set of elements that appear in an odd number of sets. We say that a list of sets has distance , if any non-empty collection of sets has a symmetric difference of size at least , i.e., for every non-empty . In the following, we prove analogs of Lemmas 2 and 4 for function assuming that has a large distance; Theorem 4 would then follow immediately.
Lemma 6.
Suppose that is a -balanced product distribution over the domain of and has distance . Then,
Furthermore, for every and ,
We prove Lemma 6 by noting that the distribution of has exponentially small Fourier coefficients (except the degree- one) under the assumptions, and is thus close to the uniform distribution over . More concretely, our goal is to show that, for every the quantity is with probability nearly exactly . Afterwards, we will show this is sufficient to guarantee that the distribution of is close to the uniform distribution.
Proof of Lemma 6.
Since , we have for every , where is the symmetric difference of the corresponding sets. Since has distance , for every non-empty and thus is the XOR of at least independent bits. Note that . By Fact 3 and ,
(2) |
Note that for , we have . Therefore, for every ,
(expansion of product and linearity) | ||||
(empty product equals ) | ||||
(triangle inequality and linearity) | ||||
() | ||||
(Inequality (2)) |
The proof of the “furthermore” part is the same, except that after conditioning on , each is now the XOR of at least independent bits, and the remaining proof goes through. ∎
We note that the proof of Lemma 4 depends on the definition of only through the application of Lemma 2. Thus, Lemma 6 directly implies the following analog of Lemma 4:
Lemma 7.
Fix and . Let and , where is chosen as in Lemma 1. The following holds for every function such that and has distance : For any -impurity function , -balanced product distribution and restriction of size that only contains the memory bits of , the purity gain is maximized by a memory bit.
Finally, we prove Theorem 4 by showing the existence of a set family with a good distance.
Proof of Theorem 4.
Fix . The Gilbert–Varshamov bound for binary linear codes implies that for some , there exists a binary linear code with rate and relative distance . It follows that for every sufficiently large , there exists such that each and has distance . This can be done by using the -th basis of the linear code as the indicator vector of subset for each .
We prove Theorem 4 using functions . Since each can be represented by a decision tree of depth , it remains to prove that impurity-based algorithms fail to learn . Lemma 7 guarantees that the tree returned by such algorithms either has depth , or never queries any addressing bits. In the latter case, by the same calculation as in the proof of the weaker version, the decision tree must have an error on distribution . ∎
6 Proof of Theorem 5
We prove Theorem 5 using the construction of in Section 5, where is a list of subsets of and each specifies how the i-th bit of the address, , is computed from the addressing bits through . Note that itself depends on input bits and is thus not an -junta. Nevertheless, we will show that, after we fix most of the memory bits of , the function is indeed close to a -junta with relevant inputs being the addressing bits. Then, as in the proof of Theorem 4, we will argue that impurity-based heuristics still query the (unfixed) memory bits before querying any of the addressing bits, resulting in a tree that is either exponentially deep or far from the target function.
Proof of Theorem 5.
As in the proof of Theorem 4, we can find functions for some such that each has distance . We fix a sufficiently large integer and shorthand for in the following.
Partition into three sets , and such that and . Consider the restriction of function such that the memory bit is fixed to be for every and fixed to be for every ; the memory bits with addresses in are left as “free” variables. We will prove the theorem using as the -th function in the family.
is close to a junta.
Consider the function defined as , where denotes and each . Clearly, only depends on and is thus a -junta. Furthermore, for every input such that (resp. ), both and evaluate to (resp. ). Thus, and may disagree only if . It follows that for every -balanced product distribution ,
(Lemma 6) | ||||
() |
Therefore, is -close to an -junta (namely, ) with respect to distribution .
Impurity-based algorithms fail to learn .
Let be the decision tree returned by an -impurity based algorithm when learning on distribution . By Lemma 7, must query all the free memory bits with addresses in before querying any of the addressing bits. Thus, either has depth , or only queries the free memory bits of .
In the latter case, we may again assume without loss of generality that queries all the free memory bits before reaching any of its leaves, i.e., is a full binary tree of depth . Then, every leaf naturally specifies bits defined as
Let . Again, the minimum possible error conditioning on reaching leaf is , achieved by labeling with . On the other hand, we have
(Lemma 6) | ||||
() | ||||
() |
and a similar calculation shows . We conclude that the error of the decision tree over distribution is at least . ∎
7 Conclusion
We have constructed target functions for which greedy decision tree learning heuristics fail badly, even in the smoothed setting. Our lower bounds complement and strengthen the parity-of-two-features example discussed in the introduction, which showed that these heuristics fail badly in the non-smoothed setting.
It can be reasonably argued that real-world data sets do not resemble the target functions considered in this paper or the parity-of-two-features example. Perhaps the sought-for guarantee , while false for certain target functions even in the smoothed setting, is nonetheless true for broad and natural classes of targets? It would be interesting to reexamine, through the lens of smoothed analysis, provable guarantees for restricted classes of functions that have been established. For example, can the guarantees of [BLT20b, BLT20a] for monotone target functions and product distributions be further strengthened in the smoothed setting? The target functions considered in this paper, as well as the parity-of-two-features example, are non-monotone.
Acknowledgements
We thank the anonymous reviewers, whose suggestions have helped improved this paper.
References
- [BDM19] Alon Brutzkus, Amit Daniely, and Eran Malach. On the Optimality of Trees Generated by ID3. ArXiv, abs/1907.05444, 2019.
- [BDM20] Alon Brutzkus, Amit Daniely, and Eran Malach. ID3 learns juntas for smoothed product distributions. In Proceedings of the 33rd Annual Conference on Learning Theory (COLT), pages 902–915, 2020.
- [BFJ+94] Avirm Blum, Merrick Furst, Jeffrey Jackson, Michael Kearns, Yishay Mansour, and Steven Rudich. Weakly learning DNF and characterizing statistical query learning using Fourier analysis. In Proceedings of the 26th Annual ACM Symposium on Theory of Computing (STOC), pages 253–262, 1994.
- [BFSO84] Leo Breiman, Jerome Friedman, Charles Stone, and Richard Olshen. Classification and regression trees. Wadsworth International Group, 1984.
- [BGLT20] Guy Blanc, Neha Gupta, Jane Lange, and Li-Yang Tan. Universal guarantees for decision tree induction via a higher-order splitting criterion. In Proceedings of the 34th Conference on Neural Information Processing Systems (NeurIPS), 2020.
- [BLT20a] Guy Blanc, Jane Lange, and Li-Yang Tan. Provable guarantees for decision tree induction: the agnostic setting. In Proceedings of the 37th International Conference on Machine Learning (ICML), 2020. Available at https://arxiv.org/abs/2006.00743.
- [BLT20b] Guy Blanc, Jane Lange, and Li-Yang Tan. Top-down induction of decision trees: rigorous guarantees and inherent limitations. In Proceedings of the 11th Innovations in Theoretical Computer Science Conference (ITCS), volume 151, pages 1–44, 2020.
- [Blu92] Avrim Blum. Rank- decision trees are a subclass of -decision lists. Inform. Process. Lett., 42(4):183–185, 1992.
- [Bsh93] Nader Bshouty. Exact learning via the monotone theory. In Proceedings of 34th Annual Symposium on Foundations of Computer Science (FOCS), pages 302–311, 1993.
- [CM19] Sitan Chen and Ankur Moitra. Beyond the low-degree algorithm: mixtures of subcubes and their applications. In Proceedings of the 51st Annual ACM Symposium on Theory of Computing (STOC), pages 869–880, 2019.
- [DKM96] Tom Dietterich, Michael Kearns, and Yishay Mansour. Applying the weak learning framework to understand and improve C4.5. In Proceedings of the 13th International Conference on Machine Learning (ICML), pages 96–104, 1996.
- [EH89] Andrzej Ehrenfeucht and David Haussler. Learning decision trees from random examples. Information and Computation, 82(3):231–246, 1989.
- [FP04] Amos Fiat and Dmitry Pechyony. Decision trees: More theoretical justification for practical algorithms. In Proceedings of the 15th International Conference on Algorithmic Learning Theory (ALT), pages 156–170, 2004.
- [GKK08] Parikshit Gopalan, Adam Kalai, and Adam Klivans. Agnostically learning decision trees. In Proceedings of the 40th ACM Symposium on Theory of Computing (STOC), pages 527–536, 2008.
- [Han93] Thomas Hancock. Learning decision trees on the uniform distribution. In Proceedings of the 6th Annual Conference on Computational Learning Theory (COT), pages 352–360, 1993.
- [HJLT96] Thomas Hancock, Tao Jiang, Ming Li, and John Tromp. Lower bounds on learning decision lists and trees. Information and Computation, 126(2):114–122, 1996.
- [HKY18] Elad Hazan, Adam Klivans, and Yang Yuan. Hyperparameter optimization: A spectral approach. Proceedings of the 6th International Conference on Learning Representations (ICLR), 2018.
- [JS06] Jeffrey C. Jackson and Rocco A. Servedio. On learning random dnf formulas under the uniform distribution. Theory of Computing, 2(8):147–172, 2006.
- [Kea96] Michael Kearns. Boosting theory towards practice: recent developments in decision tree induction and the weak learning framework (invited talk). In Proceedings of the 13th National Conference on Artificial intelligence (AAAI), pages 1337–1339, 1996.
- [KM93] Eyal Kushilevitz and Yishay Mansour. Learning decision trees using the fourier spectrum. SIAM Journal on Computing, 22(6):1331–1348, December 1993.
- [KM96] Michael Kearns and Yishay Mansour. On the boosting ability of top-down decision tree learning algorithms. In Proceedings of the 28th Annual Symposium on the Theory of Computing (STOC), pages 459–468, 1996.
- [KM99] Michael Kearns and Yishay Mansour. On the boosting ability of top-down decision tree learning algorithms. Journal of Computer and System Sciences, 58(1):109–128, 1999.
- [KS06] Adam Klivans and Rocco Servedio. Toward attribute efficient learning of decision lists and parities. Journal of Machine Learning Research, 7(Apr):587–602, 2006.
- [KST09] Adam Kalai, Alex Samorodnitsky, and Shang-Hua Teng. Learning and smoothed analysis. In Proceedings of the 50th Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 395–404, 2009.
- [Lee09] Homin Lee. On the learnability of monotone functions. PhD thesis, Columbia University, 2009.
- [MR02] Dinesh Mehta and Vijay Raghavan. Decision tree approximations of boolean functions. Theoretical Computer Science, 270(1-2):609–623, 2002.
- [OS07] Ryan O’Donnell and Rocco Servedio. Learning monotone decision trees in polynomial time. SIAM Journal on Computing, 37(3):827–844, 2007.
- [Qui86] Ross Quinlan. Induction of decision trees. Machine learning, 1(1):81–106, 1986.
- [Qui93] Ross Quinlan. C4.5: Programs for Machine Learning. Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, 1993.
- [Riv87] Ronald Rivest. Learning decision lists. Machine learning, 2(3):229–246, 1987.
- [ST04] Daniel A. Spielman and Shang-Hua Teng. Smoothed analysis of algorithms: Why the simplex algorithm usually takes polynomial time. Journal of the ACM, 51(3):385–463, 2004.
- [WFHP16] Ian Witten, Eibe Frank, Mark Hall, and Christopher Pal. Data Mining: Practical machine learning tools and techniques. Morgan Kaufmann, 2016.
- [WKQ+08] Xindong Wu, Vipin Kumar, J Ross Quinlan, Joydeep Ghosh, Qiang Yang, Hiroshi Motoda, Geoffrey J McLachlan, Angus Ng, Bing Liu, S Yu Philip, et al. Top 10 algorithms in data mining. Knowledge and information systems, 14(1):1–37, 2008.