The Role of Entropy
in Guiding a Connection Prover
Abstract
In this work we study how to learn good algorithms for selecting reasoning steps in theorem proving. We explore this in the connection tableau calculus implemented by leanCoP where the partial tableau provides a clean and compact notion of a state to which a limited number of inferences can be applied. We start by incorporating a state-of-the-art learning algorithm – a graph neural network (GNN) – into the plCoP theorem prover. Then we use it to observe the system’s behavior in a reinforcement learning setting, i.e., when learning inference guidance from successful Monte-Carlo tree searches on many problems. Despite its better pattern matching capability, the GNN initially performs worse than a simpler previously used learning algorithm. We observe that the simpler algorithm is less confident, i.e., its recommendations have higher entropy. This leads us to explore how the entropy of the inference selection implemented via the neural network influences the proof search. This is related to research in human decision-making under uncertainty, and in particular the probability matching theory. Our main result shows that a proper entropy regularization, i.e., training the GNN not to be overconfident, greatly improves plCoP’s performance on a large mathematical corpus.
Keywords:
automated theorem proving machine learning reinforcement learning graph neural networks connection calculus entropy regularization1 Introduction
Automated Theorem Proving (ATP) and Interactive Theorem Proving (ITP) are today increasingly benefiting from combinations with Machine Learning (ML) methods [44]. A number of learning-based inference guiding methods have been developed recently, starting with the leanCoP [34, 33] style connection tableaux setting [46, 23, 12, 25, 51, 35, 32], later expanding into the E prover’s [38, 39] and Vampire’s [28] superposition setting [19, 29, 9, 18, 42], and HOL’s [15, 16, 41], Coq’s [10] and Isabelle’s [48] tactical settings [13, 31, 17, 50, 4, 6, 14].
The connection tableau calculus as implemented by leanCoP is a very good framework for studying combinations of ML and ATP methods [5]. leanCoP has a compact Prolog implementation that is both easy to modify and surprisingly efficient. At the same time, unlike in the superposition and tactical setting, the partial tableau provides a clean and compact notion of a state to which a limited number of inferences (actions) can be applied. This has recently allowed the first experiments with AlphaZero [40] style Monte-Carlo tree search (MCTS) [12] and Reinforcement Learning (RL) [43] of theorem proving in the rlCoP [25], graphCoP [32] and plCoP [51] systems.
In this work, we start by extending plCoP with a state-of-the-art learning algorithm –- a graph neural network (GNN) [32] – which was designed for processing logical formulae and exhibits several useful invariance properties, namely invariance under renaming of symbols, negation and reordering of clauses and literals. Despite its better pattern matching capability, the GNN initially performs worse than a simpler previously used learning algorithm based on gradient boosted trees (XGBoost [8]). We observe that the simpler algorithm is less confident about the inferences that should be applied to the proof states, i.e., its recommendations have higher entropy, leading to greater exploration of different inferences.
This leads us to analyze how the entropy of the inference selection implemented via the neural network influences the proof search. We try increasingly high penalties for overconfidence (low entropy) during the training of the GNN, using an approach called Maximum Entropy Reinforcement Learning [49]. For this, we need to be able to compare the entropy of proof states with different numbers of possible inferences (actions). We do that by introducing normalized entropy, which allows for comparing discrete distributions of different lengths. We make a rather surprising discovery that replacing the particular trained predictors by arbitrary (random) but entropy-normalized predictors that respect the inference ordering yields only slightly weaker ATP performance. This suggests that the correct ordering of possible inferences according to their utility in a given state plus the right amount of entropy capture most of the benefits of the learned guidance. In summary, our contributions are:
-
1.
We integrate a fast logic-aware graph neural network into the plCoP system, allowing its use for guiding the choice of inferences (policy) and for estimating the provability of a partial connection tableau (value).
-
2.
We adapt the graph construction algorithm to support the paramodulation inferences used in plCoP.
-
3.
We analyze the entropy of the policy and its role in plCoP’s performance.
-
4.
We show that random policies with the right ordering and normalized entropy perform already quite well.
-
5.
We do several smaller and larger evaluations over the standard corpus extracted from the Mizar Mathematical Library (MML) and show that the best entropy regularized GNN greatly improves over other learning-guided connection tableaux systems. In particular, we report improvement on the Mizar40 evaluation set over rlCoP, the best previously published result.
The rest of the paper is structured as follows. Section 2 introduces in more detail the necessary background such as neural guidance of provers, the leanCoP setting, reinforcement learning and Monte-Carlo tree search, and Maximum Entropy learning. Section 3 discusses in more depth the use of Maximum Entropy learning in guiding MCTS-based theorem proving. Section 4 describes our new implementation and Section 5 experimentally evaluates the methods.
2 Background and Related Work
2.1 Neural Feature Extraction for Guiding Theorem Provers
Learning based ATP systems have for a long time explored suitable characterizations of mathematical objects, leading to solutions that process text directly (e.g. [2, 29, 4]) and solutions that rely on manually engineered features (e.g. [26, 18]). Graph neural networks (GNN) [37] provide an alternative to both approaches: the graph representation allows for retaining the syntactic structure of mathematical objects, while also allowing for end-to-end (i.e., involving no manually designed features) training. However, improving over learning based on manual feature extraction has proven to be challenging with GNNs, especially in real time, as noted in several works (e.g. [9, 11]). Usually it required high level of technical engineering. The GNN presented in [32] was designed to preserve many useful invariance properties of logical formulae and has demonstrated impressive improvement in guiding the leanCoP connection calculus compared with gradient boosted trees. We refer to this system as graphCoP.
2.2 Systems Guiding the leanCoP Theorem Prover
leanCoP [34] is a compact theorem prover for first-order logic, implementing connection tableau search. The proof search starts with a start clause as a goal and proceeds by building a connection tableau by applying extension steps and reduction steps. leanCoP uses iterative deepening to ensure completeness.
A series of learning systems guiding the leanCoP connection calculus have been developed recently. Of these, we highlight three that use roughly the same reinforcement learning setup: rlCoP [25], plCoP [51] and graphCoP [32]. These systems search for proofs using Monte Carlo Tree Search [7] and they train the value (the proof state quality) and the policy (the inference quality in a proof state) functions similarly to systems like AlphaZero [40, 3]. rlCoP and plCoP use manually developed features [26] and gradient boosted trees (XGBoost [8]) for learning while graphCoP employs a GNN for end-to-end feature extraction and learning. This graph neural network was designed for processing mathematical formulae and has several useful invariance properties: the graph structure is invariant under renaming of symbols, negation and reordering of clauses and literals. In this paper, we incorporate the GNN of [32] into plCoP.
The plCoP system extends leanCoP with paramodulation steps that can handle equality predicates more efficiently. Let denote the subterm of at position and denote the term obtained after replacing in at position by term . Given a goal and an input clause 111Assuming Disjunctive Normal Form. , such that, for some position there is a substitution such that , the paramodulation step changes to . Rewriting is allowed in both directions, i.e., the roles of and can be switched.
2.3 Reinforcement Learning (RL)
Reinforcement learning (RL) [43] aims to find the optimal behaviour in an environment defined as a Markov Decision Process (MDP). An describes a dynamic process and consists of the following components: is the set of possible states, is the set of possible actions, is a reward function, is the state transition function and is the discount factor. We assume that an agent interacts with this MDP, generating sequences of state-action-reward tuples, called trajectories. The agent is equipped with a policy function which determines which action it selects in a particular state. The policy is often stochastic, i.e., it defines a probability distribution over inferences that are possible in a given state. The aim of the agent is to maximize its total accumulated reward . Several components, such as the reward and transition functions, can be stochastic, in which case the aim of the agent it to find the policy that maximizes its cumulative expected reward, where future rewards are discounted with the discount factor:
2.4 Monte Carlo Tree Search (MCTS)
MCTS is a simple RL algorithm, which builds a search tree whose nodes are states and where edges represent actions. The aim of the search algorithm is to find trajectories (branches in the search tree) that yield high accumulated rewards. The search starts from a single root node, and new nodes are added iteratively. In each node , we maintain the number of visits , the total reward , and the prior probability (estimated typically by a trained predictor) of all its possible successors (in our case produced by the possible inferences). Each iteration, also called playout, involves the addition of a new leaf node.
The policy used for selecting the actions of the playout is based on the standard UCT [27] (Upper Confidence Trees) formula (1): in each state we select the action with the maximal UCT value in the successor state. Once a new leaf state is created, we observe its reward and update its ancestors: visit counts are increased by 1 and rewards are increased by the reward of the leaf.
(1) |
In (1), is the number of visits of the parent, and is a parameter that determines the balance between nodes with high reward (exploitation) and rarely visited nodes (exploration). In [3, 40] MCTS is augmented with two learned functions. The value function estimates the accumulated reward obtainable from a state, and the leaf nodes are initialized with this value estimates. The second function predicts the prior probability of state-action pairs, which is usually referred to as the policy, with a slight abuse of terminology. When it can lead to confusion, we refer to this policy as .
The plCoP, rlCoP and graphCoP systems use the MaLARea/DAgger [45, 36] meta-learning algorithm to learn the policy and value functions. They interleave ATP runs based on the current policy and value (data collection phase) with a training phase, in which these functions are updated to fit the collected data. Such iterative interleaving of proving and learning has been used successfully in ATP systems such as MaLARea [45] and ENIGMA [20].
During the proof search we build a Monte Carlo tree for each training problem. Its nodes are the proof states (partial tableaux), and the edges represent inferences. Note that the Monte Carlo tree is thus different from the tableau trees. A branch of this Monte Carlo tree leading to a node with a closed tableau is a valid proof. Initially, the three leanCoP-based systems use somewhat different heuristic value and policy functions, later to be replaced with the learned guidance. To enforce deeper exploration, we perform a bigstep after a fixed number of playouts: the starting node of exploration is moved one level down towards the child with the highest value (called the bigstep node). Later MCTS steps thus only extend the subtree under the bigstep node. This in practice means no backtracking of the bigsteps, which in turn involves giving up completeness.
2.5 Maximum Entropy Reinforcement Learning
When training the policy, directly maximizing the expected utility on the action sequences observed by an RL agent (i.e., the training examples) can lead to instability. The policy can get stuck in local minima and become overconfident, preventing it from exploring the search space sufficiently when necessary to make good decisions. This has motivated using stochastic policies and several regularization (i.e., encouraging generality) techniques that ensure that all actions have a chance of being selected. Another motivation for properly regularized stochastic policy learning comes from experiments on humans and animals, suggesting that biological agents do not deterministically select the action with the greatest expected utility [47]: instead they randomly select actions with probability proportional to the expected utility, called probability matching. Consequently, action sequences that generate similar rewards tend to be similarly probable, i.e., we avoid making strong commitments whenever it is possible. Maximum Entropy Reinforcement Learning (MaxEnt RL), achieves probability matching by adding an entropy term to the loss founction when the policy is trained:
where is the Shannon entropy of the probability distribution over valid actions in state :
and is the entropy coefficient. This means that the training of the policy will be maximizing a weighted sum of the (discounted) rewards and of the entropy of the resulting distribution, thus discouraging overconfidence. The entropy term in the objective was first used in [49] and since then its benefit has been empirically demonstrated in several domains. It is particularly useful in dynamic environments, where some uncertainty remains, irrespective of the amount of exploration.
2.6 Kullback-Leibler divergence
Shannon’s entropy measures the uniformity of a single distribution. However, when contrasting different policies, we will need a measure to compare different distributions. For this, one of the most widely used options is the Kullback-Leibler (KL) divergence, also called relative entropy, which is a measure of how one probability distribution differs from a given reference distribution. For a discrete target distribution and a reference distribution , the KL divergence is defined as:
This measure is zero exactly when and are identical, otherwise it is positive. It can be infinite if there is some such that and . A small means that the two distributions are similar on the domain where most of the probability mass of lies. Note that . For example, consider the following distributions:
and . When the summed terms are weighted according to in , the first two terms get large weight, while only the first term gets large weight in . When both KL divergences are small, it is a good indicator of similarity of the two distributions.
3 Maximum Entropy for MCTS and Theorem Proving
In this section we discuss the entropy of the inference policy and its potential influence on the proof search. We also introduce a normalized entropy to correctly handle probability vectors of different lengths and argue that MaxEnt RL is more targeted and powerful than previously used temperature-based entropy control. Section 4 then describes our implementation.
3.1 Exploration and Entropy in MCTS
The MCTS implemented via the UCT formula has a built-in mechanism for balancing the exploitation of proof states that already have high rewards and the exploration of inferences whose effect is not yet known. This balancing serves to mitigate errors in the proof state value estimates. However, we need to do another balancing within exploration, between the different under-explored inference branches. This is estimated by the policy predictor as the prior probabilities of the possible inferences. Hence, besides the ordering of the inferences that are possible from a given state, their exact prior probabilities are important as they determine how the exploration budget is split between them. This observation directs our attention to the entropy (uncertainty) of and its relation to the theorem proving performance.
We argue that finding the right level of (un)certainty is particularly important for theorem proving. The goal of learning is to acquire inductive biases that allow the system to perform well on novel problems.222In this sense, theorem proving can be considered as a meta learning task. In many situations, however, it is not realistic to extract enough knowledge from the training data that justifies a high level of confidence. Sometimes, there is just not enough analogy between a new problem and the training problems, and we would like our guidance to be more conservative so that we can explore all directions equally. This makes a strong case for using MaxEnt RL, which gives tools for shaping the entropy (uncertainty) profile of our learned policy predictor.
3.2 Normalized Entropy
In this work, we empirically demonstrate the importance of including the “right amount” of entropy when training the policy that guides the theorem prover. To the best of our knowledge, this is the first time that the effect of entropy regularization for MCTS in general and for theorem proving in particular is examined.
Using standard entropy for comparing probability vectors of different length would, however, be misleading. The same entropy value can mean very different uncertainty if the length of the vector changes. For example, consider the vectors
Their entropy is roughly the same (1.1), despite the fact that the first is nearly uniform and the second centers around its first value. To make the uncertainty of these two vectors comparable, we introduce normalized entropy:
Definition 1
Given a discrete probability vector of length , let denote the normalized entropy of .
Here, is the entropy of the uniform distribution when the length is , hence it is the upper bound of . Consequently, . Furthermore, it is dimensionless, i.e., it does not depend on the base of the logarithm. The difference between the two distributions in the example above is better captured by their normalized entropy, which is 1 and 0.48.
3.3 Temperature-based and Regularization-based Entropy Control
An alternative mechanism for injecting entropy into the policy is through the softmax temperature parameter . Our policy predictors (both XGBoost and GNN) output an unconstrained logit vector , which is normalized to a probability vector using the softmax function:
Increasing the temperature flattens the probability curve, approaching the uniform distribution in the limit. On the other hand, if the temperature gets close to , then most of the probability mass concentrates on the most likely action.
While both higher temperature and entropy regularization increase the ultimate entropy of the policy, they work differently. The temperature acts globally and uniformly, flattening all inference probabilities estimated by the trained policy predictor. Entropy regularization, on the other hand, is part of the training process and it allows the neural network to learn distinguishing between situations with low and high uncertainty. In obvious situations, entropy regularization does not prevent the neural network from acquiring great certainty, while it will drive the network to more uniform predictions when there is no strong evidence against that in the training data. Hence, we expect entropy regularization to be more targeted and powerful than the temperature optimization. This is empirically demonstrated in Section 5.
4 Entropy Regularized Neural Guidance for plCoP
This section gives an overview of the training procedure, including how data is extracted from tableaux and Monte Carlo trees.
4.1 Neural Representation of the State and Inference Steps
The proof state in the leanCoP setting is roughly described by the partial tableau and by the set of input clauses corresponding to the initial axioms and conjecture. Each time we choose an extension step, we map the state into a hypergraph, as described in [32]. In more detail, we use the current goal (single literal), the active path leading to the current goal (set of literals), the set of all open goals (set of literals), and the input clauses in the hypergraph construction. The GNN processes the hypergraph and outputs a value prediction for the proof state in the range . It also outputs a probability distribution over all the literals of the axiom clauses that can be used for extension steps, i.e., that can be unified with the negation of the current goal literal.
The above method is used already in graphCoP, but the hypergraph construction algorithm used by graphCoP was designed to guide only the extension steps. Hence it expects the set of all clauses together with the information that identifies literals within the clauses that unify with the negation of the current goal. We adapt this to paramodulation by temporarily creating a clause for each valid paramodulation step that “simulates” the latter as an extension step. Suppose that the current goal is and there is an input clause , such that, for some position there is a substitution such that . There is a valid paramodulation step that replaces with . We simulate this step as an extension by adding clause to the input clauses, when constructing the graph.
4.2 Training the Policy and Value Guidance for MCTS
As in plCoP, the value and policy estimates are stored in the MCTS nodes and are used for guiding proof search. The training data for learning policy (inference probabilities) and value (state quality) are also handled as in plCoP. They are extracted from the tableau states of the bigstep nodes. For each bigstep state, the value target is 1333A discount factor of 0.99 is applied to positive rewards to favor shorter proofs. if it leads to a proof and 0 otherwise. The policy targets at a particular proof state are the relative frequencies of the possible inferences, i.e., the children in the search tree.
For graphCoP, a single GNN was jointly trained to predict both the value and the policy [32]. However, we observed that training separate predictors yields a small improvement, hence we conduct our experiments in this setup. Consider a tableau state for which we want to learn its target value and target policy . Suppose that the partially trained value predictor outputs and the policy predictor outputs , then the objectives that we minimize are:
-
•
value objective:
-
•
policy objective:
For more details of the graph construction and neural training we refer to [32]. In summary, we use the same setting as there, except for (i) extending guidance to paramodulation steps, (ii) training separate policy and value GNNs, (iii) increasing the number of neural layers in the GNN from 5 to 10,444This is motivated by the experiments with the ENIGMA-GNN system [18], where 8-10 layers produce better results than 5 layers. and (iv) changing the policy training to encourage policies with higher entropy.
5 Experiments
We first introduce our datasets and other experimental settings (Section 5.1). Then we show the impact of entropy regularization (Section 5.2) and experimentally compare the entropies and other characteristics of the XGBoost and GNN predictors (Section 5.3). In Section 5.4, we show that random policies with the right ordering and normalized entropy perform already quite well. Section 5.5 then compares temperature-based entropy control with our approach. Finally, Section 5.6 evaluates the methods in a train/test scenario on the full Mizar40 dataset using several iterations of proving and learning.
5.1 Datasets and Common Settings
We evaluate our system555The new extensions described here and the experimental configuration files are publicly available at plCoP’s repository: https://github.com/zsoltzombori/plcop. using the same datasets as those in [25]. The Mizar40 dataset [22] consists of 32524 problems from the Mizar Mathematical Library that have been proven by several state-of-the-art ATPs used with many strategies and high time limits in the experiments described in [24]. Based on the proofs, the axioms were ATP-minimized, i.e., only those axioms were kept that were needed in any of the ATP proofs found. The smaller M2k dataset [21] consists of 2003 Mizar40 problems that come from related Mizar articles. Finally, we use the bushy (small) problems from the MPTP2078 benchmark [1], which contains just an article-based selection of Mizar problems, regardless of their solvability by a particular ATP system.
Unless otherwise specified, we use the same hyperparameters as described in [51], with the following important exceptions. To allow for faster experiments and put more emphasis on guidance instead of search, we reduce the per problem inference limit from 200000 to 20000 and the bigstep frequency from 2000 to 200. Hence the overall search budget is reduced by a factor of 10. We use a very large CPU time limit (300 sec) intended to ensure that proof search terminates after exhausting the inference limit.
5.2 Experiment 1: Influence of Entropy Regularization
In this experiment, we examine the effect of regularizing the entropy of our policy predictor. We produce several variants of the GNN policy predictor which differ in the entropy coefficient used in its training. Table 1 summarizes our results. We find that the entropy coefficient has a big impact on performance. By the 10th iteration, the best GNN predictor with is better than the unregularized GNN and better than XGBoost. Table 1 also shows the average entropy of the policies generated by the predictor during the proof search. Note that the average entropy of the best predictor in most iterations is reasonably close (often the closest) to the entropy of the XGBoost predictor. This suggests that one key strength of the XGBoost predictor is that it hits the “right” amount of entropy. Matching this entropy in the GNN with adequate regularization allows for matching and even surpassing XGBoost in performance.
Iter 1 | Iter 2 | Iter 4 | Iter 6 | Iter 8 | Iter 10 | ||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Model | Ent | Succ | Ent | Succ | Ent | Succ | Ent | Succ | Ent | Succ | Ent | Succ | |
XGB | 1.41 | 790 | 1.29 | 956 | 1.22 | 1061 | 1.19 | 1119 | 1.17 | 1147 | 1.14 | 1171 | |
GNN | 0 | 0.91 | 746 | 0.56 | 850 | 0.37 | 938 | 0.34 | 992 | 0.31 | 1021 | 0.32 | 1050 |
GNN | 0.1 | 0.86 | 787 | 0.6 | 867 | 0.43 | 933 | 0.37 | 996 | 0.37 | 1031 | 0.38 | 1070 |
GNN | 0.2 | 1.11 | 769 | 0.71 | 878 | 0.51 | 976 | 0.51 | 1045 | 0.49 | 1077 | 0.46 | 1114 |
GNN | 0.3 | 1.05 | 736 | 0.8 | 868 | 0.7 | 991 | 0.73 | 1071 | 0.69 | 1109 | 0.78 | 1170 |
GNN | 0.5 | 1.31 | 781 | 1.14 | 884 | 1.17 | 1015 | 1.13 | 1085 | 1.12 | 1144 | 1.06 | 1191 |
GNN | 0.6 | 1.37 | 759 | 1.25 | 889 | 1.26 | 1040 | 1.21 | 1098 | 1.18 | 1150 | 1.19 | 1197 |
GNN | 0.7 | 1.41 | 727 | 1.32 | 854 | 1.27 | 1057 | 1.22 | 1132 | 1.24 | 1184 | 1.2 | 1228 |
GNN | 0.8 | 1.42 | 757 | 1.37 | 912 | 1.35 | 1029 | 1.32 | 1079 | 1.29 | 1111 | 1.3 | 1144 |
GNN | 1.0 | 1.53 | 742 | 1.41 | 911 | 1.38 | 1032 | 1.35 | 1102 | 1.36 | 1144 | 1.35 | 1173 |
GNN | 2.0 | 1.59 | 725 | 1.57 | 782 | 1.53 | 894 | 1.5 | 1007 | 1.5 | 1047 | 1.5 | 1086 |
5.3 Experiment 2: Relative Entropy on the Same Proof States
Table 1 reveals that there is a reasonable match in average entropy between XGBoost policies and our best GNN policies. Note, however, that this is in general measured on different proof states as the policies themselves determine what proof states the prover explores. To gain a deeper understanding, we create a separate dataset of proof states and compare the different GNNs from Experiment 1 with XGBoost on these proof states using the following four metrics: 1) fraction of proof states where the two predictors have the same most probable inference (Best), 2) fraction of proof states where the two predictors yield the same inference ordering (Order), and the average KL divergence (relative entropy) between the predictors in both directions: 3) and 4) .666 and stand for the probability distributions predicted by XGBoost and GNN, respectively. We contrast these metrics with the number of problems solved (Succ) by the corresponding entropy regularized GNN.
We perform this comparison using two datasets. These are the set of states visited by an unguided prover on the 1) M2k dataset and the 2) MPTP2078 bushy benchmark. The first set is part of the training corpus, while the second was never seen by the predictors before. The results can be seen in Table 2.
M2K | MPTP2078b | |||||||||
---|---|---|---|---|---|---|---|---|---|---|
Succ | Best | Order | Succ | Best | Order | |||||
0 | 1050 | 0.81 | 0.43 | 0.52 | 2.9 | 230 | 0.56 | 0.22 | 0.97 | 4.5 |
0.1 | 1070 | 0.8 | 0.44 | 0.5 | 2.37 | 245 | 0.58 | 0.24 | 0.91 | 3.83 |
0.2 | 1114 | 0.81 | 0.42 | 0.47 | 1.66 | 256 | 0.56 | 0.24 | 0.88 | 2.82 |
0.3 | 1170 | 0.82 | 0.42 | 0.36 | 0.58 | 276 | 0.56 | 0.23 | 0.61 | 0.9 |
0.5 | 1191 | 0.82 | 0.42 | 0.24 | 0.28 | 335 | 0.59 | 0.23 | 0.41 | 0.43 |
0.6 | 1197 | 0.82 | 0.4 | 0.22 | 0.23 | 359 | 0.59 | 0.23 | 0.36 | 0.36 |
0.7 | 1228 | 0.82 | 0.4 | 0.22 | 0.21 | 399 | 0.58 | 0.22 | 0.34 | 0.32 |
0.8 | 1144 | 0.81 | 0.39 | 0.22 | 0.21 | 357 | 0.58 | 0.22 | 0.34 | 0.31 |
1.0 | 1173 | 0.82 | 0.4 | 0.24 | 0.21 | 363 | 0.58 | 0.22 | 0.33 | 0.29 |
2.0 | 1086 | 0.81 | 0.39 | 0.34 | 0.26 | 362 | 0.58 | 0.21 | 0.37 | 0.3 |
Changing the entropy coefficient mostly does not change the order of actions, as expected. For the two datasets, the GNN and the XGBoost predictors select the same best inference in around and of the states and yield exactly the same inference ordering in around and of the states. This reveals a significant diversity among the two predictor families, suggesting potential in combining them. We leave this direction for future work.
We find that the same level of entropy regularization () is the best when running both on the familiar (M2k) and the previously unseen (MPTP2078) dataset. This is where the two directional KL divergences (relative entropies) roughly coincide and their sum is roughly minimal. These results make a stronger case for the hypothesis from Experiment 1, that the best GNN performance is obtained when the policy distributions are statistically close to those of the XGBoost predictor.
5.4 Experiment 3: Order and Entropy Are Largely Sufficient
Tables 1 and 2 demonstrate the importance of the entropy of the inference policy in ATP performance. To make this even more apparent, we design an experiment in which we remove a large part of the information contained in the inference policy, only preserving the inference ordering and the normalized entropy. The top two lines of Table 3 show the normalized entropy across iterations of the XGBoost and GNN predictors. Note that it is very stable. We select a target normalized entropy and for each length we generate a fixed random discrete probability of length whose normalized entropy is . Finally, we run an MCTS evaluation in which each time our policy predictor emits a probability vector of length , we replace it with , permuted so that the ordering remains the same as in the original policy. Table 3 shows the ATP performance for the differently normalized entropy targets.
We find that the performance of this predictor is surprisingly good: its performance (1154) is only worse than XGBoost (1171), better than unregularized GNN (1050) and worse than the best GNN (1228). This suggests that the right inference ordering plus the right amount of entropy capture most of the benefits of the learned inference guidance.
Iteration | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
---|---|---|---|---|---|---|---|---|---|---|---|
XGBoost | 1 | 0.73 | 0.71 | 0.69 | 0.68 | 0.67 | 0.67 | 0.67 | 0.66 | 0.67 | 0.66 |
GNN () | 1 | 0.83 | 0.79 | 0.8 | 0.79 | 0.79 | 0.78 | 0.78 | 0.78 | 0.8 | 0.78 |
GNN | 523 | 700 | 782 | 849 | 909 | 956 | 984 | 1019 | 1037 | 1059 | 1083 |
GNN | 523 | 702 | 800 | 856 | 922 | 954 | 995 | 1040 | 1077 | 1110 | 1129 |
GNN | 523 | 693 | 832 | 938 | 1023 | 1054 | 1086 | 1077 | 1115 | 1129 | 1154 |
5.5 Experiment 4: Temperature vs. Entropy Regularization
As noted in Section 3, tuning the softmax temperature is an alternative to entropy regularization. For XGBoost, the temperature was previously optimized to be and all reported experiments use this number. For the GNN predictors, we used the default . In Table 4, we show how the ATP performance of the GNN changes after it has been trained for 10 iterations on the M2k dataset (without entropy regularization). Increasing the temperature brings some improvement, however, this is much smaller than the benefit of entropy regularization. This is true even if we take the best temperature () and perform a full 10 iteration training with this temperature, as shown in Table 5. We obtain improvement via the temperature optimization, compared with improvement via the entropy regularization. We conclude that the effect of entropy regularization is much more refined and powerful than just flattening the probability curve.
Model | ||||||
---|---|---|---|---|---|---|
GNN | 1036 | 1050 | 1057 | 1066 | 1068 | 1061 |
Iteration | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
---|---|---|---|---|---|---|---|---|---|---|---|
GNN | 523 | 746 | 850 | 899 | 938 | 971 | 992 | 1012 | 1021 | 1023 | 1050 |
GNN | 523 | 705 | 800 | 864 | 894 | 931 | 993 | 1017 | 1049 | 1065 | 1079 |
5.6 Experiment 5: Final Large Train/Test Evaluation on Mizar40
Finally, we perform a large evaluation of plCoP using XGBoost and GNN on the full Mizar40 dataset, and we compare its performance with rlCoP and graphCoP. This evaluation, including the training of the GNNs on the growing sets of proofs generated from the successive proving/learning iterations, takes over 10 days on a large multicore server, and the number of training policy/value examples extracted from the MCTS proof searches goes over 5M in the last iteration.
The 32524 problems are randomly split using a 9:1 ratio into 29272 training problems and 3252 evaluation problems. For consistency, we employ the same split that was used in [25]. Successive predictors are only trained on data extracted from the training problems, but we always evaluate the predictors on both the training and evaluation sets and report the number of proofs found for them. Our results can be found in Table 6, with plCoP/GNN solving in the last iteration 16906 training problems and 1767 evaluation problems. These are the highest numbers obtained so far with any learning-guided leanCoP-based system.
For training the GNN policy predictor, we use the entropy regularization coefficient () that worked best on M2k, without its further tuning on this larger dataset. Note that the resource limits are higher in Table 6 for rlCoP and also a bit different for graphCoP (which was run only for a few iterations), as we took their published results from [25, 32] rather than rerunning the systems. Also, the evaluation of plCoP with GNN was stopped after iteration 8 due to our resource limits and the clear flattening of the performance on the evaluation set (1767 vs 1758 in the 8th vs 7th iteration).
Iteration | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
---|---|---|---|---|---|---|---|---|---|---|---|
Train set | |||||||||||
rlCoP | 7348 | 12325 | 13749 | 14155 | 14363 | 14403 | 14431 | 14342 | 14498 | 14481 | 14487 |
graphCoP | 4595 | 11978 | 12648 | 12642 | |||||||
plCoP XGB | 4904 | 8917 | 10600 | 11221 | 11536 | 11627 | 11938 | 11999 | 12085 | 12063 | 12151 |
plCoP GNN | 4888 | 8704 | 12630 | 14566 | 15449 | 16002 | 16467 | 16745 | 16906 | ||
Eval set | |||||||||||
rlCoP | 804 | 1354 | 1519 | 1566 | 1595 | 1624 | 1586 | 1582 | 1591 | 1577 | 1621 |
graphCoP | 510 | 1322 | 1394 | 1360 | |||||||
plCoP XGB | 554 | 947 | 1124 | 1158 | 1177 | 1204 | 1217 | 1210 | 1212 | 1213 | 1204 |
plCoP GNN | 554 | 969 | 1375 | 1611 | 1650 | 1730 | 1742 | 1758 | 1767 |
In particular, plCoP was given 20000 inferences per problem, i.e., one tenth of the inference limit used for rlCoP in [25]. For a fair comparison with rlCoP, we thus take the predictors (both XGBoost and GNN) used in the last plCoP iteration (iteration 10 for XGBoost and iteration 8 for GNN) and run plCoP with them on the evaluation set with 200000 inference limit and 2000 bigstep frequency, which corresponds to the limits used for rlCoP in [25]. To ensure that the system has enough time to exhaust its inference limit, we increase the timeout to 6000 seconds. plCoP with XGBoost then solves 1499 of the evaluation problems while plCoP with GNN solves 1907 (58.6%) of them. This is our final evaluation result, which is 17.4% higher than the 1624 evaluation problems solved by rlCoP in the best previously published result so far [25].
6 Conclusion
We have extended the plCoP learning-based connection prover with a fast, logic-aware graph neural network (GNN) and explored how the GNN can learn good guidance for selecting inferences in this setting. We have identified the entropy of the inference selection predictor as a key driver of the ATP performance and shown that Maximum Entropy Reinforcement Learning largely improves the performance of the trained policy network, outperforming simpler temperature-based entropy increasing methods. To the best of our knowledge, this is the first time that the role of entropy in guiding a theorem prover has been analyzed.
We have discovered that replacing the particular trained predictors by arbitrary (random) but entropy-normalized predictors that respect the inference ordering yields only slightly weaker theorem proving performance than the best methods. This suggests that the right inference ordering plus the right amount of entropy capture most of the benefits of the learned inference guidance. In the large final train/test evaluation on the full Mizar40 benchmark our system improves by over the best previously published result achieved by rlCoP.
7 Acknowledgments
ZZ was supported by the European Union, co-financed by the European Social Fund (EFOP-3.6.3-VEKOP-16-2017-00002), the Hungarian National Excellence Grant 2018-1.2.1-NKP-00008 and by the Hungarian Ministry of Innovation and Technology NRDI Office within the framework of the Artificial Intelligence National Laboratory Program. JU was funded by the AI4REASON ERC Consolidator grant nr. 649043 and the European Regional Development Fund under the Czech project AI&Reasoning CZ.02.1.01/0.0/0.0/15_003/0000466. MO was supported by the ERC starting grant no. 714034 SMART. We thank the TABLEAUX’21 reviewers for their thoughtful reviews and comments.
References
- [1] J. Alama, T. Heskes, D. Kühlwein, E. Tsivtsivadze, and J. Urban. Premise selection for mathematics by corpus analysis and kernel methods. J. Autom. Reasoning, 52(2):191–213, 2014.
- [2] A. A. Alemi, F. Chollet, N. Een, G. Irving, C. Szegedy, and J. Urban. Deepmath - Deep Sequence Models for Premise Selection. In Proceedings of the 30th International Conference on Neural Information Processing Systems, NIPS’16, pages 2243–2251, USA, 2016. Curran Associates Inc.
- [3] T. Anthony, Z. Tian, and D. Barber. Thinking fast and slow with deep learning and tree search. CoRR, abs/1705.08439, 2017.
- [4] K. Bansal, S. M. Loos, M. N. Rabe, C. Szegedy, and S. Wilcox. HOList: An environment for machine learning of higher order logic theorem proving. In K. Chaudhuri and R. Salakhutdinov, editors, International Conference on Machine Learning, ICML 2019, volume 97 of Proceedings of Machine Learning Research, pages 454–463. PMLR, 2019.
- [5] W. Bibel. A vision for automated deduction rooted in the connection method. In R. A. Schmidt and C. Nalon, editors, Automated Reasoning with Analytic Tableaux and Related Methods - 26th International Conference, TABLEAUX 2017, volume 10501 of LNCS, pages 3–21. Springer, 2017.
- [6] L. Blaauwbroek, J. Urban, and H. Geuvers. Tactic learning and proving for the Coq proof assistant. In E. Albert and L. Kovács, editors, LPAR 2020: 23rd International Conference on Logic for Programming, Artificial Intelligence and Reasoning, volume 73 of EPiC Series in Computing, pages 138–150. EasyChair, 2020.
- [7] C. Browne, E. J. Powley, D. Whitehouse, S. M. Lucas, P. I. Cowling, P. Rohlfshagen, S. Tavener, D. P. Liebana, S. Samothrakis, and S. Colton. A survey of monte carlo tree search methods. IEEE Transactions on Computational Intelligence and AI in Games, 4:1–43, 2012.
- [8] T. Chen and C. Guestrin. XGBoost: A scalable tree boosting system. In Proceedings of the 22Nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’16, pages 785–794, New York, NY, USA, 2016. ACM.
- [9] K. Chvalovský, J. Jakubuv, M. Suda, and J. Urban. ENIGMA-NG: efficient neural and gradient-boosted inference guidance for E. In P. Fontaine, editor, Automated Deduction - CADE 27 - 27th International Conference on Automated Deduction, Natal, Brazil, August 27-30, 2019, Proceedings, volume 11716 of Lecture Notes in Computer Science, pages 197–215. Springer, 2019.
- [10] The Coq Proof Assistant. http://coq.inria.fr.
- [11] M. Crouse, I. Abdelaziz, B. Makni, S. Whitehead, C. Cornelio, P. Kapanipathi, K. Srinivas, V. a Thost, M. Witbrock, and A. Fokoue. A deep reinforcement learning approach to first-order logic theorem proving. arXiv: Artificial Intelligence, 2019.
- [12] M. Färber, C. Kaliszyk, and J. Urban. Machine learning guidance for connection tableaux. J. Autom. Reason., 65(2):287–320, 2021.
- [13] T. Gauthier, C. Kaliszyk, and J. Urban. TacticToe: Learning to reason with HOL4 tactics. In T. Eiter and D. Sands, editors, LPAR-21. 21st International Conference on Logic for Programming, Artificial Intelligence and Reasoning, volume 46 of EPiC Series in Computing, pages 125–143. EasyChair, 2017.
- [14] T. Gauthier, C. Kaliszyk, J. Urban, R. Kumar, and M. Norrish. TacticToe: Learning to prove with tactics. J. Autom. Reason., 65(2):257–286, 2021.
- [15] M. J. C. Gordon and T. F. Melham, editors. Introduction to HOL: A theorem proving environment for higher order logic. Cambridge University Press, 1993.
- [16] J. Harrison. HOL Light: A tutorial introduction. In M. K. Srivas and A. J. Camilleri, editors, FMCAD, volume 1166 of LNCS, pages 265–269. Springer, 1996.
- [17] D. Huang, P. Dhariwal, D. Song, and I. Sutskever. Gamepad: A learning environment for theorem proving. In 7th International Conference on Learning Representations, ICLR 2019. OpenReview.net, 2019.
- [18] J. Jakubuv, K. Chvalovský, M. Olsák, B. Piotrowski, M. Suda, and J. Urban. ENIGMA anonymous: Symbol-independent inference guiding machine (system description). In N. Peltier and V. Sofronie-Stokkermans, editors, Automated Reasoning - 10th International Joint Conference, IJCAR 2020, volume 12167 of LNCS, pages 448–463. Springer, 2020.
- [19] J. Jakubuv and J. Urban. ENIGMA: efficient learning-based inference guiding machine. In H. Geuvers, M. England, O. Hasan, F. Rabe, and O. Teschke, editors, Intelligent Computer Mathematics - 10th International Conference, CICM 2017, Edinburgh, UK, July 17-21, 2017, Proceedings, volume 10383 of Lecture Notes in Computer Science, pages 292–302. Springer, 2017.
- [20] J. Jakubuv and J. Urban. Hammering Mizar by learning clause guidance. In J. Harrison, J. O’Leary, and A. Tolmach, editors, 10th International Conference on Interactive Theorem Proving, ITP 2019, September 9-12, 2019, Portland, OR, USA, volume 141 of LIPIcs, pages 34:1–34:8. Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 2019.
- [21] C. Kaliszyk and J. Urban. M2K dataset. https://github.com/JUrban/deepmath/blob/master/M2k_list.
- [22] C. Kaliszyk and J. Urban. Mizar40 dataset. https://github.com/JUrban/deepmath.
- [23] C. Kaliszyk and J. Urban. FEMaLeCoP: Fairly efficient machine learning connection prover. In M. Davis, A. Fehnker, A. McIver, and A. Voronkov, editors, Logic for Programming, Artificial Intelligence, and Reasoning - 20th International Conference, LPAR-20 2015, Suva, Fiji, November 24-28, 2015, Proceedings, volume 9450 of Lecture Notes in Computer Science, pages 88–96. Springer, 2015.
- [24] C. Kaliszyk and J. Urban. MizAR 40 for Mizar 40. J. Autom. Reasoning, 55(3):245–256, 2015.
- [25] C. Kaliszyk, J. Urban, H. Michalewski, and M. Olsák. Reinforcement learning of theorem proving. In NeurIPS, pages 8836–8847, 2018.
- [26] C. Kaliszyk, J. Urban, and J. Vyskočil. Efficient semantic features for automated reasoning over large theories. In Q. Yang and M. Wooldridge, editors, IJCAI’15, pages 3084–3090. AAAI Press, 2015.
- [27] L. Kocsis and C. Szepesvári. Bandit based monte-carlo planning. In J. Fürnkranz, T. Scheffer, and M. Spiliopoulou, editors, Machine Learning: ECML 2006, pages 282–293, Berlin, Heidelberg, 2006. Springer Berlin Heidelberg.
- [28] L. Kovács and A. Voronkov. First-order theorem proving and Vampire. In N. Sharygina and H. Veith, editors, CAV, volume 8044 of LNCS, pages 1–35. Springer, 2013.
- [29] S. M. Loos, G. Irving, C. Szegedy, and C. Kaliszyk. Deep network guided proof search. In 21st International Conference on Logic for Programming, Artificial Intelligence, and Reasoning (LPAR), 2017.
- [30] O. A. Mohamed, C. A. Muñoz, and S. Tahar, editors. Theorem Proving in Higher Order Logics, 21st International Conference, TPHOLs 2008, Montreal, Canada, August 18-21, 2008. Proceedings, volume 5170 of LNCS. Springer, 2008.
- [31] Y. Nagashima and Y. He. PaMpeR: proof method recommendation system for Isabelle/HOL. In M. Huchard, C. Kästner, and G. Fraser, editors, Proceedings of the 33rd ACM/IEEE International Conference on Automated Software Engineering, ASE 2018, Montpellier, France, September 3-7, 2018, pages 362–372. ACM, 2018.
- [32] M. Olsák, C. Kaliszyk, and J. Urban. Property invariant embedding for automated reasoning. In G. D. Giacomo, A. Catalá, B. Dilkina, M. Milano, S. Barro, A. Bugarín, and J. Lang, editors, ECAI 2020 - 24th European Conference on Artificial Intelligence, 29 August-8 September 2020, Santiago de Compostela, Spain, August 29 - September 8, 2020 - Including 10th Conference on Prestigious Applications of Artificial Intelligence (PAIS 2020), volume 325 of Frontiers in Artificial Intelligence and Applications, pages 1395–1402. IOS Press, 2020.
- [33] J. Otten. leanCoP 2.0 and ileanCoP 1.2: High performance lean theorem proving in classical and intuitionistic logic (system descriptions). In A. Armando, P. Baumgartner, and G. Dowek, editors, Automated Reasoning, 4th International Joint Conference, IJCAR 2008, Sydney, Australia, August 12-15, 2008, Proceedings, volume 5195 of Lecture Notes in Computer Science, pages 283–291. Springer, 2008.
- [34] J. Otten and W. Bibel. leanCoP: lean connection-based theorem proving. J. Symb. Comput., 36:139–161, 2003.
- [35] M. Rawson and G. Reger. Automated theorem proving, fast and slow. EasyChair Preprint no. 4433, EasyChair, 2020.
- [36] S. Ross, G. Gordon, and D. Bagnell. A reduction of imitation learning and structured prediction to no-regret online learning. In G. Gordon, D. Dunson, and M. Dudík, editors, Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, volume 15 of Proceedings of Machine Learning Research, pages 627–635, Fort Lauderdale, FL, USA, 11–13 Apr 2011. PMLR.
- [37] F. Scarselli, M. Gori, A. C. Tsoi, M. Hagenbuchner, and G. Monfardini. The graph neural network model. Trans. Neur. Netw., 20(1):61–80, Jan. 2009.
- [38] S. Schulz. E - A Brainiac Theorem Prover. AI Commun., 15(2-3):111–126, 2002.
- [39] S. Schulz. System description: E 1.8. In K. L. McMillan, A. Middeldorp, and A. Voronkov, editors, LPAR, volume 8312 of LNCS, pages 735–743. Springer, 2013.
- [40] D. Silver, J. Schrittwieser, K. Simonyan, I. Antonoglou, A. Huang, A. Guez, T. Hubert, L. Baker, M. Lai, A. Bolton, Y. Chen, T. Lillicrap, F. Hui, L. Sifre, G. van den Driessche, T. Graepel, and D. Hassabis. Mastering the game of go without human knowledge. Nature, 550:354–, Oct. 2017.
- [41] K. Slind and M. Norrish. A brief overview of HOL4. In Mohamed et al. [30], pages 28–32.
- [42] M. Suda. New techniques that improve Enigma-style clause selection guidance. In International Conference on Automated Deduction, CADE 2021, 2021.
- [43] R. S. Sutton and A. G. Barto. Reinforcement Learning: An Introduction. The MIT Press, second edition, 2018.
- [44] J. Urban. ERC project AI4Reason final scientific report, 2021. http://grid01.ciirc.cvut.cz/~mptp/ai4reason/PR_CORE_SCIENTIFIC_4.pdf.
- [45] J. Urban, G. Sutcliffe, P. Pudlák, and J. Vyskočil. MaLARea SG1 - Machine Learner for Automated Reasoning with Semantic Guidance. In IJCAR, pages 441–456, 2008.
- [46] J. Urban, J. Vyskočil, and P. Štěpánek. MaLeCoP: Machine learning connection prover. In K. Brünnler and G. Metcalfe, editors, TABLEAUX, volume 6793 of LNCS, pages 263–277. Springer, 2011.
- [47] N. Vulkan. An economist’s perspective on probability matching. Journal of Economic Surveys, 14(1):101–118, 2000.
- [48] M. Wenzel, L. C. Paulson, and T. Nipkow. The Isabelle framework. In Mohamed et al. [30], pages 33–38.
- [49] R. J. Williams and J. Peng. Function optimization using connectionist reinforcement learning algorithms. Connection Science, 3(3):241–268, 1991.
- [50] K. Yang and J. Deng. Learning to prove theorems via interacting with proof assistants. In K. Chaudhuri and R. Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA, volume 97 of Proceedings of Machine Learning Research, pages 6984–6994. PMLR, 2019.
- [51] Z. Zombori, J. Urban, and C. E. Brown. Prolog technology reinforcement learning prover - (system description). In N. Peltier and V. Sofronie-Stokkermans, editors, Automated Reasoning - 10th International Joint Conference, IJCAR 2020, Paris, France, July 1-4, 2020, Proceedings, Part II, volume 12167 of Lecture Notes in Computer Science, pages 489–507. Springer, 2020.