This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Does Pre-training Induce Systematic Inference?
How Masked Language Models Acquire Commonsense Knowledge

Ian Porada Mila, McGill University Alessandro Sordoni Microsoft Research Montréal Jackie Chi Kit Cheung Mila, McGill University
Abstract

Transformer models pre-trained with a masked-language-modeling objective (e.g., BERT) encode commonsense knowledge as evidenced by behavioral probes; however, the extent to which this knowledge is acquired by systematic inference over the semantics of the pre-training corpora is an open question. To answer this question, we selectively inject verbalized knowledge into the minibatches of a BERT model during pre-training and evaluate how well the model generalizes to supported inferences. We find generalization does not improve over the course of pre-training, suggesting that commonsense knowledge is acquired from surface-level, co-occurrence patterns rather than induced, systematic reasoning.

1 Introduction

Pre-trained Transformers, such as BERT, encode knowledge about the world (Petroni et al., 2019; Zhou et al., 2020); e.g., BERT assigns relatively high probability to “fly” appearing in the context “robins can       .” In this work, we investigate whether such knowledge is acquired during pre-training through systematic inference over the semantics of the pre-training corpora; e.g., can models systematically infer “robins can fly” from the premises “birds can fly” and “robins are birds?”

Resolving how models acquire commonsense knowledge has important implications. If models learn to make systematic inferences through pre-training, then scaling up pre-training is a promising direction for commonsense knowledge acquisition. If, instead, models only ever generalize based on surface-level patterns, then the majority of commonsense knowledge, which is only supported implicitly, will never be acquired (Gordon and Van Durme, 2013; Forbes and Choi, 2017).

On the one hand, there is cursory evidence that pre-training might induce the ability to systematically reason about the world. When fine-tuned on supervised training sets, pre-trained models can classify valid inferences better than strong baselines (Clark et al., 2020; Talmor et al., 2020b); and, in zero-shot evaluations, pre-trained models perform relatively well on reasoning tasks that may require systematic reasoning, such as number comparison (Talmor et al., 2020a) and Winograd schemas Sakaguchi et al. (2021).

On the other hand, existing works have argued that pre-training does not generalize by systematic inference over semantics on the basis of theoretical or synthetic results Bender and Koller (2020); Merrill et al. (2021); Traylor et al. (2021). Referring to physical commonsense knowledge acquired by BERT, Forbes et al. (2019) conclude that “neural language representations still only learn associations that are explicitly written down.”

Our main contribution is a direct evaluation of the training dynamics of BERT’s reasoning ability. We inject verbalized knowledge, such as “robins can fly” (where the masked token is the predicate, e.g., “fly”), into the minibatches of BERT throughout pre-training. We then consider how well BERT generalizes to supported inferences; e.g., how does the likelihood of “robins are      \rightarrow{} “birds” change?

We find generalization does not improve over the majority of pre-training which supports the hypothesis that commonsense knowledge is not acquired by systematic inference. Rather, our findings suggest knowledge is acquired from surface-level, co-occurrence patterns.

2 Related Work

Commonsense knowledge acquisition is a longstanding challenge in natural language processing (Charniak, 1973; Hwang et al., 2021; Zhang et al., 2021), and current approaches rely on knowledge acquired by pre-training Transformer language models (Bosselut et al., 2019; Zhang et al., 2020; West et al., 2021). The commonsense reasoning ability of pre-trained models has been evaluated using behavioral probes (Ettinger, 2020; Misra et al., 2021; He et al., 2021) and downstream, fine-tuned evaluations (Banerjee et al., 2021; Zhou et al., 2021). While these works only consider the ability of a fully pre-trained model, we evaluate how knowledge acquisition develops throughout pre-training.

When fine-tuned on supervised datasets, pre-trained models can learn to make systematic inferences to some extent (Clark et al., 2020; Tafjord et al., 2021; Gontier et al., 2020; Shaw et al., 2021; Li et al., 2021). Here, by systematic inferences, we refer to the ability to learn general rules and apply them in novel settings (Fodor and Pylyshyn, 1988; Lake and Baroni, 2018; Bahdanau et al., 2019) as opposed to learning only some particular instances of the rule.

As in our experiments, recent work has also considered the training dynamics of pre-trained models (Brown et al., 2020; Kaplan et al., 2020). Liu et al. (2021) specifically consider zero-shot performance of RoBERTa on the oLMpics reasoning tasks (Talmor et al., 2020a), but find the knowledge studied is never learned. In contrast, we explore how learned knowledge is acquired.

Close in spirit to our work, Kassner et al. (2020) pre-train a masked language model on a synthetic dataset to isolate reasoning ability. Wei et al. (2021) also intervene on BERT’s pre-training data in a syntactic evaluation and conclude that subject-verb agreement is inferred from rules to some extent.

Finally, De Cao et al. (2021) explore how knowledge encoded in BERT is affected by gradient updates when fine-tuning on a downstream task. Hase et al. (2021) build on this work and explore how updates on premises affect supported knowledge. Our work is unique, however, in that we focus on pre-training itself which we contrast with fine-tuning evaluations (§5.1).

3 Method

The purpose of our evaluation is to answer the question: does BERT systematically infer commonsense knowledge from premises present in the pre-training corpus?

We focus on commonsense knowledge that BERT is known to encode, namely simple entity properties such as those annotated in ConceptNet (Speer et al., 2017). This knowledge can be represented abstractly as (subject, predicate, object) triples. We verify BERT’s encoding of knowledge by the ability to predict the object conditioned on a verbalization of the knowledge containing the subject and predicate; e.g., for (robin, capable-of, fly), we evaluate the ability to predict “fly” appearing in the context “robins can       .”

Type Example
Super-statement A bird can       . \rightarrow{} fly
Sub-statement A robin can       . \rightarrow{} fly
Class Relation A robin is a       . \rightarrow{} bird
Table 1: Illustrative example of the three knowledge types as masked-token prediction.

This knowledge may be supported by simple co-occurrence patterns (such as “robin” and “fly” having high co-occurrence), but we are interested in the extent to which knowledge might also be supported by induced, systematic inference. We focus on the inference of downward monotonicity (A is-a B \land B has-property C \vDash A has-property C). We refer to the hypernym property (B has-property C) as the super-statement, the hyponym property (A has-property C) as the sub-statement, and the hypernymy relation (A is-a B) simply as the class-relation (Table 1).

We can then evaluate, for example, whether “robins can fly” is influenced by the inference “robins are birds” \land “birds can fly” \vDash “robins can fly.” For this evaluation, we inject a supporting premise into a pre-training minibatch (i.e., we replace one of the sentences in the minibatch with the premise) and then evaluate knowledge of the supported inference after updating on the premise.

We run this evaluation across time, evaluating BERT at several checkpoints throughout pre-training. If pre-training induces the ability to systematically make the downward monotonicity inference, one would expect that generalization from premise to inference will improve during pre-training.

3.1 Metrics

Let θi\theta_{i} be the parameterization of BERT at pre-training iteration ii, and let be w=(x,y,z)w=(x,y,z) be a knowledge instance where xx is the corresponding super-statement, yy the sub-statement, and zz the class-relation.

We take xx to be a logical premise. Let θix\theta_{i}^{x} be θi\theta_{i} after one gradient update on a minibatch containing xx. For a hypothesis hh (which is a possible inference and could be xx, yy, or zz), we consider:

  1. (1)

    Prior log-probability: logp(h|θi)\log p(h|\theta_{i})

  2. (2)

    Posterior log-probability: logp(h|θix)\log p(h|\theta_{i}^{x})

  3. (3)

    PMI: logp(h|θix)logp(h|θi)\log p(h|\theta_{i}^{x})-\log p(h|\theta_{i})

Intuitively, (1) describes the model’s prior knowledge of hh at step ii, and (3) describes how a pre-training update on xx affects the knowledge of hh. We also consider standard informational retrieval metrics such as mean reciprocal rank (MRR).

4 Experiments

4.1 Inference Dataset

We evaluate on the Leap-of-Thought dataset presented by Talmor et al. (2020b). This is a dataset of 30k true or false downward-monotonic inferences. The hypernymy relations are taken from WordNet (Miller, 1995), and the properties are taken from both WordNet and ConceptNet (Speer et al., 2017).

We reformulate this supervised, classification dataset as a zero-shot, cloze-style task. First, we filter the dataset by removing examples where one type of knowledge is withheld. Then, we filter out the randomly-generated negative examples (e.g. “a car cannot fly”), and those where the predicate is longer than one word-piece. (This last step follows the procedure of the LAMA evaluation (Petroni et al., 2019) and allows us to evaluate BERT in a zero-shot setting.) The filtered dataset consists of 711 examples. For each example, the knowledge is converted into a cloze task by masking the predicate.

To evaluate relative performance, we also generate a control for each example by randomly sampling a WordNet sibling of the super-statement hypernym as a pseudo-negative; e.g., controls are of the form “A robin is a       .” \rightarrow{} “fish.” We use a property of the control entity as a control predicate for the super and sub-statements.

4.2 Model

We evaluate the training dynamics of a BERT-base model (Devlin et al., 2019) with whole-word masking and sentence-order prediction (Lan et al., 2020). We pre-train the model for 1 million steps on a concatenation of English Wikipedia and the Toronto Book Corpus (Zhu et al., 2015) as released by Huggingface datasets (Lhoest et al., 2021). The training corpora are sentence tokenized using NLTK Punkt tokenizer (Bird and Loper, 2004), and these sentences used as training sequences instead of random spans of text as used in the original BERT.

Refer to caption
Figure 1: Accuracy on Talmor et al. (2020b)’s original Leap-of-Thought evaluation across pre-training iterations (from 50k to 1M).

Closely following the original hyperparameters of BERT, we use a batch size of 256, sequence length of 128, and train for 1 million steps. We linearly warmup the learning rate to 1e-4 over the first 10,000 steps of pre-training, and then linearly decay the learning rate. Additional hyperparameters can be found in Appendix A. Our code builds on the Huggingface Transformers (Wolf et al., 2020) and MegatronLM (Shoeybi et al., 2019) implementations of BERT. For evaluating training dynamics, we save a checkpoint of the model and optimizer throughout pre-training.

At each pre-training checkpoint, we perform the pre-training intervention. Specifically, we inject 20 random super-statements into a minibatch and perform one gradient update on this minibatch using the saved optimizer and a constant learning rate of 1e-4 (to control for the effects of the learning rate scheduler). We then evaluate change in likelihood of the inferences supported by the injected super-statements. We continue this evaluation 200 times so that each of the 711 Leap-of-Thought examples has been evaluated in five separate minibatches. We evaluate 20 pre-training checkpoints with this procedure.

Refer to caption
(a) Prior Log-Prob.
Refer to caption
(b) Super-statement->Sub-statement
Refer to caption
(c) Super-statement->Sub-statement
Refer to caption
(d) Super-statement->Super-statement
Refer to caption
(e) Super-statement->Class-relation
Figure 2: Evaluations of BERT’s knowledge and generalization across pre-training iterations. (a) shows the prior log-probability of sub-statements over time. (c-e) are the PMI for injecting supporting super-statements into the minibatch and evaluating on the labelled knowledge type. (b) is the change in MRR specifically for evaluating on sub-statements.

5 Results

5.1 Fine-tuning

We first run Talmor et al. (2020b)’s original fine-tuning evaluation on our pre-training checkpoints in order to validate our trained model and contrast fine-tuned evaluations with our pre-training interventions.

The explicit reasoning test requires classifying a sub-statement as true or false given the supporting super-statement and class-relation. The implicit reasoning test requires classifying a sub-statement given only the super-statement (and thus requires reasoning over implicit knowledge of the relation). Both test sets consist of disjoint entities from the training set. We fine-tune for 4 epochs following Talmor et al. and otherwise use default hyperparameters. The final implicit reasoning accuracy of our model is 0.89, slightly higher than Talmor et al. (2020b) report for RoBERTa-large.

We find performance increases log-linearly with pre-training iterations in the implicit reasoning test, but interestingly performance of the explicit reasoning evaluation peaks at just 15% of pre-training (Figure 1). Numerical results are presented in Appendix B.

5.2 Pre-training Interventions

Figure 2a shows the prior log-probability for our BERT model predicting sub-statement predicates during pre-training. The fact that the difference between the correct and control predicates increases during pre-training suggests knowledge of the sub-statement is acquired by BERT. Interestingly, probability of the correct predicate initially peaks at just 50k pre-training steps.

In Figures 2b, 2c, 2d and 2e, we visualize the results of our pre-training interventions. At each pre-training checkpoint, we update BERT on a minibatch with injected super-statements and then evaluate on predicting the predicate of the labelled knowledge type. For Figures 2c, 2d and 2e, we consider PMI; i.e., how does updating on a super-statement affect the likelihood of supported knowledge?

When BERT is updated on a pre-training minibatch containing a super-statement, this unsurprisingly increases the probability of the super-statements predicate (Figure 2d).

However, the PMI of the correct sub-statement predicate is the same as for the control predicate during the final iterations of pre-training (Figure 2c). Even more pronounced, the PMI of the class-relation control predicate is higher than the correct predicate during the entire second half of pre-training (Figure 2e).

In other words, updating on statements such as “birds can      \rightarrow{} “fly” increases the BERT’s estimate that “robins are      \rightarrow{} “fish” more than “birds.” We find a similar pattern in the reverse case: updating on class-relations improves BERT’s knowledge of super-statements and sub-statements less than the control baselines.

If knowledge is supported by the appropriate inference rules, we would expect that updating on a premise would improve the knowledge of inferences. Furthermore, if this reasoning was being induced, we would expect this generalization from premise to inference to improve over time. And yet, we find the opposite to be true. It follows that the studied inferences are not inferred by BERT using the rule of downward monotonicity.

For Figure 2b, we consider the change in MRR of the sub-event predicate after updating on a minibatch containing the super-statement. In this case, the difference between predicting the correct and control predicate seems indiscernible across pre-training checkpoints.

6 Conclusion

We show that BERT does not acquire commonsense knowledge from premises and learned inferences. This highlights limitations of scaled pre-training and suggests developments in commonsense knowledge acquisition may require explicit reasoning mechanisms.

References

  • Bahdanau et al. (2019) Dzmitry Bahdanau, Shikhar Murty, Michael Noukhovitch, Thien Huu Nguyen, Harm de Vries, and Aaron Courville. 2019. Systematic generalization: What is required and can it be learned? In International Conference on Learning Representations.
  • Banerjee et al. (2021) Pratyay Banerjee, Swaroop Mishra, Kuntal Kumar Pal, Arindam Mitra, and Chitta Baral. 2021. Commonsense reasoning with implicit knowledge in natural language. In 3rd Conference on Automated Knowledge Base Construction.
  • Bender and Koller (2020) Emily M. Bender and Alexander Koller. 2020. Climbing towards NLU: On meaning, form, and understanding in the age of data. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 5185–5198, Online. Association for Computational Linguistics.
  • Bird and Loper (2004) Steven Bird and Edward Loper. 2004. NLTK: The natural language toolkit. In Proceedings of the ACL Interactive Poster and Demonstration Sessions, pages 214–217, Barcelona, Spain. Association for Computational Linguistics.
  • Bosselut et al. (2019) Antoine Bosselut, Hannah Rashkin, Maarten Sap, Chaitanya Malaviya, Asli Celikyilmaz, and Yejin Choi. 2019. COMET: Commonsense transformers for automatic knowledge graph construction. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 4762–4779, Florence, Italy. Association for Computational Linguistics.
  • Brown et al. (2020) Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel Ziegler, Jeffrey Wu, Clemens Winter, Chris Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. 2020. Language models are few-shot learners. In Advances in Neural Information Processing Systems, volume 33, pages 1877–1901. Curran Associates, Inc.
  • Charniak (1973) Eurene Charniak. 1973. Jack and janet in search of a theory of knowledge. In Proceedings of the 3rd International Joint Conference on Artificial Intelligence, IJCAI’73, page 337–343, San Francisco, CA, USA. Morgan Kaufmann Publishers Inc.
  • Clark et al. (2020) Peter Clark, Oyvind Tafjord, and Kyle Richardson. 2020. Transformers as soft reasoners over language. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI-20, pages 3882–3890. International Joint Conferences on Artificial Intelligence Organization. Main track.
  • De Cao et al. (2021) Nicola De Cao, Wilker Aziz, and Ivan Titov. 2021. Editing factual knowledge in language models. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 6491–6506, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
  • Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics.
  • Ettinger (2020) Allyson Ettinger. 2020. What BERT Is Not: Lessons from a New Suite of Psycholinguistic Diagnostics for Language Models. Transactions of the Association for Computational Linguistics, 8:34–48.
  • Fodor and Pylyshyn (1988) Jerry A. Fodor and Zenon W. Pylyshyn. 1988. Connectionism and cognitive architecture: A critical analysis. Cognition, 28(1):3–71.
  • Forbes and Choi (2017) Maxwell Forbes and Yejin Choi. 2017. Verb physics: Relative physical knowledge of actions and objects. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 266–276, Vancouver, Canada. Association for Computational Linguistics.
  • Forbes et al. (2019) Maxwell Forbes, Ari Holtzman, and Yejin Choi. 2019. Do neural language representations learn physical commonsense? Proceedings of the 41st Annual Conference of the Cognitive Science Society.
  • Gontier et al. (2020) Nicolas Gontier, Koustuv Sinha, Siva Reddy, and Chris Pal. 2020. Measuring systematic generalization in neural proof generation with transformers. In Advances in Neural Information Processing Systems, volume 33, pages 22231–22242. Curran Associates, Inc.
  • Gordon and Van Durme (2013) Jonathan Gordon and Benjamin Van Durme. 2013. Reporting bias and knowledge acquisition. In Proceedings of the 2013 Workshop on Automated Knowledge Base Construction, AKBC ’13, page 25–30, New York, NY, USA. Association for Computing Machinery.
  • Hase et al. (2021) Peter Hase, Mona Diab, Asli Celikyilmaz, Xian Li, Zornitsa Kozareva, Veselin Stoyanov, Mohit Bansal, and Srinivasan Iyer. 2021. Do language models have beliefs? methods for detecting, updating, and visualizing model beliefs.
  • He et al. (2021) Weinan He, Canming Huang, Yongmei Liu, and Xiaodan Zhu. 2021. WinoLogic: A zero-shot logic-based diagnostic dataset for Winograd Schema Challenge. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 3779–3789, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
  • Hwang et al. (2021) Jena D. Hwang, Chandra Bhagavatula, Ronan Le Bras, Jeff Da, Keisuke Sakaguchi, Antoine Bosselut, and Yejin Choi. 2021. (comet-) atomic 2020: On symbolic and neural commonsense knowledge graphs. Proceedings of the AAAI Conference on Artificial Intelligence, 35(7):6384–6392.
  • Kaplan et al. (2020) Jared Kaplan, Sam McCandlish, T. J. Henighan, Tom B. Brown, Benjamin Chess, Rewon Child, Scott Gray, Alec Radford, Jeff Wu, and Dario Amodei. 2020. Scaling laws for neural language models. ArXiv, abs/2001.08361.
  • Kassner et al. (2020) Nora Kassner, Benno Krojer, and Hinrich Schütze. 2020. Are pretrained language models symbolic reasoners over knowledge? In Proceedings of the 24th Conference on Computational Natural Language Learning, pages 552–564, Online. Association for Computational Linguistics.
  • Lake and Baroni (2018) Brenden M. Lake and Marco Baroni. 2018. Generalization without systematicity: On the compositional skills of sequence-to-sequence recurrent networks. In ICML.
  • Lan et al. (2020) Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. 2020. Albert: A lite bert for self-supervised learning of language representations. In International Conference on Learning Representations.
  • Lhoest et al. (2021) Quentin Lhoest, Albert Villanova del Moral, Yacine Jernite, Abhishek Thakur, Patrick von Platen, Suraj Patil, Julien Chaumond, Mariama Drame, Julien Plu, Lewis Tunstall, Joe Davison, Mario Šaško, Gunjan Chhablani, Bhavitvya Malik, Simon Brandeis, Teven Le Scao, Victor Sanh, Canwen Xu, Nicolas Patry, Angelina McMillan-Major, Philipp Schmid, Sylvain Gugger, Clément Delangue, Théo Matussière, Lysandre Debut, Stas Bekman, Pierric Cistac, Thibault Goehringer, Victor Mustar, François Lagunas, Alexander Rush, and Thomas Wolf. 2021. Datasets: A community library for natural language processing. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 175–184, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
  • Li et al. (2021) Belinda Z. Li, Maxwell Nye, and Jacob Andreas. 2021. Implicit representations of meaning in neural language models. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 1813–1827, Online. Association for Computational Linguistics.
  • Liu et al. (2021) Zeyu Liu, Yizhong Wang, Jungo Kasai, Hannaneh Hajishirzi, and Noah A. Smith. 2021. Probing across time: What does RoBERTa know and when? In Findings of the Association for Computational Linguistics: EMNLP 2021, pages 820–842, Punta Cana, Dominican Republic. Association for Computational Linguistics.
  • Merrill et al. (2021) William Merrill, Yoav Goldberg, Roy Schwartz, and Noah A. Smith. 2021. Provable Limitations of Acquiring Meaning from Ungrounded Form: What Will Future Language Models Understand? Transactions of the Association for Computational Linguistics, 9:1047–1060.
  • Miller (1995) George A. Miller. 1995. Wordnet: A lexical database for english. Commun. ACM, 38(11):39–41.
  • Misra et al. (2021) Kanishka Misra, Allyson Ettinger, and Julia Rayz. 2021. Do language models learn typicality judgments from text? In Proceedings of the 43rd Annual Conference of the Cognitive Science Society.
  • Petroni et al. (2019) Fabio Petroni, Tim Rocktäschel, Sebastian Riedel, Patrick Lewis, Anton Bakhtin, Yuxiang Wu, and Alexander Miller. 2019. Language models as knowledge bases? In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 2463–2473, Hong Kong, China. Association for Computational Linguistics.
  • Sakaguchi et al. (2021) Keisuke Sakaguchi, Ronan Le Bras, Chandra Bhagavatula, and Yejin Choi. 2021. Winogrande: An adversarial winograd schema challenge at scale. Commun. ACM, 64(9):99–106.
  • Shaw et al. (2021) Peter Shaw, Ming-Wei Chang, Panupong Pasupat, and Kristina Toutanova. 2021. Compositional generalization and natural language variation: Can a semantic parsing approach handle both? In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 922–938, Online. Association for Computational Linguistics.
  • Shoeybi et al. (2019) Mohammad Shoeybi, Mostofa Patwary, Raul Puri, Patrick LeGresley, Jared Casper, and Bryan Catanzaro. 2019. Megatron-lm: Training multi-billion parameter language models using model parallelism.
  • Speer et al. (2017) Robyn Speer, Joshua Chin, and Catherine Havasi. 2017. Conceptnet 5.5: An open multilingual graph of general knowledge. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, AAAI’17, page 4444–4451. AAAI Press.
  • Tafjord et al. (2021) Oyvind Tafjord, Bhavana Dalvi, and Peter Clark. 2021. ProofWriter: Generating implications, proofs, and abductive statements over natural language. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 3621–3634, Online. Association for Computational Linguistics.
  • Talmor et al. (2020a) Alon Talmor, Yanai Elazar, Yoav Goldberg, and Jonathan Berant. 2020a. oLMpics-on what language model pre-training captures. Transactions of the Association for Computational Linguistics, 8:743–758.
  • Talmor et al. (2020b) Alon Talmor, Oyvind Tafjord, Peter Clark, Yoav Goldberg, and Jonathan Berant. 2020b. Leap-of-thought: Teaching pre-trained models to systematically reason over implicit knowledge. In Advances in Neural Information Processing Systems, volume 33, pages 20227–20237. Curran Associates, Inc.
  • Traylor et al. (2021) Aaron Traylor, Roman Feiman, and Ellie Pavlick. 2021. AND does not mean OR: Using formal languages to study language models’ representations. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pages 158–167, Online. Association for Computational Linguistics.
  • Wei et al. (2021) Jason Wei, Dan Garrette, Tal Linzen, and Ellie Pavlick. 2021. Frequency effects on syntactic rule learning in transformers. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 932–948, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
  • West et al. (2021) Peter West, Chandra Bhagavatula, Jack Hessel, Jena D. Hwang, Liwei Jiang, Ronan Le Bras, Ximing Lu, Sean Welleck, and Yejin Choi. 2021. Symbolic knowledge distillation: from general language models to commonsense models.
  • Wolf et al. (2020) Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, Remi Louf, Morgan Funtowicz, Joe Davison, Sam Shleifer, Patrick von Platen, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven Le Scao, Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander Rush. 2020. Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 38–45, Online. Association for Computational Linguistics.
  • Zhang et al. (2020) Hongming Zhang, Daniel Khashabi, Yangqiu Song, and Dan Roth. 2020. Transomcs: From linguistic graphs to commonsense knowledge. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI-20, pages 4004–4010. International Joint Conferences on Artificial Intelligence Organization. Main track.
  • Zhang et al. (2021) Hongming Zhang, Xin Liu, Haojie Pan, Haowen Ke, Jiefu Ou, Tianqing Fang, and Yangqiu Song. 2021. Aser: Towards large-scale commonsense knowledge acquisition via higher-order selectional preference over eventualities. arXiv preprint arXiv:2104.02137.
  • Zhou et al. (2021) Pei Zhou, Rahul Khanna, Seyeon Lee, Bill Yuchen Lin, Daniel Ho, Jay Pujara, and Xiang Ren. 2021. RICA: Evaluating robust inference capabilities based on commonsense axioms. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 7560–7579, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
  • Zhou et al. (2020) Xuhui Zhou, Yue Zhang, Leyang Cui, and Dandan Huang. 2020. Evaluating commonsense in pre-trained language models. Proceedings of the AAAI Conference on Artificial Intelligence, 34(05):9733–9740.
  • Zhu et al. (2015) Yukun Zhu, Ryan Kiros, Rich Zemel, Ruslan Salakhutdinov, Raquel Urtasun, Antonio Torralba, and Sanja Fidler. 2015. Aligning books and movies: Towards story-like visual explanations by watching movies and reading books. In The IEEE International Conference on Computer Vision (ICCV).

Appendix A BERT Hyperparameters

We follow the BERT-base architecture (12 layers, 12 attention heads, hidden size of 768) and train with the Adam optimizer. We use a sequence length of 128 throughout pre-training, and MegatronLM pre-processing. Training takes four days on eight V100 GPUs.

Appendix B Leap-of-Thought Fine-tuning Results

Iteration Implicit Explicit
0 0.507 0.493
5000 0.507 0.493
10000 0.490 0.490
15000 0.571 0.621
20000 0.625 0.636
30000 0.710 0.763
40000 0.798 0.900
50000 0.814 0.965
100000 0.838 0.971
150000 0.860 0.992
200000 0.843 0.953
250000 0.855 0.973
300000 0.870 0.958
350000 0.863 0.978
400000 0.850 0.931
450000 0.867 0.937
500000 0.859 0.933
550000 0.874 0.951
600000 0.867 0.943
650000 0.880 0.931
700000 0.877 0.937
750000 0.874 0.929
800000 0.872 0.949
850000 0.877 0.979
900000 0.875 0.967
950000 0.894 0.945