This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

JointLK: Joint Reasoning with Language Models and Knowledge Graphs for Commonsense Question Answering

Yueqing Sun, Qi Shi, Le Qi, Yu Zhang
Research Center for Social Computing and Information Retrieval
Harbin Institute of Technology, Harbin, China
{yqsun,qshi,lqi,zhangyu}@ir.hit.edu.cn
  Corresponding author.
Abstract

Existing KG-augmented models for commonsense question answering primarily focus on designing elaborate Graph Neural Networks (GNNs) to model knowledge graphs (KGs). However, they ignore (i) the effectively fusing and reasoning over question context representations and the KG representations, and (ii) automatically selecting relevant nodes from the noisy KGs during reasoning. In this paper, we propose a novel model, JointLK, which solves the above limitations through the joint reasoning of LM and GNN and the dynamic KGs pruning mechanism. Specifically, JointLK performs joint reasoning between LM and GNN through a novel dense bidirectional attention module, in which each question token attends on KG nodes and each KG node attends on question tokens, and the two modal representations fuse and update mutually by multi-step interactions. Then, the dynamic pruning module uses the attention weights generated by joint reasoning to prune irrelevant KG nodes recursively. We evaluate JointLK on the CommonsenseQA and OpenBookQA datasets, and demonstrate its improvements to the existing LM and LM+KG models, as well as its capability to perform interpretable reasoning111Our code is available at: https://github.com/Yueqing-Sun/JointLK.

Refer to caption
Figure 1: Our knowledge-augmented joint reasoning model framework with an example from CommonsenseQA. The subgraph is retrieved from ConceptNet.

1 Introduction

Commonsense question answering (CSQA) requires systems to acquire different types of commonsense knowledge and reasoning skills, which is normal for humans, but challenging for machines Talmor et al. (2019). Recently, large pre-trained language models (LMs) have achieved remarkable success in many QA tasks and appear to use implicit (factual) knowledge encoded in their model parameters during fine-tuning Liu et al. (2019); Raffel et al. (2020). Nevertheless, commonsense knowledge is self-evident to humans and is rarely expressed clearly in natural language Gunning (2018), which makes it difficult for LMs to learn commonsense knowledge from the pre-training text corpus alone.

An extensive research path is to elaborately design graph neural networks (GNNs) Scarselli et al. (2008) to perform reasoning over explicit structural common sense knowledge from external knowledge bases Vrandečić and Krötzsch (2014); Speer et al. (2017). Related methods usually follow a retrieval-and-modeling paradigm. First, the knowledge subgraphs or paths related to a given question are retrieved by string matching or semantic similarity; such retrieved structured information indicates the relation between concepts or implies the process of multi-hop reasoning. Second, the retrieved subgraphs are modeled by a well-designed graph neural network module Lin et al. (2019); Feng et al. (2020); Yasunaga et al. (2021) to perform reasoning over knowledge graphs.

However, these approaches have two main issues. First, the retrieved knowledge subgraph contains many noisy nodes. Whether through simple string matching or semantic matching, in order to retrieve sufficient relevant knowledge, noise knowledge graph nodes will inevitably be included Lin et al. (2019); Yasunaga et al. (2021). Especially with the increase of hop count, the number of irrelevant nodes will expand dramatically, raising the burden of the model. As the example in Figure 1, some graph nodes such as “wood", “burn", and “gas", although related to some entities in the questions and choice, can mislead the global understanding of the question. Second, there are limited interactions between language representation and knowledge graph representation. Specifically, existing LM+KG methods Lin et al. (2019); Feng et al. (2020) model question context and knowledge subgraphs in isolation by LMs and GNNs, and perform only one interaction in a shallow manner to fuse their representations at the output for prediction. We argue that the limited interaction between the two modalities is the main bottleneck that may prevent the model from understanding the complex question-knowledge relations necessary to answer the question correctly.

Based on the above consideration, we propose JointLK, a model that performs the fine-grained modal fusion and multi-layer joint reasoning between the language model and the knowledge graph (see Figure 2). Specifically, given a question and retrieved subgraphs, JointLK first obtain the representations of the two modalities by using an LM encoder and a GNN encoder respectively. Then we design a joint reasoning module to generate fine-grained bidirectional attention maps between each question token and each KG node to fuse the information from each modality to the other. Guided by the attention generated in the interaction process, the dynamic pruning module deletes irrelevant nodes to make the model reason along the correct knowledge path. Multiple JointLK layers are stacked to form a hierarchy that supports multi-step interactions and recursive pruning. In summary, our contributions are three-fold:

  • We propose JointLK, a novel model that supports multi-step joint reasoning between LM and KG. It uses dense bidirectional attention to simultaneously update query-aware knowledge graph representation and knowledge-aware query representation, bridging the gap between the two information modalities.

  • We design a dynamic graph pruning module that recursively removes irrelevant graph nodes at each JointLK layer to ensure that the model reasons correctly with complete and appropriate evidence.

  • Experimental results show that JointLK is superior to current LM+KG methods, and the refined evidence is interpretable. Furthermore, through the multi-layer fusion of these two modalities, JointLK exhibits strong performance over previous state-of-the-art LM+KG methods in performing complex reasoning, such as solving questions with negation and complex questions with more entities.

Refer to caption
Figure 2: Overall architecture of our proposed JointLK model, which takes a query (question + choice) and a retrieved knowledge subgraph as input, and outputs a scalar that represents the plausibility score of this query. JointLK mainly consists of four modules the Query Encoder, the Graph Layer, the Joint Reasoning Module and the Dynamic Pruning Module, of which the latter three form a stack of N identical layers.

2 Related Work

Commonsense question answering is challenging because the required commonsense knowledge is rarely given in the context of questions and answer choices or encoded in the parameters of pre-trained LMs. Therefore, many works obtain the required knowledge from external sources (e.g., KGs, corpus) to augment CSQA models. Due to the heterogeneity between structured knowledge and unstructured text questions, there are currently two main research methods. Some works Lv et al. (2020); Bian et al. (2021); Xu et al. (2021) unify the two modalities during model input, such as transforming structured knowledge into plain text through templates or transforming question context into structured graphs. However, the original structural/textual information will inevitably be lost during the conversion process. Other works Lin et al. (2019); Feng et al. (2020); Yan et al. (2021) use LM and GNN to model the two modalities separately, and perform shallow interactions in the latter model stage, such as attentive pooling or simple concatenation of the two modal representations. Although this method can retain the original information of question context and KGs, the limited interaction will affect the flow of information between the two modalities, so we mainly improve on this point.

Recently, QA-GNN Yasunaga et al. (2021) explicitly views the QA context as an additional node, connects it and KG to form a joint graph, and mutually updates their representations through graph-based message passing. However, it pools the representation of the question context into a single node, which limits the updating of the text representation and fine-grained interaction between LM and GNN. Compared with prior works, we retain the individual structure of both modalities, consider fine-grained interaction between any token in question and any entity in KG through dense bidirectional attention, and perform multi-step joint reasoning by stacking several interaction layers. Furthermore, we gradually prune the KG size in each stacked model layer under the guidance of attention weights generated in the interactions, making the reasoning path transparent and interpretable.

3 Methodology

In this section, we introduce the task definition (§ 3.1) and our JointLK model. The model framework is shown in Figure 2. JointLK takes the query and the retrieved knowledge subgraph as input, and outputs a real value as the correctness score of the answer. The model is mainly composed of four parts: query encoder, GNN layer, joint reasoning module and dynamic pruning module, of which the latter three form a stack of N identical layers. We use a pre-trained language model to learn the query representation (§ 3.2), and use the GNN layer to learn the graph representation (§ 3.3). The Joint Reasoning Module receives these two modalities’ representations and then apply dense bidirectional attention to make information fusion and representation update for each token and node (§ 3.4). The LM-to-KG attention weights generated in reasoning represents the global importance of each node in the graph, so the dynamic pruning module prunes the graph layer by layer according to this weights and finally retains the most relevant nodes (§ 3.5). After N layers of iteration, the query representation and the trimmed graph representation are used to predict the answer (§ 3.6).

3.1 Task Definition

The CSQA task in this paper is a multiple-choice problem with some answer choices. Given a commonsense question qq and a set of answer choices {a1,a2,,an}\{a_{1},a_{2},...,a_{n}\}, our task is to measure the plausibility score between qq and each answer choice aa then select the answer with the highest plausibility score. In general, questions do not contain any reference to answer choices, so the external knowledge graph provides the necessary background knowledge. We extract from the external KG a subgraph g=(V,R)g=(V,R) with the guidance of question and choice. Here V is a subset of entity nodes retrieved from the external KG. EV×R×VE\subseteq{V\times R\times V} is the set of edges that connect nodes in VV, where RR is a set of relations types. We describe the detailed extraction process in Appendix A.

3.2 Query Encoder

We follow baselines to use pre-trained language models to encode the query {wi}i=1M\{w_{i}\}_{i=1}^{M} (question and choice) into a sequence of vectors {qi0}i=1M\{q_{i}^{0}\}_{i=1}^{M}:

{q~10,,q~M0}=𝐄𝐧𝐜𝐋𝐌({w1,,wM})\{\widetilde{q}_{1}^{0},...,\widetilde{q}_{M}^{0}\}=\mathbf{Enc_{LM}}\left(\{w_{1},...,w_{M}\}\right) (1)

Here {q~i0}i=1MT\{\widetilde{q}_{i}^{0}\}_{i=1}^{M}\in{\mathbb{R}^{T}} is the last hidden layer vector of each token in the query. Then we feed the representation of tokens into a non-linear layer so that the text representation space is aligned to the entity representation space:

qi0=σ(fs(q~i0))q_{i}^{0}=\sigma\left(f_{s}\left(\widetilde{q}_{i}^{0}\right)\right) (2)

where fs:TDf_{s}:\mathbb{R}^{T}\rightarrow\mathbb{R}^{D} is a linear transformation, and σ\sigma is the activation function. The representations of tokens 𝐐0={qi0}i=1MD\mathbf{Q}^{0}=\{{q}_{i}^{0}\}_{i=1}^{M}\in{\mathbb{R}^{D}} will be provided to the joint reasoning module for further interaction with the graph entities representations.

3.3 GNN Layer

After obtaining token representations by the query encoder, we further model the subgraph to obtain entity representations. First, We use the BERT model with average pooling to get the initial representation for each entity 𝐗𝟎={xi0}i=0|V|D\mathbf{X^{0}}=\{x_{i}^{0}\}_{i=0}^{|V|}\in{\mathbb{R}^{D}}. Then, we apply GNN Layer to update node representation through iterative message passing between neighbors on the graph, while GNN is built on the RGAT Wang et al. (2020a) and is a simplification of Yasunaga et al. (2021). For brevity, we formulate the entire computation in one layer as:

{x~1l,,x~|V|l}=𝐆𝐍𝐍({x1l1,,x|V|l1})\{\widetilde{x}_{1}^{l},...,\widetilde{x}_{|V|}^{l}\}=\mathbf{GNN}(\{x_{1}^{l-1},...,x_{|V|}^{l-1}\}) (3)

The output representation xilx_{i}^{l} is computed by

α^ji\displaystyle\hat{\alpha}_{ji} =(xil1Wq)(xjl1Wk+rji)T,\displaystyle=(x_{i}^{l-1}W_{q})(x_{j}^{l-1}W_{k}+r_{ji})^{T}, (4)
αji\displaystyle{\alpha}_{ji} =𝐬𝐨𝐟𝐭𝐦𝐚𝐱(α^ji/D),\displaystyle=\mathbf{softmax}(\hat{\alpha}_{ji}/{\sqrt{D}}), (5)
x^il1\displaystyle\hat{x}_{i}^{l-1} =jNi{i}αji(xjl1Wv+rji),\displaystyle=\sum_{j\in N_{i}\cup\{i\}}{{\alpha}_{ji}(x_{j}^{l-1}W_{v}+r_{ji})}, (6)
x~il\displaystyle\widetilde{x}_{i}^{l} =𝐋𝐚𝐲𝐞𝐫𝐍𝐨𝐫𝐦(xil1+x^il1Wo)\displaystyle=\mathbf{LayerNorm}(x_{i}^{l-1}+\hat{x}_{i}^{l-1}W_{o}) (7)

where matrices Wq,Wk,Wv,WoD×DW_{q},W_{k},W_{v},W_{o}\in\mathbb{R}^{D\times D} are trainable parameters, NiN_{i} is the neighbor of node ii. rji=ψ(eji,uj,ui)r_{ji}=\psi(e_{ji},u_{j},u_{i}) is the relation feature vector, where ejie_{ji} is a one-hot vector denoting the relation type of the edge (j,i)(j,i) and uj,uiu_{j},u_{i} are one-hot vectors denoting the node types of jj and ii. The following joint reasoning module will further fuse x~il\widetilde{x}_{i}^{l} and qil1q_{i}^{l-1} to obtain their updated representations.

3.4 Joint Reasoning Module

To reduce the gap of query and knowledge graph features, we fuse them in the joint reasoning module by the dense bidirectional attention mechanism that connects two encoding layers of query and knowledge graph and captures the fine-grained interplay between them.

The module takes the query and KG representations 𝐐\mathbf{Q} and 𝐗\mathbf{X} as inputs and then outputs their updated versions. We denote the inputs to the joint reasoning module in the l-st fusion layer by 𝐐l1={qil1}i=1M\mathbf{Q}^{l-1}=\{q_{i}^{l-1}\}_{i=1}^{M} and 𝐗~l={x~il}i=1|V|\mathbf{\widetilde{X}}^{l}=\{\widetilde{x}_{i}^{l}\}_{i=1}^{|V|}. Given qil1q_{i}^{l-1} and x~il\widetilde{x}_{i}^{l}, an affinity matrix is first constructed via:

Sijl=WST[qil1;x~jl;qil1x~jl]S_{ij}^{l}=W_{S}^{T}[q_{i}^{l-1};\widetilde{x}_{j}^{l};q_{i}^{l-1}\circ\widetilde{x}_{j}^{l}] (8)

where WSTW_{S}^{T} is a learnable weight matrix, \circ is elementwise multiplication, [;] is vector concatenation across row. We normalize SijlS_{ij}^{l} in row-wise to derive KG-to-LM attention maps on query tokens conditioned by each entity in KG as

Sqil=𝐬𝐨𝐟𝐭𝐦𝐚𝐱(Sij)S_{q_{i}}^{l}=\mathbf{softmax}\left(S_{ij}\right) (9)

and also normalize SijlS_{ij}^{l} in column-wise to derive LM-to-KG attention maps on entities conditioned by each query token as

Sxjl=𝐬𝐨𝐟𝐭𝐦𝐚𝐱(SijT)S_{x_{j}}^{l}=\mathbf{softmax}\left(S_{ij}^{T}\right) (10)

The attended representations are computed as follows:

q^ij=qil1Sqil,x^ij=x~jlSxjl\hat{q}_{ij}=q_{i}^{l-1}\otimes S_{q_{i}}^{l},\\ \hat{x}_{ij}=\widetilde{x}_{j}^{l}\otimes S_{x_{j}}^{l} (11)

where \otimes represents matrix multiplication. The attended features are fused with the original features of the other modality by concatenation and then compressed to low-dimensional space by:

qil\displaystyle q_{i}^{l} =WQ[qil1;x^ij;qil1x^ij;qil1q^ij],\displaystyle=W_{Q}[q_{i}^{l-1};\hat{x}_{ij};q_{i}^{l-1}\circ\hat{x}_{ij};q_{i}^{l-1}\circ\hat{q}_{ij}], (12)
x¯jl\displaystyle\bar{x}_{j}^{l} =WX[x~jl;q^ij;x~jlq^ij;x~jlx^ij]\displaystyle=W_{X}[\widetilde{x}_{j}^{l};\hat{q}_{ij};\widetilde{x}_{j}^{l}\circ\hat{q}_{ij};\widetilde{x}_{j}^{l}\circ\hat{x}_{ij}] (13)

where WQ,WXW_{Q},W_{X} are learnable weights. Then the updated query representation 𝐐l={qil}i=1M\mathbf{Q}^{l}=\{q_{i}^{l}\}_{i=1}^{M} will be input to the next ll-th stacked JointLK layer of to continue participating in joint reasoning, and the updated KG representation 𝐗¯l={x¯il}i=1|V|\mathbf{\bar{X}}^{l}=\{\bar{x}_{i}^{l}\}_{i=1}^{|V|} will be input to the next module of the current JointLK layer for pruning.

3.5 Dynamic Pruning Module

In Equation 10, the LM-to-KG attention value implies the importance of different nodes in the subgraph for question answering. Inspired by SAGPool Lee et al. (2019), under the guidance of query, we retain relevant nodes and cut out irrelevant nodes according to the LM-to-KG attention. Then, We define a hyperparameter, the Retention ratio K(0,1]K\in(0,1], which determines the number of nodes to be retained. We choose the top K|V|\left\lceil K\cdot|V|\right\rceil nodes according to the value of LM-to-KG attention:

idx=𝐭𝐨𝐩𝐫𝐚𝐧𝐤(Z,K|V|),\displaystyle idx=\mathbf{top-rank}\left(Z,\left\lceil K\cdot|V|\right\rceil\right), (14)
Zmask=Zidx\displaystyle Z_{mask}=Z_{idx} (15)

where top-rank is a function that returns the index of top K|V|\left\lceil K\cdot|V|\right\rceil value, idx\cdot_{idx} is an indexing operation, and ZmaskZ_{mask} is corresponding attention mask. Next, the subgraph is formed by pooling out the less essential entity nodes as:

𝐗l=𝐗¯idx,:lZmask,𝐀l=𝐀¯idx,idxl\begin{split}\mathbf{X}^{l}&=\mathbf{\bar{X}}_{idx,:}^{l}\odot Z_{mask},\\ \mathbf{A}^{l}&=\mathbf{\bar{A}}_{idx,idx}^{l}\end{split} (16)

where 𝐗¯idx,:l\mathbf{\bar{X}}_{idx,:}^{l} is the row-wise indexed representation matrix of 𝐗¯l\mathbf{\bar{X}}^{l}, \odot is the broadcasted elementwise product, and 𝐀¯idx,idxl\mathbf{\bar{A}}_{idx,idx}^{l} is the row-wise and col-wise of indexed adjacency matrix. 𝐗l=(x1l,x2l,,xk|V|l)\mathbf{X}^{l}=(x_{1}^{l},x_{2}^{l},\ldots,x_{\left\lceil k|V|\right\rceil}^{l}), 𝐀l\mathbf{A}^{l} and K|V|\left\lceil K\cdot|V|\right\rceil are the representation matrix, the adjacency matrix and the number of graph nodes in the next JointLK layer.

3.6 Answer Prediction

After N layers of iteration, we finally obtain the query representation 𝐐N\mathbf{Q}^{N} that fuses knowledge information and the graph representation 𝐗N\mathbf{X}^{N} that fuses question information. We compute the score of aa being the correct answer as:

p=(a|q)=𝐌𝐋𝐏([s;g])p=\left(a|q\right)=\mathbf{MLP}\left([{s};{g}]\right) (17)

where ss is the mean pooling of 𝐐N\mathbf{Q}^{N}, and gg is the attention-based pooling of 𝐗N\mathbf{X}^{N}. We get the final probability by normalize all question-choice pairs with softmax.

4 Experimental Setup

4.1 Datasets

We evaluate our model on two typical commonsense question answering datasets CommonsenseQA Talmor et al. (2019) and OpenBookQA Mihaylov et al. (2018). CommonsenseQA is a 5-way multiple-choice question answering dataset that requires commonsense for reasoning and contains 12,102 questions. We experiment and report the accuracy on the in-house dev (IHdev) and test (IHtest) splits used by Lin et al. (2019), and report the accuracy of our final system on the official test set. OpenBookQA is a 4-way multiple choice question answering dataset that requires reasoning with elementary science knowledge. It contains 5,957 questions along with an open book of scientific facts. We use the official data split.

4.2 Implementation Details

Following previous work Yasunaga et al. (2021), we use ConceptNet Speer et al. (2017), a commonsense knowledge graph, as our structured knowledge source for both of the above tasks. Given each query, we follow the preprocessing steps described in Feng et al. (2020) to retrieve the subgraph from ConceptNet, and the max hop size is 3 (see Appendix A for the detail). We use cross-entropy loss and RAdam optimizer Liu et al. (2020). In training, we set the maximum input sequence length to text encoders to 100, batch size to 128, and perform early stopping. We set the dimension (D = 200) and number of layers (N = 5) of our GNN module, with dropout rate 0.2 applied to each layer Srivastava et al. (2014). We use separate learning rates for the LM encoder and the graph encoder. We choose the LM encoder learning rate from{1×105, 2×105, 3×105}\{1\times{10}^{-5},\ 2\times{10}^{-5},\ 3\times{10}^{-5}\}, and choose the graph encoder learning rate from{1×103, 2×103}\{1\times{10}^{-3},\ 2\times{10}^{-3}\}. Each model is trained using one GPU (Tesla_v100-sxm2-16gb), which takes 20 hours on average.

4.3 Compared Method

Although text corpus can provide complementary knowledge except for knowledge graphs, our model focuses on improving the use of KG and the joint reasoning between LM and KG, so we choose LM and LM+KG as the comparison methods.

To investigate the role of KGs, we compare with the benchmark model RoBERTa-large Liu et al. (2019) for CommonsenseQA, and compare with RoBERTa-large and AristoRoBERTa Clark et al. (2020) for OpenBookQA. For LM+KG methods, they share a similar high-level framework with our methods, that is, LM is used as a text encoder, GNN or RN is used as a KG encoder, but the way of using knowledge or reasoning is different: (1) Relationship network (RN) Santoro et al. (2017), (2) RGCN Schlichtkrull et al. (2018), (3) GconAttn Wang et al. (2019), (4)KagNet Lin et al. (2019) and (5)MHGRN Feng et al. (2020), (6) QA-GNN Yasunaga et al. (2021). (1), (2) and (3) are the relational perception GNNs for KGs, and (4), (5) and (6) are further model paths in KGs. To be fair, we use the same LM for all comparison methods.

Methods IHdev-Acc.(%) IHtest-Acc.(%)
RoBERTa-large(w/o KG) 73.07 (±\pm0.45) 68.69 (±\pm0.56)
+ RGCN 72.69 (±\pm0.19) 68.41 (±\pm0.66)
+ GconAttn 71.61 (±\pm0.39) 68.59 (±\pm0.96)
+ KagNet 73.47 (±\pm0.22) 69.01 (±\pm0.76)
+ RN 74.57 (±\pm0.91) 69.08 (±\pm0.21)
+ MHGRN 74.45 (±\pm0.10) 71.11 (±\pm0.81)
+ QA-GNN 76.54 (±\pm0.21) 73.41 (±\pm0.92)
+ JointLK (Ours) 77.88 (±\pm0.25) 74.43 (±\pm0.83)
Table 1: Performance comparison on CommonsenseQA in-house split. We follow the data division method of Lin et al. (2019) and report the in-house Dev (IHdev) and Test (IHtest) accuracy(mean and standard deviation of four runs).
Methods Test
RoBERTa Liu et al. (2019) 72.1
Albert Lan et al. (2020) (ensemble) 76.5
RoBERTa + FreeLB Zhu et al. (2020) (ensemble) 73.1
RoBERTa + HyKAS Ma et al. (2019) 73.2
RoBERTa + KE (ensemble) 73.3
RoBERTa + KEDGN (ensemble) 74.4
XLNet + GraphReason Lv et al. (2020) 75.3
RoBERTa + MHGRN Feng et al. (2020) 75.4
Albert + PG Wang et al. (2020b) 75.6
RoBERTa + QA-GNN Yasunaga et al. (2021) 76.1
RoBERTa + JointLK (Ours) 76.6
Table 2: Performance comparison on the CommonsenseQA official leaderboard. Our model has achieved state-of-the-art under the setting of RoBERTa-large.
Methods RoBERTa-large AristoRoBERTa
Fine-tuned LMs (w/o KG) 64.80 (±\pm2.37) 78.40 (±\pm1.64)
+ RGCN 62.45 (±\pm1.57) 74.60 (±\pm2.53)
+ GconAttn 64.75 (±\pm1.48) 71.80 (±\pm1.21)
+ RN 65.20 (±\pm1.18) 75.35 (±\pm1.39)
+ MHGRN 66.85 (±\pm1.19) 80.6
+ QA-GNN 67.80 (±\pm2.75) 82.77 (±\pm1.56)
+ JointLK (Ours) 70.34 (±\pm0.75) 84.92 (±\pm1.07)
Table 3: Test accuracy on OpenBookQA. Methods with AristoRoBERTa use the textual evidence by Clark et al. (2020) as an additional input to the QA context.
Methods Test
Careful Selection Banerjee et al. (2019) 72.0
AristoRoBERTa 77.8
KF + SIR Banerjee and Baral (2020) 80.0
AristoRoBERTa + PG Wang et al. (2020b) 80.2
AristoRoBERTa + MHGRN Feng et al. (2020) 80.6
ALBERT + KB 81.0
AristoRoBERTa + QA-GNN Yasunaga et al. (2021) 82.8
T5* Raffel et al. (2020) 83.2
UnifiedQA(11B)* Khashabi et al. (2020) 87.2
AristoRoBERTa + JointLK (Ours) 85.6
Table 4: Test accuracy on OpenBookQA leaderboard. All listed methods use the provided science facts as an additional input to the language context. The previous top 2 systems, UnifiedQA (11B params) and T5 (3B params) are 30x and 8x larger than our model.
Methods IHdev-Acc. (%)
JointLK (NN=5) 77.88
- Dynamic Pruning Module 77.38
- Joint Reasoning Module 76.61
Table 5: Ablation study on model components using RoBERTa-large as the text encoder. We report the IHdev accuracy on CommonsenseQA.
Refer to caption
Refer to caption
Figure 3: Ablation study on stacked of JointLK layers (a) and the retention ratio in pruning (b).

5 Results and Analysis

5.1 Main Results

The results on CommonsenseQA in-house split dataset and official test dataset are shown in Table 1 and Table 2. The results on OpenBookQA test dataset and leaderboard are shown in Table 3 and Table 4. We can observe that JointLK performs best among all fine-tuned LMs and existing LM+KG models. On CommonsenseQA, our model’s test performance improves by 5.74% over fine-tuned LMs and 1.02% over the prior best LM+KG model, QA-GNN. On OpenbookQA, our model’s test performance improves by 6.52% over fine-tuned AristoRoBERTa, and 2.15% over QA-GNN. Additionally, we also submit our best model to the leaderboards, and our JointLK (with the text encoder being RoBERTa-large) ranks first among comparable approaches. Compared with the previous best model MHGRN and QA-GNN, the boost over them suggests the effectiveness of our proposed joint reasoning between LM and KG and the dynamic pruning mechanism.

In particular, we do not compare with the higher ranking models on the leaderboard, such as unified QA Khashabi et al. (2020), Albert + DESC-KCR Xu et al. (2021), because they either use a stronger text encoder or use additional data resources, while our model focuses on improving the joint reasoning between LM and KG.

Methods IHdev-Acc (Overall) IHdev-Acc (Questions w/ negation) IHdev-Acc (Questions w/ \leq7 entities) IHdev-Acc (Questions w/ >7 entities)
Number 1221 133 723 498
QA-GNN 76.99 72.18 76.63 77.51
JointLK(Ours) 78.38 75.18 (\uparrow3.00) 77.59 (\uparrow0.96) 79.52 (\uparrow2.01)
Table 6: Performance on questions with negative words and fewer/more entities. The questions are retrieved from the CommonsenseQA IHdev set.

5.2 Ablation Studies

We further conduct in-depth analyses to investigate the effectiveness of different components in our model. We show the accuracy of JointLK on the CommonsenseQA IHdev set.

Impact of JointLK components We assess the impact of the joint reasoning module (§ 3.4) and the dynamic pruning module (§ 3.5), shown in Table 5. Disabling the dynamic pruning module results in 0.5% drop in performance, showing that some nodes in subgraph are not conducive to reasoning. Especially, when we disable the joint reasoning module, the corresponding dynamic pruning module will also be removed, because the latter depends on the attention value in the former. Then the results have a significant drop: 77.88%76.61%77.88\%\rightarrow 76.61\%, suggesting that the joint reasoning between LM and KG is critical.

Impact of stacked of JointLK Layers We investigate the impact of the number of JointLK layers (shown in Figure 3 (a)). The increase of layers continues to bring benefits until layers N=5N=5. However, performance begins to drop when N>5N>5. As the number of layers increases, the model changes from underfitting to overfitting.

Impact of the Retention Ratio in Pruning The retention ratio KK is a hyperparameter of the dynamic pruning module. Since it is recursively pruning in each stacked layer of JointLK, the percentage of graph nodes that the model ultimately retains is also related to the number of layers of JointLK, that is, KNK^{N}, where N=5N=5. Experiments show that if the retention ratio is too high, there may be almost no pruning effect (for example, K=0.98, 90% of the nodes are retained in the last layer); otherwise, useful nodes may be deleted. As shown in Figure 3 (b), when the number of JointLK layers N=5N=5, K=0.92K=0.92 (about 66% of the original nodes remain in the last layer) works the best on the CommonsenseQA dev set.

5.3 Quantitative Analysis

Considering the overall performance improvement of our model on these two datasets, we analyze whether the improvement is reflected in questions that require more complex reasoning, such as questions with negation and complex questions with more entities. We compare our model with the prior best LM+KG model, QA-GNN in Table 6.

Questions with negation Large LMs do well due to memorizing subject and filler co-occurrences but are easily distracted by elements like negation Zagoury et al. (2021). To investigate the reasoning ability of the model on negation, we retrieved 133 questions with negation terms (e.g., no, not, nothing, never, unlikely, don’t, doesn’t, didn’t, can’t, couldn’t) from the CommonsenseQA IHdev set. JointLK exhibits a big boost (\uparrow3.00%) over QA-GNN, suggesting its strength in negation reasoning. The fine-grained joint inference of LM and GNN allows the model to pay attention to the semantic nuances of language expressions.

Questions with fewer/more entities When the question contains many entities, the size and noise of the retrieved KG may limit the model’s performance because the model needs to understand the complex relationship between entities. According to statistics (see Appendix A), questions contain an average of 7 entities, so we divide the question into two categories: containing fewer entities (\leq7) and more entities(>>7). Compared with QA-GNN, JointLK has a bigger boost on questions with more entities (\uparrow2.01%) than those with fewer entities (\uparrow0.96%), suggesting that our model can reduce the reasoning difficulty of complex questions because it can remove irrelevant nodes in reasoning.

Refer to caption
Figure 4: Case study of our model reasoning and pruning process. The question and answer choices corresponding to this case are: "What do people typically do while playing guitar? A.cry B. hear sounds C. singing D. arthritis E. making music".

5.4 Interpretability: A Case Study

We aim to interpret JointLK’s reasoning process by analyzing the pruning of the knowledge subgraph. Figure 4 shows an example from CommonsenseQA where our model correctly answers the question and finally retains reasonable reasoning paths by pruning the subgraph. The flow from (a) to (b) to (c) represents the recursive pruning of the subgraph according to the LM-to-KG attention weight at each GNN update layer. From (a) to (b), although the nodes wood and burn bridge the reasoning gap between question entity and answer entity, their semantics are very different from the question. From (b) to (c), “play_guitarusedforfunplay\_guitar\stackrel{{\scriptstyle usedfor}}{{\longrightarrow}}fun" and “funrelatedtogasrelatedtosingefun\stackrel{{\scriptstyle relatedto}}{{\longrightarrow}}gas\stackrel{{\scriptstyle relatedto}}{{\longrightarrow}}singe" are both reasonable, but the former is related to the semantics of the question, and the latter is not. Two paths are reserved in (c), “play_guitarhassubeventtake_lessonshassubeventdancerelatedtosinging"play\_guitar\stackrel{{\scriptstyle hassubevent}}{{\longrightarrow}}take\_lessons\stackrel{{\scriptstyle hassubevent}}{{\longrightarrow}}dance\stackrel{{\scriptstyle relatedto}}{{\longrightarrow}}singing" and “play_guitarrelatedtoactionrelatedtosingerrelatedtosingingplay\_guitar\stackrel{{\scriptstyle relatedto}}{{\longrightarrow}}action\stackrel{{\scriptstyle relatedto}}{{\longrightarrow}}singer\stackrel{{\scriptstyle relatedto}}{{\longrightarrow}}singing". These two paths describe two possible scenarios that support answering the question.

5.5 Error Analysis

In order to understand why our model fails in some cases, we randomly select 100 error cases and group them into several categories. There are three main types of errors, and we show some examples in the Appendix C.

Miss important evidence (39/100) Although we can retrieve many nodes related to questions and choices from ConceptNet, due to the incompleteness of the knowledge graph, there may be missing essential evidence nodes in the reasoning paths to answer the question. For example, although “eating_dinner" will cause “sleepiness" or “indigestion", knowledge such as “lactose intolerance causes indigestion" is essential to answer the question (Wikipedia: Lactose intolerance is a common condition caused by a decreased ability to digest lactose, a sugar found in dairy products.). However, ConceptNet does not cover such knowledge or not is retrieved.

Indistinguishable knowledge (25/100) Several choices of the question may be correct, difficult to distinguish, and which one is correct may vary from person to person. For example, “human" and “cat" may be at location “bed" or “comfortable chair", and the knowledge provided by ConceptNet is also the same. The model may choose bed because the bed appears more frequently in the pre-trained corpus.

Incomprehensible questions (23/100) This type of error often occurs when the question is particularly long, involving various events and changes in the characters’ emotions. The model is difficult to understand the scene described by the question. Some questions may require reasoning based on events, but the knowledge in ConceptNet is more based on entities and attributes.

The above three types of errors show that selecting complete, accurate, and context-sensitive knowledge is vital for more effective KG-augmented models.

6 Conclusion

In this work, we propose JointLK and provide a set of experiments to prove that (i) LM and KG interactive fusion can reduce the semantic gap between the two information modalities and make better use of KG for joint reasoning with LM. (ii) Dynamic pruning module can recursively delete irrelevant subgraph nodes at each layer of JointLK to provide fine appropriate evidence. Our results on CommonsenseQA and OpenBookQA demonstrate the superiority of JointLK over other methods using external knowledge and the strong performance in performing complex reasoning. In addition, our research results can be broadly extended to other tasks that require KGs as additional background knowledge to augment LMs, such as entity linking, KG completion and the recommendation system.

Acknowledgements

We would like to thank the anonymous reviewers for their helpful comments. This work was supported by the Key Development Program of the Ministry of Science and Technology (No.2019YFF0303003), the National Natural Science Foundation of China (No.61976068) and "Hundreds, Millions" Engineering Science and Technology Major Special Project of Heilongjiang Province (No.2020ZX14A02).

Ethical Impact

This paper proposes a general approach to fuse language models and external knowledge graphs for commonsense reasoning. We worked within the purview of acceptable privacy practices and strictly followed the data usage policy. In all the experiments, we use public datasets and consist of their intended use. We neither introduce any social/ethical bias to the model nor amplify any bias in the data, so we do not foresee any direct social consequences or ethical issues.

References

  • Banerjee and Baral (2020) Pratyay Banerjee and Chitta Baral. 2020. Knowledge fusion and semantic knowledge ranking for open domain question answering. arXiv preprint arXiv:2004.03101.
  • Banerjee et al. (2019) Pratyay Banerjee, Kuntal Kumar Pal, Arindam Mitra, and Chitta Baral. 2019. Careful selection of knowledge to solve open book question answering. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 6120–6129, Florence, Italy. Association for Computational Linguistics.
  • Bian et al. (2021) Ning Bian, Xianpei Han, Bo Chen, and Le Sun. 2021. Benchmarking knowledge-enhanced commonsense question answering via knowledge-to-text transformation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 12574–12582.
  • Clark et al. (2020) Peter Clark, Oren Etzioni, Tushar Khot, Daniel Khashabi, Bhavana Mishra, Kyle Richardson, Ashish Sabharwal, Carissa Schoenick, Oyvind Tafjord, Niket Tandon, et al. 2020. From ‘f’to ‘a’ on the NY regents science exams: An overview of the aristo project. AI Magazine, 41(4):39–53.
  • Feng et al. (2020) Yanlin Feng, Xinyue Chen, Bill Yuchen Lin, Peifeng Wang, Jun Yan, and Xiang Ren. 2020. Scalable multi-hop relational reasoning for knowledge-aware question answering. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1295–1309, Online. Association for Computational Linguistics.
  • Gunning (2018) David Gunning. 2018. Machine common sense concept paper. arXiv preprint arXiv:1810.07528.
  • Khashabi et al. (2020) Daniel Khashabi, Sewon Min, Tushar Khot, Ashish Sabharwal, Oyvind Tafjord, Peter Clark, and Hannaneh Hajishirzi. 2020. UNIFIEDQA: Crossing format boundaries with a single QA system. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 1896–1907, Online. Association for Computational Linguistics.
  • Lan et al. (2020) Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. 2020. Albert: A lite bert for self-supervised learning of language representations. In International Conference on Learning Representations.
  • Lee et al. (2019) Junhyun Lee, Inyeop Lee, and Jaewoo Kang. 2019. Self-attention graph pooling. In Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pages 3734–3743. PMLR.
  • Lin et al. (2019) Bill Yuchen Lin, Xinyue Chen, Jamin Chen, and Xiang Ren. 2019. KagNet: Knowledge-aware graph networks for commonsense reasoning. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 2829–2839, Hong Kong, China. Association for Computational Linguistics.
  • Liu et al. (2020) Liyuan Liu, Haoming Jiang, Pengcheng He, Weizhu Chen, Xiaodong Liu, Jianfeng Gao, and Jiawei Han. 2020. On the variance of the adaptive learning rate and beyond. In International Conference on Learning Representations.
  • Liu et al. (2019) Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692.
  • Lv et al. (2020) Shangwen Lv, Daya Guo, Jingjing Xu, Duyu Tang, Nan Duan, Ming Gong, Linjun Shou, Daxin Jiang, Guihong Cao, and Songlin Hu. 2020. Graph-based reasoning over heterogeneous external knowledge for commonsense question answering. In Proceedings of the AAAI Conference on Artificial Intelligence, pages 8449–8456.
  • Ma et al. (2019) Kaixin Ma, Jonathan Francis, Quanyang Lu, Eric Nyberg, and Alessandro Oltramari. 2019. Towards generalizable neuro-symbolic systems for commonsense question answering. In Proceedings of the First Workshop on Commonsense Inference in Natural Language Processing, pages 22–32, Hong Kong, China. Association for Computational Linguistics.
  • Mihaylov et al. (2018) Todor Mihaylov, Peter Clark, Tushar Khot, and Ashish Sabharwal. 2018. Can a suit of armor conduct electricity? a new dataset for open book question answering. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 2381–2391, Brussels, Belgium. Association for Computational Linguistics.
  • Raffel et al. (2020) Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. 2020. Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research, 21(140):1–67.
  • Santoro et al. (2017) Adam Santoro, David Raposo, David G Barrett, Mateusz Malinowski, Razvan Pascanu, Peter Battaglia, and Timothy Lillicrap. 2017. A simple neural network module for relational reasoning. In Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc.
  • Scarselli et al. (2008) Franco Scarselli, Marco Gori, Ah Chung Tsoi, Markus Hagenbuchner, and Gabriele Monfardini. 2008. The graph neural network model. IEEE transactions on neural networks, 20(1):61–80.
  • Schlichtkrull et al. (2018) Michael Schlichtkrull, Thomas N Kipf, Peter Bloem, Rianne Van Den Berg, Ivan Titov, and Max Welling. 2018. Modeling relational data with graph convolutional networks. In European semantic web conference, pages 593–607. Springer.
  • Speer et al. (2017) Robyn Speer, Joshua Chin, and Catherine Havasi. 2017. Conceptnet 5.5: An open multilingual graph of general knowledge. In Thirty-first AAAI conference on artificial intelligence.
  • Srivastava et al. (2014) Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. 2014. Dropout: A simple way to prevent neural networks from overfitting. Journal of Machine Learning Research, 15(56):1929–1958.
  • Talmor et al. (2019) Alon Talmor, Jonathan Herzig, Nicholas Lourie, and Jonathan Berant. 2019. CommonsenseQA: A question answering challenge targeting commonsense knowledge. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4149–4158, Minneapolis, Minnesota. Association for Computational Linguistics.
  • Vrandečić and Krötzsch (2014) Denny Vrandečić and Markus Krötzsch. 2014. Wikidata: A free collaborative knowledgebase. Commun. ACM, 57(10):78–85.
  • Wang et al. (2020a) Kai Wang, Weizhou Shen, Yunyi Yang, Xiaojun Quan, and Rui Wang. 2020a. Relational graph attention network for aspect-based sentiment analysis. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 3229–3238, Online. Association for Computational Linguistics.
  • Wang et al. (2020b) Peifeng Wang, Nanyun Peng, Filip Ilievski, Pedro Szekely, and Xiang Ren. 2020b. Connecting the dots: A knowledgeable path generator for commonsense question answering. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 4129–4140, Online. Association for Computational Linguistics.
  • Wang et al. (2019) Xiaoyan Wang, Pavan Kapanipathi, Ryan Musa, Mo Yu, Kartik Talamadupula, Ibrahim Abdelaziz, Maria Chang, Achille Fokoue, Bassem Makni, Nicholas Mattei, et al. 2019. Improving natural language inference using external knowledge in the science questions domain. In Proceedings of the AAAI Conference on Artificial Intelligence, pages 7208–7215.
  • Xu et al. (2021) Yichong Xu, Chenguang Zhu, Ruochen Xu, Yang Liu, Michael Zeng, and Xuedong Huang. 2021. Fusing context into knowledge graph for commonsense question answering. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 1201–1207, Online. Association for Computational Linguistics.
  • Yan et al. (2021) Jun Yan, Mrigank Raman, Aaron Chan, Tianyu Zhang, Ryan Rossi, Handong Zhao, Sungchul Kim, Nedim Lipka, and Xiang Ren. 2021. Learning contextualized knowledge structures for commonsense reasoning. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 4038–4051, Online. Association for Computational Linguistics.
  • Yasunaga et al. (2021) Michihiro Yasunaga, Hongyu Ren, Antoine Bosselut, Percy Liang, and Jure Leskovec. 2021. QA-GNN: Reasoning with language models and knowledge graphs for question answering. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 535–546, Online. Association for Computational Linguistics.
  • Zagoury et al. (2021) Avishai Zagoury, Einat Minkov, Idan Szpektor, and William W Cohen. 2021. What’s the best place for an ai conference, vancouver or _: Why completing comparative questions is difficult. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 14292–14300.
  • Zhu et al. (2020) Chen Zhu, Yu Cheng, Zhe Gan, Siqi Sun, Tom Goldstein, and Jingjing Liu. 2020. Freelb: Enhanced adversarial training for natural language understanding. In International Conference on Learning Representations.

Appendix A Extracting subgraph from External KG

We choose ConceptNet as the external knowledge base, and we follow the process of Feng et al. (2020) and Yasunaga et al. (2021) to retrieve the knowledge subgraph.

Given the question and choice, we identify the concepts that appear in ConceptNet in question and choice, respectively, and get the initial node set VqV_{q} and VaV_{a}, which form the initial node set Vq,aV_{q,a}. For example, in the question “What do people typically do while playing guitar?" and choice “singing", VqV_{q} = {guitar, people, play, play_guitar, playing, playing_guitar, typically}, VaV_{a} = {singe, singing}. Then, in order to extract the subgraph related to question and choice, we add the bridge entities on the 1 and 2 hop paths between any pair of entities in Vq,aV_{q,a}, thus obtaining the retrieved entity set VV.

There may be many nodes in VV, especially long questions contain many concepts. We follow the preprocessing method of Yasunaga et al. (2021), connect the nodes with question + choice, and calculate the relevant scores of the nodes through a pre-trained LM. We only retain the top 200 scoring nodes (It is worth noting that this is the preprocessing of the retrieval process, which is different from the dynamic pruning in section 3.5. The former is to score only one node and separate from the whole subgraph where the node is located, while the latter is recursive pruning in the updating process of the modeling subgraph).

Finally, we get the relation set RR by merging the relation types in ConceptNet and adding reverse relation. We retrieve all the edges in RR of any two nodes in VV. In addition, we add question as a node qq to VV, and add the bidirectional edges of qq to VqV_{q} and qq to VaV_{a}. The relation types are shown in Table 7, and the statistics of the retrieved nodes are shown in Table 8.

Relation Merged Relation
AtLocation AtLocation
LocatedNear
Causes Causes
CausesDesire
MotivatedByGoal
Antonym Antonym
DistinctFrom
HasSubevent HasSubevent
HasFirstSubevent
HasLastSubevent
HasPrerequisite
Entails
MannerOf
IsA IsA
InstanceOf
DefinedAs
PartOf PartOf
HasA
RelatedTo RelatedTo
SimilarTo
Synonym
CapableOf CapableOf
CreatedBy CreatedBy
Desires Desires
UsedFor UsedFor
HasContext HasContext
HasProperty HasProperty
MadeOf MadeOf
NotCapableOf NotCapableOf
NotDesires NotDesires
ReceivesAction ReceivesAction
qVqq\rightarrow V_{q} qVqq\rightarrow V_{q}
qVaq\rightarrow V_{a} qVaq\rightarrow V_{a}
Table 7: Relation types after preprocessing. *RelationX indicates the reverse relation of RelationX. There are 19 kinds of merged relations. We consider the reverse edge of each relation during training and testing, so there are 38 relation types in total.
Datesets Split Average |Vq||V_{q}| Average |Va||V_{a}| Average |V||V|
CommomsenseQA Train set 7.43 2.07 107.96
Dev set 7.20 2.05 106.55
Test set 7.38 2.05 106.22
OpenBookQA Train set 6.59 2.85 100.14
Dev set 6.48 3.41 108.15
Test set 6.42 3.08 101.60
Table 8: Statistics on the number of retrieved subgraph nodes corresponding to each piece of data. VqV_{q} is the set of entities included in a question. VaV_{a} is the set of entities included in a choice. VV contains VqV_{q}, VaV_{a}, and any bridging entity with no more than two hops between any pair of entities in VqV_{q} and VaV_{a}.

Appendix B Node Initialization

For each entity in the subgraph, we need to obtain its feature representation. Following Feng et al. (2020), we first use the template to convert the knowledge triples in ConceptNet into sentences, and feed them into BERT-Large, obtaining a sequence of tokens embeddings from the last layer. For each entity, we perform mean pooling over the tokens of the entity’s occurrences across all the sentences to form the initial embeddings xi0x_{i}^{0}.

Appendix C Error Types and Examples

In Table 9, we present examples for each error type in the Commonsense IHdev set. Because the average number of subgraph nodes corresponding to each case is about 100, we cannot list them all. Only some important nodes are shown here.

Error type Example
Missing important evidence (39/100) Question He has lactose intolerant, but was eating dinner made of cheese, what followed for him?
Answer choices digestive×|\times| feel better ×|\times| sleepiness×|\times| indigestion |\checkmark| illness×\times
Subgraph for correct answer eating_dinnercausesindigestioneating\_dinner\stackrel{{\scriptstyle causes}}{{\longrightarrow}}indigestion, intolerantrelatedtopainisasymptomisaindigestionintolerant\stackrel{{\scriptstyle relatedto}}{{\longrightarrow}}pain\stackrel{{\scriptstyle isa}}{{\longrightarrow}}symptom\stackrel{{\scriptstyle isa}}{{\longleftarrow}}indigestion, …
Subgraph for predicted answer lactoserelatedtofoodhassubeventeating_dinnercausessleepinesslactose\stackrel{{\scriptstyle relatedto}}{{\longrightarrow}}food\stackrel{{\scriptstyle hassubevent}}{{\longrightarrow}}eating\_dinner\stackrel{{\scriptstyle causes}}{{\longrightarrow}}sleepiness, intolerantrelatedtobearrelatedtosleepintolerant\stackrel{{\scriptstyle relatedto}}{{\longrightarrow}}bear\stackrel{{\scriptstyle relatedto}}{{\longrightarrow}}sleep, …
Indistinguishable knowledge (25/100) Question Where would a cat snuggle up with their human?
Answer choices floor×|\times| humane society×|\times| bed×|\times| comfortable chair|\checkmark| window sill×\times
Subgraph for correct answer catatlocationchaircat\stackrel{{\scriptstyle atlocation}}{{\longrightarrow}}chair, humanatlocationchairhuman\stackrel{{\scriptstyle atlocation}}{{\longrightarrow}}chair, …
Subgraph for predicted answer catatlocationbedcat\stackrel{{\scriptstyle atlocation}}{{\longrightarrow}}bed, humanatlocationbedhuman\stackrel{{\scriptstyle atlocation}}{{\longrightarrow}}bed, …
Incomprehensible questions (23/100) Question The man tried to break the glass in order to make his escape in time, but he could not. The person in the car, trying to kill him, did what?
Answer choices accelerate|\checkmark| putting together×|\times| working×|\times| construct×|\times| train×\times
Subgraph for correct answer escapeisabreakantonymaccelerateescape\stackrel{{\scriptstyle isa}}{{\longleftarrow}}break\stackrel{{\scriptstyle antonym}}{{\longrightarrow}}accelerate, killrelatedtoattackrelatedtoacceleratekill\stackrel{{\scriptstyle relatedto}}{{\longrightarrow}}attack\stackrel{{\scriptstyle relatedto}}{{\longrightarrow}}accelerate, manrelatedtobreakrelatedtofallinghassubeventaccelerateman\stackrel{{\scriptstyle relatedto}}{{\longrightarrow}}break\stackrel{{\scriptstyle relatedto}}{{\longrightarrow}}falling\stackrel{{\scriptstyle hassubevent}}{{\longrightarrow}}accelerate, …
Subgraph for predicted answer breakisaactionrelatedtoworkbreak\stackrel{{\scriptstyle isa}}{{\longrightarrow}}action\stackrel{{\scriptstyle relatedto}}{{\longrightarrow}}work, escapeisabreakhassubeventworkescape\stackrel{{\scriptstyle isa}}{{\longleftarrow}}break\stackrel{{\scriptstyle hassubevent}}{{\longleftarrow}}work, killcausesdiehassubeventworkkill\stackrel{{\scriptstyle causes}}{{\longrightarrow}}die\stackrel{{\scriptstyle hassubevent}}{{\longleftarrow}}work, …
Table 9: Several error cases of JointLK model on CommonsenseQA dev dataset. Because there are many nodes in the subgraph, we represent some nodes and relationships in the subgraph in the form of links.