\ul
Commonsense-Focused Dialogues for Response Generation: An Empirical Study
Abstract
Smooth and effective communication requires the ability to perform latent or explicit commonsense inference. Prior commonsense reasoning benchmarks (such as SocialIQA and CommonsenseQA) mainly focus on the discriminative task of choosing the right answer from a set of candidates, and do not involve interactive language generation as in dialogue. Moreover, existing dialogue datasets do not explicitly focus on exhibiting commonsense as a facet. In this paper, we present an empirical study of commonsense in dialogue response generation. We first auto-extract commonsensical dialogues from existing dialogue datasets by leveraging ConceptNet, a commonsense knowledge graph. Furthermore, building on social contexts/situations in SocialIQA, we collect a new dialogue dataset with 25K dialogues aimed at exhibiting social commonsense in an interactive setting. We evaluate response generation models trained using these datasets and find that models trained on both extracted and our collected data produce responses that consistently exhibit more commonsense than baselines. Finally we propose an approach for automatic evaluation of commonsense that relies on features derived from ConceptNet and pretrained language and dialog models, and show reasonable correlation with human evaluation of responses’ commonsense quality. We are releasing a subset of our collected data, Commonsense-Dialogues111The released version was manually validated by in-house crowd workers and hence is slightly different from the initial version used in the paper. URL: github.com/alexa/commonsense-dialogues, containing about 11K dialogs.
1 Introduction
Open-domain dialogue response generation (RG) models aim to provide human-like natural language responses given dialogue histories Chen et al. (2017). To improve generated response quality, many studies have been conducted to develop knowledge-grounded RG Ghazvininejad et al. (2018); Gopalakrishnan et al. (2019), personalized dialogue agents Zhang et al. (2018), empathetic response Rashkin et al. (2019), etc. For all the above-mentioned directions for RG, large-scale dialogue data geared towards the specific goals is crucial, since most current state-of-the-art neural RG models require training on appropriate and large data. Therefore several datasets have been collected to support such research efforts such as knowledge-grounded dialogues Ghazvininejad et al. (2018); Gopalakrishnan et al. (2019), PersonaChat Zhang et al. (2018), and EmpatheticDialogues Rashkin et al. (2019). Producing natural and logically-coherent responses given dialogue contexts involves making commonsense inferences during the communication. For example, if someone says “I’m going to perform in front of a thousand people tomorrow…” the listener is likely to conclude that the speaker is probably feeling nervous and respond by comforting them: “Relax, you’ll do great!” In contrast to other efforts to make RG models more empathetic or knowledgeable, there is a lack of commonsense focused dialogue data for both training neural models and evaluation. An ideal dataset for studying commonsense in RG needs to simulate how humans have multi-turn conversations as much as possible. Existing commonsense-focused work in RG uses extracted post-response pairs from Reddit Zhou et al. (2018), which are single-turn and rough approximations for real-life conversations.
Aiming to bridge the gap in commonsense for dialogue response generation, we collect a large-scale multi-turn open-domain dialogue dataset that is focused on commonsense knowledge. We first consider extracting commonsense-focused dialogues from three existing dialogue datasets by identifying responses that contain commonsense inferences using ConceptNet Liu and Singh (2004). This filtering results in 21k dialogues. Then we collect 25k new dialogues focusing on social commonsense inferences, where prompts are context sentences describing an event in the SocialIQA data Sap et al. (2019b).
To study commonsense in RG, we train large generative language models on our datasets and compare with models trained on existing datasets. We find through sampled human evaluation that our dataset helps to generate more commonsensical responses (average score of 6.9 out of 10 compared to 4.8 using other data), and automatically generated responses still have a large gap in comparison to human performances (9.2 out of 10). To help lower the evaluation cost and increase the efficiency of evaluating commonsense in RG, we further propose an automatic metric using combined neural and symbolic features derived from ConceptNet, and show that this metric has reasonable correlation with human annotations and symbolic features contribute positively to system performance.
Our contributions are as follows: (1) We create the first large-scale open-domain dialogue dataset focusing on social commonsense inferences. This includes a new collection of 25k dialogues based on SocialIQA event prompts, and ConceptNet filtered data from some existing data sets. We are also releasing the ConceptNet filtered portion of our data collection, Commonsense-Dialogues, which contains about 11K dialogs. (2) We benchmark our dataset and show that models trained on our dataset helps make models produce more commonsensical responses. (3) We propose the first automatic metric for evaluating the commonsense plausibility in response generation that reaches statistically significant correlation with human annotations.
2 Task Introduction and Motivations
2.1 Commonsense-Focused Dialogue Response Generation
We study commonsense-focused response generation for dialogues. Commonsense can be defined as “the basic level of practical knowledge and reasoning concerning everyday situations and events that are commonly shared among most people” Sap et al. (2020). Dialogue response generation is the task of generating a response turn in a conversational setting given previous history turns . Thus by combining these two together, we want to examine models’ ability to produce responses that make sense or is plausible in terms of commonsense.
2.2 Motivations
Lack of Commonsense-Focused Analysis on Existing Dialogue Datasets
Numerous dialogue data has been collected for training RG models and other dialogue-related tasks. As mentioned before, many different aspects of RG have been explored, such as knowledge-grounded Ghazvininejad et al. (2018); Gopalakrishnan et al. (2019) and empathy Rashkin et al. (2019), whereas, to the best of our knowledge, there is no study or large-scale multi-turn data for analyzing whether model-generated responses present the ability to communicate with commonsense knowledge or reasoning.
Lack of real-life interactive setting for Commonsense Reasoning Benchmarks
Current commonsense reasoning (CSR) benchmarks mostly target models’ ability to choose a right answer from several candidates given a question. We argue that this is a highly artificial scenario as models do not get options to choose from in real-life, and often they need to generate utterances. Recent work such as CommonGen Lin et al. (2020) has started to explore generative settings to examine commonsense in natural language processing (NLP) models. This line of work, however, is still far from real use cases as it does not consider a real-life interaction task setup such as conversations. Thus we argue that existing commonsense benchmarks in NLP are not enough to train a language agent that produces smooth interpersonal communications, nor evaluate whether models have such capabilities.
3 Commonsense Focused Dialogue Collection
To collect more commonsense focused dialogues for response generation model training and evaluation, our effort is along two directions: filtering existing data to collect dialogues with responses that consist of commonsense (Section 3.1), and curating new data using prompts from a commonsense reasoning multiple-choice benchmark SocialIQA (Section 3.2).
3.1 Filtering Based on Existing Dialogue Datasets
We propose a simple process for filtering commonsense in dialogues and present our analysis of three dialogue datasets with different focuses: DailyDialog Li et al. (2017), EmpatheticDialogues Rashkin et al. (2019), and MuTual Cui et al. (2020). The general idea is to refer to a commonsense knowledge graph (CSKG) such as ConceptNet Liu and Singh (2004) to identify potential commonsense triples () expressing a commonsense assertion between turns in a dialogue. The following describes the detailed process.
Identify Candidate Concepts The first step is to identify potential candidates for concept entities in the commonsense triples. For a turn in a dialogue, we use a part-of-speech (POS) tagger to find the nouns, verbs, and adjectives that are not stopwords and then construct a set of potential concepts by including the lemmatized version of these words. We use the POS tagger, lemmatizer, and stopword list from the Natural Language Toolkit (NLTK) package Bird et al. (2009). This step results in a set of concept words for each turn of a dialogue. For example, consider an exchange between two participants in a conversation: “Hi, I want to find a doctor”, “What kind of doctor are you looking for? A general doctor or a specialist?”, the concept sets for the two turns are “want, find, doctor” and “look, general, doctor, specialist”, respectively.
Query ConceptNet for Neighboring Entities With a set of concepts we extract for every dialogue turn, we then identify a list of candidate triples expressing commonsense assertions about each concept such that we can later check if some of those assertions indeed appear in this dialogue. We rely on the widely-used ConceptNet Liu and Singh (2004) as the knowledge resource, which consists of commonsense knowledge about various concepts. Specifically we use the ConceptNet containing single-word concepts pre-processed by Zhou et al. (2018). For each concept we identified in a turn, we store all triples in ConceptNet that contain this concept, either as subject or object. Using the above example, example triples about “doctor” include “doctor LocateAt hospital”, “patient RelatedTo doctor”, and “specialist TypeOf doctor”.
Search Entities in the Next Turn After getting a list of commonsense triples containing concepts in a particular turn using ConceptNet, we next examine if any of the other entity in the triples appears in the concept set of the next turn. In the example dialogue exchange above, where “doctor” is a concept appearing in a turn, for the triple “specialist TypeOf doctor”, we search if “specialist” is in the concept set of the next turn. Since we find such a match, we record this triple to be a commonsense assertion that might be implied in the response.
Filtering Results We filter dialogues using the above-mentioned approach: if we can successfully find a matching triple between two adjacent turns, we keep the dialogue as it might contain commonsense assertions identified from ConceptNet. We consider three dialogue datasets in this study:
-
•
DailyDialog(DD) Li et al. (2017). It includes general-domain day-to-day dialogues crawled from various English learning websites.
-
•
EmpatheticDialogues (ED) Rashkin et al. (2019). It is an empathy-focused dialogue dataset crowdsourced from Amazon Mechanical Turk (MTurk).
-
•
MuTual Cui et al. (2020). It is a reasoning-focused response selection dataset based on English listening comprehension exams for Chinese students.
We choose these three datasets to examine three different types of focuses in dialogue datasets: general-domain, empathy, and general reasoning (but not specifically on commonsense).
After the process, we find that in the training sets, around 7k out of the 11k dialogues (63%) from Dailydialogue contain at least one matched triple between their turns, and 9.5k out of the 18k for EmpatheticDialogues (53%), and 5k out of 7k (73%) for MuTual dialogues. For the valid and test sets, the proportion of such dialogues is similar to that in the training sets for these three data sets.
Note that there are some limitations in our ConceptNet based data selection approach. First, we match concept entities based on just surface form, rather than semantic meaning or word senses in the context. Second, we are only using single word concepts, not phrases. Third, we are only considering one-hop concept relation identified in ConceptNet. The first one may affect the precision of the selected dialogues, and the other two reasons affect the recall. Without human annotated commonsense reasoning for dialog turns, we can not compute the exact performance of our filtering method. We plan to conduct some human annotation in our future work. Among the three data sets used in this study, the fact that there is a higher percentage of dialogues selected in MuTual may indicate that data focuses more on reasoning and thus is more likely to contain commonsense relations.
3.2 New Data Collection Using SocialIQA Prompts
To facilitate commonsense-guided response generation training, we collect more dialogues with a focus on getting responses that require commonsense. Specifically, we make use of an existing commonsense multiple-choice benchmark SocialIQA Sap et al. (2019b) to crowdsource dialogues. This section provides background on SocialIQA, the crowdsourcing process, and the resulting dialogues.
Background and motivation We collect dialogues by prompting crowdsourcing workers on Amazon Mechanical Turk (MTurk) with context sentences from SocialIQA that describe an event in everyday social scenarios. SocialIQA Sap et al. (2019b) is a large-scale commonsense reasoning benchmark about social situations. It contains around 38k multiple-choice questions, each consisting of a context sentence, a question, and three answer choices. Context was generated by rewriting events from ATOMIC Sap et al. (2019a), a large knowledge graph (KG) that contains inferential knowledge about the causes and effects of 24k short events. An example event in ATOMIC is “PersonX spills all over the floor”, which crowd workers were asked to turn into a sentence by adding names, fixing potential grammar errors, and filling in placeholders, resulting in a context like “Alex spilled food all over the floor.”
We choose to use SocialIQA contexts because of three reasons: (1) they are specific instantiations of the event phrases found in the knowledge graph ATOMIC, which guarantees that there is at least one potential commonsense inference that can be made from the event; (2) ATOMIC covers a wide range of commonsense motivations and reactions and thus the contexts also embed diverse commonsense; (3) the rewriting process from SocialIQA ensures that the context sentences are well-formed and similar to natural sentences, which we expect is not hard for crowd workers to come up with a dialogue.
Prompt selection We inspected around 200 contexts trying to write a dialogue and found that the contexts that we had the most difficulty with are the ones that are too short or do not contain an interesting event to start a conversation. For example, contexts such as “Robin stopped eating the food to save room for dessert” might not be an interesting event to talk about in a dialogue. To select appropriate contexts as prompts for dialog writing, we apply a simple heuristic criteria: the context has to be either longer than 15 words or contains a punctuation such as a comma or a period in the middle. The intuition is that longer contexts are easier to write a dialogue with because they contain more information and a punctuation often indicates a development in the narrative of the event (e.g., “Tracy performed her function. Their employer gave them a raise”). This makes the event more complicated, and thus avoids too trivial events. We also filter out context sentences that do not contain any person names. As a result of this preprocessing, we kept 12.8k out of 33k contexts in the training set and 754 out of 2k contexts in the development set, adding up to 13.5k contexts from SocialIQA.
Prompts | Dialogue Examples | |||||||||
---|---|---|---|---|---|---|---|---|---|---|
Tracy performed her function. |
|
|||||||||
|
||||||||||
Addison wanted to go on a trip to Mexico, and messaged all of his friends to set up a schedule. |
|
|||||||||
|
Dialogue Collection Using selected contexts from SocialIQA, we ran a task on MTurk asking each worker to write a dialogue with 4 to 6 turns between two friends about the event described in the context. Note that, this is a ‘self-talk’ dialog collection. Specifically, since there is a name appearing in the context, we ask a worker to write a dialogue by first imagining that they are the person mentioned in the context and are talking with their friend about the event described. For example, consider the context above (“Tracy performed her function. Their employer gave them a raise”), we ask a worker to imagine themselves to be “Tracy” and that they are talking to a friend (also played by themselves) about getting a raise.
We pose three requirements for turkers in order to work on our task: locate in US, UK, or Canada; successful HITS are over 1000, and with more than 95% HIT acceptance rate. We pay MTurk workers $0.5 for each instance, roughly translating to 10 dollars per hour, well above the minimum wage of US.
To account for multiple plausible dialogues expanded from the context event, we assign each context to five different MTurk workers. We randomly sample 5k context sentences out of 13.5k filtered ones and collect five dialogues for each context, resulting in 25k dialogues. Examples of our collected dialogues are shown in Table 1.
For our collected data, we follow the same filtering steps as used for other existing data (Section 3.1). This ConceptNet filtering identifies about 11K dialog from the entire collection.222Our released data is based on these ConceptNet filtered conversations. We recruited in-house crowd workers to manually check the dialogs for profanity, speaker name, and other issues in the data. Note the experiments conducted in this paper used the initial collection, not this released version. Though we expect the SocialIQA contexts are from ATOMIC and may trigger more commonsensical dialogue, we find this is not the case since the percentage of dialogues containing ConceptNet triples is even lower than what we observed for the other existing data sets. This may be because of the limitations of the filtering method we are using as described earlier: matching to ConceptNet is based on surface textual form and concepts are on word-level, which omits deeper and more contextual commonsense relationships.
4 Experiment Setup and Evaluation Methods
The focus of this study is to examine how commonsense plays a role in dialogue response generation. In previous sections, we propose a simple filtering method to obtain commonsense-focused dialogues from existing three datasets and crowdsource more dialogues based on the SocialIQA commonsense reasoning benchmark. Here we aim to evaluate response generation models’ ability to produce responses that follow commonsense and if training on commonsense-focused dialogue data helps boost model performance. In addition to using automatic referenced metrics and human evaluation, we also propose a new automatic unreferenced metric aiming to evaluate responses for commonsense quality.
4.1 Experiment Settings
For response generation models, we take one of the state-of-the-art pre-trained language models, GPT2 Radford et al. (2019), and further train it on our training data sets. Specifically, the model is trained in a multitask fashion that minimizes the LM loss as well as the multiple choice loss following Wolf et al. (2019), and generates responses for a given dialog history.
We consider the follow three types of training data setups.
-
•
Existing data sets, including DailyDialog Li et al. (2017) (DD), EmpatheticDialogues Rashkin et al. (2019)(ED), and Topical-Chat Gopalakrishnan et al. (2019), a knowledge-grounded open-domain dataset with around 11k dialogues. MuTual Cui et al. (2020) is not included since it is designed for response selection.
-
•
As described in Section 3.1, we use ConceptNet to search for potential triples in response turns and filter three dialogue datasets, DD, ED, and MuTual. We combine the three filtered dialogues from these datasets to form our training data, named ‘filter existing’ (FE, total around 21K dialogues).
-
•
The third category includes our collected dialogues using SocialIQA contexts. This is used along with the FE data above: FE and all of the 25k collected dialogues (FE+new crowdsourced), and FE plus the 11K filtered dialogues of our collected data (FE+filtered crowdsourced).
To evaluate models’ response generation capabilities, we sample 10% of the FE+new data, resulting in 4.6k testing dialogues with no overlap with the training set of any of the settings above. We use GPT2 trained on different versions of dialogue data (6 trained GPT2 models in total) to generate a randomly sampled response for each turn of our test set dialogues.
4.2 Evaluation Metrics
We perform automatic evaluation on the test set and human evaluation on sampled dialogs.
Automatic Evaluation We consider several widely-used automatic metrics for evaluating response generation: perplexity of the reference responses in the data, Meteor score Banerjee and Lavie (2005), ROUGE score Lin (2004), and BERTScore Zhang et al. (2019). Note that these metrics (except perplexity) provide general evaluation of the generated responses, but do not specifically focus on commonsense plausibility.
Human Evaluation Since there is no existing evaluation method that reliably examines whether a response follows commonsense and correlates with human judgements, we ask humans to score system generated responses as well as the reference response given a dialogue history. We sample 300 history-response pairs from dialogues in our test set to perform human evaluation. All the model-generated responses from the 6 trained models above and the original response (human response) (around 2100 responses in total) are scored in terms of commonsense plausibility by MTurkers. We specifically asked workers to score the responses in terms of commonsense plausibility using a scale of 1 to 10. We also instructed them that criteria such as grammatical correctness and fluency should not be taken into much account and they should focus on evaluating the commonsense aspect of the response. Three annotators evaluated each response. We calculate the average human scores and variance to measure the performances of different responses.
4.3 Proposed Automatic Metric for Commonsense
Human evaluation is expensive to obtain, especially when the dataset is large. In addition, they are also subjective and hard to reproduce. Aiming to provide a reliable and scalable automatic metric focusing on commonsense in response generation, we propose an unreferenced automatic metric, which is a regression model trained from the human annotation scores for different responses. The metric is reference-free, meaning that it does not require human ground truth response when scoring a model-generated response, unlike referenced metrics such as BLEU, ROUGE, Meteor.
Regressor model We use a simple multi-layer perceptron (MLP) as our regressor and consider both neural and symbolic features to train the MLP model. For symbolic features, we consider the number of one-hop and two-hop triples that can be found between the dialogue history and the response turn from ConceptNet. The triple identifying process is the same as our filtering process described earlier (Section 3.1). That is, we first identify a set of concepts in the response turn and query ConceptNet for potential triples and match those with the other concepts appearing in the dialogue history. Two-hop triples are searched in a similar manner, with the only difference being that the number of potential triples will be much larger. We also include the length of the response as an additional feature. As for neural features, we use the scores from a dialogue-focused language model DialoGPT Zhang et al. (2020) on both the response itself and the dialogue history concatenated with the response. The score from DialoGPT can be considered as the plausibility of the sentence. We train this MLP model using the human evaluation scores for different responses.
5 Results and Analysis
5.1 Automatic Evaluation Results
Table 2 shows results according to automatic metrics on our 4.6K testing dialogues. We find that perplexity scores for the GPT2 model trained on filtered existing dialogue data (FE), or plus new collected data (FE+Crowdsourced), are much lower than that just trained on existing datasets as is. There are several reasons for this. One is that since the testing dialogues are from the filtered version, training on those better matches the evaluation scenario. In addition, the test set is a sample of multiple data sets, and thus training on just one data set does not perform well. Finally the combined data (the last three rows in the table) is larger in size (see training size in Table 3). However, note the gain from the increasing training data size decreases in comparison to the difference between using the filter data settings and those single data sets. Meteor and ROUGE scores for all the trained models are quite low, and show less differences, probably indicating the limitation of these metrics for dialog response evaluation. BERTScore shows a similar pattern as perplexity in terms of model quality.
Data | Perplexity | Meteor | ROUGE | BERTScore |
---|---|---|---|---|
DD | 31.25 | 0.06 | 0.06 | 0.12 |
ED | 24.80 | 0.08 | 0.08 | 0.14 |
TC | 28.48 | 0.09 | 0.08 | 0.11 |
Filtered Existing (FE) | 13.20 | 0.09 | 0.08 | 0.16 |
FE+Crowdsourced | 11.31 | 0.09 | 0.08 | 0.17 |
FE+Filtered Crowdsourced | 12.27 | 0.09 | 0.08 | 0.17 |
5.2 Human Evaluation Results
Table 3 shows the human evaluation scores on 300 responses for models trained with different types of data. The most obvious and perhaps expected finding is that GPT2, no matter trained on what types of data, is still way behind human performance (6.86 with high variance versus 9.3 with low variance). By analyzing different variables that cause performance difference, we find the following patterns, some of which are similar to using automatic metrics. (1) Using the Filtered Existing dialogue data (FE) helps improve the average of commonsense scores (more than 1 point improvement comparing to using individual data sets), but variance remains high; (2) Including our collected dialogues further increases the average (FE+Crowdsourced), and also decreases the variance in response quality in terms of commonsense plausibility; (3) Regarding our collected data, using the filter subset of it yields slightly better performance than using the entire data collection. This suggests that even though our data is collected using SocialIQA events, some dialogues may not be commonsense rich, which is also reflected by the percentage of dialogues that contain ConceptNet triples as discussed earlier. In addition, it shows that though overall increasing training data size benefits model performance, the quality of data plays a more important role. We plan to perform more sophisticated data selection and commonsense annotation for our data set in the future. We include examples of responses from humans and models trained on these different types of data as well as annotation scores in Appendix A Table 5. It shows some different characteristics of the responses, for example, empathy in the responses using ED model, and richer information (though inappropriate since they are off topic) using TC model.
Data | Training Size | Avg. Score | Variance |
---|---|---|---|
DD | 11k | 4.677 | 11.977 |
ED | 18k | 4.998 | 12.233 |
TC | 10k | 4.558 | 11.562 |
Filtered Existing (FE) | 21k | 5.968 | 12.426 |
FE+Crowdsourced | 46k | 6.767 | 9.067 |
FE+Filtered Crowdsourced | 31k | 6.865 | 8.684 |
Human response | N/A | 9.298 | 2.544 |
5.3 Proposed Commonsense Automatic Evaluation Results
We now examine the correlation of our proposed automatic metric (MLP regressor) with human scores on the testing portion of our annotations. We cross-validate on the collected dialogues with 0.8/0.1/0.1 proportions. For comparison, we consider three baselines: our MLP with only symbolic features, our MLP with only neural features, and FED Mehri and Eskenazi (2020a), which uses DialoGPT to score how likely the next turn after the response expresses confusion. It requires no training nor human references, and has been shown to correlate with humans judgements on different criteria (commonsense not included). Table 4 shows the Spearman’s correlation of the system computed scores and human annotation scores using all the annotated data in a cross-validation setup. We can see that our simple MLP-based regressor reaches the highest spearman’s correlation with human scores, outperforming other baselines significantly. However, such a correlation result still suggests a large gap for a reliable scorer targeting commonsense evaluation for dialogue response generation. We also notice that FED performs poorly in terms of commonsense evaluation. Furthermore, there is a large correlation drop when considering either symbolic or neural features alone in our model, indicating that they might each capture a different aspect for evaluating commonsense.
Metrics | Spearman’s Correlation | p-Value | |
---|---|---|---|
FED | -0.00797 | 0.80569 | |
Symbolic | 0.12336 | 1.27E-08 | |
Ours | Neural | 0.06176 | 0.00450 |
All features | 0.20789 | 4.53E-22 |
6 Related Work
6.1 Commonsense Reasoning
The majority of recent commonsense reasoning benchmarks Zellers et al. (2018); Talmor et al. (2019); Bisk et al. (2020); Sap et al. (2019b) test a model’s ability to choose the correct option given a context and a question; pre-trained language models have reached high performance on these benchmarks after fine-tuning. There have been many benchmarks that focus on reasoning abilities in multiple tasks such as reading comprehension Huang et al. (2019); Yu et al. (2020), dialogue systems Cui et al. (2020), and natural language inference Williams et al. (2018), which involve inferences on language. Recent work also aims to probe models in these tasks to see if reasoning is actually achieved Richardson and Sabharwal (2020); Richardson et al. (2020); Zhou et al. (2020). In this study we tackle the response generation problem in dialogues, with a focus on collecting commonsense rich dialog data and evaluating commonsense quality of model responses.
6.2 Open Domain Dialogue Response Generation
Recently open domain dialog systems have been modeled using end-to-end approaches, more specifically encoder-decoder architectures (Sordoni et al., 2015; Serban et al., 2017, 2016; Vinyals and Le, 2015). Recent work focused on finetuning large pre-trained transformer models Radford et al. (2019); Zhang et al. (2020) on dialog data. Many dialog datasets have been collected with different focuses such as incorporating knowledge (Gopalakrishnan et al., 2019; Dinan et al., 2018), empathy Rashkin et al. (2019), task completion Budzianowski et al. (2018), consistency Nie et al. (2020), personality Zhang et al. (2018) and reasoning Cui et al. (2020) within dialog systems. There has also been work on combining a variety of datasets to exhibit multiple attributes Roller et al. (2020).
6.3 Dialog Response Evaluation
Due to the diverse responses that a dialog system can output, referenced automatic metrics (such as BLEU, ROUGE, Perplexity) do not correlate well with human judgement of these systems (Deriu et al., 2020; Liu et al., 2016). As a result, human evaluation has become the de-facto standard to evaluate dialog systems. However human evaluation is costly. Recently model-based metrics have been proposed with good correlation with human annotations (Zhang et al., 2019; Sellam et al., 2020; Mehri and Eskenazi, 2020b, a; Tao et al., 2018; Lowe et al., 2017). Most metrics focus on evaluating the coherence or appropriateness of a response with respect to its dialog context. Mehri and Eskenazi (2020a) identified 18 different dialog qualities such as interesting and topic depth. However none of these metrics evaluate the commonsense of a response, which is the focus of this work.
7 Conclusion
We present our empirical study on commonsense in dialogue response generation. To obtain data for commonsense-focused analysis in open domain response generation, we use two strategies: filtering existing dialogue data using a commonsense knowledge graph ConcepetNet, and collecting new dialogues using prompts from multiple-choice commonsense benchmark. Our data has a few limitations such as our filtering process focuses on word-level matching to ConceptNet, which might omit more complex commonsense relations embedded in dialogues. We leave deeper analysis of how implicit commonsense is represented in dialogues and how to elicit multi-hop granular reasoning steps during communications to future work. Our experimental results show that our newly collected data helps boost response generation model performance based on human evaluation of commonsense. To close the gap in automatic evaluation metric for response generation, we propose a simple MLP regressor using both neural and symbolic features, and show promising correlation with human judgements.
We are releasing the ConceptNet filtered portion of our data collection, with further manual examination, Commonsense-Dialogues, which consists of about 11K dialogs. We hope our work and this new data will help with future attempts to make models produce responses with more commonsense, which is a challenging but crucial task to tackle in dialog systems.
Ethics and Broader Impact
Our work uses ConceptNet to filter for commonsense-focused dialogues, but Mehrabi et al. (2021) have found representational harms in common sense resources. We acknowledge that the generated responses from models we use might contain biases. All of the dialogue datasets and models are in English, which benefits English speakers more. We used Amazon Mechanical Turks for data collection. We pay turkers around $14 per hour, well above the highest state minimum wage and engage in constructive discussions if they have concerns about the process. We also give each annotation instance enough time so that we do not pressure annotators.
References
- Banerjee and Lavie (2005) Satanjeev Banerjee and Alon Lavie. 2005. Meteor: An automatic metric for mt evaluation with improved correlation with human judgments. In Proceedings of the acl workshop on intrinsic and extrinsic evaluation measures for machine translation and/or summarization, pages 65–72.
- Bird et al. (2009) Steven Bird, Ewan Klein, and Edward Loper. 2009. Natural language processing with Python: analyzing text with the natural language toolkit. ” O’Reilly Media, Inc.”.
- Bisk et al. (2020) Yonatan Bisk, Rowan Zellers, Ronan Le Bras, Jianfeng Gao, and Yejin Choi. 2020. Piqa: Reasoning about physical commonsense in natural language. AAAI.
- Budzianowski et al. (2018) Paweł Budzianowski, Tsung-Hsien Wen, Bo-Hsiang Tseng, Iñigo Casanueva, Stefan Ultes, Osman Ramadan, and Milica Gasic. 2018. Multiwoz-a large-scale multi-domain wizard-of-oz dataset for task-oriented dialogue modelling. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 5016–5026.
- Chen et al. (2017) Hongshen Chen, Xiaorui Liu, Dawei Yin, and Jiliang Tang. 2017. A survey on dialogue systems: Recent advances and new frontiers. Acm Sigkdd Explorations Newsletter, 19(2):25–35.
- Cui et al. (2020) Leyang Cui, Yu Wu, Shujie Liu, Yue Zhang, and Ming Zhou. 2020. Mutual: A dataset for multi-turn dialogue reasoning. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 1406–1416.
- Deriu et al. (2020) Jan Deriu, Alvaro Rodrigo, Arantxa Otegi, Guillermo Echegoyen, Sophie Rosset, Eneko Agirre, and Mark Cieliebak. 2020. Survey on evaluation methods for dialogue systems. Artificial Intelligence Review, pages 1–56.
- Dinan et al. (2018) Emily Dinan, Stephen Roller, Kurt Shuster, Angela Fan, Michael Auli, and Jason Weston. 2018. Wizard of wikipedia: Knowledge-powered conversational agents. arXiv preprint arXiv:1811.01241.
- Ghazvininejad et al. (2018) Marjan Ghazvininejad, Chris Brockett, Ming-Wei Chang, Bill Dolan, Jianfeng Gao, Wen-tau Yih, and Michel Galley. 2018. A knowledge-grounded neural conversation model. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 32.
- Gopalakrishnan et al. (2019) Karthik Gopalakrishnan, Behnam Hedayatnia, Qinglang Chen, Anna Gottardi, Sanjeev Kwatra, Anu Venkatesh, Raefer Gabriel, and Dilek Hakkani-Tür. 2019. Topical-chat: Towards knowledge-grounded open-domain conversations. In INTERSPEECH, pages 1891–1895.
- Huang et al. (2019) Lifu Huang, Ronan Le Bras, Chandra Bhagavatula, and Yejin Choi. 2019. Cosmos qa: Machine reading comprehension with contextual commonsense reasoning. arXiv preprint arXiv:1909.00277.
- Li et al. (2017) Yanran Li, Hui Su, Xiaoyu Shen, Wenjie Li, Ziqiang Cao, and Shuzi Niu. 2017. Dailydialog: A manually labelled multi-turn dialogue dataset. In Proceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 986–995.
- Lin et al. (2020) Bill Yuchen Lin, Wangchunshu Zhou, Ming Shen, Pei Zhou, Chandra Bhagavatula, Yejin Choi, and Xiang Ren. 2020. Commongen: A constrained text generation challenge for generative commonsense reasoning. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings, pages 1823–1840.
- Lin (2004) Chin-Yew Lin. 2004. Rouge: A package for automatic evaluation of summaries. In Text summarization branches out, pages 74–81.
- Liu et al. (2016) Chia-Wei Liu, Ryan Lowe, Iulian V Serban, Michael Noseworthy, Laurent Charlin, and Joelle Pineau. 2016. How not to evaluate your dialogue system: An empirical study of unsupervised evaluation metrics for dialogue response generation. arXiv preprint arXiv:1603.08023.
- Liu and Singh (2004) Hugo Liu and Push Singh. 2004. Conceptnet—a practical commonsense reasoning tool-kit. BT technology journal, 22(4):211–226.
- Lowe et al. (2017) Ryan Lowe, Michael Noseworthy, Iulian V Serban, Nicolas Angelard-Gontier, Yoshua Bengio, and Joelle Pineau. 2017. Towards an automatic turing test: Learning to evaluate dialogue responses. arXiv preprint arXiv:1708.07149.
- Mehrabi et al. (2021) Ninareh Mehrabi, Pei Zhou, Fred Morstatter, Jay Pujara, Xiang Ren, and Aram Galstyan. 2021. Lawyers are dishonest? quantifying representational harms in commonsense knowledge resources. arXiv preprint arXiv:2103.11320.
- Mehri and Eskenazi (2020a) Shikib Mehri and Maxine Eskenazi. 2020a. Unsupervised evaluation of interactive dialog with dialogpt. arXiv preprint arXiv:2006.12719.
- Mehri and Eskenazi (2020b) Shikib Mehri and Maxine Eskenazi. 2020b. Usr: An unsupervised and reference free evaluation metric for dialog generation. arXiv preprint arXiv:2005.00456.
- Nie et al. (2020) Yixin Nie, Mary Williamson, Mohit Bansal, Douwe Kiela, and Jason Weston. 2020. I like fish, especially dolphins: Addressing contradictions in dialogue modelling. arXiv preprint arXiv:2012.13391.
- Radford et al. (2019) Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. Language models are unsupervised multitask learners. OpenAI Blog 1.8 (2019): 9.
- Rashkin et al. (2019) Hannah Rashkin, Eric Michael Smith, Margaret Li, and Y-Lan Boureau. 2019. Towards empathetic open-domain conversation models: A new benchmark and dataset. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 5370–5381.
- Richardson et al. (2020) Kyle Richardson, Hai Hu, Lawrence S Moss, and Ashish Sabharwal. 2020. Probing natural language inference models through semantic fragments. In AAAI, pages 8713–8721.
- Richardson and Sabharwal (2020) Kyle Richardson and Ashish Sabharwal. 2020. What does my qa model know? devising controlled probes using expert knowledge. Transactions of the Association for Computational Linguistics, 8:572–588.
- Roller et al. (2020) Stephen Roller, Emily Dinan, Naman Goyal, Da Ju, Mary Williamson, Yinhan Liu, Jing Xu, Myle Ott, Kurt Shuster, Eric M Smith, et al. 2020. Recipes for building an open-domain chatbot. arXiv preprint arXiv:2004.13637.
- Sap et al. (2019a) Maarten Sap, Ronan Le Bras, Emily Allaway, Chandra Bhagavatula, Nicholas Lourie, Hannah Rashkin, Brendan Roof, Noah A Smith, and Yejin Choi. 2019a. Atomic: An atlas of machine commonsense for if-then reasoning. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 3027–3035.
- Sap et al. (2019b) Maarten Sap, Hannah Rashkin, Derek Chen, Ronan Le Bras, and Yejin Choi. 2019b. Social iqa: Commonsense reasoning about social interactions. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 4453–4463.
- Sap et al. (2020) Maarten Sap, Vered Shwartz, Antoine Bosselut, Yejin Choi, and Dan Roth. 2020. Commonsense reasoning for natural language processing. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: Tutorial Abstracts, pages 27–33.
- Sellam et al. (2020) Thibault Sellam, Dipanjan Das, and Ankur P Parikh. 2020. Bleurt: Learning robust metrics for text generation. arXiv preprint arXiv:2004.04696.
- Serban et al. (2016) Iulian V Serban, Alessandro Sordoni, Yoshua Bengio, Aaron Courville, and Joelle Pineau. 2016. Building end-to-end dialogue systems using generative hierarchical neural network models. In Thirtieth AAAI Conference on Artificial Intelligence.
- Serban et al. (2017) Iulian Vlad Serban, Alessandro Sordoni, Ryan Lowe, Laurent Charlin, Joelle Pineau, Aaron Courville, and Yoshua Bengio. 2017. A hierarchical latent variable encoder-decoder model for generating dialogues. In Thirty-First AAAI Conference on Artificial Intelligence.
- Sordoni et al. (2015) Alessandro Sordoni, Michel Galley, Michael Auli, Chris Brockett, Yangfeng Ji, Margaret Mitchell, Jian-Yun Nie, Jianfeng Gao, and Bill Dolan. 2015. A neural network approach to context-sensitive generation of conversational responses. arXiv preprint arXiv:1506.06714.
- Talmor et al. (2019) Alon Talmor, Jonathan Herzig, Nicholas Lourie, and Jonathan Berant. 2019. Commonsenseqa: A question answering challenge targeting commonsense knowledge. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4149–4158.
- Tao et al. (2018) Chongyang Tao, Lili Mou, Dongyan Zhao, and Rui Yan. 2018. Ruber: An unsupervised method for automatic evaluation of open-domain dialog systems. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 32.
- Vinyals and Le (2015) Oriol Vinyals and Quoc Le. 2015. A neural conversational model. arXiv preprint arXiv:1506.05869.
- Williams et al. (2018) Adina Williams, Nikita Nangia, and Samuel Bowman. 2018. A broad-coverage challenge corpus for sentence understanding through inference. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 1112–1122. Association for Computational Linguistics.
- Wolf et al. (2019) Thomas Wolf, Victor Sanh, Julien Chaumond, and Clement Delangue. 2019. Transfertransfo: A transfer learning approach for neural network based conversational agents. CoRR.
- Yu et al. (2020) Weihao Yu, Zihang Jiang, Yanfei Dong, and Jiashi Feng. 2020. Reclor: A reading comprehension dataset requiring logical reasoning. arXiv preprint arXiv:2002.04326.
- Zellers et al. (2018) Rowan Zellers, Yonatan Bisk, Roy Schwartz, and Yejin Choi. 2018. Swag: A large-scale adversarial dataset for grounded commonsense inference. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (EMNLP).
- Zhang et al. (2018) Saizheng Zhang, Emily Dinan, Jack Urbanek, Arthur Szlam, Douwe Kiela, and Jason Weston. 2018. Personalizing dialogue agents: I have a dog, do you have pets too? In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2204–2213.
- Zhang et al. (2019) Tianyi Zhang, Varsha Kishore, Felix Wu, Kilian Q Weinberger, and Yoav Artzi. 2019. Bertscore: Evaluating text generation with bert. arXiv preprint arXiv:1904.09675.
- Zhang et al. (2020) Yizhe Zhang, Siqi Sun, Michel Galley, Yen-Chun Chen, Chris Brockett, Xiang Gao, Jianfeng Gao, Jingjing Liu, and William B Dolan. 2020. Dialogpt: Large-scale generative pre-training for conversational response generation. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: System Demonstrations, pages 270–278.
- Zhou et al. (2018) Hao Zhou, Tom Young, Minlie Huang, Haizhou Zhao, Jingfang Xu, and Xiaoyan Zhu. 2018. Commonsense knowledge aware conversation generation with graph attention. In IJCAI, pages 4623–4629.
- Zhou et al. (2020) Pei Zhou, Rahul Khanna, Seyeon Lee, Bill Yuchen Lin, Daniel Ho, Jay Pujara, and Xiang Ren. 2020. Rica: Evaluating robust inference capabilities based on commonsense axioms. arXiv preprint arXiv:2005.00782.
Appendix A Example Responses and Scores
Table 5 shows responses to some dialogue histories from both human and models. We also include scores from human annotations described in 5.2.
History |
|
|
|
|||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Human and Model Responses with Annotation Scores (1-10) | Human Response |
|
|
|
||||||||||||||||
DD | B: I hope so. [3] |
|
A: You are so wonderful! [3] | |||||||||||||||||
ED |
|
|
|
|||||||||||||||||
TC |
|
|
|
|||||||||||||||||
Filtered Existing (FE) |
|
|
|
|||||||||||||||||
FE+Crowdsourced |
|
|
|
|||||||||||||||||
FE+Filtered Crowdsourced |
|
|
|