This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

UniDU: Towards A Unified Generative Dialogue Understanding Framework

Zhi Chen1, Lu Chen1, Bei Chen2, Libo Qin2, Yuncong Liu1,
Su Zhu3, Jian-Guang Lou2
The corresponding authors are Lu Chen and Kai Yu.
   Kai Yu111footnotemark: 1
1X-LANCE Lab, Department of Computer Science and Engineering
MoE Key Lab of Artificial Intelligence, AI Institute, Shanghai Jiao Tong University
State Key Lab of Media Convergence Production Technology and Systems, Beijing, China
2Microsoft Research Asia
3AISpeech Co., Ltd., Suzhou, China
Abstract

With the development of pre-trained language models, remarkable success has been witnessed in dialogue understanding (DU). However, current DU approaches usually employ independent models for each distinct DU task without considering shared knowledge across different DU tasks. In this paper, we propose a unified generative dialogue understanding framework, named UniDU, to achieve effective information exchange across diverse DU tasks. Here, we reformulate all DU tasks into a unified prompt-based generative model paradigm. More importantly, a novel model-agnostic multi-task training strategy (MATS) is introduced to dynamically adapt the weights of diverse tasks for best knowledge sharing during training, based on the nature and available data of each task. Experiments on ten DU datasets covering five fundamental DU tasks show that the proposed UniDU framework largely outperforms task-specific well-designed methods on all tasks. MATS also reveals the knowledge-sharing structure of these tasks. Finally, UniDU obtains promising performance in the unseen dialogue domain, showing the great potential for generalization.

1 Introduction

The development of the conversational system plays an important role in the spread of the intelligence devices, such as intelligence assistants and car play. In recent years, there has been a growing interest in neural dialogue system Wen et al. (2017); Ultes et al. (2017); Li et al. (2017); Chen et al. (2018a, 2019, 2020b); Bao et al. (2020); Adiwardana et al. (2020); Ham et al. (2020); Peng et al. (2020); Chen et al. (2022). Dialogue understanding is a core technology and hot topic in the dialogue system, aiming to analyze a dialogue from different fine-grained angles accurately.

There are five classical dialogue understanding tasks: dialogue summary (DS) Liu et al. (2019a), dialogue completion (DC) Su et al. (2019); Quan et al. (2020), intent detection (ID) Kim et al. (2016); Casanueva et al. (2020); Qin et al. (2021a), slot filling (SF) Zhang et al. (2017); Qin et al. (2021b); Haihong et al. (2019) and dialogue state tracking (DST) Kim et al. (2020); Chen et al. (2020a); Hosseini-Asl et al. (2020); Xu et al. (2020); Liao et al. (2021). Dialogue summary aims to generate a concise description of given dialogue content, which is normally formulated as a sequence-to-sequence generation problem Wu et al. (2021). Dialogue completion eliminates the co-reference and information ellipsis in the latest utterance, which is also a generation task Chen et al. (2021b). Intent detection and slot filling are two traditional spoken language understanding tasks that aim to map natural language to logical form. Intent detection is typically treated as a classification problem Liu and Lane (2016) and slot filling is usually formulated as a sequence labeling task Zhang et al. (2017); Qin et al. (2019); Coope et al. (2020). The dialogue state tracking task is to extract the user’s constraints on the predefined dialogue domains and slots Budzianowski et al. (2018). The five different tasks aim to interpret a dialogue from five different perspectives. To date, these DU tasks are still learned independently due to different task formats. However, they are intuitively related. For example, the dialogue completion task should have a positive effect on the dialogue state tracking task Han et al. (2020). On the other hand, it is usually very expensive to collect dialogue data and annotate them, which constraints the scale of annotated dialogue corpora. It is important and imperative to study how to enhance dialogue understanding capability with the existing diverse dialogue corpora.

There are two main challenges in knowledge sharing across DU tasks: data annotation diversity and task nature diversity. It is necessary to employ a unified DU model to allow all types of DU data to be used together. In this paper, we propose a Unifined Dialogue Uderstanding (UniDU) framework, in which the five fundamental DU tasks are modelled by a unified sequence-to-sequence generative model. The second challenge is related to the nature of diverse tasks. Since the output label dynamic ranges and the goals of the DU tasks are different, tasks may not be well suited to be trained together with straightforward multi-task learning. It is then a nontrivial problem to effectively weight diverse tasks for the unified model with different dialogue corpora. In this paper, we propose a novel adaptive weighting approach and compare it with other different training strategies under the UniDU framework.

The main contributions of this paper are summarized below:

  • To the best of our knowledge, we are the first to formulate different dialogue understanding tasks as a unified generation task spanned five DU tasks. The proposed UniDU outperforms well-designed models on five well-studied dialogue understanding benchmarks.

  • We propose a model-agnostic adaptive weighting approach for multitask learning to address the task nature diversity problem. We find that the intuitive multitask mixture training method makes the unified model bias convergence to more complex tasks. The proposed model-agnostic training method can efficiently relieve this problem.

  • Experimental results show that the proposed UniDU method has excellent generalization ability, which achieves advanced performance both on few-shot and zero-shot setups.

2 Dialogue Understanding Tasks

We denote dialogue context as C=(Hn,Un)C=(H_{n},U_{n}), where Hn=(U1,U2,,Un1)H_{n}=(U_{1},U_{2},\dots,U_{n-1}) represents the dialogue history containing the first n1n-1 turns of utterances. UnU_{n} is nn-th turn utterance, which may consist of multiple sentences stated by one speaker. For the task-oriented dialogue, the domain scope is restricted by the dialogue ontology, which the dialogue expert designs. The ontology OO is composed of dialogue domains D={d}D=\{d\} (like hotel), domain slots (like price) S={s}S=\{s\} and user intent candidates I={i}I=\{i\} (like find_hotel). There are five fundamental tasks to interpret a dialogue from different perspectives.

Dialogue Summary (DS) aims to extract important information of the dialogue. It is a typical generation problem, which takes the whole dialogue context CC as input and generates the summary description. DS requires the model to focus on the whole dialogue flow and the important concepts.

Dialogue Completion (DC) purposes to relieve the co-reference and information ellipsis problems, which frequently occur in the dialogue context. It is also a typical generation task, which inputs the dialogue history HnH_{n} and the current utterance UnU_{n} and then infers the semantic-completed statement of the current utterance UnU_{n}. DC requires the model to focus on the connection between current utterance and dialogue history.

Slot Filling (SF) is to extract the slot types SS of the entities mentioned by the user. It is a word tagging problem where the utterance is labeled in the IOB (Inside, Outside, and Beginning) format. The input is only the current utterance UnU_{n}.

Intent Detection (ID) is to recognize the intent from predefined abstracted intent expresses II. It is normally formulated as a classification problem. The input is the current utterance UnU_{n}, and the output is the possible distribution of all the intent candidates II.

Dialogue State Tracking (DST) aims to record the user’s constraints, which consists of the triple set of domain-slot-value. For example, hotel-price-cheap means the user wants a cheap hotel. The input of DST at the nn-th turn is the first nn turns (U1,,Un)(U_{1},\dots,U_{n}).

Refer to caption
Figure 1: Overview of UniDU. Under UniDU framework, the input consists of three parts: task identification, dialogue content and task query, where \oplus means concatenation. The output has two components: task identification and query answer. We train the UniDU model with different multitask learning strategies.

3 UniDU

In this section, we first introduce the unified sequence-to-sequence data format for the five DU tasks. Then we introduce the formulation of each task in detail, especially how to reformulate the intent detection, slot filling and dialogue state tracking as the generation task.

There are three components in the input of UniDU: task identification, dialogue content, and task query. The task identification represents with a special token, e.g., dialogue summary identified by “[DS]”. The dialogue content means the task-dependent input, such as dialogue history for dialogue summary. The task query can be regarded as the task-specific prompt, which includes the task definition and domain-related information. There are two elements in the output of UniDU: task identification and query answer. The query answer is the understanding result of the task query given by the dialogue content. The unified input and output can be formalized as:

where “[C]” is separate character and “[TI]” is task identification (replaced by “[DS]”, “[DC]”, “[SF]”,“[ID]” and “[DST]”, which correspond to dialogue summary, dialogue completion, slot filling, intent detection and dialogue state tracking respectively). At inference time, the UniDU model must first predict the task identification.

Dialogue summary and dialogue completion are originally generative tasks. The dialogue contents in the input are the whole dialogue context CC and multi-turn utterances HnH_{n} respectively. Since these two tasks are independent of the dialogue domain, there is no domain information in the task query. For dialogue summary, the task query is “what is the summary of this dialogue?”. For dialogue completion, the query is ‘‘what is the semantic completion statement of UnU_{n}?”, where UnU_{n} is the tt-th utterance. Their understanding answers are annotated dialogue summaries and rewritten utterances in the output.

The original slot filling task demands the model to extract all the mentioned slot values and their slot types in an utterance UnU_{n}. In this paper, the UniDU model predicts the value slot by slot, which is an iterated generation process on the slot candidate list. Two different slot filling formats are shown below:

[Uncaptioned image]

To be clear, we do not list all the candidate slots here. In general, for each sample, it can be formalized as:

where ss and dd are predefined slots and domains. If ss has no value in UnU_{n}, slot value will be “not mentioned”. If ss has multiple values, they will be separated by a comma in the slot value. When the value is “not mentioned”, we call it a negative sample. Otherwise, it is a positive sample. To balance the ratio of negative and positive samples during the training process, we set the ratio to 2:1. If the number of negative samples exceeds the threshold, we randomly sample twice as many negative instances as positive ones.

For dialogue state tracking tasks, the classification methods always achieve better performance than generative methods. However, under the UniDU framework, we also formulate DST as a slot-wise value generation task similar to the slot filling task. The DST task formats are shown below:

[Uncaptioned image]

where the output of the original DST model is the distribution of all the candidate values of the slot. The input and output of the DST task under UniDU can be formalized as follows:

where (Hn,Un)(H_{n},U_{n}) is dialogue context. If slot ss of the domain dd is not in the dialogue state, its value is “not mentioned”, which is a negative sample. Note that different utterances are separated by the special token “[T]” in the input. During the training process, the ratio of negative and positive samples is also set below 2:1.

For the intent detection task, the original methods formulate it as the intent classification problem and output the distribution of all the candidate intents. The UniDU model directly generates the intent name of the current utterance, which can be formalized as:

where domain dd is normally known in advance. The specific examples of original and UniDU formats are shown below:

[Uncaptioned image]

where we do not list all the intents. To integrate the generalization capability into the UniDU model, we also construct negative samples for the intent detection task. The intent name of the negative sample is “not defined”, where the input utterances UnU_{n} are sampled from out-of-domain dialogues. The ratio of negative and positive samples is set to 2:1. Until now, all the five dialogue understanding tasks have been formulated as the unified sequence-to-sequence generation task. The specific examples are shown in Figure 1.

4 Multitask Training Strategies

Although the five DU tasks can be formulated as a unified generative task, straightforward multitask training may not work due to the different natures of these tasks. In this section, we discuss multitask training strategies and propose a novel model-agnostic adaptive weighting strategy.

4.1 Multitask Learning Classification

The existing multitask training strategies can be classified into three categories: average sum method, manual scheduled method, and learnable weight method.

Average Sum method distributes all the samples with the same weight. In other words, the losses from different samples are directly averaged, formulated as =1Tt=1Tt\mathcal{L}=\frac{1}{T}\sum_{t=1}^{T}\mathcal{L}_{t}, where TT is the number of the tasks and t\mathcal{L}_{t} is the loss of the tt-th task.

Manual Schedule method designs a heuristic training schedule for planning the learning process of different tasks. For example, curriculum learning Bengio et al. (2009) is a kind of typical manual scheduled method, which first trains the easier samples and then adds the more complicated cases. The manual scheduled method can be formulated as =1𝕀(t)t=1T𝕀(t)t\mathcal{L}=\frac{1}{\sum\mathbb{I}(t)}\sum_{t=1}^{T}\mathbb{I}(t)\cdot\mathcal{L}_{t}, where 𝕀(t)\mathbb{I}(t) is indicator function, whose value is 0 or 1.

Learnable Weight method aims to parameterize the loss weights of different tasks. The target of the parameterized weights is to balance the effects of task instances, which prevents the model from slanting to one or several tasks and achieves global optimization. There are two classical learnable weight algorithms: homoscedastic uncertainty weighting (HUWKendall et al. (2018) and gradient normalization (GradNormChen et al. (2018b). For the tasks, the loss function is formulated as =t=1TWtt\mathcal{L}=\sum_{t=1}^{T}W_{t}\cdot\mathcal{L}_{t}, where WtW_{t} is learnable weights and greater than 0. In the HUW algorithm, the weights update as the following loss function:

HUW=t=1T(tWtlog(Wt)),\mathcal{L}_{\rm HUW}=\sum_{t=1}^{T}(\mathcal{L}_{t}\cdot W_{t}-\log(W_{t})), (1)

where log(Wt)\log(W_{t}) is to regularize weights, which is adaptive to regression tasks and classification tasks. The motivation of the GradNorm method is to slow down the learning scale of the task that has a larger gradient magnitude and faster convergence rate.

4.2 Model-Agnostic Training Strategy

In Equation 1, the learnable weight WtW_{t} is only dependent on the corresponding task. Thus, we can regard the weight as the function of task Wϕ(t)W_{\phi}(t), where ϕ\phi are parameters shared among five tasks. Under the UniDU framework, five tasks share the same encoder-decoder model, which is a constant in weight function Wϕ(t)W_{\phi}(t). The task format depends on task attributes, such as input, output, and data scale. To extract the characters of five tasks, we manually design a vector as the task feature to represent a task. Each dimension in the task feature has its physical meaning related to the model-agnostic setting. In this paper, we design 14 dimensional vector 𝐟t\mathbf{f}_{t} for each task introduced in detail in Appendix B. Since the model-agnostic training strategy (MATS) formulates the weight as the task-related function and may share the function parameters among different tasks, the weights are no longer independent as in the original learnable weight method. The MATS improved from Equation 1 is formalized as:

MATS=t=1T(tWϕ(𝐟t)log(Wϕ(𝐟t))).\mathcal{L}_{\rm MATS}=\sum_{t=1}^{T}(\mathcal{L}_{t}\cdot W_{\phi}(\mathbf{f}_{t})-\log(W_{\phi}(\mathbf{f}_{t}))). (2)

5 Experiments

We conduct the experiments on ten dialogue understanding corpora. Each task has two corpora. We evaluate the UniDU framework with eight different training strategies. Compared with well-designed models, our proposed UniDU can get better performance in five benchmarks. Then we deeply analyze different factors affecting the UniDU model’s performance, including DU tasks, unified format, and pre-trained language models. Last but not least, we conduct few-shot experiments to validate the generalization ability of UniDU.

Methods DS(SAMSUM) DC(TASK) ID(BANKING77) SF(RESTAURANTS8K) DST(WOZ2.0)
R-1 R-L EM BLEU ACC. F1F_{1} JGA
Baselines 49.67 48.95 74.2 89.4 93.44 96.00 91.4
 Wu et al. (2021)  Chen et al. (2021b)  Mehri et al. (2020)  Coope et al. (2020)  Tian et al. (2021)
Eight Training Strategies under UniDU Framework
ST 49.74 47.10 76.4 89.0 91.49 95.76 89.8
TT 51.24 48.59 76.1 89.2 91.94 95.12 91.0
MIX 50.98 48.13 76.2 90.8 91.91 96.43 90.8
G2S 51.13 48.75 76.3 90.1 90.12 94.81 86.8
CL 51.04 48.36 77.2 89.8 92.17 96.02 90.8
GradNorm 51.33 48.69 77.4 90.4 92.07 96.69 90.5
HUW 50.31 47.69 76.2 90.4 93.14 97.43 91.9
MATS 50.53 47.97 76.6 90.6 93.60 97.61 92.3
Finetune 51.93 49.01 76.1 91.0 93.54 97.19 92.1
Table 1: The results on five DU tasks trained with eight learning strategies. Finetune means that the best model (according to underlined metric values) of each task continues to be fine-tuned on separate task corpus. means that we run their released code with BART-base instead of BART-large to fairly compare with our model.

5.1 Corpora&Metrics

There are ten dialogue understanding corpora in total spanned five tasks: dialogue summary (DS), dialogue completion (DC), slot filling (SF), intent detection (ID), and dialogue state tracking (DST). We choose two well-studied corpora for each task: one is the evaluation corpus, and the other is the auxiliary corpus. The dataset statistics are shown in Appendix A.

Dialogue Summary: We choose SAMSUM Gliwa et al. (2019) and DIALOGSUM Chen et al. (2021a) datasets. The common metrics for the summary task are ROUGE scores, which measure the overlap of nn-grams in the generated summary against the reference summary.

Dialogue Completion: TASK Quan et al. (2019) and CANARD Elgohary et al. (2019) are used. The metrics are BLEU score and exact match (EM) accuracy. BLEU measures how similar the rewritten sentences are to golden ones. Exact match means the rate of the generated totally equaled to the golden.

Intent Detection: We conduct the experiments on BANKING77 Casanueva et al. (2020) and HWU64 Liu et al. (2019c), where 77 and 64 means the number of predefined intents. The evaluation metric is detection accuracy (ACC.).

Slot Filling: We choose to conduct the experiments on RESTAURANTS8K Coope et al. (2020) and SNIPS Coucke et al. (2018). We report F1F_{1} scores for extracting the correct span per user utterance. Note that the correct predictions on negative samples are not calculated in the F1F_{1} score, which is comparable with traditional methods.

Dialogue State Tracking: WOZ2.0 Wen et al. (2017) and MULTIWOZ2.2 Zang et al. (2020) are used. The metric is joint goal accuracy (JGA), which measures the percentage of success in all dialogue turns, where a turn is considered a success if and only if all the slot values are correctly predicted. Note that we only use “hotel” domain data of MULTIWOZ2.2 in the training phase.

5.2 Eight Training Strategies

As introduced in Section 4, the multitask training strategies can be divided into three categories: average sum, manual schedule, and learnable weight. Before introducing MTL training methods, there is an intuitive baseline trained on its own data named single training (ST). In ST, the sequence-to-sequence models are only trained on five evaluated datasets, respectively. In average sum method, there are two types of training strategies: task transfer learning (TTTorrey and Shavlik (2010); Ruder et al. (2019) and mixture learning (MIXWei et al. (2021). The task transfer learning aims to enhance the performance using external data from the auxiliary corpus that has the same task setup. This is the main reason that we select two corpora for each task. The mixture learning directly mixes up all the training samples from ten corpora together. In these two methods, the learning weight for each sample is equally distributed. In the manual schedule method, we test two training routes according to the curriculum learning method. From the input perspective, five tasks can be divided into three classes: utterance-level input on intent detection and slot filling, turn-level input on dialogue completion, and dialogue state tracking and dialogue-level input on dialogue summary. The inputs gradually become more complex in order: utterance-level, turn-level, and dialogue-level. Thus, the intuitive method (named CL) trains five tasks in this order. Note that the previous data are kept in the next training phase. From the task setup perspective, dialogue summary and dialogue completion belong to domain-independent tasks. The other three tasks are domain-dependent tasks. There is another training route (G2S): from general tasks to domain-specific tasks. In learnable weight method, we evaluate three methods introduced in Section 4: GradNorm, HUW and our proposed MATS.

5.3 Experimental Setup

In this paper, we set BART-base as the backbone of the unified encoder-decoder model. The BART model is implemented with HuggingFace library Wolf et al. (2019). We conduct all the experiments on the 2080TI GPU with 11G memory. we run every experiment for 60 epochs spent 72 hours. The batch size is 32 with the gradient accumulation strategy (updated per 8 steps). The learning rates of the unified model and learnable weights are 1e-5 and 1e-4, respectively. In the MATS method, the weight function consists of two linear layers with the ReLU activation function, whose hidden sizes are 64.

5.4 Results

In Table 1, we report the best evaluation performance on five tasks with eight training strategies. The well-designed models as baselines are introduced in Section 1. The experimental results show that different training strategies greatly affect the performance of five tasks under the UniDU framework. Our proposed MATS achieves the best or near best performance except on dialogue summary. On the atypical generation tasks (intent detection, slot filling, and dialogue state tracking), the UniDU with MATS methods can achieve promising improvement compared to well-designed models. The simple task transfer learning method (TT) can not largely increase the performance compared with single training. The mixture operation leads to consistent performance improvement on five tasks. However, compared with TT, the improvement is still limited except for dialogue completion. Compared with our proposed MATS, MIX biases convergence to more complex DU tasks (dialogue summary and dialogue completion). Two manual schedule methods (G2S and CL) do not have any distinct advantages. In learnable weight methods, GradNorm only achieves excellent performance on dialogue summary. HUW achieves performance gain on intent detection, slot filling, and dialogue state tracking. We continue fine-tuning the best UniDU models (signed with underline) on the corresponding corpus. We find that only the dialogue summary and dialogue completion have obvious performance gain, which reflects the necessity of the UniDU framework for simpler generative tasks.

Methods DS DC ID SF DST Overall
(R-L) (BLEU) (ACC.) (F1F_{1}) (JGA)
MIX 48.04 90.40 91.9 96.43 90.1 83.23
HUW 47.63 89.95 93.0 97.43 91.8 83.97
MATS 47.57 90.43 93.5 97.46 91.9 84.16
Table 2: The best overall performance of MIX, HUW and MATS methods.
Method DS DC ID SF DST
(R-L) (BLEU) (ACC.) (F1F_{1}) (JGA)
MATS 47.97 90.6 93.60 97.61 92.3
- DS - 90.2\blacktriangledown\blacktriangledown0.4 93.20\blacktriangledown\blacktriangledown0.4 97.35\blacktriangledown\blacktriangledown0.26 92.8\blacktriangle\blacktriangle0.5
- DC 47.77\blacktriangledown\blacktriangledown0.20 - 93.41\blacktriangledown\blacktriangledown0.19 97.39\blacktriangledown\blacktriangledown0.22 91.8\blacktriangledown\blacktriangledown0.5
- ID 47.81\blacktriangledown\blacktriangledown0.16 90.5\blacktriangledown\blacktriangledown0.1 - 97.45\blacktriangledown\blacktriangledown0.16 92.3<0.0
- SF 47.77\blacktriangledown\blacktriangledown0.20 90.5\blacktriangledown\blacktriangledown0.1 93.60<0.0 - 92.0\blacktriangledown\blacktriangledown0.3
- DST 47.85\blacktriangledown\blacktriangledown0.12 90.6<0.0 93.47\blacktriangledown\blacktriangledown0.13 97.58\blacktriangledown\blacktriangledown0.03 -
Table 3: Ablation study on effects of each task corpora.

In Table 1, we report the task-specific performance of the UniDU model, whose checkpoints are selected by the task-specific metric. Table 2 shows unified performance on five tasks with MIX, HUW, and MATS methods. We evaluate the single checkpoint of UniDU model, which has the highest evaluated overall score, on the five tasks. The overall score is the average value of the five main metrics shown in Table 2. We can see that our proposed MATS gets the highest overall performance and the best performance on four DU tasks.

5.5 Analysis

In this subsection, we analyze factors to affect the performance of UniDU model including DU tasks, unified format and pre-trained language models.

Backbone DS DC ID SF DST
(R-L) (BLEU) (ACC.) (F1F_{1}) (JGA)
Trans.-B 34.84 74.2 86.36 83.01 72.5
BART-B 47.97 90.6 93.60 97.61 92.3
T5-S 41.63 85.9 87.04 96.94 89.9
Trans.-L 34.10 67.4 86.46 71.65 71.0
BART-L 48.89 88.6 93.44 97.12 92.6
T5-B 48.89 90.7 93.90 98.14 92.6
Table 4: Ablation study on effects of different pre-trained language models with encoder-decoder architecture.

5.5.1 Effects of DU Tasks

To validate the effects of the dialogue understanding tasks, we directly remove one of five DU corpora and train the UniDU model with the MATS method shown in Table 3. In general, the five DU tasks benefit each other, except that dialogue summary has negative effects on the dialogue state tracking task. We guess the general dialogue summary task summarizes a dialogue into a sentence, ignoring the domain-specific information. On the other hand, we find that the dialogue completion task has the most significant effect on the other four DU tasks. It indicates that the co-reference and information ellipsis are still the main factors to impact the dialogue understanding ability. The phenomenon can facilitate the dialogue understanding community to pay more attention to dialogue completion. For example, when pre-training a scaling dialogue model, the pre-trained tasks should be close to the dialogue completion task.

Unseen Dialogue Content UniDUMATS
[DS] USER : I’d like a taxi to take me to ruskin gallery [T] SYSTEM : Sure! What is your departure site? [T] USER : I will depart from saffron brasserie at 7:15. What is the car type and contact number so I know who and where you will pick me up? [T] SYSTEM : Booking completed! A grey ford will be picking you up. The contact number is 07689877132. [T] USER : That is all I needed, thank you. [C] what’s the summary of this dialogue? [DS] a grey ford will take USER to ruskin gallery at 7:15.
[DC] USER : Please reserve for me a taxi that will pick me up at cambridge arts theatre after 09:30 [T] SYSTEM : And where will you be going? [T] USER : I’m going to restaurant one seven. [T] SYSTEM : Your booking is complete, a black audi will be picking you up. [T] USER : Thank you. I need the contact number, as well. [C] what is the semantic completion statement of “Thank you. I need the contact number, as well.”? [DC] I need the contact number of a black audi to pick me up at cambridge arts theatre
[ID] help me get a taxi to the cambridge museum of technology please. [C] what is the user’s intent on the taxi? [ID] transport taxi
[SF] I need a taxi to pick me up at Ashley Hotel to leave after 10:45. [C] what is leaving time of taxi? [SF] 10:45
[DST] USER : I need a taxi. I am going to avalon and I need to leave after 16:15 [C] what is the user’s constraint about the destination of the taxi? [DST] avalon
Table 5: Case study of the zero-shot performance of the best unified model trained with MATS method. The input dialogue contents are sampled from unseen “Taxi” domain.

5.5.2 Effects of Unified Format

As introduced in Section 3, we formulate dialogue understanding tasks in QA format. There is an intuitive alternative: prefix format, where the task query is concatenated on the decoder side. At inference time, the decoder is directly fed with task query and then generates the answer. As shown in Figure 2, the QA format achieves a performance boost on four of five DU tasks (except for dialogue summary) compared to the prefix format.

5.5.3 Effects of PLMs

To validate the effects of the different pre-trained backbones, we initialize the encoder-decoder of UniDU model with random mechanism, BART Lewis et al. (2020) and T5 Raffel et al. (2020). The Trans.-B and Trans.-L in Table 4 mean the random-initialized Transformer trained from scratch, which has the same parameters with BART-base model (BART-B) and BART-large model (BART-L). T5-S and T5-B mean T5-small and T5-base respectively. We can see that the pre-trained language models get absolute performance gain compared to random-initialized models. BART-B can get better performance than T5-S. When the parameter scale increases, T5-base achieves the best performance than other models. The results show that the large PLMs can improve the complex dialogue summary by a large margin.

Refer to caption
Figure 2: Ablation study of different unified understanding format.

5.6 Generalization Ability

To further evaluate the generalization ability of the UniDU model, we first conduct few-shot learning experiments on the domain-dependent slot filling task. We test the zero-shot capability of UniDU on unseen dialogue data.

Few-shot Learning: We select the UniDU model that gets the best evaluation of overall performance on five tasks learned with the MATS method. For the slot filling task, we extend another dialogue corpus DSTC8 Rastogi et al. (2020). We choose the “Bus” domain data in DSTC8, which is unseen in the training process of UniDU. Compared with vanilla BART, UniDU has obvious advantages, especially in the extremely resource-limited situation. When there is only 1% training data, the vanilla BART is disabled to learn, as shown in Figure 3. The few-shot experiment on the DST task is shown in Appendix C.

Zero-shot Performance: We validate UniDU model trained with MATS method on unseen “Taxi” domain dialogue data collected from MULTIWOZ2.2 corpus. UniDU model can get 18.24% accuracy on ID, 39.69% F1 score on SF and 1.6% JGA on DST.

Refer to caption
Figure 3: Few-shot learning results on slot filling fine-tuned on BART and UniDU. 1%, 2% and 5% are the percents of the training data on unseen “Bus” domain.

6 Case Study

We directly validate the UniDU model trained with the MATS method on unseen “Taxi” domain dialogue data collected from MULTIWOZ2.2 corpus. As shown in Table 5, we find that the UniDU model can generate reasonable dialogue summary and completion. Note that the UniDU model did not see any task-oriented dialogue in these two tasks. For domain-specific tasks, the UniDU model can still generate accurate query answers in some cases. It indicates that our proposed generative UniDU model has excellent generalization ability, which not only can adapt to unseen dialogue and also directly generate reasonable answers on five DU tasks in the zero-shot setting.

Refer to caption
Figure 4: The reduce-dimension map of task embeddings collected from UniDU model trained by MDTS. The task embedding is the final decoder representation of the task identification token.

To further explore the relations among five tasks, we plot the reduced-dimension map of the task embeddings of five tasks with the t-SNE algorithm shown in Figure 4. The task embeddings are the final decoder layer representation of the task identification token, whose model is trained with MDTS. The dialogue data is from the above unseen “Taxi” domain to eliminate the impacts of the dialogue context. We find that the embeddings of dialogue summary, dialogue completion, and intent detection cluster together. These three tasks under the UniDU framework are more general than slot filling and dialogue state tracking, whose task queries are slot-wise. The task formats between slot filling and dialogue state tracking are close. However, the UniDU model can still have good performance to distinguish between these two tasks.

7 Related Work

Our work relates to several broad research areas, including prompting, dialogue modeling, and multitask learning. Due to the content limitation, here we describe one subarea: multitask learning in NLP applications that relate most closely to our work.  Luong et al. (2016) apply a sequence-to-sequence model on three general NLP tasks and study different parameter-sharing strategies.  Kumar et al. (2016); McCann et al. (2018) try to cast NLP tasks as QA over a context. The main topics in this work are how to design an efficient model to integrate the knowledge between question and context. Liu et al. (2019b) combine four natural language understanding tasks, which utilize BERT as the shared representation model. The model corresponding to each task still has a well-designed part of solving the intrinsic problem. It hampers the analysis of the interaction among the different tasks.

Recently, Wei et al. (2021) formulated the NLP tasks as the generation task by directly mixing scaling annotated data up. They only focus on zero-shot and few-shot ability on the NLP tasks and ignore the impacts of the different multitask training strategies, which can not achieve better performance on general NLP tasks compared to supervised learning methods on well-designed models. In task-oriented dialogue (TOD) modelling,  Peng et al. (2020); Su et al. (2021) reformulate the pipeline TOD model as the sequential end-to-end generation problem. The end-to-end model needs to generate dialogue state, dialogue action, and response at the same time, which is not scalable when the number of tasks increases. The sequential format needs all the annotations of the same context, which is unavailable in the DU area. Most recently, PPTOD Su et al. (2021) unifies the TOD task as multiple generation tasks, including intent detection, DST, and response generation. However, they focus on the response generation ability and ignore the effects of different tasks. In this paper, we deep dive into analyzing the effects of five DU tasks.

8 Conclusion&Future Work

In this paper, we propose a unified generative dialogue understanding framework (UniDU) to share the knowledge across five typical dialogue understanding tasks. We introduce a model-agnostic adaptive weight learning method for multitask training to alleviate the biased generation problem. Our proposed UniDU method achieves better performance compared to well-designed models on a total of five DU tasks. We further deep dive into studying the affected factors. Finally, experimental results indicate that our proposed UniDU model can also get excellent performance under few-shot and zero-shot settings. In the future, we will increase the scale of the DU corpora and integrate the unsupervised dialogue pre-training tasks. We will further examine the task-level transferability of the UniDU model.

Acknowledgements

We sincerely thank the anonymous reviewers for their valuable comments. We thank the SIGDIAL mentors Stefan Ultes and Ondrej Dusek to help us prepare our final submission. This work has been supported by the China NSFC Projects (No.62120106006 and No. 62106142), Shanghai Municipal Science and Technology Major Project (2021SHZDZX0102), CCF-Tencent Open Fund and Startup Fund for Youngman Research at SJTU (SFYR at SJTU).

References

  • Adiwardana et al. (2020) Daniel Adiwardana, Minh-Thang Luong, David R So, Jamie Hall, Noah Fiedel, Romal Thoppilan, Zi Yang, Apoorv Kulshreshtha, Gaurav Nemade, Yifeng Lu, et al. 2020. Towards a human-like open-domain chatbot. arXiv preprint arXiv:2001.09977.
  • Bao et al. (2020) Siqi Bao, Huang He, Fan Wang, Hua Wu, and Haifeng Wang. 2020. Plato: Pre-trained dialogue generation model with discrete latent variable. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 85–96.
  • Bengio et al. (2009) Yoshua Bengio, Jérôme Louradour, Ronan Collobert, and Jason Weston. 2009. Curriculum learning. In Proceedings of the 26th annual international conference on machine learning, pages 41–48.
  • Budzianowski et al. (2018) Paweł Budzianowski, Tsung-Hsien Wen, Bo-Hsiang Tseng, Iñigo Casanueva, Stefan Ultes, Osman Ramadan, and Milica Gasic. 2018. Multiwoz-a large-scale multi-domain wizard-of-oz dataset for task-oriented dialogue modelling. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 5016–5026.
  • Casanueva et al. (2020) Inigo Casanueva, Tadas Temcinas, Daniela Gerz, Matthew Henderson, and Ivan Vulic. 2020. Efficient intent detection with dual sentence encoders. ACL 2020, page 38.
  • Chen et al. (2018a) Lu Chen, Cheng Chang, Zhi Chen, Bowen Tan, Milica Gašić, and Kai Yu. 2018a. Policy adaptation for deep reinforcement learning-based dialogue management. In 2018 IEEE international conference on acoustics, speech and signal processing (ICASSP), pages 6074–6078. IEEE.
  • Chen et al. (2019) Lu Chen, Zhi Chen, Bowen Tan, Sishan Long, Milica Gašić, and Kai Yu. 2019. Agentgraph: Toward universal dialogue management with structured deep reinforcement learning. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 27(9):1378–1391.
  • Chen et al. (2020a) Lu Chen, Boer Lv, Chi Wang, Su Zhu, Bowen Tan, and Kai Yu. 2020a. Schema-guided multi-domain dialogue state tracking with graph attention neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 7521–7528.
  • Chen et al. (2021a) Yulong Chen, Yang Liu, and Yue Zhang. 2021a. DialogSum challenge: Summarizing real-life scenario dialogues. In Proceedings of the 14th International Conference on Natural Language Generation, pages 308–313, Aberdeen, Scotland, UK. Association for Computational Linguistics.
  • Chen et al. (2018b) Zhao Chen, Vijay Badrinarayanan, Chen-Yu Lee, and Andrew Rabinovich. 2018b. Gradnorm: Gradient normalization for adaptive loss balancing in deep multitask networks. In International Conference on Machine Learning, pages 794–803. PMLR.
  • Chen et al. (2022) Zhi Chen, Jijia Bao, Lu Chen, Yuncong Liu, Da Ma, Bei Chen, Mengyue Wu, Su Zhu, Jian-Guang Lou, and Kai Yu. 2022. Dialogzoo: Large-scale dialog-oriented task learning. arXiv preprint arXiv:2205.12662.
  • Chen et al. (2021b) Zhi Chen, Lu Chen, Hanqi Li, Ruisheng Cao, Da Ma, Mengyue Wu, and Kai Yu. 2021b. Decoupled dialogue modeling and semantic parsing for multi-turn text-to-sql. In Findings of ACL 2021.
  • Chen et al. (2020b) Zhi Chen, Lu Chen, Xiaoyuan Liu, and Kai Yu. 2020b. Distributed structured actor-critic reinforcement learning for universal dialogue management. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 28:2400–2411.
  • Coope et al. (2020) Samuel Coope, Tyler Farghly, Daniela Gerz, Ivan Vulić, and Matthew Henderson. 2020. Span-convert: Few-shot span extraction for dialog with pretrained conversational representations. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 107–121.
  • Coucke et al. (2018) Alice Coucke, Alaa Saade, Adrien Ball, Théodore Bluche, Alexandre Caulier, David Leroy, Clément Doumouro, Thibault Gisselbrecht, Francesco Caltagirone, Thibaut Lavril, et al. 2018. Snips voice platform: an embedded spoken language understanding system for private-by-design voice interfaces. arXiv preprint arXiv:1805.10190.
  • Elgohary et al. (2019) Ahmed Elgohary, Denis Peskov, and Jordan Boyd-Graber. 2019. Can you unpack that? learning to rewrite questions-in-context. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 5918–5924.
  • Gliwa et al. (2019) Bogdan Gliwa, Iwona Mochol, Maciej Biesek, and Aleksander Wawer. 2019. Samsum corpus: A human-annotated dialogue dataset for abstractive summarization. EMNLP-IJCNLP 2019, page 70.
  • Haihong et al. (2019) E Haihong, Peiqing Niu, Zhongfu Chen, and Meina Song. 2019. A novel bi-directional interrelated model for joint intent detection and slot filling. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 5467–5471.
  • Ham et al. (2020) Donghoon Ham, Jeong-Gwan Lee, Youngsoo Jang, and Kee-Eung Kim. 2020. End-to-end neural pipeline for goal-oriented dialogue systems using gpt-2. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 583–592.
  • Han et al. (2020) Ting Han, Ximing Liu, Ryuichi Takanobu, Yixin Lian, Chongxuan Huang, Dazhen Wan, Wei Peng, and Minlie Huang. 2020. Multiwoz 2.3: A multi-domain task-oriented dialogue dataset enhanced with annotation corrections and co-reference annotation. arXiv preprint arXiv:2010.05594.
  • Hosseini-Asl et al. (2020) Ehsan Hosseini-Asl, Bryan McCann, Chien-Sheng Wu, Semih Yavuz, and Richard Socher. 2020. A simple language model for task-oriented dialogue. Advances in Neural Information Processing Systems, 33:20179–20191.
  • Kendall et al. (2018) Alex Kendall, Yarin Gal, and Roberto Cipolla. 2018. Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 7482–7491.
  • Kim et al. (2016) Joo-Kyung Kim, Gokhan Tur, Asli Celikyilmaz, Bin Cao, and Ye-Yi Wang. 2016. Intent detection using semantically enriched word embeddings. In 2016 IEEE Spoken Language Technology Workshop (SLT), pages 414–419. IEEE.
  • Kim et al. (2020) Sungdong Kim, Sohee Yang, Gyuwan Kim, and Sang-Woo Lee. 2020. Efficient dialogue state tracking by selectively overwriting memory. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 567–582.
  • Kumar et al. (2016) Ankit Kumar, Ozan Irsoy, Peter Ondruska, Mohit Iyyer, James Bradbury, Ishaan Gulrajani, Victor Zhong, Romain Paulus, and Richard Socher. 2016. Ask me anything: Dynamic memory networks for natural language processing. In International conference on machine learning, pages 1378–1387. PMLR.
  • Lewis et al. (2020) Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Veselin Stoyanov, and Luke Zettlemoyer. 2020. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 7871–7880.
  • Li et al. (2017) Xiujun Li, Yun-Nung Chen, Lihong Li, Jianfeng Gao, and Asli Celikyilmaz. 2017. End-to-end task-completion neural dialogue systems. In Proceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 733–743.
  • Liao et al. (2021) Lizi Liao, Le Hong Long, Yunshan Ma, Wenqiang Lei, and Tat-Seng Chua. 2021. Dialogue state tracking with incremental reasoning. Transactions of the Association for Computational Linguistics, 9:557–569.
  • Liu and Lane (2016) Bing Liu and Ian Lane. 2016. Attention-based recurrent neural network models for joint intent detection and slot filling. arXiv preprint arXiv:1609.01454.
  • Liu et al. (2019a) Chunyi Liu, Peng Wang, Jiang Xu, Zang Li, and Jieping Ye. 2019a. Automatic dialogue summary generation for customer service. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 1957–1965.
  • Liu et al. (2019b) Xiaodong Liu, Pengcheng He, Weizhu Chen, and Jianfeng Gao. 2019b. Multi-task deep neural networks for natural language understanding. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 4487–4496.
  • Liu et al. (2019c) Xingkun Liu, Arash Eshghi, Pawel Swietojanski, and Verena Rieser. 2019c. Benchmarking natural language understanding services for building conversational agents. In 10th International Workshop on Spoken Dialogue Systems Technology 2019.
  • Luong et al. (2016) Thang Luong, Quoc V. Le, Ilya Sutskever, Oriol Vinyals, and Lukasz Kaiser. 2016. Multi-task sequence to sequence learning. In International Conference on Learning Representations.
  • McCann et al. (2018) Bryan McCann, Nitish Shirish Keskar, Caiming Xiong, and Richard Socher. 2018. The natural language decathlon: Multitask learning as question answering. arXiv preprint arXiv:1806.08730.
  • Mehri et al. (2020) Shikib Mehri, Mihail Eric, and Dilek Hakkani-Tur. 2020. Dialoglue: A natural language understanding benchmark for task-oriented dialogue. arXiv preprint arXiv:2009.13570.
  • Peng et al. (2020) Baolin Peng, Chunyuan Li, Jinchao Li, Shahin Shayandeh, Lars Liden, and Jianfeng Gao. 2020. Soloist: Building task bots at scale with transfer learning and machine teaching. arXiv preprint arXiv:2005.05298.
  • Qin et al. (2019) Libo Qin, Wanxiang Che, Yangming Li, Haoyang Wen, and Ting Liu. 2019. A stack-propagation framework with token-level intent detection for spoken language understanding. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 2078–2087.
  • Qin et al. (2021a) Libo Qin, Tailu Liu, Wanxiang Che, Bingbing Kang, Sendong Zhao, and Ting Liu. 2021a. A co-interactive transformer for joint slot filling and intent detection. In ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 8193–8197. IEEE.
  • Qin et al. (2021b) Libo Qin, Tianbao Xie, Wanxiang Che, and Ting Liu. 2021b. A survey on spoken language understanding: Recent advances and new frontiers. arXiv preprint arXiv:2103.03095.
  • Quan et al. (2019) Jun Quan, Deyi Xiong, Bonnie Webber, and Changjian Hu. 2019. Gecor: An end-to-end generative ellipsis and co-reference resolution model for task-oriented dialogue. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 4547–4557.
  • Quan et al. (2020) Jun Quan, Shian Zhang, Qian Cao, Zizhong Li, and Deyi Xiong. 2020. Risawoz: A large-scale multi-domain wizard-of-oz dataset with rich semantic annotations for task-oriented dialogue modeling. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 930–940.
  • Raffel et al. (2020) Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J Liu. 2020. Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research, 21(140):1–67.
  • Rastogi et al. (2020) Abhinav Rastogi, Xiaoxue Zang, Srinivas Sunkara, Raghav Gupta, and Pranav Khaitan. 2020. Schema-guided dialogue state tracking task at dstc8. arXiv preprint arXiv:2002.01359.
  • Ruder et al. (2019) Sebastian Ruder, Matthew Peters, Swabha Swayamdipta, and Thomas Wolf. 2019. Transfer learning in natural language processing tutorial. NAACL HTL 2019, page 15.
  • Su et al. (2019) Hui Su, Xiaoyu Shen, Rongzhi Zhang, Fei Sun, Pengwei Hu, Cheng Niu, and Jie Zhou. 2019. Improving multi-turn dialogue modelling with utterance rewriter. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 22–31.
  • Su et al. (2021) Yixuan Su, Lei Shu, Elman Mansimov, Arshit Gupta, Deng Cai, Yi-An Lai, and Yi Zhang. 2021. Multi-task pre-training for plug-and-play task-oriented dialogue system. arXiv preprint arXiv:2109.14739.
  • Tian et al. (2021) Xin Tian, Liankai Huang, Yingzhan Lin, Siqi Bao, Huang He, Yunyi Yang, Hua Wu, Fan Wang, and Shuqi Sun. 2021. Amendable generation for dialogue state tracking. In Proceedings of the 3rd Workshop on Natural Language Processing for Conversational AI, pages 80–92.
  • Torrey and Shavlik (2010) Lisa Torrey and Jude Shavlik. 2010. Transfer learning. In Handbook of research on machine learning applications and trends: algorithms, methods, and techniques, pages 242–264. IGI global.
  • Ultes et al. (2017) Stefan Ultes, Lina M Rojas Barahona, Pei-Hao Su, David Vandyke, Dongho Kim, Inigo Casanueva, Paweł Budzianowski, Nikola Mrkšić, Tsung-Hsien Wen, Milica Gasic, et al. 2017. Pydial: A multi-domain statistical dialogue system toolkit. In Proceedings of ACL 2017, System Demonstrations, pages 73–78.
  • Wei et al. (2021) Jason Wei, Maarten Bosma, Vincent Y Zhao, Kelvin Guu, Adams Wei Yu, Brian Lester, Nan Du, Andrew M Dai, and Quoc V Le. 2021. Finetuned language models are zero-shot learners. arXiv preprint arXiv:2109.01652.
  • Wen et al. (2017) Tsung-Hsien Wen, David Vandyke, Nikola Mrkšić, Milica Gasic, Lina M Rojas Barahona, Pei-Hao Su, Stefan Ultes, and Steve Young. 2017. A network-based end-to-end trainable task-oriented dialogue system. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 1, Long Papers, pages 438–449.
  • Wolf et al. (2019) Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, Rémi Louf, Morgan Funtowicz, et al. 2019. Huggingface’s transformers: State-of-the-art natural language processing. arXiv preprint arXiv:1910.03771.
  • Wu et al. (2021) Chien-Sheng Wu, Linqing Liu, Wenhao Liu, Pontus Stenetorp, and Caiming Xiong. 2021. Controllable abstractive dialogue summarization with sketch supervision. arXiv preprint arXiv:2105.14064.
  • Xu et al. (2020) Zihan Xu, Zhi Chen, Lu Chen, Su Zhu, and Kai Yu. 2020. Memory attention neural network for multi-domain dialogue state tracking. In CCF International Conference on Natural Language Processing and Chinese Computing, pages 41–52. Springer.
  • Zang et al. (2020) Xiaoxue Zang, Abhinav Rastogi, Srinivas Sunkara, Raghav Gupta, Jianguo Zhang, and Jindong Chen. 2020. Multiwoz 2.2: A dialogue dataset with additional annotation corrections and state tracking baselines. ACL 2020, page 109.
  • Zhang et al. (2017) Yuhao Zhang, Victor Zhong, Danqi Chen, Gabor Angeli, and Christopher D Manning. 2017. Position-aware attention and supervised data improve slot filling. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pages 35–45.

Appendix

 

Appendix A Dialogue Understanding Corpora

Corpora #Sample I(Token) I(Turn) O(Token) Task
SAMSUM 14732 104.95 11.16 20.31 DS
DIALOGSUM 12460 140.48 9.49 22.86 DS
TASK 2205 34.92 2.75 10.84 DC
CANARD 31526 102.67 9.80 11.55 DC
BANKING77 12081 21.64 1 3.14 ID
HWU64 25715 17.69 1 2.05 ID
RESTAURANTS8K 15270 14.44 1 3.38 SF
SNIPS 35748 15.31 1 1.77 SF
WOZ2.0 7608 78.96 4.63 1.30 DST
MULTIWOZ2.2 35119 115.80 5.99 1.45 DST
Table 6: The ten DU corpora trained on UniDU model. I(Token) and I(Turn) mean the average length of the split tokens and the average turns of the input dialogue content. O(Token) means the average length of the split tokens of the task-specific output.

In this paper, we train our proposed unified generative model on ten dialogue understanding corpora, as shown in Table 6. For each DU tasks, we select two well-studied datasets. The first one is used to evaluate and the second one is an auxiliary corpus. The main reason to select two datasets for each task is to compare the multitask learning with the task transfer learning. We aim to know whether the knowledge sharing between different dialogue understanding data is only happening in the same DU task rather than all the DU tasks. The experimental results show that the annotated data from the other DU tasks are also important to enhance the performance, which indicates that it is an efficient way to transfer the knowledge among all the DU tasks. Note that the selected DU data are from different corpora, which means that the distribution of the input dialogue content is totally different. As shown in Table 6, the inputs and the outputs of the five DU tasks are greatly different from each other. The longest average input reaches to 140.48 and the shortest is only 14.44. The longest output is 22.86 from dialogue summary and the shortest is 1.30 from dialogue state tracking. These characters lead a big challenge to train all the dialogue understanding data in multitask learning way. The experimental results show that the intuitive mixture learning method makes UniDU model bias convergence to the more complex tasks like dialogue summary and dialogue completion. In this paper, we compare eight multitask training strategies. Our proposed MATS method can achieve the best overall performance on the five tasks under UniDU framework.

Refer to caption
Figure 5: Overview of model-agnostic training strategy.

Appendix B Model-Agnostic Training Strategy

In traditional HWU algorithm, the learnable weight WtW_{t} is only dependent on the corresponding task. Thus, we can regard the weight function of task Wϕ(t)W_{\phi}(t), where ϕ\phi are parameters shared among five tasks. Generally, the task is associated with two factors: its corresponding model and task format. Under UniDU framework, five tasks share the same encoder-decoder model, which can be regarded as a constant in weight function Wϕ(t)W_{\phi}(t). The task format dependents on model-agnostic task setting, such as input, output and data scale. To distinguish the five tasks under UniDU framework, we manually design a vector as the task feature to represent a task. Each dimension in the task feature has its physical meaning related to model-agnostic setting. In this paper, we design 14 dimensional vector 𝐟t\mathbf{f}_{t}, as shown in Figure 5. For input and output, we add the average length of token, the average sentence number, the n-grams and the perplexity (PPL) as the attributes of the DU tasks. Especially for input, the average turn number is also an important character. The last attribute is training scale for each task. Since the model-agnostic training strategy (MATS) formulates the weight as the task-related function and may share the function parameters among different tasks, the weights are not longer independent to each other as in original learnable weight method.

Refer to caption
Figure 6: Few-shot learning results on DST fine-tuned on BART and UniDU. 1%, 2% and 5% are the percents of the training data on unseen “Taxi” domain.

Appendix C Few-shot Learning

We select UniDU model that gets the best evaluation overall performance on five tasks learned with MATS method. For dialogue state tracking, we utilize the “Train” domain data in MULTIWOZ2.2, which is unseen in MTL training phase. Compared with vanilla BART, UniDU has obvious advantages, especially on extremely resource-limited situation. When there is only 1% and 2% training data, the vanilla BART is disable to learn. UniDU model warmed up by MATS method can quickly adapt the model on the unseen domain.