This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Improving Conversational Recommender System
via Contextual and Time-Aware Modeling with
Less Domain-Specific Knowledge

Lingzhi Wang, Shafiq Joty, Wei Gao, Xingshan Zeng, Kam-Fai Wong Lingzhi Wang and Kam-Fai Wong are with The Chinese University of Hong Kong, Hong Kong.
E-mail: [email protected], [email protected] Shafiq Joty is with Nanyang Technological University, Singapore.
E-mail: [email protected] Wei Gao is with Singapore Management University, Singapore.
E-mail: [email protected] Xingshan Zeng is with Huawei Noah’s Ark Lab, Hong Kong.
E-mail: [email protected] Manuscript received April 19, 2005; revised August 26, 2015.
Abstract

Conversational Recommender Systems (CRS) has become an emerging research topic seeking to perform recommendations through interactive conversations, which generally consist of generation and recommendation modules. Prior work on CRS tends to incorporate more external and domain-specific knowledge like item reviews to enhance performance. Despite the fact that the collection and annotation of the external domain-specific information needs much human effort and degenerates the generalizability, too much extra knowledge introduces more difficulty to balance among them. Therefore, we propose to fully discover and extract the internal knowledge from the context. We capture both entity-level and contextual-level representations to jointly model user preferences for the recommendation, where a time-aware attention is designed to emphasize the recently appeared items in entity-level representations. We further use the pre-trained BART to initialize the generation module to alleviate the data scarcity and enhance the context modeling. In addition to conducting experiments on a popular dataset (ReDial), we also include a multi-domain dataset (OpenDialKG) to show the effectiveness of our model. Experiments on both datasets show that our model achieves better performance on most evaluation metrics with less external knowledge and generalizes well to other domains. Additional analyses on the recommendation and generation tasks demonstrate the effectiveness of our model in different scenarios.

Index Terms:
Recommender System, Conversational Recommendation, Pre-trained Language Model.

1 Introduction

Conversational Recommender Systems (CRS) [1, 2, 3, 4] have recently attracted many researchers due to the booming of e-commerce platforms. A CRS aims to provide high-quality recommendations to users through conversations. Different from the traditional recommender systems, it focuses on learning users’ preferences through natural language interaction with users, and has a high impact in e-commerce.

An effective CRS is expected to be able to clarify user intents, learn user preferences, recommend high-quality items and reply to users with suitable responses. As we can see from Table I, the recommender (CRS) recommends items (movies, books, etc.) to the seeker (user) by using natural and fluent sentences. Therefore, most of the studies on the CRS field generally divide the system into two parts: a recommendation module to recommend items with top probabilities and a generation module to generate responses containing the recommended items.

For the recommendation module, previous works [3, 4] tend to include more and more external knowledge into the system, as most of the available CRS datasets are relatively small (due to the expensive annotation process) [1, 5] and hard to extract meaningful features based on the context alone. For example, to improve the performance in conversational movie recommendation, external knowledge like entity-level knowledge graph [2], word-level knowledge graph [3] and item reviews [4] is successively introduced into the system. However, there are three issues existing in the development of including more external knowledge. (i) Though the performance is improved by introducing more external knowledge, how to balance them in a single system becomes a new challenge. (ii) The collection and annotation of external knowledge need much human effort. (iii) Some of the collected external knowledge may be domain-specific and lack generalizability when facing broader application scenarios, e.g., the item reviews introduced by [4] (some recommended items like sports or unpopular items might lack reviews). On the other hand, the previous work largely ignores the timing or sequence order information by regarding the appeared items in context as a set rather than a sequence, which is contrary to the goal of the system. As one of the goals of CRS is to guide the users to express their preferences explicitly through conversations, we believe that the timing or sequence information would be a crucial factor in further improving the performance of a CRS.

Therefore, instead of exploring more domain-specific external knowledge to assist the learning of user preferences, we choose to fully discover and extract the internal knowledge from the context and propose a time-aware user preference modeling. Concretely, we capture both entity-level and contextual-level representations to model the user preferences. The entity-level representations summarize user preferences with the appeared items in context by our proposed time-aware modeling. As for the contextual-level representations extracted by a context encoder, they reflect the semantic- and discourse-level user preferences, which cannot be captured by entity-level ones. These two representations complement each other to enhance the recommendation results.

To elaborate the motivation of our recommendation module design, Table I shows two conversation examples from CRS datasets, where the upper conversation focuses on recommending movies and the lower one is about recommending books. It is observed that the context contains rich and key information. For example, “my kids don’t like Power Rangers (2017)” in the first example of Table I indicates the user shows negative opinion to Power Rangers (2017) and the CRS is supposed to not recommend items that are similar to the negative item. Only focusing on the appeared items (e.g., Power Rangers (2017)) like the previous work [1, 2, 3, 4] would misunderstand the users’ true preference. In addition to the potential misunderstanding of the user intention, there is also rich and direct information that is vital for a recommendation. For example, “Do you know any other books she wrote.” in the second example of Table I implies that the user needs a specific book recommendation from the aforementioned author Virginia Woolf. Such kind of context information provides precise user preference, which should not be ignored by a CRS. Therefore, in addition to the entity-level representation, it’s essential to consider the contextual-level user interest representation.

TABLE I: Two conversation examples from ReDial (the upper row) and OpenDialKG (the lower row) datasets. Seeker is a user who asks movie (or book) recommendation and Recommender is supposed to do chit-chat and recommendation. The mentioned items (i.e., movie and book) are in blue.
Seeker: Can you recommend some newer released family friendly movies?
Recommender: Yes you should watch Power Rangers (2017), Captain Underpants: The First Epic Movie was great too
Seeker: Well that is a good suggestion, but my kids don’t like Power Rangers (2017) much.
Recommender: How old are they?
Seeker: They are 14 years old and 10 years old.
Recommender: Despicable Me (2010) and Wonder Woman (2017) are great flicks
Seeker: Can you recommend a good book by Virginia Woolf?
Recommender: Sure! Have your read the To The Lighthouse?
Seeker: No I have not. When was it released?
Recommender: It came out in 1927.
Seeker: Do you know any other books she wrote?
Recommender: She wrote many. You may be interested in The Waves or Mrs. Dalloway.

As for the generation module, most of the previous methods [1, 2, 3, 4] employ a general encoder-decoder framework and train the model from scratch. It suffers from the overfitting issue on the relatively small dataset, as most of the CRS datasets only contain about 10k conversations and most of the syntactic structure of the utterances is simple and boring. Therefore, we use the pre-trained BART [6] model to initialize our generation module and thus alleviate the data scarcity effect in capturing meaningful context features.

Different from most previous works [1, 2, 3, 4, 7] that only test the effectiveness of their models on a single domain data, we conduct experiments on two public and popular CRS datasets ReDial [1] and OpenDialKG [5] under the multi-domain scenario. The results show that our model can achieve better recommendation and generation performance when assisted with less domain-specific external knowledge, and validate that our model generalizes well to other domains. Further analyses also demonstrate the effectiveness of our proposed modules.

The main contributions of this work can be summarized as follows:

  • We propose to combine both entity-level and contextual-level representations for modeling user preference in conversational recommendation, which achieves better performance with less domain-specific external knowledge compared to the previous works and generalizes well to other domains.

  • We point out the limitation of the previous entity modeling method and contribute a time-aware user preference modeling method to enhance the recommendation.

  • We adopt the pre-trained language model BART to enhance the semantic learning and diversity of our generated responses.

The remainder of this paper is organized as follows. The related work is surveyed in Section 2. Section 3 presents the proposed approach. And Section 4 and Section 5 present the experimental setup and the corresponding results respectively. Finally, conclusions are drawn in Section 6.

2 Related Work

In this section, we provide an in-depth review of the related research work from three different aspects, conversational recommender system, traditional recommender system and pre-trained languages model.

2.1 Conversational Recommender System

Conversational Recommender System (CRS) has attracted many researchers’ interest in recent years. Various task formulations with different hypotheses and application scenarios have been proposed. We summarize them into three categories and introduce them in detail as below.

2.1.1 Question Driven Systems

As the rating or click feedback in traditional recommender systems is limited in that they do not exactly tell why the users like or dislike an item, the feedback from the users could be very sparse. Therefore, question driven systems are proposed to effectively understand users’ preferences and improve the recommendations over time by asking clarifying questions. [8] proposes question-based video recommender system. [9] builds systems based on aspect-centered questions. [10] formulates the task of asking clarifying questions in open-domain information-seeking conversational systems. More recent works focus on asking attribute-central questions and developing reinforcement learning based approaches [11, 12, 13] or graph based approaches [14, 15, 16, 17].

2.1.2 Strategies Learning in Multi-turn CRS

Some works in this category focus on balancing the trade-off between exploration (i.e., asking questions) and exploitation (i.e., making recommendations), especially for cold-start users. They study the trade-off strategies to achieve engaging and successful recommendations. Some of them [18, 19, 20, 21] leverage bandit online recommendation methods and focus on cold-start scenarios, while others work on strategically asking clarification questions with fewer turns [11, 15, 22].

2.1.3 Open-ended CRS

An open-ended CRS aims to make recommendation in a more natural and casual way compared with the task-oriented CRS. Many datasets have been collected or built to push forward the research of CRS, including ReDial [1], TG-ReDial (Chinese) [23], GoRecDial [24], DuRecDial (Chinese) [25], INSPIRED [26] and OpenDialKG [5] datasets. Most of them consist of around 10,000 conversations that are focused on recommendation and chit-chat on different domains. For example, ReDial is about movie recommendation, while OpenDialKG is concerned with several domains, including movie, book, sports and music. The follow-up studies on the ReDial dataset generally divide the CRS into recommendation and generation modules. For the recommendation module, the previous works tend to apply more and more external knowledge to improve the recommendation performance, e.g., entity-level knowledge graph [2], word-level knowledge graph [3] and item reviews [4]. However, it’s challenging to manage so much external knowledge via an end-to-end model. What’s more, some (like item reviews) might need much effort to collect and annotate, and are not generally applicable for all kinds of domains (e.g., some items might lack reviews). For generation module, most of the previous works adopt encoder-decoder framework and train the generation model from scratch. However, it’s difficult to learn diverse and valuable patterns from relatively small datasets. Our work further explores approaches and improves quality of CRS for this category.

2.2 Recommender System

Traditional recommender system is different from the conversational recommender system, as it generally does not focus on dialogue interactions with users. It makes more efforts on the information filtering techniques to handle information overflow problems in recent years.

The problems in the traditional recommender system can be divided into two categories [27]: recommendation with explicit feedback (i.e., the ratings given to products) and recommendation with implicit feedback (purchase history or click-through history) [28]. The former problem can be evaluated directly by the differences between ground truths and predictions while the evaluation of the latter problem is usually formulated as Top-N ranking problems.

The methods proposed in the conventional recommender system can be divided into three categories: collaborative filtering based recommendation [29], content-based recommendation [30] and the hybrid of them [31]. Collaborative filtering has been the most popular recommendation technique in recent years. It hypothesizes that people who have had similar interests before would also prefer similarly in the future [32]. Many methods are proposed to identify the similarity of users or items, including memory-based methods [33], clustering methods [34], Bayesian models [35] and matrix factorization methods [36]. However, collaborative filtering methods may perform poorly when handling new items that lack history information. As for content-based recommendation methods, they usually first compare the similarity between the descriptive characteristics of items and the user profiles, and then recommend items that are similar to what users like before. Therefore, the content-based methods are usually efficient in recommending new items. But they are limited in terms of diversity and novelty [37]. The hybrid of collaborative filtering and content-based methods take advantages of both of them [31].

2.3 Pre-trained Language Model

The pre-trained language models [38, 39, 6, 40] become more and more popular in recent years. The idea behind them is first pre-training the models on large-scale unlabeled corpus (which can be easily crawled on the Internet) and then apply them to many downstream tasks (e.g., QA system, dialogue system, summarization and machine translation). The architecture of these pre-trained language models has been developed from shallow to deep (the parameter scale is advancing from million-level [38, 40] to trillion-level [41]) along with the improvement of computational capability and the enhancement of training mechanisms. Pre-training tasks such as nearby word prediction [42], next word prediction [43, 44], masked language model [38] are widely explored in language modeling. These tasks do not need any annotated labels, so they can be trained on a huge amount of unlabeled data from the Internet, e.g., Wiki and Reddit.

We introduce how these pre-trained language models are applied to the downstream tasks. The development of utilizing pre-trained language models can be divided into two stages, embedding-based approaches and finetuning-based approaches. In embedding-based approaches, the word-, sentence- or paragraph-level embeddings from the pre-trained language models is utilized as features in the downstream tasks. As for finetuning-based approaches, they benefit the knowledge transferring between the pre-trained language models and the downstream tasks. A common finetuning procedure would fix (or apply smaller learning rate to) the original parameters of the pre-trained language model and add some finetunable adaptation modules for the downstream tasks. Besides, prompt tuning is getting more attention in all of finetuning methods. Discrete [45, 46] or continuous [47] prompts based finetuning are proposed to help bridge the gap between the pre-training and finetuning and reduce the computational cost. Different from the domain-specific knowledge (e.g., review of items) introduced by the previous work [4, 48], the information learned by pre-trained language models is general and not task-specific. The finetuning methods based on pre-trained language model can be well generalized to other domains while the training scheme based on domain-specific knowledge limits the model generalizability.

3 Methodology

Refer to caption
Figure 1: Our framework for conversational recommendation. It consists of two modules, a Recommendation Module to rank the candidate items and a Generation Module to generate fluent responses to support interaction with users. The input of our framework can be two parts: a knowledge graph and context CC. The outputs are responses with recommended items.

In this section, we first formulate the task of CRS in Section 3.1, followed by the description of our generation module in Section 3.2, which is finetuned with a pre-trained sequence-to-sequence framework BART [6]. Then we introduce how we do recommendation based on conversation context and entity-level information in Section 3.3. Finally, we describe how we integrate the above two modules and produce the final responses in Section 3.4.

3.1 Problem Formulation

A classical architecture of conversational recommender system generally consists of two modules, named generation module and recommendation module. It takes the history context of a conversation C=(t1,t2,,tm)C=(t_{1},t_{2},\ldots,t_{m}) as input, where tit_{i} (1im1\leq i\leq m) is an utterance from the seeker (i.e. the user) or the CRS itself and mm is the number of context utterances. Each utterance tit_{i} contains word tokens ti=(wi,1,wi,2,,wi,ni)t_{i}=(w_{i,1},w_{i,2},\ldots,w_{i,n_{i}}), where nin_{i} is the number of tokens in utterance tit_{i}. The CRS uses its recommendation module to recommend items i\mathcal{I}_{i} from a candidate item set \mathcal{I} based on the context CC and some related external knowledge (e.g., knowledge graphs or item reviews). Then it embeds them into a response R=(y1,y2,,yn)R=(y_{1},y_{2},\ldots,y_{n}), a sequence of nn tokens generated by the generation module based on the context CC to produce natural responses with recommendations.

3.2 BART-based Response Generation

Our response generation module follows a general Transformer [49] sequence-to-sequence framework. We first briefly introduce the Transformer framework followed by the BART pre-trained language model. Then we explain how we finetune the BART model to enable generating responses with recommended items.

Transformer. A Transformer layer typically consists of two kinds of sub-layers: multi-head attention layers and fully connected feed-forward networks. In a multi-head attention layer, for each head jj, the output vector is computed with the following equation:

Attnj(𝑸,𝑲,𝑽)=softmax(𝑸j𝑲jTdk)𝑽j{\mathop{\mathrm{Attn}}}_{j}(\bm{Q},\bm{K},\bm{V})=\mathop{\mathrm{softmax}}(\frac{\bm{Q}_{j}\bm{K}_{j}^{T}}{\sqrt{d_{k}}})\bm{V}_{j} (1)

where 𝑸,𝑲,𝑽\bm{Q},\bm{K},\bm{V} are the matrices of queries, keys and values, dkd_{k} is the dimension of keys, and 𝑸j=𝑸𝑾jQ\bm{Q}_{j}=\bm{Q}\bm{W}_{j}^{Q}, 𝑲j=𝑲𝑾jK\bm{K}_{j}=\bm{K}\bm{W}_{j}^{K}, 𝑽j=𝑽𝑾jV\bm{V}_{j}=\bm{V}\bm{W}_{j}^{V} are project matrices for head jj. The output of each head is concatenated and then mapped to the final output field. For the fully connected feed-forward network (FFN{\mathop{\mathrm{FFN}}}), it consists of two linear transformations with an ReLU activation in between:

FFN(𝑿)=ReLU(𝑿𝑾1+𝒃1)𝑾2+𝒃2\mathop{\mathrm{FFN}}(\bm{X})=\mathop{\mathrm{ReLU}}(\bm{X}\bm{W}_{1}+\bm{b}_{1})\bm{W}_{2}+\bm{b}_{2} (2)

A Transformer encoder layer is composed of a multi-head self-attention layer (takes the same input as 𝑸,𝑲,𝑽\bm{Q},\bm{K},\bm{V}) and a feed-forward network, while a Transformer decoder layer additionally attends to the encoder output (as 𝑲\bm{K} and 𝑽\bm{V}) for cross attention (see the bottom part of Fig. 1).

In summary, the whole process for a general Transformer sequence-to-sequence framework is as follows:

𝑯C\displaystyle\bm{H}^{C} =Transformer_Encoder(x)\displaystyle=\mathop{\mathrm{Transformer\_Encoder}}(x) (3)
yk\displaystyle y_{k} =Transformer_Decoder(y<k,𝑯C)\displaystyle=\mathop{\mathrm{Transformer\_Decoder}}(y_{<k},\bm{H}^{C}) (4)
gen\displaystyle\mathcal{L}_{gen} =k=1nlog(p(yk|y<k,𝑯C))\displaystyle=\sum\nolimits_{k=1}^{n}-\log(p(y_{k}|y_{<k},\bm{H}^{C})) (5)

where xx is the input token sequence, and y<ky_{<k} represents the target tokens before yky_{k}. The Transformer_Encoder\mathop{\mathrm{Transformer\_Encoder}} includes a word embedding layer and several Transformer encoder layers while the Transformer_Decoder\mathop{\mathrm{Transformer\_Decoder}} is similar but additionally includes an output projection layer. The model is learned to minimize the negative log-likelihood of the target sequence as in Eq. 5.

BART. BART [6] is a pre-trained language model based on Transformer trained with several denoising objectives on large-scale books and Wikipedia data, i.e., recovering the noisy input processed by token masking, token deletion, text infilling, sentence permutation and document rotation. These denoising objectives enable the model to learn contextual representations with unlabeled data, which can be crawled easily from the Internet, and then used for multiple downstream tasks. BART has been shown to be effective in many generation tasks, including abstractive QA, summarization, machine translation at the sentence and document level, and persona-based response generation.

Finetune BART for CRS. As most of the available CRS datasets are relatively small and contain only around 10K conversations [1, 5], it is difficult to learn complex semantic and discourse level dependencies only based on the training corpus. To relieve the burden and enhance context modeling, we choose to finetune a pre-trained BART [6] model for our response generation module. Though the introduction of pre-trained BART model can be regarded as a way of including more external knowledge to the CRS, there are two major differences between the domain-specific external knowledge (i.e., review of items) and the general external knowledge (i.e., pre-trained language model): (i) Difficulties in applying the knowledge. The collection of domain-specific knowledge is time-consuming and how to design efficient networks to include the domain-specific knowledge with other knowledge becomes a new challenge, while the finetuning scheme based on the pre-trained language model is easier and can follow some regular solutions. (ii) Generalizability. The pre-trained language models can be employed in most of the tasks while the domain-specific knowledge may only exist in a few tasks or be hard to collect for some tasks.

To enable the BART model to generate the item-aware responses, we extend its original vocabulary 𝒱\mathcal{V} with the item set \mathcal{I} to be 𝒱=𝒱\mathcal{V}^{\prime}=\mathcal{V}\cup\mathcal{I} (in this case, we do not tokenize the item names in an utterance and recognize them as single tokens), and then use a CRS training corpus to finetune the model. Specifically, we concatenate the utterances tit_{i} in context CC with an appended EOT\langle EOT\rangle token in their chronological order as the input (i.e., x=[t1;EOT;t2;..;EOT;tm]x=[t_{1};\langle EOT\rangle;t_{2};..;\langle EOT\rangle;t_{m}]), and maximize the probability of the ground truth response RR (i.e., minimize the objective in  Eq. 5).

Finally, an integration operation is added to the generation output to make the generation aware of the recommendation. We will discuss it later in Section 3.4.

3.3 Context-Time-Aware Recommendation

To fully understand user preferences over items from a given context CC, we propose to extract two kinds of information for the recommendation. The first is entity-level information, where we extract the mentioned entities (including the items in the item set \mathcal{I}) from CC and apply them to an external relational knowledge graph to perform entity linking [50]. Then the learned entity representations are summarized by a proposed time-aware attention to produce entity-level user representation. The second is contextual information produced by BART model, which is expected to capture information from the perspectives of semantic and conversational discourse. We describe the details in the following.

3.3.1 Entity-level Representation

Entity-level representation is captured and represented by extracted items as well as their related entities (e.g., if the items are the movies, the related entities can be the authors, directors, etc.) from the observed context CC. Because of the data scarcity, it is not promising to summarize the extracted entities merely based on the training corpus [2]. We follow previous work and employ a relational knowledge graph (e.g. DBpedia) to enhance entity modeling.

Entity Linking. Specifically, we denote a triplet in the relational knowledge graph with e1,r,e2\langle e_{1},r,e_{2}\rangle, where e1,e2e_{1},e_{2}\in\mathcal{E} are entities from the entity set \mathcal{E} and rr is a kind of entity relation from the relation set \mathcal{R}. For the ReDial dataset, we utilize the linked knowledge graph, a subset of DBpedia, which is released by [2]. For the OpenDialKG dataset, we also extract a subset of DBpedia by matching each item in the item set to entities in DBpedia and additionally linking them on dialogue contents by following [50].

We use an R-GCN [51] framework to encode relation-aware entity representations. The intuition behind this is that the features of an entity can be indicated and summarized by its neighboring nodes in the knowledge graph. An R-GCN computes the embeddings of the entities with multi-edge encoding with several iteration layers. Formally, the representation of an entity ee at the (l+1)(l+1)-th layer is:

𝒉e(l+1)=ReLU(reer1Ze,r𝑾r(l)𝒉e(l))\bm{h}_{e}^{(l+1)}=\mathop{\mathrm{ReLU}}(\sum_{r\in\mathcal{R}^{\prime}}\sum_{e^{\prime}\in\mathcal{E}_{e}^{r}}\frac{1}{Z_{e,r}}\bm{W}_{r}^{(l)}\bm{h}_{e^{\prime}}^{(l)}) (6)

where 𝒉e(l)\bm{h}_{e}^{(l)} represents the representation of entity ee at the ll-th layer, and er\mathcal{E}_{e}^{r} denotes the set of neighboring nodes of ee under the relation rr. ={rself}\mathcal{R}^{\prime}=\mathcal{R}\cup\{r_{self}\} contains all the relations including self loop. 𝑾r(l)\bm{W}_{r}^{(l)} is a learnable relation-specific transformation matrix and Ze,rZ_{e,r} is a normalization factor. Therefore, for each node in the knowledge graph, it receives and aggregates the messages from its neighboring nodes after relation-specific transformation. Then it combines the information with the hidden representations to form its updated representation at the next layer. Finally, structural and relational information is encoded into the entity representation 𝒉eL\bm{h}_{e}^{L} for each ee\in\mathcal{E} at the last layer LL. For simplicity, we represent the representations in the final layer LL as 𝒉e\bm{h}_{e} by omitting the superscript “LL”.

User Preference Summarization. Given a context CC, we presume we can summarize the preference of the seeker with the appeared entities. We extract the appeared entities (as user preference) 𝒯u={e1,e2,,e|𝒯u|}\mathcal{T}_{u}=\{e_{1},e_{2},...,e_{|\mathcal{T}_{u}|}\} from two perspectives: item entities (i.e., entities that appear in item set \mathcal{I}) and other relevant contextual entities (mentioned in utterances but not an item in \mathcal{I}, e.g., an actor of a movie item. We denote them as text entities). For items that are not covered by \mathcal{E}, we ignore them by following the previous work [2, 3, 4]. The entities eie_{i}\in\mathcal{E} are sorted in the order of appearance. After looking up the entities in 𝒯u\mathcal{T}_{u} from 𝑯={𝒉e}e=1||\bm{H}=\{\bm{h}_{e}\}_{e=1}^{|\mathcal{E}|}, we get (𝒉1,𝒉2,,𝒉|𝒯u|)(\bm{h}_{1},\bm{h}_{2},...,\bm{h}_{|\mathcal{T}_{u}|}).

To summarize the entity-level user representation, previous work mainly depends on a self-attention mechanism [3, 4], where a learnable matrix is used to learn and derive each entity’s importance. Such a naive mechanism might be sub-optimal for the following two reasons: (i) no supervised signals are used to guide the model to learn knowledge about entity importance which makes it difficult to summarize accurately, especially when the training corpus is limited; (ii) it ignores the order of appearance which might affect the conversation trend (e.g. the appeared entities A,B,CA,B,C and C,B,AC,B,A probably lead to different next items in a conversation, but we will get the same summarized representation with such a mechanism). Therefore, we propose Time-aware Attention to address the limitations, where the entity-level user representation 𝒉E\bm{h}^{E} is calculated as follows:

𝒉E=i=1|𝒯u|λi1i=1|𝒯u|λi1𝒉i\bm{h}^{E}=\sum_{i=1}^{|\mathcal{T}_{u}|}\frac{\lambda^{i-1}}{\sum_{i=1}^{|\mathcal{T}_{u}|}\lambda^{i-1}}\bm{h}_{i} (7)

where λ\lambda is a hyper-parameter to control the recency effect. The value of λ\lambda is usually larger than 11, which means that the recently appeared items will contribute more to the next item prediction. The intuition behind the time-aware attention can be twofold: (i) The currently available conversational recommendation datasets are quite small and the previous proposed self-attention based item representation could not show promising performance (see discussion in Section 5.1.1). Our simple but effective modeling mechanism can handle multi-domain datasets regardless of their sizes. (ii) One of the goals of conversational recommender system is to guide the users to better express their preferences explicitly through the ongoing conversations. Therefore, it’s naturally to put more weights on the most recently appeared items. We will discuss the effect of the value of λ\lambda in Section 5.5.1.

Finally, the time-aware entity-level recommendation probability can be computed as follows:

𝒑e=softmax(mask(𝒉𝑬𝑯))\bm{p}_{e}=\mathop{\mathrm{softmax}}(\mathop{\mathrm{mask}}(\bm{h^{E}}\bm{H}^{\top})) (8)

where mask\mathop{\mathrm{mask}} is an operation that sets all non-item entities to -\infty (to make the recommendation focus on the candidate items in item set \mathcal{I}), and 𝒑e||\bm{p}_{e}\in\mathbb{R}^{|\mathcal{E}|}. 𝒑e\bm{p}_{e} will be used for recommendation together with contextual-level representation. We introduce the latter one in the next subsection.

3.3.2 Contextual-level Representation

Though the time-aware entity-level representation can model the user preference from the perspective of the mentioned items in context CC, it may lead to misunderstanding of user preferences since it ignores the content from the context. For example, if a user says “I do not like A!”. We cannot capture such a negative opinion towards “A” through entity-level representation alone.

To partially address the problem and to incorporate more semantic- and discourse-level context for recommendation, we further use the context representation 𝑯C\bm{H}^{C} computed in Eq. 3 from generation module to yield semantic-aware prediction. Specifically, we average the context representations over the sequence as contextual-level representation for CC and put it through an MLP layer to give the prediction:

𝒑c=softmax(MLP(1|C|j=1|C|𝒉jC))\bm{p}_{c}=\mathop{\mathrm{softmax}}(\mathop{\mathrm{MLP}}(\frac{1}{|C|}\sum_{j=1}^{|C|}\bm{h}_{j}^{C})) (9)

where 𝒉jC\bm{h}_{j}^{C} denotes the representation for the jj-th token in the context representation 𝑯C\bm{H}^{C}, |C||C| is the total length of context, and 𝒑c||\bm{p}_{c}\in\mathbb{R}^{|\mathcal{E}|}. 𝒑c\bm{p}_{c} reflects the effects from the content of context for recommendation, which can be a complement for entity-level representation. We next describe how we combine these two kinds of representations.

3.3.3 Combination of the Entity- and Contextual-Level Recommendation

Different from the previous work, our work considers both entity-level representation and contextual-level representation for recommendation. The final recommendation based on the above two components is defined as:

𝒑rec=μ𝒑e+(1μ)𝒑c\bm{p}_{rec}=\mu\cdot\bm{p}_{e}+(1-\mu)\cdot\bm{p}_{c} (10)

where μ\mu (0μ10\leq\mu\leq 1) is a hyper-parameter to balance between the two kinds of recommendations. Therefore, the learning objective for the recommendation module is:

rec=i=1Mlogpe(ri)+logpc(ri)\mathcal{L}_{rec}=-\sum_{i=1}^{M}\log p_{e}(r_{i})+\log p_{c}(r_{i}) (11)

where MM is the number of items extracted from the training corpus that need to be recommended, while rir_{i}\in\mathcal{I} is the target item in the ii-th recommendation. pe(ri)p_{e}(r_{i}) and pc(ri)p_{c}(r_{i}) are the corresponding prediction probabilities of the target item from the entity-level and contextual-level recommendation components, respectively.

3.4 Module Integration

So far, the responses generated by the generation module (introduced in Section 3.2) are not aware of the recommendation (introduced in Section 3.3) results. We introduce an integration mechanism to incorporate the recommendation module’s results and guide the generation module to generate responses that are consistent with the user’s preference. Some of the previous works [3, 4] use copy mechanism [52, 53] to incorporate recommendation results into the generated responses. However, as it needs additional networks to predict when to switch between generation mode and copy mode, it may sometimes perform wrongly due to insufficient data to learn the pattern.

Inspired by [2], we use a simpler but effective integration mechanism. We directly add a vocabulary bias to the top decoder predictions. Different from their work, our vocabulary bias directly comes from the recommendation probabilities 𝒑rec\bm{p}_{rec} in Eq. 10:

𝒃u=[𝟎;𝒢(𝒑rec)]\bm{b}_{u}=[\bm{0};\mathcal{G}(\bm{p}_{rec})] (12)

where 𝟎\bm{0} is a 𝒱\mathcal{V}-dimensional zero vector, 𝒢()\mathcal{G}(\cdot) is an index selection operation to select items from entities (𝒢:||||\mathcal{G}:\mathbb{R}^{|\mathcal{E}|}\rightarrow\mathbb{R}^{|\mathcal{I}|}), and [;][;] means concatenation. This makes the bias has the same dimension as our generation output (i.e. |𝒱|=|𝒱|+|||\mathcal{V}^{\prime}|=|\mathcal{V}|+|\mathcal{I}|).

We dynamically add the bias 𝒃u\bm{b}_{u} during generation based on the top predictions in each time stamp tt. Only when the top predictions of that time stamp include items in item set \mathcal{I} will we add the bias to them. The effect of this operation is that for a generation token yty_{t} that is probably an item in \mathcal{I}, we add its recommendation probability prec(yt)p_{rec}(y_{t}) to the original generation probability p(yt)p(y_{t}). In this way, the results of the two modules are integrated so that our system can generate the recommendation-aware responses.

The total objective for our model is:

=gen+γrec\mathcal{L}=\mathcal{L}_{gen}+\gamma\mathcal{L}_{rec} (13)

where γ\gamma is a hyper-parameter that balances the two objectives. We jointly train the two objectives so that the two modules can benefit each other.

4 Experimental Setup

TABLE II: Statistics of ReDial and OpenDialKG datasets.
ReDial OpenDialKG
Number of conversations 10,006 13,802
Number of utterances 182,150 91,209
Avg utterance number per conv 18.2 6.6
Knowledge Graph DBpedia DBpedia
Domains Movie Movie, Book, Sports, Music

In this section, we first show the basic statistics of the two used datasets and the evaluation metrics for recommendation and generation modules in Section 4.1. Then we describe the baseline models together with some variants to our main model in Section 4.2. Finally, we list the parameter settings of our model in Section 4.3.

4.1 Datasets and Evaluation

Datasets. To empirically evaluate the proposed approach, we conduct experiments on two popular CRS datasets, namely, ReDial [1] and OpenDialKG [5]. ReDial is centered around movie recommendation and is constructed on Amazon Mechanical Turk (AMT) platform with pair crowd-workers (i.e., Seeker and Recommender). OpenDialKG consists of conversations that are mainly in four domains: movie, book, sports and music. The construction process of OpenDialKG is described as follows, which can reveal why the recommendation on OpenDialKG is much easier than on ReDial: The first agent initiates a conversation by giving a seed entity. The second agent is given with a list of facts that are relevant to the seed entity, and asked to choose the most relevant facts and use them to frame a free-form conversational response. Table II presents basic statistics about the two datasets. As can be seen, the both datasets have around 10K conversations. We can notice that the average conversation length of ReDial dataset is larger than that of OpenDialKG. For the experiments, we split the two datasets into training, validation and test at ratios of 80%:10%:10% and 75%:15%:15% by following [1] and [5], respectively.

Evaluation Metrics. We evaluate the performance of recommendation and generation separately. For recommendation, we adopt Recall@K scores where K = 1, 10, 50 for ReDial by following [2], and K = 1, 3, 5, 10, 25 for OpenDialKG by following [5]. Recall@K indicates whether the predicted top-K items contain the ground truth recommendation items. For generation, apart from Dist-n (n=2, 3, 4) and perplexity (PPL) scores reported in previous work, we also report the case-insensitive BLEU-n (n=2, 4) scores111We use NLTK package (https://www.nltk.org) to calculate the BLEU scores.. For a fair comparison, we calculate the PPL scores via a widely used off-the-shelf package KenLM222https://kheafield.com/code/kenlm/, as the PPL scores may be very different when using different vocabulary.

4.2 Baselines and Variants of Our Model

For the ReDial dataset, we compare our model with six competitive baselines proposed by the previous work:
\bullet ReDial [1]: The model consists of a dialog generation module which is based on HRED [54], a recommendation module that is based on an auto-encoder [55] and a sentiment analysis module. Switching Mechanism is adopted to combine the recommendation and generation results.
\bullet KBRD [2]: It utilizes DBpedia knowledge graph to enhance the semantic information of items (or entities) to do the recommendation and adopts a transformer based generation module, where knowledge graph information serves as word bias to assist the generation.
\bullet CRWalker [56]: The method walks on the knowledge graph to form a reasoning tree at each turn for recommendation, and then maps to dialog acts to guide response generation. The learned recommendation-oriented dialog policy on the knowledge graph enhances the mutual benefit between recommendation and conversation.
\bullet KGSF [3]: It uses mutual information maximization (MIM) [57] to align the semantic spaces of word- and entity-level KGs for the recommendation module. Its generation module includes a transformer encoder followed by a fused knowledge-enhanced decoder.
\bullet RevCore [4]: It performs review-enriched and entity-based recommendation and uses a review-attentive encoder-decoder for generation. We use the re-implementation results in [48] for a fair comparison.
\bullet C2C^{2}-CRS [48]: This method extracts and represents the associated multi-grained semantic units from different data signals, and then align the corresponding semantic units from different data signals in a coarse-to-fine way.

For the OpenDialKG dataset, we compare the models that are compared in [5] and introduce them as follows.
\bullet seq2seq [58]: It applies a seq2seq approach for entity path generation, given all of the dialog contexts.
\bullet Tri-LSTM [59]: It encodes each utterance and all of its related facts within 1-hop from a KG to retrieve a response from a small (N=10) pre-defined sentence bank.
\bullet Ext-ED [60]: It generates response entity token at its final softmax layer, and does not utilize structural information from knowledge graph.
\bullet DialKG Walker [5]: It has an attention-based graph decoder that penalizes decoding of unnatural paths which prunes candidate entities and paths from a large search space.

We also test the performance of the following variants to our model as an ablation study.
\bullet BART (ContextM): It only uses contextual-level representation (described in Section 3.3.2) to do the recommendation. No entity-level representations and extra knowledge graph based knowledge are employed.
\bullet EntityM-SelfA: EntityM-SelfA refers to entity modeling with self-attention used in previous work. It only uses the entity-level representations to do the recommendation.
\bullet EntityM-TimeA: EntityM-TimeA refers to entity modeling with our designed time-aware modeling. It also only uses entity-level representations.
\bullet EntityM-TimeA-ContextM: Our full model that adopts both entity-level and contextual-level representation.

Besides, to examine the effectiveness of adding text entities to input (defined in Section 3.3.1) for recommendation, we also evaluate our model variants with and without text entities (i.e., TextEntity in table III) as input.

4.3 Parameter Setting

For the RGCN-based recommendation module, we set both the entity embedding size and the hidden representation dimension to 128. The layer number for R-GCN is 1 and the normalization factor Ze,rZ_{e,r} is set to 1 following [2]. For the BART-based generation module, we finetune on the pre-trained BART-Base model, which consists of 6 layers of encoder and decoder, respectively. The hidden dimension, feed-forward network size and attention head number are 768, 3072 and 12 for both encoder and decoder layers. The total parameter number of our model is 269M. The time recency effect λ\lambda in Eq. 7, the balance factor between entity- and contextual-level representations in Eq. 10 and the tradeoff between two training objectives in Eq. 13 are set to 1.5, 0.5 and 1.0, respectively. We adopt diverse beam search [61] mechanism in generation with a beam size of 4 and the diverse beam group number is set to 2. All the hyper-parameters are determined by grid-search based on the validation performance.

We implement our models based on FAIRSEQ framework333https://github.com/pytorch/fairseq [62], and train on an NVIDIA 3090 GPU. During training, we set the max tokens of each batch to 4096 with an update frequency of 4. We adopt Adam optimizer with a 5e-3 learning rate (and 5e-5 learning rate for BART-related modules as they are pre-trained parameters) and 1000 warm-up updates followed by a polynomial decay scheduler. The training time of one epoch is around 22 minutes. The model needs around 5 epochs to achieve the best performance on the validation set.

TABLE III: Recommendation results (in %) on ReDial dataset. Our full model achieves the best performance with less external domain-specific knowledge as input, compared to the previous work.
Models Input Rec@1 Rec@10 Rec@50
Context TextEntity DBPedia ConceptNet Review
Baselines
ReDial 2.4 14.0 32.0
KBRD 3.1 15.0 33.6
CRWalker 3.1 15.5 36.5
KGSF 3.9 18.3 37.8
RevCore 4.6 22.0 39.6
C2C^{2}-CRS 5.3 23.3 40.7
Our Models
BART(ContextM) 3.0 16.4 35.0
EntityM-TimeA 4.6 18.3 34.1
EntityM-TimeA-ContextM 5.7 22.6 40.4
EntityM-SelfA 3.3 16.3 32.6
EntityM-TimeA 5.2 18.2 34.6
EntityM-TimeA-ContextM 5.9 24.0 41.3

5 Experimental Results

In this section, we first report the main comparison results on recommendation and generation in Section 5.1 and Section 5.2, respectively. Then we analyze the effectiveness of our proposed modules in Section 5.3 followed by a case study in Section 5.4. Finally, we give more analysis in Section 5.5.

5.1 Recommendation Result Comparisons

In this subsection, we present the main comparison results on ReDial dataset in Section 5.1.1. Then, to further verify the effectiveness of our model in multi-domain dataset, we conduct experiments on OpenDialKG dataset and report the comparison results in Section 5.1.2.

5.1.1 Results on ReDial

Table III shows the recommendation results of the baselines and the variants of our model on ReDial Dataset. We can draw the following observations from the results:

\bullet Previous work tends to add more external knowledge, which can improve the recommendation performance, but ignores fully extracting the internal information. We can see that among the baselines, KBRD and CRWalker add the entity-level knowledge (EntityKG, i.e., DBPedia and the text entities extracted from it) as input. KGSF further introduces word-level knowledge (WordKG, i.e., ConceptNet), while RevCore and C2C^{2}-CRS further incorporate item reviews (Review). All the added external knowledge introduces considerable improvement, demonstrating the effectiveness of the external knowledge. However, we can also notice that our EntityM-TimeA (Entity Modeling with Time-aware Attention) model only with the input of “Context” and “DBPedia” can also get a 4.6%4.6\% Recall@1 score which is comparable with the results of RevCove, who utilizes two more external knowledge (Text Entity and Review). This shows that the previous designed methods cannot fully extract the internal knowledge from context, but our proposed model can capture more useful information for recommendation.

\bullet Time-aware attention can better summarize user preference than self-attention mechanism. Our models with time-aware attention perform significantly better than the model with self-attention. For example, EntityM-TimeA achieves 1.9%1.9\% higher Recall@1 than EntityM-SelfA with the same input (Context + TextEntity + DBPedia). This validates our intuition that the self-attention mechanism which lacks explicit guidance lead to sub-optimal recommendations and the recently appeared items are more important in reflecting user preference. It also shows the effectiveness of our designed time-aware attention. More analysis about it can be found in Section 5.5.1.

\bullet BART-based contextual-level representations are helpful. We are the first to finetune a pre-trained BART model and utilize its representations for recommendation. As we can see in Table III, the simplest BART(ContextM) model with only context as input achieves 35.0%35.0\% Recall@50 while the KBRD model that incorporates DBPedia and text entities gets 33.6%33.6\% Recall@50. We can also find that our models with time-aware attention show good improvements in all metrics after being enhanced with BART representations (i.e., with ContextM). Both indicate that contextual-level representations extracted by the finetuned BART models can reflect user preferences that can not be captured by entity-level representations, and so complement them well. Fig. 2(a) in Section 5.3.1 shows more detailed analysis.

\bullet Text entities are effective in capturing most relative items. Our model EntityM-TimeA with additional text entities (i.e., TextEntity) as input can achieve better Recall@1 compared with the same model without text entities (5.2%5.2\% vs 4.6%4.6\%), while keeping similar Recall@10 and Recall@50. This means that text entities help re-rank the top predictions and find the most relative items for recommendation.

5.1.2 Results on OpenDialKG

Apart from the ReDial that focuses on movie recommendation, we also examine our recommendation performance in a multi-domain dataset, OpenDialKG, to show the generalizability of our model. The results are displayed in Table IV. Two observations can also be drawn as follows (We skip some of the same observations as Redial here).

TABLE IV: Recommendation results (in %) on OpenDialKG.
Models Rec@1 Rec@3 Rec@5 Rec@10 Rec@25
Baselines
seq2seq 3.1 18.3 29.7 44.1 60.2
Tri-LSTM 3.2 14.2 22.6 36.3 56.2
Ext-ED 1.9 5.8 9.0 13.3 19.0
DialKG Walker 13.2 26.1 35.3 47.9 62.2
Our Models
BART(ContextM) 5.8 19.7 31.0 45.5 57.8
EntityM-SelfA 10.9 21.2 30.3 41.6 53.2
EntityM-TimeA 16.0 28.9 34.3 45.1 57.9
EntityM-TimeA-ContextM 18.0 33.5 41.5 50.0 64.8

\bullet The results on OpenDialKG are much higher than those on ReDial. As can be seen, our model achieves 18%18\% Recall@1 score on OpenDialKG, which is much higher than 5.9%5.9\% on ReDial. The reasons can be twofold. First, OpenDialKG is constructed based on the knowledge graph by some walking strategies [5], which means that the item prediction would be easier when incorporating entity-level knowledge. Second, we can see that the conversations on OpenDialKG are much shorter than those on ReDial (see Table II). This indicates that it takes less turns leading to the final recommendation needed by the seekers and so the modeling of the appeared items on OpenDialKG may be simpler.

\bullet Our model generalizes well to multiple domains. As we can see from Table IV, our model with time-aware attention and BART-enhanced representations (i.e., EntityM-TimeA-ContextM) achieves the best performance compared with all the other methods. We can observe similar trends among the different variants as those on ReDial, e.g., time-aware attention is better than self-attention and BART representations help improve all of the metrics. This validates that our method generalizes well to other domains.

5.2 Generation Result Comparisons

In this subsection, we first discuss the automatic evaluation results in Section 5.2.1 and then adopt a human evaluation to examine the generation results from a different perspective in Section 5.2.2.

TABLE V: Generation results (in %) on the ReDial dataset. “BS” refers to beam search, “DBS” refers to diverse beam search.
Models Dist-2 Dist-3 Dist-4 BLEU-2 BLEU-4 Perplexity\downarrow
Transformer 14.8 15.1 13.7 - - -
ReDial 22.5 23.6 22.8 17.8 7.4 61.7
KBRD 26.3 36.8 42.3 18.5 7.4 58.8
KGSF 28.9 43.4 51.9 16.4 7.4 131.1
RevCore 42.4 55.8 61.2 - - -
Ours+BS 35.8 49.9 57.7 19.1 9.3 52.1
   - w/o BART Pre-train 8.5 10.9 12.3 18.6 8.7 30.7
Ours+DBS 45.7 65.3 76.1 19.1 8.9 54.8
   - w/o BART Pre-train 13.9 19.8 23.8 18.6 8.2 43.9

5.2.1 Automatic Evaluation

We show the automatic evaluation comparison results on ReDial in Table V. To investigate the performance in different scenarios, we display the results of our models with conventional beam search (BS) and diverse beam search (DBS), respectively, together with their results without BART pre-training. Our model yields significantly better results on all of the evaluation metrics (including Dist-n, BLEU-n and perplexity) than the baselines, which indicates that our model tends to produce more diverse, coherent and fluent responses. In the following, we detail our observations:

\bullet Our model is able to generate more diverse, coherent and fluent responses than the baselines. As can be seen, our model with beam search achieves the best BLEU scores and perplexity, and our model with diverse beam search yields the highest Dist-n while maintaining comparable BLEU scores and perplexity. Comparing our models with and without BART pre-training, the Dist-n metrics are affected severely, which validates the effectiveness of BART pre-training in enhancing the diversity of generated responses.

\bullet It is challenging to balance between diversity and fluency. The baselines tend to perform differently in terms of Dist-n and perplexity, e.g., KGSF achieves higher Dist-n than KBRD, but its perplexity is worse. We presume that higher diversity requires the models to extract more diverse patterns to express their content, but organizing them into a fluent response may be challenging. Another example is our model without BART pre-training achieves poor diversity (i.e., very low Dist-n scores), as it may overfit in the training corpus and tend to generate simple responses. But this also results in its lowest perplexity. Our model with diverse beam search achieves consistently better Dist-n scores, and maintains relatively lower perplexity, demonstrating the superiority of our model design.

\bullet The Dist-n scores highly depend on the searching strategy. Our model performs much better in terms of Dist-n when applying diverse beam search rather than conventional beam search. What is more, different configurations (e.g., length penalty in Section 5.5.2) applied during response generation also affects the scores. This shows that it is not enough to evaluate the generation performance of these models only based on the metric, so we also include BLEU scores here and human evaluation (Section 5.2.2) for better evaluation.

5.2.2 Human Evaluation

TABLE VI: Human evaluation of the generation results on the ReDial dataset. “BS” refers to beam search, “DBS” refers to diverse beam search. All the metrics are in the scale of [0, 2]. The overall Cohen’s kappa coefficient is larger than 0.65.
Models Fluency Informativeness Coherence
HUMAN 1.95 1.71 1.71
ReDial 1.92 1.32 1.23
KBRD 1.95 1.39 1.31
KGSF 1.91 1.02 0.95
Ours+BS 1.95 1.54 1.66
Ours+DBS 1.92 1.64 1.64

To conduct our human evaluation, we randomly sampled 100 context-response pairs from the test set and collected the corresponding generation results of our models as well as the baselines. We then employ two crowd-workers to score the results on the scale of [0, 1, 2], where higher scores indicate better quality. Following [63], we evaluate the following three aspects of the results:

  • Fluency: whether a response is in a proper English grammar and easy to understand.

  • Informativeness: whether a response contains meaningful information. The “safe responses” are treated as uninformative as they may be repetitive and meaningless.

  • Coherence: whether a response is coherent with its previous context, i.e., the discussed content should be consistent.

We display the human evaluation results in Table VI. As can be seen, all the models generate responses with high fluency, but perform differently regarding informativeness and coherence. The baselines are more likely to produce safe responses (short and repetitive), resulting in lower informativeness. Also, their coherence scores are lower since they might produce some meaningless responses no matter what the previous contexts are. Nevertheless, our model can generate more informative and coherent responses.

5.3 Effectiveness of the Proposed Mechanisms

In this subsection, we first test the recommendation results with varying mentioned items in Section 5.3.1 to show the effectiveness of the introduced contextual-level representation. Then we discuss the effects of BART pre-training for our model in Section 5.3.2.

5.3.1 Effectiveness of Contextual-level Representation

Refer to caption
(a) Recall@50 (in %) over Number of Mentioned Items
Refer to caption
(b) Number of cases over Number of Mentioned Items
Figure 2: Recall@50 of the models and number of cases over the number of mentioned items in context.
TABLE VII: Recommendation results (in %) of our models when using BART pre-training (PT) or not. The values in parentheses correspond to the percentage increase after adding PT.
Models Inputs Pre-training R@1 R@10 R@50
BART (ContextM) Context 3.0 (15.4%) 16.4 (13.1%) 35.0 (10.4%)
2.6 14.5 31.7
EntityM-TimeA-ContextM Context+DBPedia 5.7 (9.6%) 22.6 (5.6%) 40.4 (1.3%)
5.2 21.4 39.9
EntityM-TimeA-ContextM Context+TextEntity+DBPedia 5.9 (7.3%) 24.0 (9.1%) 41.3 (2.5%)
5.5 22.0 40.3

To investigate how contextual-level representations influence recommendation, we show the Recall@50 scores with varying mentioned items numbers in context in Fig. 2(a) for our three model variants, ContextM, EntityM-TimeA and EntityM-TimeA-ContextM. The mentioned item number means how many items (e.g., movie, books) appear in the history utterances. Specifically, we divide the test samples into several groups based on the mentioned item number and calculate Recall@50 score for each group.

We can observe from Fig. 2(a) that the ContextM model performs better than EntityM-TimeA when the mentioned number is 0 or 11. Such a phenomenon is desired since when the item history information is rare or even missing, the entity-level representations are not sufficient to produce reliable recommendations while the contextual-level representations can capture useful information from the text. Then we find EntityM-TimeA performs consistently better than ContextM when the mentioned number increases. Because the increasing mentioned item number also means the context becomes longer, and the model might not be able to handle the long context, especially when the mentioned item number is larger than 1010. This means that contextual-level representations are useful in the cold-start scenario (not a rare situation as shown in Fig. 2(b)), which is a shortcoming for entity-level recommendation. Therefore, our full model (i.e., EntityM-TimeA-ContextM) combines the both representations and yields the best performance in all the situations.

5.3.2 Effects of BART Pre-training

We elaborate the effects of BART pre-training (PT) here. For recommendation, we list the recall scores of our different model variants with and without BART PT in Table VII. As can be seen, all the models perform worse when removing pre-training, as pre-training on large-scale monolingual datasets helps the models learn better semantic features. When joint recommending with entity-level representations (i.e., EntityM-TimeA-ContextM model), the performance degradation becomes less, demonstrating the two kinds of information can complement each other.

For generation, we have listed the ablation without pre-training in Table V. As we can see, models without pre-training show poor performance on the Dist-n metrics as models tend to be overfitting and cannot generate diverse responses based on a small dataset.

5.4 Case Study

TABLE VIII: A conversation example on movie item recommendation. The responses of ReDial, KBRD and KGSF are from the test results of the corresponding models. Human responses are the ground-truths from the corpus.
Seeker: I have seen The Conjuring (2013) I very much enjoyed this movie. When I was younger I liked movies like Friday the 13th (1980).
Human: I liked that one too!
ReDial: i have n’t seen that one . i ’ve heard good things about it .
KBRD: i have n’t seen that one . i ’ve heard Halloween (1978) was good .
KGSF: How about The Descent (2005) ?
OUR: i haven’t seen that one, but i’ve heard good things about it.
Seeker: I also liked to watch the A Nightmare on Elm Street (1984) Series.
Human: I loved that as well mainly because Johnny Depp is in it
ReDial: i have n’t seen that one yet . i ’ve seen It (2017)
KBRD: i have n’t seen that one . i have heard Halloween (1978) was good .
KGSF: I also recommend Insidious (2011)
OUR: I liked that one too! i also recommend The Last Exorcism (2010).
Seeker: I love Johnny Depp! So handsome, such a great actor. I love johnny Depp in Alice in Wonderland (2010).
Human: Me too one of my favorites from him is The Astronaut’s Wife (1999)
ReDial: i have n’t seen that one yet . i ’ve seen It (2017) .
KBRD: i have n’t seen that one . i will check it out .
KGSF: I hope you enjoy them !
OUR: Yes, I love Johnny Depp too ! Have you seen the new It (2017) movie?
Refer to caption
(a) Effects of Time-Aware Parameter λ\lambda
Refer to caption
(b) Effects of Length Penalty
Refer to caption
(c) Effects of the Trade-Off Parameter μ\mu
Figure 3: Effects of three hyper-parameters: parameter λ\lambda to control the time-aware user preference modeling, length penalty to influence the generation, and parameter μ\mu to balance the effects of entity-level and contextual-level user representations.

To give an intuitive assessment of the quality of the generated sentences and recommended items, we present a conversation example in Table VIII. Apart from the results generated by our model and the human ground-truths from the corpus, we list the results of ReDial, KBRD and KGSF models. The reason why we do not list the results of RevCore and C2C^{2}-CRS is their released codes are not complete to re-implement the generation results.

From Table VIII, we can easily find that our model tends to generate responses with more diversity and coherence to the context, which may thank to the pre-trained BART model. For example, in the last turn, our model also talks about “Johnny Depp” which is mentioned in the turn of the seeker. This means that our model can capture the topic of discussion well while the baselines tend to repeat the boring responses with the same syntactic structure (e.g., “I haven’t seen that one….”). On the other hand, the baseline models like ReDial and KBRD tend to recommend the same items regardless of the history context, while our model can produce different recommendations based on different contexts thanks to our improved recommendation mechanism.

5.5 Further Analysis

In this subsection, we extensively explore the effects of some hyper-parameters, together with in-depth discussions.

5.5.1 Influence of recency effect in Time-Aware Model

We present the Recall@1 and Recall@10 (Recall@50 shows a similar trend with Recall@10 so we omit here) scores with varying values of λ\lambda in Fig. 3(a), which controls the effect of recency introduced in our proposed Time-Aware Attention (see Eq. 7). λ<1\lambda<1 indicates the earlier appeared items are more important, which results in quick performance drops. And the performance increases when λ>1\lambda>1 compared to λ=1\lambda=1. This validates the intuition that more recently appeared items contribute more to the next item recommendation. On the other hand, Recall@1 and Recall@10 present different trends when λ\lambda increases – Recall@1 consistently increases while Recall@10 begins to drop when λ>2\lambda>2. This is because when λ\lambda is too large, time-aware attention tends to concern with the most recently appeared item. This is helpful in finding the most relative items but hurts the overall recommendation.

5.5.2 Limitation of Dist-n Metrics.

As we find that search strategies seriously affect the Dist-n metrics (see Section 5.2.1), we present more analysis on them by exploring the length penalty (a hyper-parameter that can control the lengths of final generated results) when generating the responses. We display the results of Dist-2 and BLEU-2 for our EntityM-TimeA-ContextM model with different length penalty in Fig. 3(b) (other metrics are in similar trends). We can find that the generated lengths also affect a lot the Dist-n scores since longer responses allow more different tokens to be generated. However, this is not expected as the longer is not necessarily the better. Therefore, other metrics, including human evaluation, are desired to explicitly evaluate generation performance.

5.5.3 Trade-off between Entity-level and Contextual-level Representations.

We examine the effects of the trade-off hyper-parameter μ\mu in Eq. 10 by setting its value from 0 (only entity-level representations) to 1 (only contextual-level representations) and display the results of our EntityM-TimeA-ContextM model in Fig. 3(c). As can be seen, Recall@50 score is significantly improved when the value of μ\mu changes from 0 to 0.1 (or 1 to 0.9). This validates that the two representations can capture user preferences from different perspectives and complement each other for better recommendation performance. Besides, the trends of Recall@1 and Recall@10 are similar to that of Recall@50. The best results are achieved around μ=0.5\mu=0.5, showing that both levels of representation are important.

6 Conclusion and Future Work

In this work, we focus on conversational recommender systems and propose to capture both entity-level and contextual-level representations to improve the recommendation performance, where a time-aware user preference modeling is designed to better capture the user’s interests and a pre-trained BART model is used to enhance contextual-level user preference modeling and improve the diversity of the generated responses. Experiments on two publicly available datasets show that the proposed model can achieve better performance with less external domain-specific knowledge and generalizes well to other domains. Further analyses also examine the effectiveness of our model in different scenarios.

In future work, we would like to explore more accurate user preference modeling mechanisms. Though our current time-aware modeling method shows promising improvement on both datasets, it is based on a simple assumption and cannot deal with more complicated real-world scenarios. A combined mechanism that concerns time effects and other features like sentiment is desired. However, it is hard to train a powerful user preference learning framework with the currently available datasets, which are human-created and relatively small. A large-scale, diverse, and multi-domain dataset might be needed in the future.

References

  • [1] R. Li, S. Kahou, H. Schulz, V. Michalski, L. Charlin, and C. Pal, “Towards deep conversational recommendations,” arXiv preprint arXiv:1812.07617, 2018.
  • [2] Q. Chen, J. Lin, Y. Zhang, M. Ding, Y. Cen, H. Yang, and J. Tang, “Towards knowledge-based recommender dialog system,” in Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP).   Hong Kong, China: Association for Computational Linguistics, Nov. 2019, pp. 1803–1813. [Online]. Available: https://aclanthology.org/D19-1189
  • [3] K. Zhou, W. X. Zhao, S. Bian, Y. Zhou, J.-R. Wen, and J. Yu, “Improving conversational recommender systems via knowledge graph based semantic fusion,” in Proceedings of the 26th ACM SIGKDD, 2020, pp. 1006–1014.
  • [4] Y. Lu, J. Bao, Y. Song, Z. Ma, S. Cui, Y. Wu, and X. He, “Revcore: Review-augmented conversational recommendation,” arXiv preprint arXiv:2106.00957, 2021.
  • [5] S. Moon, P. Shah, A. Kumar, and R. Subba, “Opendialkg: Explainable conversational reasoning with attention-based walks over knowledge graphs,” in Proceedings of the 57th ACL, 2019, pp. 845–854.
  • [6] M. Lewis, Y. Liu, N. Goyal, M. Ghazvininejad, A. Mohamed, O. Levy, V. Stoyanov, and L. Zettlemoyer, “BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension,” in Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.   Online: Association for Computational Linguistics, Jul. 2020, pp. 7871–7880. [Online]. Available: https://aclanthology.org/2020.acl-main.703
  • [7] L. Wang, H. Hu, L. Sha, C. Xu, K.-F. Wong, and D. Jiang, “Finetuning large-scale pre-trained language models for conversational recommendation with knowledge graph,” arXiv preprint arXiv:2110.07477, 2021.
  • [8] K. Christakopoulou, A. Beutel, R. Li, S. Jain, and E. H. Chi, “Q&r: A two-stage approach toward interactive recommendation,” in Proceedings of the 24th ACM SIGKDD, 2018, pp. 139–148.
  • [9] Y. Zhang, X. Chen, Q. Ai, L. Yang, and W. B. Croft, “Towards conversational search and recommendation: System ask, user respond,” in Proceedings of the 27th acm international conference on information and knowledge management, 2018, pp. 177–186.
  • [10] M. Aliannejadi, H. Zamani, F. Crestani, and W. B. Croft, “Asking clarifying questions in open-domain information-seeking conversations,” in Proceedings of the 42nd international acm sigir conference, 2019, pp. 475–484.
  • [11] W. Lei, X. He, Y. Miao, Q. Wu, R. Hong, M.-Y. Kan, and T.-S. Chua, “Estimation-action-reflection: Towards deep interaction between conversational and recommender systems,” in Proceedings of the 13th International Conference on Web Search and Data Mining, 2020, pp. 304–312.
  • [12] X. Ren, H. Yin, T. Chen, H. Wang, N. Q. V. Hung, Z. Huang, and X. Zhang, “Crsal: Conversational recommender systems with adversarial learning,” ACM Transactions on Information Systems (TOIS), vol. 38, no. 4, pp. 1–40, 2020.
  • [13] Y. Deng, Y. Li, F. Sun, B. Ding, and W. Lam, “Unified conversational recommendation policy learning via graph-based reinforcement learning,” arXiv preprint arXiv:2105.09710, 2021.
  • [14] H. Xu, S. Moon, H. Liu, B. Liu, P. Shah, B. Liu, and P. Yu, “User memory reasoning for conversational recommendation,” in Proceedings of the 28th International Conference on Computational Linguistics.   Barcelona, Spain (Online): International Committee on Computational Linguistics, Dec. 2020, pp. 5288–5308. [Online]. Available: https://aclanthology.org/2020.coling-main.463
  • [15] W. Lei, G. Zhang, X. He, Y. Miao, X. Wang, L. Chen, and T. Chua, “Interactive path reasoning on graph for conversational recommendation,” in KDD ’20: The 26th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Virtual Event, CA, USA, August 23-27, 2020, R. Gupta, Y. Liu, J. Tang, and B. A. Prakash, Eds.   ACM, 2020, pp. 2073–2083.
  • [16] X. Ren, H. Yin, T. Chen, H. Wang, Z. Huang, and K. Zheng, “Learning to ask appropriate questions in conversational recommendation,” arXiv preprint arXiv:2105.04774, 2021.
  • [17] K. Xu, J. Yang, J. Xu, S. Gao, J. Guo, and J.-R. Wen, “Adapting user preference to online feedback in multi-round conversational recommendation,” in Proceedings of the 14th ACM International Conference on Web Search and Data Mining, 2021, pp. 364–372.
  • [18] L. Li, W. Chu, J. Langford, and R. E. Schapire, “A contextual-bandit approach to personalized news article recommendation,” in Proceedings of the 19th WWW, 2010, pp. 661–670.
  • [19] S. Li, A. Karatzoglou, and C. Gentile, “Collaborative filtering bandits,” in Proceedings of the 39th International ACM SIGIR, 2016, pp. 539–548.
  • [20] K. Christakopoulou, F. Radlinski, and K. Hofmann, “Towards conversational recommender systems,” in Proceedings of the 22nd ACM SIGKDD, 2016, pp. 815–824.
  • [21] S. Li, W. Lei, Q. Wu, X. He, P. Jiang, and T.-S. Chua, “Seamlessly unifying attributes and items: Conversational recommendation for cold-start users,” arXiv preprint arXiv:2005.12979, 2020.
  • [22] Y. Sun and Y. Zhang, “Conversational recommender system,” in The 41st international acm sigir conference on research & development in information retrieval, 2018, pp. 235–244.
  • [23] K. Zhou, Y. Zhou, W. X. Zhao, X. Wang, and J.-R. Wen, “Towards topic-guided conversational recommender system,” arXiv preprint arXiv:2010.04125, 2020.
  • [24] D. Kang, A. Balakrishnan, P. Shah, P. Crook, Y.-L. Boureau, and J. Weston, “Recommendation as a communication game: Self-supervised bot-play for goal-oriented dialogue,” arXiv preprint arXiv:1909.03922, 2019.
  • [25] Z. Liu, H. Wang, Z.-Y. Niu, H. Wu, W. Che, and T. Liu, “Towards conversational recommendation over multi-type dialogs,” arXiv preprint arXiv:2005.03954, 2020.
  • [26] S. A. Hayati, D. Kang, Q. Zhu, W. Shi, and Z. Yu, “Inspired: Toward sociable recommendation dialog systems,” arXiv preprint arXiv:2009.14306, 2020.
  • [27] Z. Batmaz, A. Yurekli, A. Bilge, and C. Kaleli, “A review on deep learning for recommender systems: challenges and remedies,” Artificial Intelligence Review, vol. 52, no. 1, pp. 1–37, 2019.
  • [28] Y. Hu, Y. Koren, and C. Volinsky, “Collaborative filtering for implicit feedback datasets,” in 2008 Eighth IEEE international conference on data mining.   Ieee, 2008, pp. 263–272.
  • [29] J. B. Schafer, D. Frankowski, J. Herlocker, and S. Sen, “Collaborative filtering recommender systems,” in The adaptive web.   Springer, 2007, pp. 291–324.
  • [30] R. Van Meteren and M. Van Someren, “Using content-based filtering for recommendation,” in Proceedings of the machine learning in the new information age: MLnet/ECML2000 workshop, vol. 30, 2000, pp. 47–56.
  • [31] T. Tran and R. Cohen, “Hybrid recommender systems for electronic commerce,” in Proc. Knowledge-Based Electronic Markets, Papers from the AAAI Workshop, Technical Report WS-00-04, AAAI Press, vol. 40, 2000.
  • [32] B. Sarwar, G. Karypis, J. Konstan, and J. Riedl, “Item-based collaborative filtering recommendation algorithms,” in Proceedings of the 10th international conference on World Wide Web, 2001, pp. 285–295.
  • [33] J. S. Breese, D. Heckerman, and C. Kadie, “Empirical analysis of predictive algorithms for collaborative filtering,” arXiv preprint arXiv:1301.7363, 2013.
  • [34] L. H. Ungar and D. P. Foster, “Clustering methods for collaborative filtering,” in AAAI workshop on recommendation systems, vol. 1.   Menlo Park, CA, 1998, pp. 114–129.
  • [35] X. Su and T. M. Khoshgoftaar, “Collaborative filtering for multi-class data using belief nets algorithms,” in 2006 18th IEEE international conference on Tools with Artificial Intelligence (ICTAI’06).   IEEE, 2006, pp. 497–504.
  • [36] Y. Koren, R. Bell, and C. Volinsky, “Matrix factorization techniques for recommender systems,” Computer, vol. 42, no. 8, pp. 30–37, 2009.
  • [37] P. Lops, M. d. Gemmis, and G. Semeraro, “Content-based recommender systems: State of the art and trends,” Recommender systems handbook, pp. 73–105, 2011.
  • [38] J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “BERT: Pre-training of deep bidirectional transformers for language understanding,” in Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers).   Minneapolis, Minnesota: Association for Computational Linguistics, Jun. 2019, pp. 4171–4186. [Online]. Available: https://aclanthology.org/N19-1423
  • [39] Y. Liu, M. Ott, N. Goyal, J. Du, M. Joshi, D. Chen, O. Levy, M. Lewis, L. Zettlemoyer, and V. Stoyanov, “Roberta: A robustly optimized bert pretraining approach,” arXiv preprint arXiv:1907.11692, 2019.
  • [40] T. Brown, B. Mann, N. Ryder, M. Subbiah, J. D. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam, G. Sastry, A. Askell et al., “Language models are few-shot learners,” Advances in neural information processing systems, vol. 33, pp. 1877–1901, 2020.
  • [41] W. Fedus, B. Zoph, and N. Shazeer, “Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity,” arXiv preprint arXiv:2101.03961, 2021.
  • [42] T. Mikolov, I. Sutskever, K. Chen, G. S. Corrado, and J. Dean, “Distributed representations of words and phrases and their compositionality,” in Advances in neural information processing systems, 2013, pp. 3111–3119.
  • [43] Y. Bengio, R. Ducharme, P. Vincent, and C. Jauvin, “A neural probabilistic language model,” Journal of machine learning research, vol. 3, no. Feb, pp. 1137–1155, 2003.
  • [44] X. Wang and A. Gupta, “Unsupervised learning of visual representations using videos,” in Proceedings of the IEEE international conference on computer vision, 2015, pp. 2794–2802.
  • [45] F. Petroni, T. Rocktäschel, S. Riedel, P. Lewis, A. Bakhtin, Y. Wu, and A. Miller, “Language models as knowledge bases?” in Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP).   Hong Kong, China: Association for Computational Linguistics, Nov. 2019, pp. 2463–2473. [Online]. Available: https://aclanthology.org/D19-1250
  • [46] T. Gao, A. Fisch, and D. Chen, “Making pre-trained language models better few-shot learners,” arXiv preprint arXiv:2012.15723, 2020.
  • [47] X. Liu, Y. Zheng, Z. Du, M. Ding, Y. Qian, Z. Yang, and J. Tang, “Gpt understands, too,” arXiv preprint arXiv:2103.10385, 2021.
  • [48] Y. Zhou, K. Zhou, W. X. Zhao, C. Wang, P. Jiang, and H. Hu, “C2-crs: Coarse-to-fine contrastive learning for conversational recommender system,” arXiv preprint arXiv:2201.02732, 2022.
  • [49] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin, “Attention is all you need,” in Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, I. Guyon, U. von Luxburg, S. Bengio, H. M. Wallach, R. Fergus, S. V. N. Vishwanathan, and R. Garnett, Eds., 2017, pp. 5998–6008.
  • [50] J. Daiber, M. Jakob, C. Hokamp, and P. N. Mendes, “Improving efficiency and accuracy in multilingual entity extraction,” in Proceedings of the 9th international conference on semantic systems, 2013, pp. 121–124.
  • [51] M. Schlichtkrull, T. N. Kipf, P. Bloem, R. Van Den Berg, I. Titov, and M. Welling, “Modeling relational data with graph convolutional networks,” in European semantic web conference.   Springer, 2018, pp. 593–607.
  • [52] J. Gu, Z. Lu, H. Li, and V. O. Li, “Incorporating copying mechanism in sequence-to-sequence learning,” in Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers).   Berlin, Germany: Association for Computational Linguistics, Aug. 2016, pp. 1631–1640. [Online]. Available: https://aclanthology.org/P16-1154
  • [53] C. Gulcehre, S. Ahn, R. Nallapati, B. Zhou, and Y. Bengio, “Pointing the unknown words,” in Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers).   Berlin, Germany: Association for Computational Linguistics, Aug. 2016, pp. 140–149. [Online]. Available: https://aclanthology.org/P16-1014
  • [54] A. Sordoni, Y. Bengio, H. Vahabi, C. Lioma, J. Grue Simonsen, and J.-Y. Nie, “A hierarchical recurrent encoder-decoder for generative context-aware query suggestion,” in proceedings of the 24th ACM international on conference on information and knowledge management, 2015, pp. 553–562.
  • [55] J. He, H. H. Zhuo, and J. Law, “Distributed-representation based hybrid recommender system with short item descriptions,” arXiv preprint arXiv:1703.04854, 2017.
  • [56] W. Ma, R. Takanobu, M. Tu, and M. Huang, “Bridging the gap between conversational reasoning and interactive recommendation,” arXiv preprint arXiv:2010.10333, 2020.
  • [57] P. Viola and W. M. Wells III, “Alignment by maximization of mutual information,” International journal of computer vision, vol. 24, no. 2, pp. 137–154, 1997.
  • [58] I. Sutskever, O. Vinyals, and Q. V. Le, “Sequence to sequence learning with neural networks,” Advances in neural information processing systems, vol. 27, 2014.
  • [59] T. Young, E. Cambria, I. Chaturvedi, M. Huang, H. Zhou, and S. Biswas, “Augmenting end-to-end dialog systems with commonsense knowledge (2017),” arXiv preprint arXiv:1709.05453.
  • [60] P. Parthasarathi and J. Pineau, “Extending neural generative conversational model using external knowledge sources,” arXiv preprint arXiv:1809.05524, 2018.
  • [61] A. K. Vijayakumar, M. Cogswell, R. R. Selvaraju, Q. Sun, S. Lee, D. Crandall, and D. Batra, “Diverse beam search: Decoding diverse solutions from neural sequence models,” arXiv preprint arXiv:1610.02424, 2016.
  • [62] M. Ott, S. Edunov, A. Baevski, A. Fan, S. Gross, N. Ng, D. Grangier, and M. Auli, “fairseq: A fast, extensible toolkit for sequence modeling,” arXiv preprint arXiv:1904.01038, 2019.
  • [63] S. Bao, H. He, F. Wang, H. Wu, H. Wang, W. Wu, Z. Guo, Z. Liu, and X. Xu, “Plato-2: Towards building an open-domain chatbot via curriculum learning,” arXiv preprint arXiv:2006.16779, 2020.
[Uncaptioned image] Lingzhi Wang is currently pursuing the Ph.D. degree with the Department of Systems Engineering and Engineering Management, The Chinese University of Hong Kong, Hong Kong. She received her bachelor degree from Harbin Institute of Technology in 2019. Her current research interests include recommender system, dialogue system and social media analysis. She has served as program committee member of NLP and AI top conferences like ACL and EMNLP.
[Uncaptioned image] Shafiq Joty is currently an Associate Professor at the Nanyang Technological University, Singapore and a Senior Research Manager at Salesforce Research. He holds a PhD in Computer Science from the University of British Columbia. His work has primarily focused on developing language analysis tools and applications including question answering, text summarization and dialog systems. A significant part of his current research focuses on multilingual, multimodal NLP, and robustness of NLP models. He is a PC co-chair for SIGDIAL’23 and previously served as a senior AC for ACL’22 and EMNLP’21, and AC for ACL’19-21 EMNLP’19 and NAACL’21. He is an action editor for ACL-RR and was an associate editor for ACM TALLIP. His research contributed to 16 patents and more than 110 papers in top-tier NLP and ML conferences and journals.
[Uncaptioned image] Wei Gao is currently an Assistant Professor at Singapore Management University. Previously, he was a senior lecturer at Victoria University of Wellington, a Scientist at Qatar Computing Research Institute, and a Research Assistant Professor in the Chinese University of Hong Kong. His research interests intersect natural language processing, information retrieval, artificial intelligence and social computing. He broadly serves in the program committees of top conferences (e.g., ACL, EMNLP, SIGIR, IJCAI) and reviews for the leading journals (e.g., ACM TOIS, IEEE TKDE) in his relevant field. He is currently an Associate Editor of ACM TALLIP and a member of standing review committees of TACL and CL.
[Uncaptioned image] Xingshan Zeng is currently a researcher in Huawei Noah’s Ark Lab. His research interests include natural language processing, social media analysis, recommender systems and speech translation. He got his PhD from the Department of Systems Engineering and Engineering Management, The Chinese University of Hong Kong in 2020. He has served as program committee member of NLP and AI top conferences like ACL, EMNLP, AAAI, IJCAI etc.
[Uncaptioned image] Kam-Fai Wong is currently a Professor in the Department of Systems Engineering and Engineering Management, The Chinese University of Hong Kong. He obtained his PhD from Edinburgh University in 1987. His research interest include Chinese computing, database and information retrieval. He is a fellow of ACL, and Senior Member of IEEE. He is the founding Editor-In-Chief of TALIP, and serves as associate editor of International Journal on CL. He is the Chair of Conference: General Chair of AACL2020, Co-Chair of NDBC2016, BigComp2016, NLPCC2015 and IJCNLP2011; the Finance Chair SIGMOD2007; and the PC Co-chair of lJCNLP2006.