User-specific Adaptive Fine-tuning for Cross-domain Recommendations ††thanks: Lei Chen and Fajie Yuan contribute equally.
Abstract
Making accurate recommendations for cold-start users has been a longstanding and critical challenge for recommender systems (RS). Cross-domain recommendations (CDR) offer a solution to tackle such a cold-start problem when there is no sufficient data for the users who have rarely used the system. An effective approach in CDR is to leverage the knowledge (e.g., user representations) learned from a related but different domain and transfer it to the target domain. Fine-tuning works as an effective transfer learning technique for this objective, which adapts the parameters of a pre-trained model from the source domain to the target domain. However, current methods are mainly based on the global fine-tuning strategy: the decision of which layers of the pre-trained model to freeze or fine-tune is taken for all users in the target domain. In this paper, we argue that users in RS are personalized and should have their own fine-tuning policies for better preference transfer learning. As such, we propose a novel User-specific Adaptive Fine-tuning method (UAF), selecting which layers of the pre-trained network to fine-tune, on a per user basis. Specifically, we devise a policy network with three alternative strategies to automatically decide which layers to be fine-tuned and which layers to have their parameters frozen for each user. Extensive experiments show that the proposed UAF exhibits significantly better and more robust performance for user cold-start recommendation.
Index Terms:
Cross-domain recommendations, Transfer learning, User-specific adaptive fine-tuning1 Introduction
The past decade has seen a remarkable progress in deep learning (DL) and their applications in recommender systems (RS). A variety of neural network models [1, 2, 3, 4, 5] with larger and deeper architectures are proposed to model user interaction behaviors from online systems. Among them, sequential recommendation models, such as the GRU4Rec [1], NextItNet [2] and SASRec [4] have become especially popular since they in general require neither much feature engineering nor explicit user embeddings when making recommendations. Despite the success, deep neural network models tend to fail in practice when their training data (i.e., user interactions) are insufficient. Such scenarios widely exist in practical RS, when a large number of new users enroll in but have fewer interactions. Fortunately, the interaction behaviors of cold-start users are likely to be accessible from other online systems. For example, a user in Amazon who has few purchase records might have hundreds of clicking interactions in YouTube. Such observed interaction feedback by YouTube could be a clue to infer her preference and make recommendations in Amazon.111In this paper, we assume data of both domains is available and put aside privacy concerns. To this end, cross-domain recommendations (CDR) that transfer knowledge from a related source domain, have been proposed and become a popular way to tackle the recommendation problem of cold users.
Transfer learning (TL) based on pre-training and fine-tuning [6, 7, 8, 9] is widely employed for domain adaptation tasks. Its basic idea is to first pre-train a large neural network model with plenty of source data and then fine-tune it in the target domain where there might be insufficient training data. Fine-tuning has become a core technique for successful TL and been widely studied in computer vision [10] and NLP [8]; however, thus far, it attracted very little attention in the recommendation field. In this paper, we aim to investigate how to perform fine-tuning for cross-domain recommendations and encourage practitioners to apply our proposed fine-tuning as a common technique to further improve the performance of their CDR models.

There are several challenges when realizing the idea of fine-tuning of deep neural networks for CDR. (i) Great strides in accuracy have often been accompanied by increasing the depth and complexity of neural networks. Fine-tuning such large models is prone to overfitting when the target domain has insufficient training data, and when all parameters are optimized. (ii) Existing fine-tuning methods do not differentiate users — i.e., treating the pre-trained parameters of each user equally in the target domain. We argue that these methods restrict the power of the pre-trained representations, because some users in the target domain may have high similarity with the source domain, s.t. routing these users through pre-trained parameters might be a better choice compared to updating them with limited target data. Ideally, for each individual user, we expect the TL model is able to decide whether to use the pre-trained parameters or to fine-tune them, as illustrated in Figure 1. For example, user 1 requires the first and last layers to be fine-tuned, while the the two intermediate layers are kept the same as their pre-trained model. On the other hand, user 3 requires more pre-trained parameters than user 1 and user 2 for the optimal recommendation accuracy in the target domain.
Motivated by the above-mentioned issues, in this paper we propose UAF, a user-specific adaptive fine-tuning method which tailors fine-tuning for effective CDR. To begin with, we first train an DL-based sequential recommendation model on a large source dataset and use it as our pre-trained model. This practice has been well verified by recent studies in [11, 12] (as well as in the NLP domains [13, 14, 15]) since sequential neural network can be trained in a self-supervised manner and thus could generate more universal user representations for effective transfer learning. In this paper, we choose the temporal CNN model NextItNet [2] as the pre-trained backbone model given its efficient network structure [16, 17], superior performance [18, 19, 2] and widespread usage in modeling sequential recommendation data [20, 11, 19, 12, 21, 22] in recent literature. Then, we fine-tune the pre-trained model on the target dataset using UAF to address the cold user problem. To be specific, for each user, we learn a policy network that makes binary decisions on whether to update or freeze the parameters of each layer in the pre-trained NextItNet.
The main contributions of this paper are as follows:
-
•
We propose a User-specific Adaptive Fine-tuning method for CDR on cold-start users. UAF allows each user in the target domain to have their own fine-tuning policy. To the best of our knowledge, UAF is the first ‘personalized’ fine-tuning method designed for recommender systems.
-
•
We propose a hard method, a soft method, and a RL method to derive the optimal fine-tuning policy without suffering from the non-differentiable problem of the discrete policy decision functions.
-
•
Experimental results on four cross-domain datasets show that the proposed method achieves impressive results, especially when the training data in the target domain are insufficient.
2 Related Work
We recapitulate important work for cross-domain recommendations. Since our pre-trained model is based on sequential recommendation model, we also briefly review some representative work based on deep learning (DL).
2.1 Sequential Recommender Systems (SRS)
In general, DL-based sequential recommendation (SR) models can be classified into three categories, namely RNN-based [1, 23], CNN-based [24, 2, 19] and self-attention [4, 25] based methods. Specifically, Hidasi et al. [1] proposed the first DL-based SR model GRU4Rec by adapting RNN from language model in the NLP domain. Following this work, many extended RNN variants were proposed, which either optimized a new ranking loss [26], incorporated more context features [27], or developed more advanced data augmentation [23]. While effective, these models rely heavily on the hidden states of the entire past, which cannot take full advantage of the parallel processing resources (e.g., GPU and TPU) [2] during training. Therefore, CNN and self-attention based models are proposed to mitigate such limitations, becoming more popular in recent literature [24, 2, 4, 25]. Among them, Tang et al. [24] proposed Caser, which embeds a sequence of user-item interactions into an “image” and learn sequential patterns as local features of the image by using wide convolutional filters. Subsequently, [2] proposed NextItNet, a deep temporal CNN-based recommendation model which particularly excels at modeling long-range item sequences. Later, GRec [19] improved NextItNet by considering two directional context information. CpRec [20] largely compressed NextItNet without hurting its recommendation accuracy. StackRec [18] and SkipRec [22] showed that NextItNet-style SR models could be stacked up to 100 layers for achieving its optimal accuracy, which distinguished from exiting work that usually applied less than 10 layers for evaluation. They also presented several methodologies to accelerate the training and inference processes of NextItNet. More recently, PeterRec [11] and Conure [12] demonstrated that the learned user representations by NextItNet were generic and could be transferred to solve various downstream recommendation problems. Meanwhile, self-attention based models, such as SASRec [4] and BERT4Rec [25], also showed competitive performance for the SRS task. However, compared with NextItNet with linear complexity, these models often require quadratic time complexity to compute the self-attention matrix. Thereby, in this paper, we plan to use the well-known temporal CNN model (i.e., NextItNet) as the backbone network, given that the performance of which has been well evaluated by PeterRec [11] and Conure [12] for the CDR tasks [28].
2.2 Cross-domain Recommendations (CDR)
While deep learning (DL) based models have shown impressive performance in providing personalized recommendations, they often suffer from the cold-start problem, when new users enroll in a system but have a few or no labeled training data [11]. To deal with the cold-user problem, cross-domain recommendations (CDR) were proposed and proved effective. CDR transfer user preference between domains based on similarity of users and items that occur in both domains [29]. CDR are particularly useful for recommendations of users who are cold in a target domain but have rich interactions from a source domain. According to existing literature, CDR methods can be broadly divided into two types: joint learning based CDR, referred to as JLCDR, and two-stage (pre-training + fine-tuning) transfer learning based CDR, referred to as TLCDR.
In terms of JLCDR, [30] proposed a DL-based domain adaptation approach, domain separation network [31], that solved the recommendation of cold-start users by jointly learning a stacked denoising autoencoder. Similarly, [32] explored an autoencoder and adversarial training to extract abstract rating patterns that were shared for same users across domains. CoNet [33] is another representative cross-domain recommendation model using deep neural networks as the base model. To enable effective knowledge transfer, CoNet introduced cross connections from the source network to the target and jointly trained objective functions of them.
Despite its effectiveness, JLCDR approaches are not efficient during training since the joint learning (e.g., multi-task learning) scheme is usually very costly. In particular, to optimize the objective of the source domain is computationally very expensive as training data of it is often at a large-scale [8]. By contrast, TLCDR does not need to train the learner from scratch for target task. Besides, compared with JLCDR, TLCDR can obtain better prediction accuracy [12]. This is because the joint learning scheme of JLCDR that trades off many objective functions usually does not guarantee the optimal performance for the target objective function [11].
However, existing literature of TLCDR concentrated mainly on pre-training, i.e., developing a more expressive or efficient base model [34, 35, 8]. Closely related to this paper, a recent work called PeterRec [11] firstly evidenced that the pre-training model based on self-supervised learning on user sequential behaviors largely improved the fine-tuning accuracy for the target task (i.e., CDR in our case). Inspired by meta learning [36], PeterRec injected a small-scale adapter network into the pre-trained model and optimized only the adapter during fine-tuning. The authors showed that PeterRec performed significantly better than fine-tuning the entire model when training data in the target domain is scarce. In this paper, we argue that both PeterRec and the typical fine-tuning strategies achieve only sub-optimal accuracy since they perform fine-tuning globally for all users, ignoring the individual difference of users in target tasks. To this end, we present an adaptive fine-tuning neural network architecture on a per-user basis. In other words, our proposed fine-tuning method is personalized — each user is assigned to a different fine-tuning policy. To the best of our knowledge, UAF is the first attempt to the CDR task using a personalized fine-tuning strategy.
This paper can be regarded as an extension of our previous conference paper published in AAAI 2021 [22]. In [22], we propose a user-adaptive layer selection framework by learning to skip inactive hidden layers on a per-user basis to address the inference problem for traditional sequential recommender systems with super deep layers. In this paper, we target at cross-domain recommendations, which is a completely different research topic. We propose a user-specific adaptive fine-tuning method which tailors fine-tuning for effective cross-domain recommendations. In addition, we add a new UAF-Soft method by employing the weighted combination of freezed pre-tained layers and fine-tuned layers, which can preserve pre-trained information and receive valid fine-tuned information simultaneously.

3 Methodologies
3.1 Problem Definition
Suppose that there are two domains: a source domain with a large number of user interactions and a target domain with limited user interactions (i.e.. the cold-user scenario). The input data takes the form of implicit feedback such as click logs and watching records. Let be the set of users shared in both domains. Each instance in contains a userID , and the user interaction sequence , where denotes the -th interaction of and (of size ) is the set of items in source domain . Similarly, each instance in consists of a userID and item (label) , i.e., , where (of size ) is the set of items in target domain .
Our objective is to recommend a set of items (a.k.a, the top- item recommendation) in the target domain to the user who has limited interactions but possesses sufficient sequential behaviors in the source domain.
3.2 The Overall Architecture
Figure 2 illustrates the overall architecture of the proposed UAF method. The training procedure of UAF consists of two stages: pre-training and fine-tuning. In the first stage, we pre-train the base network (NextItNet) on the source dataset with numerous user-item interactions. The pre-trained model is supposed to learn general user representations which are reusable for same users in the target dataset. Then, we propose a user-specific adaptive fine-tuning strategy to automatically select reusable layers from the pre-trained model on a per-user basis.
3.2.1 Pre-training Model
As mentioned before, we choose NextItNet as the pre-trained model to learn the user representations given its superior performance in recent literature [18, 2]. NextItNet is composed of a stack of dilated convolutional layers, every two of which are wrapped by a residual block structure. To be specific, each input item is converted into an embedding vector , and the item sequence can be represented as an embedding matrix . Then, we pass the item embeddings into a stack of dilated convolutional (DC) layers to learn the feature vector which is expected to capture sequential dependencies. Formally, the -th residual block with the DC operation is formalized as:
(1) |
where and are the input and output of the -th residual block considered, is the item embedding size. is a shortcut connection by element-wise addition. represents the residual mappings as:
(2) |
where and represent the casual convolution operations. and represent layer normalization functions. is ReLU activation function. Finally, a softmax output layer is applied to predict the probability distribution for the next item :
(3) |
where is the projection matrix, and is the bias term.
The joint probability of each user sequence is computed by the product of the conditional distributions over the interacted items as follows:
(4) |
where is the predicted probability for the -th item conditioned on all its previous interactions , and is the set of parameters.
The objective function of pre-trained NextItNet is to minimize the sum of negative log-likelihood of the joint probability, which is defined as:
(5) | ||||
3.2.2 User-specific Adaptive Fine-tuning
After the pre-training process, we employ UAF to identify which layers in the pre-trained model are required to be fine-tuned or freezed. For the -th residual block in the pre-trained NextItNet model, in order to decide whether or not to fine-tune it during training, we first freeze the original block and then create a new trainable block , which is initialized with the parameters of . With the additional block , the output of the -th residual block in UAF is computed as:
(6) |
where is a binary policy variable that indicates whether the residual block should be frozen or fine-tuned, conditioned on the input user sequence . During training, given an input user sequence, the frozen block trained on the source domain dataset is left unchanged, and the replicated block can be optimized towards the target domain dataset. Therefore, the input user sequence can either share the frozen block or fine-tune the residual block . is sampled from a discrete distribution with two categories (freeze or fine-tune), which is parameterized by the output of a lightweight policy network. To be more specific, if , then the -th frozen residual block is reused; otherwise, the -th replicated residual block is fine-tuned.
By performing a series of convolution operations on frozen and fine-tuned layers with the input user sequence , we obtain the final hidden vector of the last layer. Following [11], a fully-connected layer is applied to predict the score of item in the target domain by:
(7) |
where and are parameters to be learned.
During the fine-tuninng stage, we adopt the popular pairwise ranking loss (BPR) [37] as the objective function of UAF for top- recommendation:
(8) |
where is the parameters to be optimized, including pre-trained parameters (excluding the softmax matrix) and fine-tuned parameters (new softmax layer and newly added residual blocks), is the logistic sigmoid function, is the positive label and is a negative label randomly sampled from following [37], contains all positive pairs and the selected negative pairs in the target dataset.
3.3 Policy Network
To derive the optimal fine-tuning strategy given the input user sequence, we develop a policy network to output a binary policy vector, representing the actions to freeze or fine-tune each residual block in the pre-trained NextItNet. The policy network is implemented using a lightweight NextItNet model with dilations of {1, 2, 4, 8} (4 layers or 2 residual blocks). Without any restrictions, the policy network can also be implemented with other deep neural networks, e.g., a RNN model. In this paper, we propose three alternative strategies (i.e., hard, soft and RL) to optimize the policy network.
Domain | ColdRec-1 | ColdRec-2 | ColdRec-3 | ColdRec-4 | ||||
---|---|---|---|---|---|---|---|---|
Instances | Items | Instances | Items | Instances | Items | Instances | Items | |
1,649,095 | 191,022 | 1,472,428 | 645,981 | 22,360 | 11,349 | 340,858 | 52,171 | |
3,798,114 | 20,343 | 2,947,688 | 17,880 | 24,643 | 162 | 1,267,117 | 6,288 |
3.3.1 UAF-Hard Method
Since the policy is a discrete binary variable, it is intractable to optimize the policy network with backpropagation due to the non-differentiable problem. To resolve this issue, we propose using the Gumbeling Softmax sampling method [38, 39] to generate the actions (freeze or fine-tune) from a discrete distribution. We refer UAF with Gumbeling Softmax training as UAF-Hard.
UAF-Hard draws samples from a categorical distribution with class probabilities . Here, we have , indicating the freezing or fine-tuning actions. That is, for each residual block, and are scalars corresponding to the final output of the policy network.
(9) |
where is a one-hot vector and are i.i.d samples drawn from distribution. The distribution can be sampled using inverse transform sampling by drawing from a uniform distribution, i.e. and we can compute .
The operation in Eq. (9) is non-differentiable, but we can use the Gumbel Softmax distribution, which adopts softmax as a continuous relaxation to in order to allievate the non-differentiable problem. We relax the one-hot encoding of to a real-valued vector using:
(10) |
where is a temperature parameter to control the discreteness of the output vector . When becomes closer to 0, the samples from the Gumbel Softmax distribution become indistinguishable from the discrete distribution, i.e., almost the same as the one-hot vector . Here we set to 10 by default.
Sampling the fine-tuned policy from the Gumbel Softmax distribution allows us to backpropagate from the discrete binary decision samples to the policy network, as the Gumbel Softmax distribution is smooth for and has well-defined gradients with respect to the parameters . Similar to [40, 41], we generate all freezing/fine-tuning policies for all residual blocks at once for the trade-off of efficiency and accuracy. During the forward pass, we sample the fine-tuning policy using Eq. (9) for the -th residual block. As for the backward pass, we approximate the gradients of the discrete samples by computing the gradients of the continuous softmax relaxation in Eq. (10).
By using the above approach, we can easily achieve UAF-Hard method in a differentiable way and obtain the policy regarding which layers in pre-trained model should be fine-tuned. By using the objective function in Eq. (8) for the target domain, the policy network is jointly trained with the pre-trained NextItNet model in an end-to-end way.
3.3.2 UAF-Soft Method
In the UAF-Hard method, we learn a policy network to make binary decisions for fine-tuning or freezing the parameters of each layer in the pre-trained NextItNet model. That is, the decision in every residual block is either 0 or 1. As an alternative to the UAF-Hard method, we propose a UAF-Soft method to learn the policy network by employing the weighted combination of freezed pre-tained layers and fine-tuned layers. In such a UAF-Soft method, the output of each residual block is no longer from a certain layer, but from both pre-trained layers and fine-tuned layers. In this way, we can preserve the useful pre-trained information and receive valid fine-tuned information simultaneously.
Specifically, we design a gate for each residual block to control the flow of parameters in the freezed pre-trained layers and fine-tuned layers in a soft way. The weight ratio in each gate is still obtained from the policy network using the lightweight NextItNet model with dilations of {1, 2, 4, 8}. We first put the interaction sequence from source domain into the policy network, and then pass it into the residual block to obtain the final hidden vector . Correspondingly, we will get for all residual blocks during the fine-tuning stage and adopt a logistic sigmoid function to achieve the weight distribution of the fine-tuning policy .
(11) |
(12) |
where is a projection matrix, and is a bias term.
Following UAF-Hard, we can obtain the output of the -th residual block at the fine-tuning stage using Eq. (6). Since the sigmoid function is differentiable, we can directly train the policy network and the fine-tuning model in an end-to-end way.
Models | ColdRec-1 | ColdRec-2 | ColdRec-3 | ColdRec-4 | ||||
---|---|---|---|---|---|---|---|---|
MRR@5 | HR@5 | MRR@5 | HR@5 | MRR@5 | HR@5 | MRR@5 | HR@5 | |
Finetune-Zero | 0.2332 | 0.4013 | 0.3632 | 0.5464 | 0.1174 | 0.2150 | 0.1929 | 0.3257 |
Finetune–CLS | 0.2419 | 0.4124 | 0.3712 | 0.5547 | 0.0909 | 0.1657 | 0.1935 | 0.3283 |
Finetune-Last1 | 0.2546 | 0.4307 | 0.3864 | 0.5724 | 0.1062 | 0.1918 | 0.2032 | 0.3433 |
Finetune-Last2 | 0.2575 | 0.4337 | 0.3885 | 0.5753 | 0.1176 | 0.2118 | 0.2073 | 0.3485 |
Finetune-All | 0.2565 | 0.4345 | 0.3880 | 0.5758 | 0.1185 | 0.2192 | 0.2077 | 0.3495 |
MTL | 0.2410 | 0.4114 | 0.3633 | 0.5468 | 0.1169 | 0.2181 | 0.2093 | 0.3527 |
PeterRec | 0.2578 | 0.4356 | 0.3886 | 0.5763 | 0.0993 | 0.1795 | 0.2117 | 0.3550 |
UAF-Hard | 0.2594 | 0.4378 | 0.3940 | 0.5824 | 0.1190 | 0.2212 | 0.2151 | 0.3640 |
UAF-Soft | 0.2600 | 0.4397 | 0.3939 | 0.5841 | 0.1209 | 0.2189 | 0.2140 | 0.3613 |
UAF-RL | 0.2599 | 0.4386 | 0.3942 | 0.5837 | 0.1196 | 0.2197 | 0.2142 | 0.3619 |
3.3.3 UAF-RL Method
In fact, we can learn the policy network by employing the reinforcement learning algorithm (called UAF-RL) which is a popular technique for solving the non-differentiable problem. The main idea is to learn the policy network that outputs the posterior probabilities of all the binary decisions for freezing or fine-tuning each block in the fine-tuning network. The policy network is trained using curriculum learning [42] to maximize a reward that incentivizes the use of as few blocks to fine-tune as possible while preserving the prediction accuracy. In this regard, we can consider the potential trade-offs between computational cost and prediction accuracy.
Let denote the fine-tuning policies predicted by the policy network, where represents the number of residual blocks in the fine-tuning network. In particular, , where , indicating the freezing or fine-tuning actions. In order to evaluate the advantage of an action , we define the reward function as the following formula:
(13) |
where is an indicator function, represents the freezing or fine-tuning policy applied on the -th block, is a hyperparameter to penalize the wrong policy. The motivation behind the Eq. 13 is two fold: (i) enabling the policy network to control the fine-tuning network so as to generate well-recommended items in target domain; (ii) achieving a significant computational cost reduction to meet the requirement of online service. Specifically, measures the percentage of blocks that are fine-tuned. When a correct recommendation is produced, we incentivize block freezing by giving a larger reward to a policy that uses fewer blocks to fine-tune. In addition, we penalize incorrect recommendations with , which controls the trade-off between efficiency and effectiveness. is simply set to 1 in this paper.
To optimize the policy network, we adopt self-critical sequence training (SCST) [43], which is a form of the popular REINFORCE [44] algorithm, for model training. Rather than estimating a “baseline” to normalize the rewards and reduce variance, SCST applies the output of its own test-time inference algorithm to normalize the rewards it experiences. In details, the exploration action is obtained by sampling from the categorical distribution through modeling the source sequence, while the self-critical baseline is calculated by the greedy search, and the policy take action by . To learn the optimal parameters of the policy network, we minimize the following SCST loss:
(14) |
where represents the probability to sample the exploration action . Since the self-critical baseline is based on the test-time estimate under the current model, SCST is forced to improve the performance of the model under the inference algorithm used at test time. At inference stage, we obtain the actual fine-tuning policy by the greedy search .
Finally, we jointly train the policy network and fine-tuning network in an end-to-end way and optimize the weighted-sum of the BPR loss and SCST loss in the UAF-RL method:
(15) |
where and represent the standard BPR loss and the SCST loss, respectively, and is a hyperparameter to control the weight of the SCST loss, which is set to 1 in our experiments.
4 Experimental Setup
4.1 Experimental Datasets
To evaluate the effectiveness of UAF, we collect four transfer learning datasets from Tencent222https://www.tencent.com/en-us/. Each of them is either from different recommender systems or has different properties. Among them, two datasets have been made publicly available in [11]. The detailed properties of these datasets are described as follows.
ColdRec-1
ColdRec-1 includes both source and target domains. The source domain is the news recommendation dataset collected from QQ Browser333https://browser.qq.com. Each instance is formed of a sequence of watching interactions of a user, where is set to 50. Sequence length less than 50 will be padded with zero in the beginning, similar to [11]. The target dataset is collected from Kandian444https://sdi.3g.qq.com/v/2019111020060911550. All users in Kandian are cold with at most 3 interactions (including clicks of news, videos or advertisements) and half of them have only one interaction. The source and target domains are connected by userID as all users in the target dataset also have corresponding watching records in the source dataset.
ColdRec-2
It has similar characteristics with ColdRec-1 except that the source dataset contains 100 watching interactions for each user in the sequence session. In addition, the source dataset includes both news and video interactions. The target dataset has at most 5 interactions for each user.
ColdRec-3
This is a private dataset collected by Tencent Oula team. Similar to ColdRec-1, the source domain is the news recommendation dataset of QQ Browser, and the session length is set to 50. The target domain is an advertisement recommendation dataset, where 90% users in target dataset have only one clicking action.
ColdRec-4
This is also collected by Tencent Oula team. The source dataset is collected from Tencent Video555https://v.qq.com/ including TV series, movies and short videos, and the session length is set to 50. The target dataset is collected from Kandian, and each user in target dataset has at most 5 clicks on news and videos.
The statistics are summarized in Table I.
Models | ColdRec-1 | ColdRec-2 | ColdRec-3 | ColdRec-4 | ||||
---|---|---|---|---|---|---|---|---|
MRR@5 | HR@5 | MRR@5 | HR@5 | MRR@5 | HR@5 | MRR@5 | HR@5 | |
UAF-Random | 0.2562 | 0.4346 | 0.3917 | 0.5796 | 0.1172 | 0.2158 | 0.2123 | 0.3587 |
UAF-Hard | 0.2594 | 0.4378 | 0.3940 | 0.5824 | 0.1190 | 0.2212 | 0.2151 | 0.3640 |
UAF-Soft | 0.2600 | 0.4397 | 0.3939 | 0.5841 | 0.1209 | 0.2189 | 0.2140 | 0.3613 |
UAF-RL | 0.2599 | 0.4386 | 0.3942 | 0.5837 | 0.1196 | 0.2197 | 0.2142 | 0.3619 |
Models | ColdRec-1 | ColdRec-2 | ColdRec-3 | ColdRec-4 | ||||
---|---|---|---|---|---|---|---|---|
MRR@5 | HR@5 | MRR@5 | HR@5 | MRR@5 | HR@5 | MRR@5 | HR@5 | |
UAF-Hard | 0.2594 | 0.4378 | 0.3940 | 0.5824 | 0.1190 | 0.2212 | 0.2151 | 0.3640 |
UAF-Hard-GRU | 0.2591 | 0.4382 | 0.3943 | 0.5836 | 0.1186 | 0.2204 | 0.2159 | 0.3634 |
UAF-Soft | 0.2600 | 0.4397 | 0.3939 | 0.5841 | 0.1209 | 0.2189 | 0.2140 | 0.3613 |
UAF-Soft-GRU | 0.2598 | 0.4389 | 0.3941 | 0.5845 | 0.1217 | 0.2220 | 0.2151 | 0.3627 |
UAF-RL | 0.2599 | 0.4386 | 0.3942 | 0.5837 | 0.1196 | 0.2197 | 0.2142 | 0.3619 |
UAF-RL-GRU | 0.2597 | 0.4381 | 0.3940 | 0.5833 | 0.1199 | 0.2201 | 0.2146 | 0.3625 |
4.2 Baselines and Evaluation
We compare our models (UAF-Hard, UAF-Soft and UAF-RL) with several powerful approaches. All baseline models share the same NextItNet architecture with ours.
-
•
Finetune-Zero: It is only trained on the target dataset without performing pre-training on the source dataset. That is, its parameters are randomly initialized.
-
•
Finetune-CLS: It is pre-trained on the source dataset. We merely fine-tune the softmax layer of the model on the target dataset.
-
•
Finetune-Last- (1, 2): We pre-train the model on the source dataset and fine-tune the last residual blocks of the pre-trained network (with the softmax layer) on the target data following [45].
-
•
Finetune-All: We pre-train the model on the source dataset and fine-tune all parameters on the target dataset. Finetune-All is the most widely adopted fine-tuning approach, and usually performs better than Finetune-CLS and Finetune-Last- as the entire model is tuned [11].
- •
-
•
PeterRec: This is a recently proposed transfer learning framework which can be used to solve the cold-user problem [11]. The basic idea is to fine-tune a small-sized adapter neural network inserted in the pre-trained network. We evaluate PeterRec by using the official code666https://github.com/fajieyuan/sigir2020_peterrec under our setting.
Our evaluation follows [11] by adopting two popular top- metrics to measure the quality of recommended items, including MRR@ (Mean Reciprocal Rank) [1] and HR@ (Hit Ratio) [47]. Here is set to 5 for comparison.

Model | ColdRec-1 | ColdRec-2 | ||
---|---|---|---|---|
MRR@5 | HR@5 | MRR@5 | HR@5 | |
Finetune-Zero | 0.1624 | 0.2970 | 0.2766 | 0.4362 |
Finetune-All | 0.1857 | 0.3366 | 0.2915 | 0.4533 |
PeterRec | 0.1999 | 0.3562 | 0.2938 | 0.4557 |
UAF-Hard | 0.2120 | 0.3714 | 0.3025 | 0.4651 |
UAF-Soft | 0.2108 | 0.3704 | 0.3021 | 0.4649 |
UAF-RL | 0.2125 | 0.3721 | 0.3023 | 0.4655 |
4.3 Implementation Details
We randomly split each target dataset into training (80%), validation (5%) and test (15%) sets. The grid search algorithm is applied on the validation sets to tune the hyperparameters. The size of each item embedding is set to be 128 for all the models. The NextItNet architecture is implemented by using 16 dilation layers or 8 residual blocks (i.e., {1, 2, 4, 8, 1, 2, 4, 8, 1, 2, 4, 8, 1, 2, 4, 8}), which is used as the backbone of the pre-training and fine-tuning models. We employ Adam optimizer to train our models with learning rate of 0.0001. Batch size and kernel size are set to be 256 and 3, respectively. Regarding the policy network, we use dilations of {1, 2, 4, 8} (4 layers or 2 residual blocks), which is a lightweight neural network. Note that, as mentioned before, without any restrictions, the policy network can also be implemented with other deep neural networks, e.g., a RNN model. All the experiments are implemented in TensorFlow and trained on a single TITAN RTX GPU.


5 Experimental Results
As the key contribution of this work is to solve the CDR task with cold (or new) users through a user-specific adaptive fine-tuning method UAF, we evaluate UAF on four real-world datasets and conduct extensive experiments to answer the following research questions:
-
(1)
RQ1: Whether the proposed UAF (i.e., UAF-Hard, UAF-Soft and UAF-RL) perform better than the existing fine-tuning approaches?
-
(2)
RQ2: What are the effects of the policy network? Does it enable adaptive fine-tuning for each user? Can our policy network be replaced by other types of neural networks (e.g., a RNN model)?
-
(3)
RQ3: Can UAF stay robust when the training data in the target domain is insufficient?
-
(4)
RQ4: Is UAF a general method or does it work well in other tasks (e.g., user profile prediction)?
5.1 Performance Comparison (RQ1)
Table II reports the overall results of UAF and baselines on the four datasets. We can make the following observations. First, Finetune-Zero performs worst among all baselines on ColdRec-1/2/4 in terms of both MRR@5 and HR@5. In particularly, by comparing Finetune-Zero with Finetune-All on the three datasets, we can conclude that pre-training on the source domain is of great help to achieve better recommendation accuracy. On the other hand, Finetune-All obtains only 1% improvement over Finetune-Zero on ColdRec-3. The result suggests that the correlation between the source and target domains on ColdRec-3 may be not as significant as ColdRec-1/2/4. The observation is reasonable since advertisement recommendation may have some discrepancy with news recommendation in terms of user preference. Second, it is shown that Finetune-Last2 performs better than Finetune-Last1, which further performs better than Finetune-CLS. Our observations here are consistent with those in prior work. Third, MTL performs better than Finetune-Zero except on ColdRec-3, indicating the effectiveness of joint learning. Furthermore, Finetune-All outperforms MTL on ColdRec-1/2/3 but slightly underperforms it on ColdRec-4. We argue that the optimal parameters learned for two or more objectives in MTL does not guarantee the optimal performance on the objective of the target domain. By constrast, Finetune-All only cares about the objective on the target domain and thus performs generally better. Fourth, PeterRec achieves state-of-the-art results on ColdRec-1/2/4, but shows worse results on ColdRec-3. The result further indicates that the four datasets have different characteristics in terms of knowledge transfer: some datasets benefit a lot from pre-training, while others may not. Again, pre-training on ColdRec-3 is not as important as on other datasets, and as a result, PeterRec exhibits much worse performance because all pre-trained parameters in PeterRec are not allowed to be modified.
Meanwhile, we observe that UAF, including UAF-Hard, UAF-Soft and UAF-RL, performs obviously the best on ColdRec-2/3/4 and on par with PeterRec and FineAll on ColdRec-1. In fact, we further find that UAF performs significantly better than all other baselines even on ColdRec-1 with limited training data, as shown in Section 5.3. We believe the key advantage of UAF mainly comes from its adaptive fine-tuning mechanism since these models are based on the same backbone network. By accomplishing the goal of per user per fine-tuning policy, UAF achieves more intelligent transfer learning across domains.
5.2 Ablation Studies on the Policy Network (RQ2)
To answer RQ2, we explore the following ablation studies to verify the effects of the policy network in UAF. Note that we may only report partial results for clarity if similar behaviors are observed.
Domain | GenEst | AgeEst | LifeEst | |||
---|---|---|---|---|---|---|
Instances | ItemsClasses | Instances | ItemsClasses | Instances | ItemsClasses | |
1.55M | 646K | 1.55M | 646K | 1.08M | 488K | |
1.55M | 2 | 1.55M | 6 | 1.08M | 8 |
Model | GenEst | AgeEst | LifeEst | |||
---|---|---|---|---|---|---|
Accuracy | F1 | Accuracy | F1 | Accuracy | F1 | |
FinetuneCLS | 0.8882 | 0.7770 | 0.5717 | 0.2747 | 0.5165 | 0.2471 |
FinetuneAll | 0.9042 | 0.8205 | 0.6221 | 0.3457 | 0.5227 | 0.2670 |
PeterRec | 0.9040 | 0.8184 | 0.6163 | 0.3338 | 0.5342 | 0.2616 |
UAF-Hard | 0.9051 | 0.8220 | 0.6243 | 0.3481 | 0.5491 | 0.2733 |
UAF-Soft | 0.9048 | 0.8215 | 0.6259 | 0.3496 | 0.5484 | 0.2728 |
UAF-RL | 0.9049 | 0.8217 | 0.6250 | 0.3487 | 0.5478 | 0.2725 |
5.2.1 Random Policy
In order to verify the effectiveness of the policy network, we compare it with a random policy network, termed as UAF-Random, and the results are shown in the Table III. The random policy network is implemented with a Bernoulli random distribution, which assigns 0.5 to the probability that the policy equals 0 or 1. Note that our random seeds are kept fixed during training and inference. Clearly, the results in Table III show that a well-optimized policy outperforms a random policy. Interestingly, we find that UAF-Random performs comparably or slightly better than Finetune-All on the four datasets. The reason is because UAF-Random is able to combine parameters from both the pre-training and fine-tuning networks, albeit with a random way. However, if we use evaluation seeds different from training, UAF-Random shows much worse results than Finetune-All.
5.2.2 Visualization of Policies
To better understand the fine-tuning policies of UAF, we visualize the utilization rates of the fine-tuned residual blocks in Figure 3. We only depict UAF-Hard since UAF-Soft and UAF-RL show similar behaviors. The illustration indicates that different datasets have completely different fine-tuning policies. Clearly, the utilization rates of all fine-tuned residual blocks are always smaller than 1. This indicates UAF does not only utilize fine-tuned blocks, instead, it also takes advantage of the pre-trained blocks since the summation of utilization rates of them equals to 1. Hence, we conclude that UAF allows the fine-tuning model to automatically identify the right policy in determining which layers of the pre-trained model should be fine-tuned and which layers should have their parameters frozen on a per-user basis, which would be infeasible through a manual approach.
5.2.3 Policy Network with GRU
As mentioned before, the architecture of our policy network can be implemented with other deep neural networks without any restrictions. In order to verify the flexibility of it, we replace the original lightweight NextItNet with a Gated Recurrent Unit (GRU) for the policy network, termed as UAF-Hard-GRU, UAF-Soft-GRU and UAF-RL-GRU, respectively. We report the results in the Table IV. It clearly shows that designing the policy network by GRU yields comparable performance.
5.3 Robustness Studies with Limited Training Data (RQ3)
Transfer learning is supposed to reduce the cost of labeling data. In this subsection, we investigate the effectiveness of UAF with limited training samples. Specifically, we randomly choose 10% of the training examples in the target domain to fine-tune our models and the baselines. Table V summarizes the results of all models on ColdRec-1 and ColdRec-2. As shown, our methods (including UAF-Hard, UAF-Soft and UAF-RL) achieve notably better recommendation results than the Finetune-All and PeterRec. For example, the UAF-Hard method increases the MRR@5 metric by up to 6.1% on ColdRec-1, compared with PeterRec. Therefore, we conclude that the proposed UAF methods could be more powerful than these baselines when dealing with limited training data.
It is well-known that deep models are prone to overfitting when training data is insufficient. We here investigate the effectiveness of the proposed UAF methods against overfitting. We depict the learning curves of different methods in Figure 4. As shown, we observe that UAF-Hard, UAF-Soft and UAF-RL prevent overfitting much better than the Finetune-All. For example, on ColdRec-1, the HR@5 of the standard fine-tuning (Finetune-All) starts to decrease sharply after 5 epochs, whereas UAF-Hard, UAF-Soft and UAF-RL keep improving until 9 epochs, achieving notably higher HR@5. These results show that our methods can prevent overfitting better than many other models since UAF utilizes both pre-trained and fine-tuned parameters, where pre-trained parameters will not suffer from the overfitting issue.
5.4 Adaptability Experiments on User Profile Prediction Tasks (RQ4)
User profiles are important features for generating accurate recommendations. In [11], authors showed the transfer learning framework such as PeterRec and FineAll can also be used for predicting user profiles, which are of great help to lessen cold-start user problem in recommendation tasks. In this subsection, we investigate the versatility of UAF in the user profile prediction task.
Following [11], we predict three types of user profile information (i.e., gender, age and life status). The statistics of the three user profile datasets are summarized in Table VI. Specifically, each instance in GenEst is a user and his gender (male or female) label obtained by the registration information. Similar to GenEst, each instance in AgeEst is a user and his age bracket label - one class represents 10 years. And each instance in LifeEst is a user and his life status label, e.g., single, married, pregnancy or parenting.
The results regarding the user profile prediction tasks are shown in Table VII. We use classification accuracy (referred to as Accuracy) and F1 score (referred to as F1) as the evaluation metrics. Similar conclusions can be made as discussed in Subsection 5.1. UAF with adaptive fine-tuning policies yields better results than baselines in all three datasets, which demonstrates the versatility of UAF while adapting it to different downstream tasks.
6 Conclusion
In this paper, we proposed a User-specific Adaptive Fine-tuning method (UAF) for the cross-domain recommendations. UAF introduces a policy network to automatically decide which layers of the pre-trained model should be fine-tuned and which layers should have their parameters frozen on a per-user basis. It combines the advantages of high-capacity pre-trained network and newly optimized fine-tuned network, leading to enhanced performance on the target task. Extensive experiments on four real-world cross-domain recommendation datasets showed that UAF not only exhibited significantly better prediction accuracy, but also was more robust to overfitting, especially when the target domain has scarce training examples.
Acknowledgement
Min Yang was partially supported by the National Natural Science Foundation of China (NSFC) (No. 61906185), Youth Innovation Promotion Association of CAS China (No. 2020357), Shenzhen Science and Technology Innovation Program (No. KQTD20190929172835662), Shenzhen Basic Research Foundation (No. JCYJ20200109113441941).
References
- [1] B. Hidasi, A. Karatzoglou, L. Baltrunas, and D. Tikk, “Session-based recommendations with recurrent neural networks,” arXiv preprint arXiv:1511.06939, 2015.
- [2] F. Yuan, A. Karatzoglou, I. Arapakis, J. M. Jose, and X. He, “A simple convolutional generative network for next item recommendation,” in WSDM. ACM, 2019, pp. 582–590.
- [3] X. He, L. Liao, H. Zhang, L. Nie, X. Hu, and T.-S. Chua, “Neural collaborative filtering,” in WWW, 2017, pp. 173–182.
- [4] W.-C. Kang and J. McAuley, “Self-attentive sequential recommendation,” in ICDM. IEEE, 2018, pp. 197–206.
- [5] J. Xu, X. He, and H. Li, “Deep learning for matching in search and recommendation,” Foundations and Trends® in Information Retrieval, vol. 14, no. 2–3, pp. 102–288, 2020.
- [6] J. Yosinski, J. Clune, Y. Bengio, and H. Lipson, “How transferable are features in deep neural networks?” in Advances in neural information processing systems, 2014, pp. 3320–3328.
- [7] A. Radford, K. Narasimhan, T. Salimans, and I. Sutskever, “Improving language understanding by generative pre-training,” 2018.
- [8] J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “Bert: Pre-training of deep bidirectional transformers for language understanding,” arXiv preprint arXiv:1810.04805, 2018.
- [9] Y. Guo, Y. Li, L. Wang, and T. Rosing, “Adafilter: Adaptive filter fine-tuning for deep transfer learning,” arXiv preprint arXiv:1911.09659, 2019.
- [10] S.-A. Rebuffi, H. Bilen, and A. Vedaldi, “Efficient parametrization of multi-domain deep neural networks,” in CVPR, 2018, pp. 8119–8127.
- [11] F. Yuan, X. He, A. Karatzoglou, and L. Zhang, “Parameter-efficient transfer from sequential behaviors for user modeling and recommendation,” in Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, 2020, pp. 1469–1478.
- [12] F. Yuan, G. Zhang, A. Karatzoglou, X. He, J. Jose, B. Kong, and Y. Li, “One person, one model, one world: Learning continual user representation without forgetting,” arXiv preprint arXiv:2009.13724, 2020.
- [13] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin, “Attention is all you need,” ICLR, 2017.
- [14] A. Radford, J. Wu, R. Child, D. Luan, D. Amodei, and I. Sutskever, “Language models are unsupervised multitask learners,” OpenAI blog, vol. 1, no. 8, p. 9, 2019.
- [15] T. B. Brown, B. Mann, N. Ryder, M. Subbiah, J. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam, G. Sastry, A. Askell et al., “Language models are few-shot learners,” arXiv preprint arXiv:2005.14165, 2020.
- [16] N. Kalchbrenner, L. Espeholt, K. Simonyan, A. v. d. Oord, A. Graves, and K. Kavukcuoglu, “Neural machine translation in linear time,” arXiv preprint arXiv:1610.10099, 2016.
- [17] A. v. d. Oord, S. Dieleman, H. Zen, K. Simonyan, O. Vinyals, A. Graves, N. Kalchbrenner, A. Senior, and K. Kavukcuoglu, “Wavenet: A generative model for raw audio,” arXiv preprint arXiv:1609.03499, 2016.
- [18] J. Wang, Q. Liu, Z. Liu, and S. Wu, “Towards accurate and interpretable sequential prediction: A cnn & attention-based feature extractor,” in Proceedings of the 28th ACM International Conference on Information and Knowledge Management, 2019, pp. 1703–1712.
- [19] F. Yuan, X. He, H. Jiang, G. Guo, J. Xiong, Z. Xu, and Y. Xiong, “Future data helps training: Modeling future contexts for session-based recommendation,” in Proceedings of The Web Conference 2020, 2020, pp. 303–313.
- [20] Y. Sun, F. Yuan, M. Yang, G. Wei, Z. Zhao, and D. Liu, “A generic network compression framework for sequential recommender systems,” arXiv preprint arXiv:2004.13139, 2020.
- [21] J. Wang, F. Yuan, J. Chen, Q. Wu, C. Li, M. Yang, Y. Sun, and G. Zhang, “Stackrec: Efficient training of very deep sequential recommender models by layer stacking,” arXiv preprint arXiv:2012.07598, 2020.
- [22] L. Chen, F. Yuan, J. Yang, X. Ao, C. Li, and M. Yang, “A user-adaptive layer selection framework for very deep sequential recommender models,” in AAAI, 2021.
- [23] Y. K. Tan, X. Xu, and Y. Liu, “Improved recurrent neural networks for session-based recommendations,” in DLRS, 2016, pp. 17–22.
- [24] J. Tang and K. Wang, “Personalized top-n sequential recommendation via convolutional sequence embedding,” in WSDM. ACM, 2018, pp. 565–573.
- [25] F. Sun, J. Liu, J. Wu, C. Pei, X. Lin, W. Ou, and P. Jiang, “Bert4rec: Sequential recommendation with bidirectional encoder representations from transformer,” in Proceedings of the 28th ACM international conference on information and knowledge management, 2019, pp. 1441–1450.
- [26] B. Hidasi and A. Karatzoglou, “Recurrent neural networks with top-k gains for session-based recommendations,” in CIKM. ACM, 2018, pp. 843–852.
- [27] P. M. Gabriel De Souza, D. Jannach, and A. M. Da Cunha, “Contextual hybrid session-based news recommendation with recurrent neural networks,” IEEE Access, vol. 7, pp. 169 185–169 203, 2019.
- [28] Z. Zeng, C. Xiao, Y. Yao, R. Xie, Z. Liu, F. Lin, L. Lin, and M. Sun, “Knowledge transfer via pre-training for recommendation: A review and prospect,” Frontiers in big Data, vol. 4, 2021.
- [29] M. M. Khan, R. Ibrahim, and I. Ghani, “Cross domain recommender systems: a systematic literature review,” ACM Computing Surveys (CSUR), vol. 50, no. 3, pp. 1–34, 2017.
- [30] H. Kanagawa, H. Kobayashi, N. Shimizu, Y. Tagami, and T. Suzuki, “Cross-domain recommendation via deep domain adaptation,” in European Conference on Information Retrieval. Springer, 2019, pp. 20–29.
- [31] K. Bousmalis, G. Trigeorgis, N. Silberman, D. Krishnan, and D. Erhan, “Domain separation networks,” in Advances in neural information processing systems, 2016, pp. 343–351.
- [32] F. Yuan, L. Yao, and B. Benatallah, “Darec: Deep domain adaptation for cross-domain recommendation via transferring rating patterns,” arXiv preprint arXiv:1905.10760, 2019.
- [33] G. Hu, Y. Zhang, and Q. Yang, “Conet: Collaborative cross networks for cross-domain recommendation,” in Proceedings of the 27th ACM International Conference on Information and Knowledge Management, 2018, pp. 667–676.
- [34] Y. Ni, D. Ou, S. Liu, X. Li, W. Ou, A. Zeng, and L. Si, “Perceive your users in depth: Learning universal user representations from multiple e-commerce tasks,” in KDD, 2018, pp. 596–605.
- [35] Z. Yang, Z. Dai, Y. Yang, J. Carbonell, R. R. Salakhutdinov, and Q. V. Le, “Xlnet: Generalized autoregressive pretraining for language understanding,” in Advances in neural information processing systems, 2019, pp. 5754–5764.
- [36] L. Bertinetto, J. F. Henriques, J. Valmadre, P. Torr, and A. Vedaldi, “Learning feed-forward one-shot learners,” in Advances in neural information processing systems, 2016, pp. 523–531.
- [37] S. Rendle, C. Freudenthaler, Z. Gantner, and L. Schmidt-Thieme, “Bpr: Bayesian personalized ranking from implicit feedback,” in Proceedings of the twenty-fifth conference on uncertainty in artificial intelligence. AUAI Press, 2009, pp. 452–461.
- [38] C. J. Maddison, A. Mnih, and Y. W. Teh, “The concrete distribution: A continuous relaxation of discrete random variables,” arXiv preprint arXiv:1611.00712, 2016.
- [39] E. Jang, S. Gu, and B. Poole, “Categorical reparameterization with gumbel-softmax,” arXiv preprint arXiv:1611.01144, 2016.
- [40] Z. Wu, T. Nagarajan, A. Kumar, S. Rennie, L. S. Davis, K. Grauman, and R. Feris, “Blockdrop: Dynamic inference paths in residual networks,” in CVPR, 2018, pp. 8817–8826.
- [41] Y. Guo, H. Shi, A. Kumar, K. Grauman, T. Rosing, and R. Feris, “Spottune: transfer learning through adaptive fine-tuning,” in CVPR, 2019, pp. 4805–4814.
- [42] Y. Bengio, “Deep learning of representations: Looking forward,” in SLSP, ser. Lecture Notes in Computer Science. Springer, 2013, pp. 1–37.
- [43] S. J. Rennie, E. Marcheret, Y. Mroueh, J. Ross, and V. Goel, “Self-critical sequence training for image captioning,” in CVPR, 2017, pp. 1179–1195.
- [44] R. J. Williams, “Simple statistical gradient-following algorithms for connectionist reinforcement learning,” Machine learning, pp. 229–256, 1992.
- [45] M. Long, Y. Cao, J. Wang, and M. I. Jordan, “Learning transferable features with deep adaptation networks,” arXiv preprint arXiv:1502.02791, 2015.
- [46] R. Caruana, “Multitask learning,” Machine learning, vol. 28, no. 1, pp. 41–75, 1997.
- [47] X. He, Z. He, X. Du, and T. Chua, “Adversarial personalized ranking for recommendation,” SIGIR, pp. 355–364, 2018.
![]() |
Lei Chen is currently a master student at Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences. He received the B.S. degree in computer science and technology from China University of Mining and Technology in 2019. His research interests include natural language processing, recommendation systems and information retrieval. |
![]() |
Fajie Yuan is currently an assistant professor at Westlake University. Prior to that, he was a senior AI researcher at Tencent working on recommender systems. He received his Ph.D. degree from the University of Glasgow. His main research interests include deep learning and transfer learning and their applications in recommender systems. |
![]() |
Jiaxi Yang is currently an undergraduate of Huazhong University of Science and Technology. He will study for a Ph.D. degree at Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences. His research interests include recommendation systems and information retrieval. |
![]() |
Xiangnan He is a professor at the University of Science and Technology of China (USTC). He received his Ph.D. in Computer Science from the National University of Singapore (NUS). His research interests span information retrieval, data mining, and multi-media analytics. He has over 80 publications that appeared in several top conferences such as SIGIR, WWW, and MM, and journals including TKDE, TOIS, and TMM. His work has received the Best Paper Award Honorable Mention in WWW 2018 and ACM SIGIR 2016. He is in the editorial board of journals including Frontiers in Big Data, AI Open etc. Moreover, he has served as the PC chair of CCIS 2019 and SPC/PC member for several top conferences including SIGIR, WWW, KDD, MM, WSDM, ICML etc., and the regular reviewer for journals including TKDE, TOIS, TMM, etc. |
![]() |
Chengming Li received the B.S. and M.S. degree in computer application technology from Dalian University of Technology in 2009 and 2011, respectively. He received Ph.D. degree in Graduate School of Information Science and Electrical Engineering from Kyushu University in 2015. He currently is an associate professor at Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences. His research interests include Data Mining, Network Security, and Big Data. |
![]() |
Min Yang is currently an associate professor with the Shenzhen Institute of Advanced Technology, Chinese Academy of Science. She received her Ph.D. degree from the University of Hong Kong in February 2017. Prior to that, she received her B.S. degree from Sichuan University in 2012. Her current research interests include recommendation systems, deep learning and natural language processing. |