Intent Disentanglement and Feature Self-supervision for Novel Recommendation
Abstract
One key property in recommender systems is the long-tail distribution in user-item interactions where most items only have few user feedback. Improving the recommendation of tail items can promote novelty and bring positive effects to both users and providers, and thus is a desirable property of recommender systems. Current novel recommendation studies over-emphasize the importance of tail items without differentiating the degree of users’ intent on popularity and often incur a sharp decline of accuracy. Moreover, none of existing methods has ever taken the extreme case of tail items, i.e., cold-start items without any interaction, into consideration.
In this work, we first disclose the mechanism that drives a user’s interaction towards popular or niche items by disentangling her intent into conformity influence (popularity) and personal interests (preference). We then present a unified end-to-end framework to simultaneously optimize accuracy and novelty targets based on the disentangled intent of popularity and that of preference. We further develop a new paradigm for novel recommendation of cold-start items which exploits the self-supervised learning technique to model the correlation between collaborative features and content features. We conduct extensive experimental results on three real-world datasets. The results demonstrate that our proposed model yields significant improvements over the state-of-the-art baselines in terms of accuracy, novelty, coverage, and trade-off.
Index Terms:
recommender systems, novel recommendation, disentangled representation, self-supervised learning.1 Introduction
With the rapid growth of the web, the users are overwhelmed with the choice of “finding the right thing” from a vast number of products. The recommender systems, which use historical data to infer the user’s preference on particular items like movies, commodities, and places, are a crucial component of many e-commerce platforms. One key property in recommender systems is the long-tail distribution in user-item interactions, where a tiny amount of popular items receive most of the user attention and a large proportion of tail items only have few user feedback. Improving the recommendation of tail items can enrich users’ experience by providing them with more chances to find interesting yet unpopular items. It can also bring positive effects to the providers since the niche products may increase the companies’ marginal profits. As a result, novel recommendation becomes a desirable property of modern recommender systems. Which, however, is a challenging task due to the conflict between accuracy and novelty.
Most of existing studies on novel recommendation [1, 2, 3, 4, 5], also known as the long-tail recommendation in the literature 111We will use the terms “novel recommendation” and “long-tail recommendation” to describe the existing work interchangeably., adopt a two stage re-ranking strategy. At the first stage, the recommender systems aim to achieving a high accuracy using a base model. At the second stage, the results from the previous stage are re-ranked towards the novelty target by introducing more tail items into the candidate top- item list. Such approaches have an inherent limitation, i.e., the popularity bias that recommenders emphasize popular items much more than tail ones remains exist at the first stage.
Several recent studies [6, 7, 8] propose the end-to-end framework where the recommendation is optimized for both the accuracy and novelty at the same time. While making progress, the PPNW and TailNet methods in [6, 7] tend to recommend tail items and result in a relatively low accuracy. More importantly, none of the re-ranking and the end-to-end based novel recommendation methods differentiate the degree of users’ intent to popularity. Indeed, though the consumers often refer to others for product choice and will comply with the group norm to some extent, an individual’s conformity is affected by many factors like intelligence, personality, and status [9, 10]. That is to say, facing the same popular item, different consumers may react in the opposite direction. Some users tend to conform the others’ action and purchase the popular item, while others may or may not undertake the same action as they are more likely to meet their own personal interests.
In view of the different levels of conformity among users, we propose to disentangle a user’s intent into conformity influence and personal interests, which drives her interaction towards popular and novel niche items, respectively. By doing this, the users’ great diversity of intent can be uncovered to enhance the expressiveness of users’ representations. The same operation is performed in the item side, i.e., disentangling items’ representation into popularity and characteristic factors. We then present a unified end-to-end framework to simultaneously optimize accuracy and novelty targets based on the learned user and item representations. In this way, the intrinsic popularity and novelty factors are introduced into the novel recommendation process and the systems can naturally achieve a balance between the recommendation of tail items and that of popular items.
We further make in-depth analyses on the relationship between the normal long-tail items and the cold-start ones. As shown in Figure 1, the cold-start items are actually the extreme case of long-tail items which do not have any interactions. Kapoor et al. [11] point out in their pioneering work that an item can be novel in three ways: (1) it is new to the system and thus for every user (cold-start), (2) it is known to the system but new to the single user, (3) it is known to the user long before but forgotten at the moment. Unfortunately, existing studies [1, 2, 3, 4, 5, 6, 7, 8] focus on novel recommendation tasks under the above definition (2) (3), and none of them has ever taken cold-start items into consideration. Meanwhile, previous research on cold-start recommendation [12, 13, 14, 15, 16, 17, 18] does not include long-tail items, either.
Based on the above analyses, we develop a new paradigm for novel recommendation of cold-start items. To address the problem of missing collaborative features for cold-start items, we exploit two types of self-supervised learning technique, including the variational autoencoder and mutual information maximization, to model the correlation between collaborative features and content features. To the best of our knowledge, this is the first attempt to exploring both the known long-tail and the unknown cold-start items in novel recommendation.
Extensive experimental results on three datasets demonstrate that our proposed model yields significant improvements over the state-of-the-art baselines in terms of the overall trade-off among accuracy, coverage, and novelty metrics on novel recommendation task for both the standard long-tail and the cold-start items.
2 Related Work
In this section, we review the related work from three perspectives, namely novel recommendation, disentangled representation learning, and self-supervised learning.
2.1 Novel recommendation
Standard Long-tail recommendation. Current research on long-tail recommendation can be categorized into two groups. The first group design different re-ranking methods. They post-process the ranking list of a standard model to account for additional objectives like coverage rather than devising a new model. Some studies [1, 4, 5] improve novelty by countering the effects of item popularity, and others [3, 2] propose clustering approaches and leverage tail items directly in a recommendation list. The second group of methods adopt end-to-end models. For example, Lo et al. [6] propose a personalized pairwise novelty weighting for BPR loss function as an end-to-end method. Liu and Zheng [7] present a network architecture for long-tail session-based recommendation by introducing an adjustable preference mechanism. Zhang et al. [8] transfer knowledge from head items to tail items for leveraging the rich user feedback in head items and the semantic connections between head and tail items.
Overall, both groups of long-tail recommendation models ignore the users’ different level of conformity. In contrast, we disentangle users’ intent into popularity and preference embeddings, which capture the inherent factors that determine users’ choice on popular or niche items.
Cold-start recommendation. Early cold-start recommendation methods [19, 20, 21] mainly exploit side information as regularization in MF objective function [19] or adopt similarity based or feature mapping technique for integrating side information. More recent studies design various types of neural models to incorporate side information, including graph-based inductive learning methods [12, 13], the model based meta-learning methods [22] and the gradient based meta-learning method [23], the active learning scheme [14, 15] which require extra costs or budgets, and graph neural models [17, 18].
While these methods achieve promising performance for cold-start recommendation, they are not designed for recommending long-tail items. For example, DropoutNet [24] and STAR-GCN [16] simulate the cold-start scenario by masking and reconstructing a part of input features in training without changing the item representations. Such methods cannot be adapted to the novel recommendation. One of our contribution is to view the cold-start items as the extreme cases of tail items, and present a new learning paradigm for cold-start novel recommendation.
2.2 Disentangled Representation Learning
Disentangled representation learning is originated from computer vision field [25, 26]. It aims to learn representations that separate explanatory factors of variations behind the data to improve the robustness and interpretability.
With the superior performance, disentangled representation learning sheds new light on recommender systems. MacridVAE [27] is the first work that introduces the disentangled representation learning into user behavior data at both a macro and a micro level. Inspired by it, DICER [28] combines the content information into the procedure of disentanglement. Besides, a few studies [29, 30] apply disentangled graph convolutional networks to reflect the fine-grained latent intents. Ma et al. [31] propose to disentangle the intents behind any given sequence of behaviors for sequential recommendation.
All the aforementioned methods need to define latent intents/channels before performing disentangling and comprehend these latent representation through post-hoc explanation. Perhaps the closest work to ours is DICE [32] which contains user interest and conformity disentanglement. However, DICE is not for long-tail recommendation and does not need to consider the balance between accuracy and novelty. In addition, it adopts the casual model and negative sampling strategy to learn corresponding embeddings. Different from it, we present an intent prototype approach to aggregate users’ and items’ features, which tallies well with consumer conformity theory [33, 9, 10].
2.3 Self-supervised Learning
Self-supervised learning models can leverage input data itself as supervision and benefits almost all types of downstream tasks. The objectives in self-supervised learning can be categorized into generative, contrastive, and adversarial types [34] such as auto-encoding [35, 36], VAE variants [37, 38], discriminative models [39], and Adversarial self-supervised learning (GANs) [40].
As the research of self-supervised learning is still in its infancy, there are only several studies incorporating it with recommender systems [41, 42, 31, 43, 44, 45]. Some of these efforts mine the self-supervision signals from sequential data [31, 43, 45] and others capture the structure properties from user-item bipartite graph [44, 41] or social graph [42]. Different from the above approaches, our work is the first to consider the correlations between the collaborative features originated from interaction behaviors and user’s/item’s inherent content features, and we adopt the generative and contrastive self-supervised learning jointly under these two types of features to tackle the cold-start problem in recommendation.
3 Proposed Model
In this section, we introduce the proposed model by first formulating the problem and then presenting the details.
3.1 Problem Formulation
Let be a set of users and be a set of items, where and denote the corresponding cardinalities. Let be the user-item interaction matrix, indicating whether the user purchased or clicked on the item. We focus on the recommendation task with implicit feedbacks [46], where the interaction matrix is defined as:
The observed entries reflect users’ interest in the item, while the unobserved entries are mixed with unknown data and negative views of the item. Given the above interaction information, our goal is to estimate users’ interest for unobserved entries and rank the candidate items according to the predicted scores such that we can recommend the top- items for the traditional recommendation task. Beyond that, we are particularly interested in novel recommendation for both long-tail and cold-start items.
Recent studies show that the item space can be divided into two parts, namely popular short-head items and niche long-tail items. The long-existing but unpopular tail items have high novelty, and they can provide more surprises for users and more profits for providers. Besides, the cold-start items without any interaction can be regarded as extreme cases of tail items. The cold-start scenario is commonly found in the real world, e.g., the newly released movies have no audience. These new items can stimulate users’ interest and avoid over-specialization in recommendation. Thus we extend the definition of standard long-tail recommendation to cover cold-start items.
The main target for novel recommendation is to optimize the trade-off between accuracy and novelty [5, 6, 7]. On one hand, the recommendation accuracy is evaluated by a matching score between the recommendation list and the ground-truth list. On the other hand, the recommendation quality is measured by the proportion of novel items and the coverage of item space of the list.
3.2 An Overview
Due to the different popularity degree of users and those of items, the users have specific intents to purchase/watch different items. Our target is to find out users’ intrinsic intents for different items and eliminate the dominance of popular items, such that we can select the relevant long-tail items and cold-start items for target users to improve the novelty of recommendation. To this end, we propose an intent disentanglement and feature self-supervision (IDS4NR) model and show its architecture in Figure 2(a).
Our model consists of four layers. We first present an input and embedding layer, which takes users’/items’ unique one-hot ID encoding and their corresponding attributes as input and and transform them into low-dimensional latent vectors. We then develop a self-supervision module to map the features from different views to tackle the cold-start problem. Next, we design an intent disentanglement layer to disentangle the latent features into popularity and preference factors. Finally, we introduce a plug-in backbone layer to model the interactions between the user and the item, which can adopt different base models to make the framework generic.
3.3 Model Architecture
3.3.1 Input and Embedding Layer
Initial collaborative features. We set up a lookup table to transform the one-hot representations of each user and item into low-dimensional vector. After transformation, and are the latent factor representation of the user and the item , respectively. We denote them as collaborative features because they are learned from user-item interaction behaviors.
Initial content features. Besides the interactions, each user/item is equipped with side information that describes its own characteristics, such as attribute information, text description, and image. To be general, we term them as content features. Specifically, in our scenario, each user or item is associated with a set of attributes from different fields. Below is an example of user attributes:
Similar to the transformation of collaborative features, we set up an attribute encoding matrix of users and items. The attribute encoding of a user and an item is denoted as and , respectively, where .
3.3.2 Feature Self-supervision Module
New released items can be a good way to surprise users. However, these cold-start items do not contain any historical interaction data, so the valuable collaborative features cannot be learned from the traditional method. In this paper, since we extend the standard long-tail recommendation to cold-start items, we have to face such a collaborative feature missing problem. Below we elaborate our new paradigm on solving this challenge.
The content information of the new item has been proven to be important in solving the cold-start problem. Different from simply fusing the auxiliary content information into the item representation, we believe that it is critical to establish the correlation between the content features and the collaborative features for cold-start items. We term this the feature self-supervision between two views of the target item. For example, animation movies are the mainstream entertainment among teenage children. The animation movies (content features) and the behaviors of their audience (collaborative features) have a strong correlation, while the love movies (other content features) and the behaviors of teenage children (collaborative features) have a weak correlation.
In light of this, we first set up the mapping between collaborative and content features, and then generate the missing collaborative features of cold-start items when making recommendations. In the field of representation learning, self-supervised learning leverages input data to construct pre-training tasks and thus can benefit most types of downstream tasks. In our recommendation scenario, we consider the self-supervision as an auxiliary task, and design a collaborative-content feature self-supervision module.
As shown in Figure 2(b), our self-supervision module has two key components: variational autoencoder (VAE) and mutual information maximization (MIM). Both of them are used to establish the strong correlation between the behaviors of items and their related contents. Specifically, we adopt VAE to encode content features into the latent space, so that the latent representation has the same distribution as the original one. We further maximize the mutual information between the generated collaborative features and the inherent content features in order to enhance the representation’s discriminative ability.
The VAE structure assumes that data are generated from underlying latent representation. Given the data, the posterior distribution over a set of unobserved variables is approximated by a variational distribution. VAE consists of a generation part and an inference part. Take the item as an example, we denote its content features as the concatenation of all of its attribute encoding, i.e., . In the generation part, the reconstructed content features is generated from its latent variables through a generation network like MLP parameterized by :
(1) |
In the inference part, variational inference approximates the true intractable posterior of the latent variable by introducing an inference network parameterized by [36].
(2) |
The objective of variational inference is to optimize the free variational parameters so that the KL-divergence is minimized. With the reparameterization trick, we sample and reparameterize . In this case, the gradient towards can be back-propagated through the sampled . In summary, the loss function in VAE is defined as follows:
(3) |
Mutual information (MI) is a basic concept in statistics, which measures dependencies between random variables. When encoding the the content features into latent space, to increase the discriminative ability, we further identify the correlation between the latent content features and its corresponding collaborative features using the MIM strategy.
Intuitively, the item ’s latent content features is more relevant to its own collaborative features , compared with the collaborative features of other items. Therefore, we devise the MIM loss function as:
(4) |
where are the anchor item attributes, the positive item, and the negative item. We uniformly sample the item as the negative one, which does not contain any attributes of from the entire item space. is the discriminator function that takes two vectors as the input and then scores the agreement between them. We simply implement it as the dot product between two representations to exclude additional parameters [42].
Finally, the whole loss function of the auxiliary feature self-supervision task can be defined as:
(5) |
3.3.3 Intent Disentanglement Module
Recall that we analyze the users’ different levels of conformity in the introduction part, which are ignored by existing novel recommendation methods. In this subsection, we present our intent disentanglement module to tackle this problem.
To distinguish the reason why users interact popular or niche items, we set the intents as popularity and preference factors and model the contribution of specific intents to the user-item interactions. In this way, we alleviate the dominance of popularity and increase the novelty naturally. Our proposed intent disentanglement module is shown in Figure 2(c), which has three unique properties.
- •
-
•
Both collaborative and content features take effects when disentangling intents, which tallies better with the conformity theory than previous casual model and negative sampling strategy [32].
-
•
Our disentangled factors are naturally integrated into the end-to-end novel recommendation framework, where the popularity and preference factors account for the accuracy and novelty, respectively.
Intent prototype strategy. We take the user as an example for illustration. Formally, the user ’s overall feature list is denoted as , where is equal to her collaborative feature and the rest are content features.
We start by defining the prototypical intent representations, i.e., the popularity prototype and the preference prototype , which are part of model parameters and reflect the position of specific intent in the latent space. We then cluster the initial features according to their distance to two intent prototypes:
(6) | |||
(7) |
where . We adopt dot product to measure the similarity between a given input feature vector and an intent prototype.
Intent aggregation based on collaborative and content features. The clustering weight describes the correlation between a specific feature vector and the intent prototype. We now aggregate all features and obtain the integrated representation of the user under each disentangled intent:
(8) | |||
(9) |
where is a feed-forward network. The item ’s popularity and characteristic representation can be obtained in the same way. So far, we have disentangled the user/item representation under specific intents. We will later explain how the disentangled intents contribute to the whole framework in the optimization subsection.
3.3.4 Backbone Layer
After obtaining the user and item representations, we take traditional recommendation methods as a plug-in backbone layer to model the relationship between user and item, which ensures the generality of our framework in enhancing recommendation novelty.
Traditional recommendation methods can be roughly grouped into two types: point-wise and pair-wise [47]. Note that some methods use the list-wise objective, but they are not widely adopted due to the expensive computational cost.
Point-wise objective. The basic idea in point-wise methods is that they only consider the individual user-item pair for ranking prediction. They do not relate other observed user-item pairs to the learning process. This kind of methods takes a triplet as the input, where indicates that the user has interacted with the item (implicit preference), otherwise . The general form of the point-wise loss function can be defined as:
(10) |
where and is a regularization term. A series of traditional methods adopt this kind of optimization strategy and we adopt LFM [48] and NCF [49] as representative base models. During this process, we feed the user/item representation under specific intents into the backbone layer, then obtain the predicted score and finally calculate the ranking loss according to these base models.
Pair-wise objective. The pair-wise approaches try to construct the set of item pairs by concerning their relative ordering. It is usually considered to be more suitable for optimizing the top- recommendation since it compares the relevance of samples instead of getting one ranking score. This kind of methods inputs a triplet , where are the implicitly liked and not observed item by the user . The general form of the pair-wise loss function can be defined as:
(11) |
We chose the classic CML [50] as a representative base model of this kind of approaches.
3.4 Optimization and Learning
3.4.1 Novelty-weighted Recommendation Loss
After feeding the disentangled user and item representations into the backbone layer, we can obtain the specific loss (). However, if simply summing the losses without distinction, the learned representation of different intents cannot well reflect the popularity and preference factors. Thus we design a novelty-weighted loss to control the impacts of different intents.
Firstly, inspired by [6], the item novelty can be measured by its popularity degree, namely an item is more likely to be novel when more users have never interacted with it. The novelty score of the item can be defined as:
(12) |
After that, the recommendation loss takes a weighted sum of two intents through the novelty score, where is sigmoid function for normalization:
(13) |
Even if it is a common practice to regard the novelty score as weight, we argue that it is an effective way to align the intents to the popularity and preference factors.
3.4.2 Training on Disentangled Representations
Below we discuss the training strategy for point-wise and pair-wise objectives based on the disentangled user and item representations. (1) For the point-wise learning, the model takes a triplet as input. Assuming that the novelty of the item is high, e.g. , the users’ interaction to is reasonably due to her preference intent. Therefore, the loss has a larger weight and it would have more impacts on . Conversely, the popularity intent becomes the leading factor when the user interacts hot items.
(2) For the pair-wise learning, the model takes a triplet as input. Assuming that the novelty of the item is higher than , i.e., , the user is more likely to interact with the item for personal preference than . Hence has a larger weight, and vice versa.
Finally, the overall loss of our IDS4NR framework contains the loss for the main task of novel recommendation and that for the auxiliary task of feature self-supervision:
(14) |
where is a constant weighting factor.
4 Experiments
4.1 Experimental Setup
4.1.1 Datasets
We use three publicly available datasets from different domains. MovieLens222https://grouplens.org/datasets/movielens/ is a widely adopted dataset in the application domain of recommending movies to users, and we employ the MovieLens-100K version. Music and Beauty are chosen from Amazon333http://jmcauley.ucsd.edu/data/amazon/links.html. Following the evaluation settings in [51, 52], we take the 5-core version for experiments, where each user or item has at least five interactions.
In order to be consistent with the implicit feedback setting [49, 53], we transform the detailed rating into a value of 0 or 1, indicating whether a user has rated an item. The statistics of the datasets are shown in Table I.
Dataset | #Users | #Items | #Interactions | Sparsity |
MovieLens | 943 | 1682 | 100000 | 93.70% |
Music | 5541 | 3568 | 64706 | 99.67% |
Beauty | 8159 | 5863 | 98566 | 99.79% |
4.1.2 Evaluation Metrics
We evaluate the performance of novel recommendation in terms of accuracy, coverage, novelty and the trade-off.
We adopt Recall@N (Rec@N) [50, 5, 7] as the accuracy metric, which considers whether the ground-truth is ranked amongst the top- items (normally @5 or @10):
(15) |
where is ground-truth list of user and is user ’s top- recommendation list.
Coverage@N (Cov@N) [2, 5, 7] is the ratio of the total number of distinct recommended items to the total items:
(16) |
Following the Pareto principle [54], the head items are the 20% most popular items, and the rest are long-tail novel items. We extend the NovAccuracy@N (Nov@N) [5, 6, 7] to measure how many novel items are in each top- recommendation list:
(17) |
where is novel item set including both long-tail and cold-start items.
Lastly, we employ F-score as a harmonic mean of conflicting accuracy (recall), novelty, and coverage:
(18) |
4.1.3 Baselines
To demonstrate the effectiveness of our proposed IDS4NR model, we compare it with the following state-of-the-art baseline methods. In addition, we adopt different classical base models as backbone layer to show the generality of our framework.
Long-tail recommendation baselines:
-
•
GANC [5] presents a re-ranking framework to integrate the learned user long-tail preference for accuracy, coverage, and novelty oriented recommendation.
-
•
PPNW [6] proposes a loss weighting approach by leveraging users’ and items’ novelty information for end-to-end novel recommendation.
-
•
TailNet [7] designs a preference mechanism which determines users’ preference on popular or niche items in session-based long-tail recommendation. We remove its GRU encoder that is irrelevant to our experiments.
-
•
MIRec [8] is a dual transfer learning framework which collaboratively transfers knowledge from both model-level and item-level and from head items to tail items.
Cold-start recommendation baselines:
-
•
DropoutNet [24] incorporates content and collaborative information with a neural model. It presents the dropout technique to tackle the cold-start issues.
-
•
HERS [17] employs users’ social relations as user-user graph and builds item-item graph based on the common tags between two items. It aggregates user and item relations for addressing cold-start problem.
-
•
MetaEmb [23] is a meta-learning approach to cold-start problem. It trains an embedding generator for new IDs through gradient-based meta-learning technique.
-
•
AGNN [18] develops a graph neural network variant for attribute graph and designs an extended VAE structure for strict cold-start recommendation.
Classical base models:
-
•
LFM [48] is a classical matrix factorization method for collaborative filtering, which learns the latent factors by alternating least squares.
-
•
NCF [49] presents a neural collaborative filtering method that combines multi-layer perceptron with generalized matrix factorization to encode non-linearities.
-
•
CML [50] is a metric learning method, which encodes the user and item into a joint metric space and measures the user-item pair by euclidean distance.
4.1.4 Settings
To simulate the strict item cold-start scenario, we choose 20% items which have the latest average interactions chronologically as cold-start items. Note all their interactions are put into the testing set such that they are unseen during training. In addition, to be consistent with normal training scenario, we randomly select 10% of historical interactions for each user and add them into the testing set, and the remaining data are treated as training set. During the test phrase, our evaluation protocol ranks all unobserved items in the training set for each user [5, 55].
For the baselines, we follow the same hyper-parameter settings if they are reported by the authors and we fine-tune the hyper-parameters if they are not reported. For our IDS4NR, we set the batch size = 128, the initial learning rate = 0.001, the number of max epoch = 40, 40, 30 for the base model LFM, NCF, and CML, respectively. We use Adam [56] as optimizer to self-adapt the learning rate. We sample 4 unobserved items as negative samples for each positive user-item pair [49]. The embedding dimension is set to 50 for both our IDS4NR and all the baselines, to ensure the comparable trainable parameters.
4.2 Performance Comparison
Dataset | Model | Accuracy | Coverage | Novelty | Trade-off | ||||
Rec@5 | Rec@10 | Cov@5 | Cov@10 | Nov@5 | Nov@10 | F1@5 | F1@10 | ||
MovieLens | GANC | 0.0686 | 0.1123 | 0.3565 | 0.4194 | 0.4445 | 0.4506 | 0.0375 | 0.0648 |
PPNW | 0.0328 | 0.0562 | 0.4349 | 0.5062 | 0.6937 | 0.6896 | 0.0256 | 0.0470 | |
TailNet | 0.0761 | 0.1247 | 0.3565 | 0.4497 | 0.2483 | 0.2386 | 0.0297 | 0.0493 | |
MIRec | 0.0713 | 0.1260 | 0.3357 | 0.4212 | 0.4627 | 0.4598 | 0.0382 | 0.0727 | |
DropoutNet | 0.0811 | 0.1387 | 0.3689 | 0.4385 | 0.2517 | 0.2614 | 0.0322 | 0.0569 | |
HERS | 0.0855 | 0.1397 | 0.3701 | 0.4569 | 0.1862 | 0.2170 | 0.0275 | 0.0510 | |
MetaEmb | 0.0826 | 0.1296 | 0.4159 | 0.4979 | 0.2430 | 0.2619 | 0.0337 | 0.0570 | |
AGNN | 0.0996 | 0.1598 | 0.3184 | 0.4147 | 0.1756 | 0.2045 | 0.0281 | 0.0522 | |
LFM | 0.0888 | 0.1455 | 0.3891 | 0.4699 | 0.1974 | 0.2283 | 0.0303 | 0.0555 | |
IDS4NR(+LFM) | 0.0870 | 0.1439 | 0.4414 | 0.5721 | 0.3870 | 0.3901 | 0.0487 | 0.0871 | |
NCF | 0.0806 | 0.1274 | 0.3951 | 0.4842 | 0.2492 | 0.2745 | 0.0328 | 0.0573 | |
IDS4NR(+NCF) | 0.0852 | 0.1386 | 0.4670 | 0.6155 | 0.3906 | 0.3853 | 0.0494 | 0.0865 | |
CML | 0.0886 | 0.1376 | 0.3755 | 0.4533 | 0.1836 | 0.2121 | 0.0283 | 0.0494 | |
IDS4NR(+CML) | 0.0752 | 0.1302 | 0.5519 | 0.6613 | 0.4296 | 0.4253 | 0.0506 | 0.0903 | |
Music | GANC | 0.0428 | 0.0581 | 0.6304 | 0.6657 | 0.7515 | 0.7773 | 0.0427 | 0.0601 |
PPNW | 0.0573 | 0.0911 | 0.4956 | 0.6010 | 0.5426 | 0.5356 | 0.0422 | 0.0717 | |
TailNet | 0.0246 | 0.0383 | 0.5850 | 0.6789 | 0.7324 | 0.7004 | 0.0236 | 0.0385 | |
MIRec | 0.0779 | 0.1297 | 0.4836 | 0.5942 | 0.5547 | 0.5312 | 0.0562 | 0.0979 | |
DropoutNet | 0.1010 | 0.1411 | 0.2723 | 0.3600 | 0.1181 | 0.1853 | 0.0198 | 0.0411 | |
HERS | 0.1284 | 0.1864 | 0.4934 | 0.6351 | 0.2733 | 0.2937 | 0.0580 | 0.0935 | |
MetaEmb | 0.0894 | 0.1355 | 0.6990 | 0.7699 | 0.3633 | 0.3701 | 0.0591 | 0.0908 | |
AGNN | 0.1121 | 0.1660 | 0.5707 | 0.6867 | 0.3169 | 0.3439 | 0.0608 | 0.0983 | |
LFM | 0.1147 | 0.1620 | 0.6203 | 0.7223 | 0.2534 | 0.2755 | 0.0547 | 0.0834 | |
IDS4NR(+LFM) | 0.0891 | 0.1366 | 0.6853 | 0.8324 | 0.5006 | 0.4983 | 0.0719 | 0.1158 | |
NCF | 0.0982 | 0.1419 | 0.6231 | 0.7321 | 0.2785 | 0.3073 | 0.0511 | 0.0810 | |
IDS4NR(+NCF) | 0.0906 | 0.1373 | 0.9030 | 0.9747 | 0.5349 | 0.5243 | 0.0859 | 0.1287 | |
CML | 0.0856 | 0.1257 | 0.5923 | 0.7433 | 0.1916 | 0.2093 | 0.0335 | 0.0544 | |
IDS4NR(+CML) | 0.1123 | 0.1643 | 0.7363 | 0.8156 | 0.4331 | 0.4330 | 0.0838 | 0.1232 | |
Beauty | GANC | 0.0041 | 0.0081 | 0.4432 | 0.4862 | 0.8850 | 0.8942 | 0.0036 | 0.0076 |
PPNW | 0.0117 | 0.0189 | 0.4374 | 0.5415 | 0.5952 | 0.5920 | 0.0088 | 0.0158 | |
TailNet | 0.0060 | 0.0089 | 0.5336 | 0.6358 | 0.7290 | 0.6713 | 0.0055 | 0.0086 | |
MIRec | 0.0220 | 0.0370 | 0.4847 | 0.5954 | 0.5729 | 0.5533 | 0.0170 | 0.0309 | |
DropoutNet | 0.0338 | 0.0510 | 0.1930 | 0.2754 | 0.0691 | 0.1214 | 0.0045 | 0.0114 | |
HERS | 0.0432 | 0.0650 | 0.4680 | 0.6232 | 0.2375 | 0.2690 | 0.0192 | 0.0341 | |
MetaEmb | 0.0272 | 0.0450 | 0.6929 | 0.7676 | 0.3885 | 0.3967 | 0.0198 | 0.0340 | |
AGNN | 0.0263 | 0.0403 | 0.5377 | 0.6518 | 0.3124 | 0.3388 | 0.0151 | 0.0259 | |
LFM | 0.0242 | 0.0401 | 0.5558 | 0.6906 | 0.1978 | 0.2278 | 0.0102 | 0.0197 | |
IDS4NR(+LFM) | 0.0270 | 0.0426 | 0.6029 | 0.7115 | 0.4006 | 0.4128 | 0.0190 | 0.0321 | |
NCF | 0.0201 | 0.0317 | 0.6286 | 0.7264 | 0.3037 | 0.3319 | 0.0121 | 0.0210 | |
IDS4NR(+NCF) | 0.0254 | 0.0413 | 0.9026 | 0.9737 | 0.5387 | 0.5363 | 0.0253 | 0.0417 | |
CML | 0.0211 | 0.0314 | 0.6510 | 0.7963 | 0.2204 | 0.2396 | 0.0102 | 0.0168 | |
IDS4NR(+CML) | 0.0293 | 0.0477 | 0.7919 | 0.8608 | 0.4958 | 0.4905 | 0.0262 | 0.0432 |
The performance comparison between our proposed IDS4NR model and the baselines on three datasets is reported in Table II. From the results, we have the following important observations.
Firstly, it is clear that our IDS4NR significantly outperforms all the baselines in terms of the trade-off F1 of accuracy, coverage, and novelty on three datasets. The relative improvement of our model over the best baselines on three datasets are 32.46%, 45.35%, 32.32% on F1@5, and 24.21%, 31.46%, 26.69% on F1@10, respectively. It verifies the superiority of our proposed framework with the novelty-regulated intent disentanglement and collaborative-content feature self-supervision module.
Secondly, compared with four long-tail recommendation baselines, our IDS4NR achieves the superior performance in terms of accuracy and coverage. Although IDS4NR cannot get enhancement on the novelty metric over these baselines, it outperforms its base model by a large margin in terms of novelty. Among four baselines, GANC, PPNW, and TailNet emphasize the impact of niche items through the re-ranking strategy, or the modified loss function and adjusted user representation, and they do not consider the inherent mechanism that drives users to interact with the items. Consequently, they achieve high novelty but sacrifice the coverage of entire item space. Meanwhile, MIRec transfers the knowledge from head items to tail ones and is better than three other baselines, but it is still inferior to our model.
Dataset | Model | Accuracy | Coverage | Novelty | Trade-off | ||||
Rec@5 | Rec@10 | Cov@5 | Cov@10 | Nov@5 | Nov@10 | F1@5 | F1@10 | ||
MovieLens | IDS4NR | 0.0752 | 0.1302 | 0.5519 | 0.6613 | 0.4296 | 0.4253 | 0.0506 | 0.0903 |
IDS4NR | 0.0793 | 0.1298 | 0.4295 | 0.5436 | 0.4025 | 0.4237 | 0.0451 | 0.0817 | |
IDS4NR | 0.0767 | 0.1303 | 0.4238 | 0.5303 | 0.3687 | 0.3825 | 0.0413 | 0.0760 | |
IDS4NR | 0.0966 | 0.1576 | 0.3196 | 0.4260 | 0.2803 | 0.3101 | 0.0373 | 0.0699 | |
Music | IDS4NR | 0.1123 | 0.1643 | 0.7363 | 0.8156 | 0.4331 | 0.4330 | 0.0838 | 0.1232 |
IDS4NR | 0.1099 | 0.1591 | 0.7615 | 0.8814 | 0.4151 | 0.4350 | 0.0810 | 0.1230 | |
IDS4NR | 0.1087 | 0.1582 | 0.7329 | 0.8548 | 0.4018 | 0.4212 | 0.0772 | 0.1192 | |
IDS4NR | 0.1521 | 0.2198 | 0.5362 | 0.6791 | 0.2406 | 0.2768 | 0.0633 | 0.1054 | |
Beauty | IDS4NR | 0.0293 | 0.0477 | 0.7919 | 0.8608 | 0.4958 | 0.4905 | 0.0262 | 0.0432 |
IDS4NR | 0.0260 | 0.0432 | 0.8040 | 0.9082 | 0.4703 | 0.4867 | 0.0227 | 0.0398 | |
IDS4NR | 0.0254 | 0.0410 | 0.7879 | 0.8995 | 0.4641 | 0.4804 | 0.0218 | 0.0374 | |
IDS4NR | 0.0478 | 0.0703 | 0.4600 | 0.6336 | 0.2238 | 0.2781 | 0.0201 | 0.0378 |
Thirdly, compared with four cold-start recommendation baselines, our IDS4NR gets remarkable improvements in terms of coverage and novelty while keeping a comparable accuracy performance. These baselines methods either learn a better representation for cold-start items by incorporating the neighbor information in the form of graph neural network (HERS and AGNN), or simulate and adapt to the cold-start scenario by dropout technique (DropoutNet) or meta-learning (MetaEmb), and thus accurately capture the user preference for cold-start items. However, they ignore the long-tail problem under normal circumstances and do not perform well in many cases of novel recommendation.
Finally, for the base model LFM, NCF, and CML, these traditional methods do not take into account the novelty factor, resulting in high accuracy but low coverage and novelty. However, when we integrate them as the backbone layer in our model, we can get dramatic trade-off F1 increases. For example, the relative improvement of F1@5 are 60.73%, 50.61%, and 78.80% on MovieLens, 31.44%, 68.10%, and 150.15% on Music, and 86.27%, 109.09%, and 156.86% on Beauty. These results prove that our IDS4NR can enhance the overall performance for various types of classical methods.
4.3 Detailed Study on IDS4NR
4.3.1 Ablation Study
Given the appealing performance of IDS4NR, we further examine how the individual components in our model contribute to the improved performance on novel recommendation. Concretely, we conduct a series of ablation studies to investigate the effects of explicit intent disentanglement and collaborative-content self-supervision:
-
•
IDS4NR: This variant removes the feature self-supervision module from the IDS4NR framework.
-
•
IDS4NR: Our IDS4NR uses an item novelty score to control the impacts of disentangled representation of popularity and preference intents. This variant is based on IDS4NR and further replaces in Eq. 13 with 0.5 for treating all samples equally.
-
•
IDS4NR: This variant removes the intent disentanglement module completely. To ensure utilizing the content information as the standard IDS4NR, we take the average of collaborative feature and all kinds of attributes of users/items as the input of backbone layer on the basis of IDS4NR.
For clarity, we choose the CML backbone as the representative to perform the ablation and other parameter studies. We first present The ablation results on three datasets are shown in Table III.
Firstly, we investigate the impacts of feature self-supervision module. We find that the overall trade-off performance for IDS4NR decline on all three datasets compared with the standard IDS4NR. This indicates that the feature self-supervision component captures the relationship between the collaborative feature and content feature and well models the user preference for cold-start items, thus it is an essential part of IDS4NR.
Next, we examine the impacts of explicit intent disentanglement module. It is clear that both the coverage and novelty of IDS4NR decrease a lot if compared with IDS4NR. The reason might be that the intent disentanglement without any constraint cannot learn the specific factors explicitly and will limit the ability of learning representations for users’ intents on popular/niche items.
Finally, the overall performance of the variant IDS4NR decreases dramatically along with a few increased accuracy. After removing the intent disentanglement, the model degrades into a naive content-based method, and focuses on the items having similar contents with users’ historical interactions. With the limited recommendation scope, the accuracy of the model may increase a bit but the coverage and novelty performance become extremely poor.
4.3.2 Parameter Study
We investigate the impacts of hyper-parameters in IDS4NR, including the dimension of the latent vector and the weighting factor for feature self-supervision loss. We first fix to 0.01 and vary , and then fix to 50 and vary . The results of IDS4NR with CML as backbone layer for the varying parameter and are shown in Figure 3 and Figure 4, respectively.
We can see from Figure 3 that accuracy and novelty show opposite trends in all datasets. This is consistent with the problem setting and the results in previous studies. Movielens is a small dataset with relatively dense user-item interactions. The model can already learn a good feature representation even with a small dimensionality (=10). In contrast, Music and Beauty are even sparser. The larger dimension improves the representation ability of features with more latent factors of users and items, and thus the model reaches a better accuracy, but its novelty declines at the same time. Hence we set as 50 for a better balance in most cases.
Our IDS4NR takes feature self-supervision learning as an auxiliary task to enhance the collaborative features of cold-start items, and the weighting factor controls the proportion of self-supervision task. As we can seen in Figure 4, with the increase of , the accuracy on different datasets gradually decreases and the novelty increases, and this trend is more obvious when exceeds 0.01. When is small, it can improve the performance for cold start items as an auxiliary task and thus enhances the recommendation novelty. However, when is too large, the self-supervision task dominates the loss and interfere the learning of the main task, leading to a sharp decline of recommendation accuracy.
4.3.3 Analysis on Computational Cost
To investigate the training process of model, we plot the model’s training curves in Figure 5. As can be seen, the loss declines quickly first and gradually slows down. It finally reaches the convergence when the training epoch increases, which proves that our model is stable and easy to train.
We further compare the model’s space complexity by presenting the trainable parameter number of our model and that of base models in Table IV. The parameter number of our IDS4NR and its different base models mainly depends on the user/item embedding matrix. Compared to the base models, our framework introduces additional user/item attributes, so the increased cost lies primarily in the attribute embedding. For the intent disentanglement and self-supervision module in the framework, they need a few weight matrix at a lower cost. Note that NCF has the largest parameter scale on Music and Beauty because it uses the independent embedding layer in its GMF and MLP module, which is equivalent to expanding the dimension of embedding vectors. When we take it as the backbone layer, these two modules are shared by the disentangled user/item representation and thus decrease the parameter number.
Model | Model size () | ||
MovieLens | Music | Beauty | |
LFM | 0.13 | 0.46 | 0.72 |
IDS4NR(+LFM) | 0.57 | 0.78 | 1.19 |
NCF | 0.28 | 0.93 | 1.42 |
IDS4NR(+NCF) | 0.60 | 0.84 | 1.24 |
CML | 0.13 | 0.46 | 0.70 |
IDS4NR(+CML) | 0.52 | 0.76 | 1.16 |
5 Conclusion
In this paper, we propose a new model IDS4NR for novel recommendation. We distinguish our model with existing novelty-oriented methods in two key issues. Firstly, we realize that the varying degrees of users’ intent to popular and niche items are the inherent factors in determining the trade-off between accuracy and novelty in novel recommendation. We then implement this idea with the intent disentanglement integrated end-to-end framework. Secondly, we extend the definition of standard long-tail recommendation to cover the cold-start items. We also propose a self-supervision strategy to solve the missing collaborative feature problem which is a big obstacle when recommending cold-start items. To validate our thoughts, we conduct extensive comparison experiments and deep analysis studies on three real-world datasets. The results prove that our proposed IDS4NR model achieves a new state-of-the-art trade-off performance on the novel recommendation task for both the standard long-tail items and the extended cold-start items.
References
- [1] G. Adomavicius and Y. Kwon, “Improving aggregate recommendation diversity using ranking-based techniques,” IEEE Trans. Knowl. Data Eng., vol. 24, no. 5, pp. 896–911, 2012.
- [2] H. Yin, B. Cui, J. Li, J. Yao, and C. Chen, “Challenging the long tail recommendation,” Proc. VLDB Endow., vol. 5, no. 9, pp. 896–907, 2012.
- [3] Y. Park, “The adaptive clustering method for the long tail problem of recommender systems,” IEEE Trans. Knowl. Data Eng., vol. 25, no. 8, pp. 1904–1915, 2013.
- [4] G. Oestreicher-Singer and A. Sundararajan, “Recommendation networks and the long tail of electronic commerce,” MIS Q., vol. 36, no. 1, pp. 65–83, 2012.
- [5] Z. Zolaktaf, R. Babanezhad, and R. Pottinger, “A generic top-n recommendation framework for trading-off accuracy, novelty, and coverage,” in ICDE, 2018, pp. 149–160.
- [6] K. Lo and T. Ishigaki, “Matching novelty while training: Novel recommendation based on personalized pairwise loss weighting,” in ICDM, 2019, pp. 468–477.
- [7] S. Liu and Y. Zheng, “Long-tail session-based recommendation,” in RecSys, 2020, pp. 509–514.
- [8] Y. Zhang, D. Z. Cheng, T. Yao, X. Yi, L. Hong, and E. H. Chi, “A model of two tales: Dual transfer learning framework for improved long-tail item recommendation,” CoRR, vol. abs/2010.15982, 2020.
- [9] D.-N. Lascu and G. Zinkhan, “Consumer conformity: Review and applications for marketing theory and practice,” Journal of Marketing Theory and Practice, vol. 7, no. 3, pp. 1–12, 1999.
- [10] J. Park and R. Feinberg, “E‐formity: consumer conformity behaviour in virtual communities,” Journal of Research in Interactive Marketing, vol. 4, no. 3, pp. 197–213, 2010.
- [11] K. Kapoor, V. Kumar, L. G. Terveen, J. A. Konstan, and P. R. Schrater, “”i like to explore sometimes”: Adapting to dynamic user novelty preferences,” in RecSys, 2015, pp. 19–26.
- [12] W. Hamilton, Z. Ying, and J. Leskovec, “Inductive representation learning on large graphs,” in NeuIPS, 2017, pp. 1024–1034.
- [13] M. Zhang and Y. Chen, “Inductive matrix completion based on graph neural networks,” in ICLR, 2020.
- [14] Y. Zhu, J. Lin, S. He, B. Wang, Z. Guan, H. Liu, and D. Cai, “Addressing the item cold-start problem by attribute-driven active learning,” IEEE Trans. Knowl. Data Eng., vol. 32, no. 4, pp. 631–644, 2020.
- [15] M. Aharon, O. Anava, N. Avigdor-Elgrabli, D. Drachsler-Cohen, S. Golan, and O. Somekh, “Excuseme: Asking users to help in item cold-start recommendations,” in RecSys, 2015, pp. 83–90.
- [16] J. Zhang, X. Shi, S. Zhao, and I. King, “Star-gcn: Stacked and reconstructed graph convolutional networks for recommender systems,” in IJCAI, 2019, pp. 4264–4270.
- [17] L. Hu, S. Jian, L. Cao, Z. Gu, Q. Chen, and A. Amirbekyan, “HERS: modeling influential contexts with heterogeneous relations for sparse and cold-start recommendation,” in AAAI, 2019, pp. 3830–3837.
- [18] T. Qian, Y. Liang, Q. Li, and H. Xiong, “Attribute graph neural networks for strict cold start recommendation,” IEEE Transactions on Knowledge and Data Engineering, pp. 1–1, 2020.
- [19] G. Guo, J. Zhang, and N. Yorke-Smith, “Trustsvd: Collaborative filtering with both the explicit and implicit influence of user trust and of item ratings,” in AAAI, 2015, pp. 123–125.
- [20] X. He and T.-S. Chua, “Neural factorization machines for sparse predictive analytics,” in SIGIR, 2017, pp. 355–364.
- [21] X. Xin, X. He, Y. Zhang, Y. Zhang, and J. Jose, “Relational collaborative filtering: Modeling multiple item relations for recommendation,” in SIGIR, 2019, pp. 125–134.
- [22] M. Vartak, A. Thiagarajan, C. Miranda, J. Bratman, and H. Larochelle, “A meta-learning perspective on cold-start recommendations for items,” in NeuIPS, 2017, pp. 6904–6914.
- [23] F. Pan, S. Li, X. Ao, P. Tang, and Q. He, “Warm up cold-start advertisements: Improving CTR predictions via learning to learn ID embeddings,” in SIGIR, 2019, pp. 695–704.
- [24] M. Volkovs, G. Yu, and T. Poutanen, “Dropoutnet: Addressing cold start in recommender systems,” in NeuIPS, 2017, pp. 4957–4966.
- [25] I. Higgins, L. Matthey, A. Pal, C. Burgess, X. Glorot, M. Botvinick, S. Mohamed, and A. Lerchner, “beta-vae: Learning basic visual concepts with a constrained variational framework,” in ICLR. OpenReview.net, 2017.
- [26] R. Hamaguchi, K. Sakurada, and R. Nakamura, “Rare event detection using disentangled representation learning,” in CVPR, 2019, pp. 9327–9335.
- [27] J. Ma, C. Zhou, P. Cui, H. Yang, and W. Zhu, “Learning disentangled representations for recommendation,” in NIPS, 2019, pp. 5712–5723.
- [28] Y. Zhang, Z. Zhu, Y. He, and J. Caverlee, “Content-collaborative disentanglement representation learning for enhanced recommendation,” in RecSys. ACM, 2020, pp. 43–52.
- [29] L. Hu, S. Xu, C. Li, C. Yang, C. Shi, N. Duan, X. Xie, and M. Zhou, “Graph neural news recommendation with unsupervised preference disentanglement,” in ACL. Association for Computational Linguistics, 2020, pp. 4255–4264.
- [30] X. Wang, H. Jin, A. Zhang, X. He, T. Xu, and T. Chua, “Disentangled graph collaborative filtering,” in SIGIR. ACM, 2020, pp. 1001–1010.
- [31] J. Ma, C. Zhou, H. Yang, P. Cui, X. Wang, and W. Zhu, “Disentangled self-supervision in sequential recommenders,” in KDD. ACM, 2020, pp. 483–491.
- [32] Y. Zheng, C. Gao, X. Li, X. He, Y. Li, and D. Jin, “Disentangling user interest and conformity for recommendation with causal embedding,” in WWW, 2021.
- [33] M. Venkatesan, “”experimental study of consumer behavior conformity and independence,” Journal of Marketing Research, vol. 3, pp. 384–387, 1966.
- [34] X. Liu, F. Zhang, Z. Hou, Z. Wang, L. Mian, J. Zhang, and J. Tang, “Self-supervised learning: Generative or contrastive,” CoRR, vol. abs/2006.08218, 2020.
- [35] P. Vincent, H. Larochelle, Y. Bengio, and P. Manzagol, “Extracting and composing robust features with denoising autoencoders,” in ICML, vol. 307, 2008, pp. 1096–1103.
- [36] D. P. Kingma and M. Welling, “Auto-encoding variational bayes,” in ICLR, 2014.
- [37] T. N. Kipf and M. Welling, “Variational graph auto-encoders,” in NeurIPS Workshop on Bayesian Deep Learning, 2016.
- [38] A. Razavi, A. van den Oord, and O. Vinyals, “Generating diverse high-fidelity images with VQ-VAE-2,” in NIPS, 2019, pp. 14 837–14 847.
- [39] I. Misra and L. van der Maaten, “Self-supervised learning of pretext-invariant representations,” in CVPR, 2020, pp. 6706–6716.
- [40] A. Radford, L. Metz, and S. Chintala, “Unsupervised representation learning with deep convolutional generative adversarial networks,” in ICLR, 2016.
- [41] J. Wu, X. Wang, F. Feng, X. He, L. Chen, J. Lian, and X. Xie, “Self-supervised graph learning for recommendation,” CoRR, vol. abs/2010.10783, 2020.
- [42] J. Yu, H. Yin, J. Li, Q. Wang, N. Q. V. Hung, and X. Zhang, “Self-supervised multi-channel hypergraph convolutional network for social recommendation,” in WWW, 2021, pp. 413–424.
- [43] K. Zhou, H. Wang, W. X. Zhao, Y. Zhu, S. Wang, F. Zhang, Z. Wang, and J. Wen, “S3-rec: Self-supervised learning for sequential recommendation with mutual information maximization,” in CIKM. ACM, 2020, pp. 1893–1902.
- [44] J. Cao, X. Lin, S. Guo, L. Liu, T. Liu, and B. Wang, “Bipartite graph embedding via mutual information maximization,” in WSDM. ACM, 2021, pp. 635–643.
- [45] X. Xia, H. Yin, J. Yu, Q. Wang, L. Cui, and X. Zhang, “Self-supervised hypergraph convolutional networks for session-based recommendation,” in AAAI, 2021, pp. 4503–4511.
- [46] Y. Hu, Y. Koren, and C. Volinsky, “Collaborative filtering for implicit feedback datasets,” in ICDM, 2008, pp. 263–272.
- [47] Y. Wu, C. DuBois, A. X. Zheng, and M. Ester, “Collaborative denoising auto-encoders for top-n recommender systems,” in WSDM. ACM, 2016, pp. 153–162.
- [48] Y. Koren, R. Bell, and C. Volinsky, “Matrix factorization techniques for recommender systems,” Computer, vol. 42, no. 8, pp. 30–37, 2009.
- [49] X. He, L. Liao, H. Zhang, L. Nie, X. Hu, and T. Chua, “Neural collaborative filtering,” in WWW, 2017, pp. 173–182.
- [50] C. Hsieh, L. Yang, Y. Cui, T. Lin, S. J. Belongie, and D. Estrin, “Collaborative metric learning,” in WWW, 2017, pp. 193–201.
- [51] Z. Cheng, Y. Ding, X. He, L. Zhu, X. Song, and M. S. Kankanhalli, “A^3ncf: An adaptive aspect attention model for rating prediction,” in IJCAI, 2018, pp. 3748–3754.
- [52] C. Chen, M. Zhang, Y. Liu, and S. Ma, “Neural attentional rating regression with review-level explanations,” in WWW, 2018, pp. 1583–1592.
- [53] C. Chen, M. Zhang, Y. Zhang, Y. Liu, and S. Ma, “Efficient neural matrix factorization without sampling for recommendation,” ACM Trans. Inf. Syst., vol. 38, no. 2, pp. 14:1–14:28, 2020.
- [54] G. Box and D. Meyer, “An analysis for unreplicated fractional factorials,” Technometrics, vol. 28, pp. 11–18, 02 1986.
- [55] C. Chen, M. Zhang, Y. Zhang, W. Ma, Y. Liu, and S. Ma, “Efficient heterogeneous collaborative filtering without negative sampling for recommendation,” in AAAI, 2020, pp. 19–26.
- [56] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” in ICLR, 2015.