This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

{textblock}

10(3,0.4)

International Workshop on Industrial Recommendation Systems

in conjunction with ACM KDD 2020

Multi-sided Exposure Bias in Recommendation

Himan Abdollahpouri [email protected] University of Colorado BoulderBoulderUSA  and  Masoud Mansoury [email protected] Eindhoven University of TechnologyEindhovenThe Netherlands
Abstract.

Academic research in recommender systems has been greatly focusing on the accuracy-related measures of recommendations. Even when non-accuracy measures such as popularity bias, diversity, and novelty are studied, it is often solely from the users’ perspective. However, many real-world recommenders are often multi-stakeholder environments in which the needs and interests of several stakeholders should be addressed in the recommendation process. In this paper, we focus on the popularity bias problem which is a well-known property of many recommendation algorithms where few popular items are over-recommended while the majority of other items do not get proportional attention and address its impact on different stakeholders. Using several recommendation algorithms and two publicly available datasets in music and movie domains, we empirically show the inherent popularity bias of the algorithms and how this bias impacts different stakeholders such as users and suppliers of the items. We also propose metrics to measure the exposure bias of recommendation algorithms from the perspective of different stakeholders.

Multi-sided platforms, Recommender systems, Popularity bias, Multi-stakeholder recommendation

1. Introduction

Popularity bias is a well-known phenomenon in recommender systems: popular items are recommended even more frequently than their popularity would warrant, amplifying the long-tail effect already present in many recommendation domains. Prior research has examined the impact of this bias on some properties of the recommenders such as aggregate diversity (aka catalog coverage) (Vargas and Castells, 2011; Adomavicius and Kwon, 2011). One of the consequences of the popularity bias is disfavoring less popular items where the recommendations are not fair in terms of the amount of exposure they give to different items with varying degree of popularity: an exposure bias. However, as we discuss in (Abdollahpouri et al., 2020), many recommender systems are multi-stakeholder environments in which the needs and interests of multiple stakeholders should be taken into account in the implementation and evaluation of such systems.

In many multi-stakeholer recommenders as described in (Abdollahpouri et al., 2020) two main stakeholders (or what often is being referred to as sides in multi-sided platforms (Hagiu and Wright, 2015) ) can be identified: consumers (aka users) and suppliers. For instance, in a music platform such as Spotify, on one side there are users who get recommendations for songs in which they are interested and, on the other side, there are artists whose songs are being recommended to different users. The popularity bias can be investigated from both sides’ perspective.

Regarding the users, not everyone has the same level of interest in popular items. In the music domain as an example, some users might be interested in internationally popular artists such as Drake, Beyoncé, or Ed Sheeran and some might be more interested in artists from their own culture that might not necessarily have the same popularity as the aforementioned artists (such as the Iranian musician Kayhan Kalhor) or generally they prefer certain type of music that might not be popular among the majority of other users (such as country music). With that being said, we expect the personalization to handle this difference in taste but as we will see in section  4.1 that is certainly not the case.

The suppliers also do not have the same level of popularity. In many recommendation domains including movies, music, or even house sharing, few suppliers have a large audience while the majority of others may not be as popular though they still might have their fair share of audience. Now the question is, do recommender systems let different suppliers with varying degree of popularity reach their desired audience? Again, the short answer is No as we will see more details in section  4.2.

Investigating the impact of recommendation algorithms on the exposure bias on both users and suppliers is the focus of this paper. We study several recommendation models in terms of their inherent popularity bias and propose metrics that can measure such impact.

2. Experimental Setting

2.1. Data

We have used two publicly available datasets for our experiments. We needed datasets that either had information about the supplier of the items or we could extract them. We found two: the first one is a sample of the Last.fm (LFM-1b) dataset (Schedl, 2016) used in (Dominik et al., 2019). The dataset contains user interactions with songs (and the corresponding albums). We used the same methodology in (Dominik et al., 2019) to turn the interaction data into rating data using the frequency of the interactions with each item (more interactions with an items will result in higher ratings). In addition, we used albums as the items to reduce the size and sparsity of the item dimension, therefore the recommendation task is to recommend albums to users. We considered the artists of each album as the supplier. Each album is associated with an artist. Artists could have multiple albums. We removed users with less than 20 ratings so only consider users for which we have enough data. The resulting dataset contains 274,707 ratings by 2,697 users to 6,006 albums from 1,998 artists.

The second dataset is the MovieLens 1M dataset111Our experiments showed similar results on MovieLens 20M, and so we continue to use ML1M for efficiency reasons.. This dataset does not have the information about the suppliers. We considered the director of each movie as the supplier of that movie and we extracted that information from the IMDB API. Total number of ratings in the MovieLens 1M data is 1,000,209 given by 6,040 users to 3,706 items. Overall, we were able to extract the director information for 3,043 movies reducing the ratings to 995,487. The total number of directors is 831.

We used 80% of each dataset as our training set and the other 20% for the test.

2.2. Algorithms

For our analysis, we have used three personalized recommendation algorithms: Biased Matrix Factorization (Biased-MF) (Koren et al., 2009), User-based Collaborative Filtering (User-CF) (Aggarwal, 2016), and Item-based Collaborative Filtering (Item-CF) (Deshpande and Karypis, 2004). All algorithms are tuned to achieve their best performance in terms of precision. The size of the recommendation lists for each user is set to 10. We also include a non-personalized Most-popular algorithm which only recommends the 10 most popular items to every user given they have not rated the items before. We used LibRec (Guo et al., 2015) and librec-auto (Mansoury et al., 2018) for running the algorithms.

3. Popularity Bias

Skew in wealth distribution is well-known: The richest 10% of adults in the world own 85% of global household wealth while the bottom half collectively owns barely 1% 222https://www.wider.unu.edu/publication/global-distribution-household-wealth; in recommender systems a similar problem exists: a small number of popular items appear frequently in user profiles and a much larger number of less popular items appear rarely. This bias can originate from two different sources: data and algorithms.

3.1. Bias in Data

\SetFigLayout

32

Refer to caption
(a) MovieLens
Refer to caption
(b) Last.fm
Figure 1. The Long-tail Distribution of Rating data. HH items are a small number of very popular items that take up around 20% of the entire ratings. TT is the larger number of less popular items which collectively take up roughly 20% of the ratings, and MM includes those items in between that receive around 60% of the ratings, collectively.

Rating data is generally skewed towards more popular items. Figure  1 shows the percentage of users who rated different items in MovieLens and Last.fm datasets: the popularity of each item. Items are ranked from the most popular to the least with the most popular item being on the far left on the x-axis. Three different groups of items can be seen in these plots: HH which represents few items that are very popular and take up around 20% of the entire ratings according to the Pareto Principle (Sanders, 1987). TT is the larger number of less popular items which collectively take up roughly 20% of the ratings, and MM includes those items in between that receive around 60% of the ratings, collectively. The curve has a long-tail shape (Anderson, 2006; Celma and Cano, 2008) indicating few popular items are taking up the majority of the ratings while many other items on the far right of the curve have not received much attention. This type of distribution can be found in many other domains such as e-commerce where few products are best-sellers, online dating where few profiles attract the majority of the attention, social networking platforms where few users have millions of followers, to name a few.

The bias in rating data could be due to two different reasons:

  • External Bias: Some items and products are inherently more popular than others even outside of the recommender systems and in the real world. For instance, even before the music streaming services emerge, there were always few artists that were nationally or internationally popular such as Shakira, Jennifer Lopez, or Enrique Iglesias. As a result of this external bias (or tendency) towards popular artists, users also often listen to those artists more on streaming services and hence they get more user interactions.

  • Feedback Loop: Since the recommendation algorithms have a higher tendency towards recommending popular items, these items have a higher chance to be recommended to the users and hence garnering a larger number of interactions from the users. When these interactions are logged and stored, the popularity of those items in the rating data increases since they get more and more interactions over time (Jiang et al., 2019; Sinha et al., 2016).

3.2. Bias in Algorithm

Due to this imbalance property of the rating data, often algorithms inherit this bias and, in many cases, intensify it by over-recommending the popular items and, therefore, giving them a higher opportunity of being rated by more users: the rich get richer and the poor get poorer (Abdollahpouri et al., 2017).

Figure  2 shows the percentage of users who have rated an item on the x-axis (the popularity of an item in the data) and the percentage of users who received that item in their recommendations using four different recommendation algorithms Biased-MF, User-CF, Item-CF, and Most-popular in both datasets. The plots aim at showing the correlation between the popularity of an item in the rating data versus how often it is recommended to different users. It is clear that in all four algorithms, many items are either never recommended or just rarely recommended. Among the three personalized algorithms, Item-CF and User-CF show the strongest evidence that popular items are recommended much more frequently than the others. In fact, they are recommended to a much greater degree than even what their initial popularity warrants. For instance, the popularity of some items have been amplified from roughly 0.4 to 0.7 indicating a 75% increase. Both Item-CF and User-CF are over-promoting popular items (items on the right side of the x-axis) while significantly hurting other items by not recommending them proportionate to what their popularity in data warrants. In fact, the vast majority of the items on the left are never recommended indicating an extreme bias of the algorithms towards popular items and against less popular ones. Biased-MF does not show a positive correlation between popularity in data and in recommendations although some items are still over-recommended (have much higher popularity in recommendations versus what they had in rating data). However, this over-recommendation is not concentrated on only popular items and some items from lower popularity values are also over-recommended. Most-popular obviously shows the strongest bias but, unlike the other three, it is not a personalized method333If an item in the top 10 is already rated by a user it will be replaced by another popular item. That is why there is an inflection point in the scatter plot for this algorithm on MovieLens dataset..

\SetFigLayout

21

Refer to caption
(a) MovieLens
Refer to caption
(b) Last.fm
Figure 2. Item popularity versus recommendation popularity

4. Multi-sided Exposure Bias

We measure the impact of popularity bias on different stakeholders in terms of exposure: how the bias in algorithms prevents users to be exposed to the appropriate range of items and also how it stops items from different suppliers to be exposed to their desired audience.

\SetFigLayout

21

Refer to caption
(a) MovieLens
Refer to caption
(b) Last.fm
Figure 3. Users’ Propensity towards item popularity

4.1. Exposure Bias From the Users’ Perspective

\SetFigLayout

21

Refer to caption
(a) MovieLens
Refer to caption
(b) Last.fm
Figure 4. Users’ centric view of popularity bias

Not every user is equally interested in popular items. In cinema, for instance, some might be interested in movies from Yasujiro Ozu, Abbas Kiarostami, or John Cassavetes, and others may enjoy more mainstream directors such as James Cameron or Steven Spielberg. Figure  3 shows the ratio of rated items for three item categories HH, MM, and TT in the profiles of different users in the MovieLens 1M and Last.fm datasets. Users are sorted from the highest interest towards popular items to the least and divided into three equal-sized bins G={G1G=\{G_{1},G2G_{2},G3}G_{3}\} from most popularity-focused to least. The y-axis shows the proportion of each user’s profile devoted to different popularity categories. The narrow blue band shows the proportion of each users profile that consists of popular items (i.e. the HH category), and its monotone decrease reflects the way the users are ranked. Note, however, that all groups have rated many items from the middle (green) and tail (red) parts of the distribution.

The plots in Figure  4 are parallel to Figure  3, with the users ordered by their popularity interest, but now the y-axis shows the proportion of recommended items using different algorithms from different item popularity categories. The difference with the original user profiles in rating data especially in the case of Most-popular, Item-CF, and User-CF is stark where the users’ profiles are rich in diverse popularity categories, the generated recommendations are nowhere close to what the user has shown interest at. In fact, in Item-CF almost 100% of the recommendations are from the head category, even for the users with the most niche-oriented profiles. Tail items do not appear at all. We demonstrated here that popularity bias in the algorithm is not just a problem from a global, system, perspective. It is also a problem from the user perspective (Abdollahpouri et al., 2019): users are not getting recommendations that reflect the diversity of their profiles and the users with the most niche tastes (G3G_{3}) are the most poorly served.

To measure the impact of popularity bias on users, we need to compare two lists together: the list of the items in a users’ profile and the list of the items recommended to the user. To do this comparison, we need to compute a discrete probability distribution PP for each user uu, reflecting the popularity of the items found in their profile ρu\rho_{u} over each item category cCc\in C (in this paper, C={H,M,T}C=\{H,M,T\}). We also need a corresponding distribution QQ over the recommendation list given to user uu, u\ell_{u}, indicating what popularity categories are found among the list of recommended items. In Steck (Steck, 2018) and Kaya et al. (Kaya and Bridge, 2019) where, unlike this paper, the calibration is done according to genres, the rating data is binarized to reflect just “liked” items. We are retaining the original ratings when computing over user profiles and, instead, using Vargas et al.’s (Vargas and Castells, 2013) measure of category propensity which has the users’ rating component in it. Note that unlike the genre labels in (Steck, 2018) where it is possible for a movie to have multiple genres, each item only has one popularity category.

(1) p(c|u)=iρur(u,i)𝟙(ic)cjCiρur(u,i)𝟙(icj)p(c|u)=\frac{\sum_{i\in\rho_{u}}r(u,i)\mathbbm{1}(i\in c)}{\sum_{c_{j}\in C}\sum_{i\in\rho_{u}}r(u,i)\mathbbm{1}(i\in c_{j})}
(2) q(c|u)=iu𝟙(ic)cjCiu𝟙(icj)q(c|u)=\frac{\sum_{i\in\ell_{u}}\mathbbm{1}(i\in c)}{\sum_{c_{j}\in C}\sum_{i\in\ell_{u}}\mathbbm{1}(i\in c_{j})}

𝟙(.)\mathbbm{1}(.) is the indicator function returning zero when its argument is False and 1 otherwise. Once we have PP and QQ, we can measure the distance using Jensen–Shannon divergence, which is a modification of KL-Divergence that has two useful properties which KL-divergence lacks: 1) it is symmetric: 𝔍(P,Q)=𝔍(Q,P)\mathfrak{J}(P,Q)=\mathfrak{J}(Q,P) and 2) it has always a finite value even when there is a zero in QQ. For our application, it is particularly important that the function be well-behaved at the zero point since it is possible for certain item categories to have zero items in them in the recommendation list.

(3) 𝔍(P,Q)=12KL(P,M)+12KL(Q,M),M=12(P+Q)\mathfrak{J}(P,Q)=\frac{1}{2}KL(P,M)+\frac{1}{2}KL(Q,M),\;\;M=\frac{1}{2}(P+Q)

where KLKL is the KL divergence.

We introduce the following metric to quantify the exposure bias from the users’ perspective.

Users’ Popularity propensity Deviation (UPD):

Having different groups of users with varying degree of interest towards popular items, we measure the average deviation of the recommendations given to the users in each group in terms of item popularity. More formally,

(4) UPD=gGug𝔍(P(ρu),Q(u))|g||G|UPD=\frac{\sum_{\textsl{g}\in G}\frac{\sum_{u\in\textsl{g}}\mathfrak{J}(P(\rho_{u}),Q(\ell_{u}))}{|\textsl{g}|}}{|G|}

where |g||\textsl{g}| is the number of users in group g and |G||G| is the number of user groups. UPDUPD can be also seen as the average miscalibration of the recommendations from the perspective of users in different groups.

4.2. Exposure Bias From the Suppliers’ Perspective

As noted above, multi-stakeholder analysis in recommendation also includes providers or as termed here suppliers, “those entities that supply or otherwise stand behind the recommended items” (Abdollahpouri et al., 2020). We can think of many different kinds of contributors standing behind a particular movie: for the purposes of this paper, we will focus on movie directors. In the music domain, often the artists are considered as the suppliers of the songs (Mehrotra et al., 2018). In this paper, we also make the same assumption on Last.fm dataset.

We create three supplier groups S={S1,S2,S3}S=\{S_{1},S_{2},S_{3}\}: S1S_{1} represents few popular suppliers whose items take up 20% of the ratings, S2S_{2} are larger number of suppliers with medium popularity whose items take up around 60% of the ratings, and S3S_{3} are the less popular suppliers whose items get 20% of the ratings. Figure  5 shows the rank of different directors in MovieLens and artists in Lats.fm datasets by popularity and the corresponding recommendation results from different algorithms. The recommendations have amplified the popularity of the popular suppliers (the ones on the extreme left) while suppressing the less popular ones dramatically. Strikingly, using Item-CF, movies from just 3 directors in S1S_{1} (less than 0.4% of the suppliers here) take up 50% of the recommendations produced, while items items from the S3S_{3} are seeing essentially zero recommendations.

To quantify the impact of popularity bias on the supplier exposure, we measure the amount of deviation different groups of suppliers experience in terms of exposure. In other words, the degree of over-recommendation or under-recommendation of suppliers from different groups.

Supplier Popularity Deviation (SPD):

(5) SPD=sS|q(s)p(s)||S|SPD=\frac{\sum_{s\in S}|q(s)-p(s)|}{|S|}
q(s)=uUiu𝟙(A(i)s)n×|U|,p(s)=uUiρu𝟙(A(i)s)uU|ρu|q(s)=\frac{\sum_{u\in U}\sum_{i\in\ell_{u}}\mathbbm{1}(A(i)\in s)}{n\times|U|},\;p(s)=\frac{\sum_{u\in U}\sum_{i\in\rho_{u}}\mathbbm{1}(A(i)\in s)}{\sum_{u\in U}|\rho_{u}|}

where q(s)q(s) is the ratio of recommendations that come from items of supplier group ss. p(s)p(s) is the ratio of ratings that come from items of supplier group ss. UU is the set of users and A(.)A(.) is a mapping function that returns the supplier for each item.

In fact, 1SPD1-SPD can be considered as Proportional Supplier Fairness (Kusner et al., 2017) since it measures how the items from different supplier groups are exposed to different users proportional to their popularity in rating data.

\SetFigLayout

21

Refer to caption
(a) MovieLens
Refer to caption
(b) Last.fm
Figure 5. Suppliers’ centric view of popularity bias
\SetFigLayout

21

Refer to caption
(a) UPD vs SPD
Refer to caption
(b) UPD vs Precision
Refer to caption
(c) SPD vs Coverage
Figure 6. The relationship between different metrics (MovieLens)
\SetFigLayout

21

Refer to caption
(a) UPD vs SPD
Refer to caption
(b) UPD vs Precision
Refer to caption
(c) SPD vs Coverage
Figure 7. The relationship between different metrics (Last.fm)

5. Discussion and Future Work

One important consideration regarding the deviation of popularity from the users’ perspective (UPDUPD) and suppliers’ perspective (SPDSPD) is how these two metrics behave with respect to one another. In other words, whether calibrating the recommendations for the users in terms of popularity would make the experience for the suppliers also better. Figures  6(a) and  7(a) show the connection between these two in the MovieLens and Last.fm datasets, respectively. We can see that, generally, the lower the UPDUPD (better calibration from the users’ perspective) the lower the SPDSPD (better proportional fairness for the suppliers) will be. This indicates the advantage of optimizing for the popularity calibration in the recommendations since it will also benefit the suppliers.

Another important finding is the fact that more accuracy does not necessarily leads to a better calibration and vice versa as can be seen from Figures  6(b) and  7(b). For instance, on Last.fm, Most-popular and Biased-MF have roughly equal precision but the UPDUPD for Biased-MF is significantly lower (better) than Most-popular. The reason is, the majority of the items are usually not very popular and, therefore, exclusively recommending popular items would not match the original distribution of the ratings (Figure  1) which leads to high UPDUPD. In addition, if an algorithm randomly recommends items, the likelihood of having items from MM and TT increases compared to many personalized recommendations where they suffer from popularity bias. However, improvement in UPDUPD by randomly recommending items would happen under the cost of having an extremely low precision. Therefore, in practice, both UPDUPD and PrecisionPrecision should be taken into account simultaneously in the optimization process.

The relationship between catalog coverage and supplier popularity deviation is also interesting to look at. Will an algorithm that covers many items necessarily have lower (better) SPDSPD? The answer is No as can be seen from Figures  6(c) and  7(c). The reason is, item coverage (aka aggregate diversity) is the number of unique items that an algorithm has recommended even if an item is only recommended only once. However, what matters in SPDSPD is giving appropriate exposure to the items from different suppliers proportional to their popularity.

The impact of UPDUPD and SPDSPD on the success of a real-world recommender system can be evaluated using online A/B testing and see how metrics such as user engagement, retention, and also the satisfaction of suppliers (e.g. artists in the music recommendation platforms) would be influenced.

6. Conclusion

Recommender systems are multi-stakeholder environments; in addition to the users, some other stakeholders such as the supplier of the items also benefit from the recommendation of their items and gaining a larger audience. The algorithmic popularity bias can negatively impact both users and suppliers on a recommender system platform. In this paper, we demonstrated the severity of the popularity bias impact on different sides of a recommender system using several recommendation algorithms on two datasets. We also proposed metrics to quantify the exposure bias from the perspective of both the users and suppliers. Our experiments showed that when the recommendations are calibrated for the users in terms of popularity (lower UPDUPD), it will also benefit the suppliers of the recommendations by giving them proportional exposure (lower SPDSPD). We believe, it is extremely crucial for the recommender systems researchers to see the implications of real-world recommenders where a single-stakeholder focus might not address all the complexities.

References

  • (1)
  • Abdollahpouri et al. (2020) Himan Abdollahpouri, Gediminas Adomavicius, Robin Burke, Ido Guy, Dietmar Jannach, Toshihiro Kamishima, Jan Krasnodebski, and Luiz Pizzato. 2020. Multistakeholder recommendation: Survey and research directions. User Modeling and User-Adapted Interaction 30 (2020), 127–158. Issue 1.
  • Abdollahpouri et al. (2017) Himan Abdollahpouri, Robin Burke, and Bamshad Mobasher. 2017. Controlling Popularity Bias in Learning to Rank Recommendation. In Proceedings of the 11th ACM conference on Recommender systems. ACM, 42–46.
  • Abdollahpouri et al. (2019) Himan Abdollahpouri, Masoud Mansoury, Robin Burke, and Bamshad Mobasher. 2019. The unfairness of popularity bias in recommendation. In RecSys Workshop on Recommendation in Multistakeholder Environments (RMSE).
  • Adomavicius and Kwon (2011) Gediminas Adomavicius and YoungOk Kwon. 2011. Improving aggregate recommendation diversity using ranking-based techniques. IEEE Transactions on Knowledge and Data Engineering 24, 5 (2011), 896–911.
  • Aggarwal (2016) Charu C Aggarwal. 2016. Neighborhood-based collaborative filtering. In Recommender systems. Springer, 29–70.
  • Anderson (2006) Chris Anderson. 2006. The long tail: Why the future of business is selling more for less. Hyperion.
  • Celma and Cano (2008) Òscar Celma and Pedro Cano. 2008. From hits to niches?: or how popular artists can bias music recommendation and discovery. In Proceedings of the 2nd KDD Workshop on Large-Scale Recommender Systems and the Netflix Prize Competition. ACM, 5.
  • Deshpande and Karypis (2004) M. Deshpande and G. Karypis. 2004. Item-Based Top-N Recommendation Algorithms. ACM Transactions on Information Systems 22, 1 (2004), 143–177.
  • Dominik et al. (2019) Kowald Dominik, Schedl Markus, and Lex Elisabeth. 2019. The Unfairness of Popularity Bias in Music Recommendation: A Reproducibility Study. arXiv preprint arXiv:1912.04696 (2019).
  • Guo et al. (2015) Guibing Guo, Jie Zhang, Zhu Sun, and Neil Yorke-Smith. 2015. LibRec: A Java Library for Recommender Systems. In UMAP Workshops.
  • Hagiu and Wright (2015) Andrei Hagiu and Julian Wright. 2015. Multi-sided platforms. International Journal of Industrial Organization 43 (2015), 162–174.
  • Jiang et al. (2019) Ray Jiang, Silvia Chiappa, Tor Lattimore, András György, and Pushmeet Kohli. 2019. Degenerate feedback loops in recommender systems. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society. 383–390.
  • Kaya and Bridge (2019) Mesut Kaya and Derek Bridge. 2019. A comparison of calibrated and intent-aware recommendations. In Proceedings of the 13th ACM Conference on Recommender Systems. 151–159.
  • Koren et al. (2009) Y. Koren, R. Bell, and C. Volinsky. 2009. Matrix factorization techniques for recommender systems. Computer 42, 8 (2009), 30–37.
  • Kusner et al. (2017) Matt J Kusner, Joshua Loftus, Chris Russell, and Ricardo Silva. 2017. Counterfactual fairness. In Advances in Neural Information Processing Systems. 4066–4076.
  • Mansoury et al. (2018) Masoud Mansoury, Robin Burke, Aldo Ordonez-Gauger, and Xavier Sepulveda. 2018. Automating recommender systems experimentation with librec-auto. In Proceedings of the 12th ACM Conference on Recommender Systems. ACM, 500–501.
  • Mehrotra et al. (2018) Rishabh Mehrotra, James McInerney, Hugues Bouchard, Mounia Lalmas, and Fernando Diaz. 2018. Towards a fair marketplace: Counterfactual evaluation of the trade-off between relevance, fairness & satisfaction in recommendation systems. In Proceedings of the 27th acm international conference on information and knowledge management. 2243–2251.
  • Sanders (1987) Robert Sanders. 1987. The Pareto principle: its use and abuse. Journal of Services Marketing (1987).
  • Schedl (2016) Markus Schedl. 2016. The lfm-1b dataset for music retrieval and recommendation. In Proceedings of the 2016 ACM on International Conference on Multimedia Retrieval. 103–110.
  • Sinha et al. (2016) Ayan Sinha, David F. Gleich, and Karthik Ramani. 2016. Deconvolving feedback loops in recommender systems. In Advances in neural information processing systems. 3243–3251.
  • Steck (2018) Harald Steck. 2018. Calibrated recommendations. In Proceedings of the 12th ACM Conference on Recommender Systems. ACM, 154–162.
  • Vargas and Castells (2011) Saúl Vargas and Pablo Castells. 2011. Rank and Relevance in Novelty and Diversity Metrics for Recommender Systems. In Proceedings of the Fifth ACM Conference on Recommender Systems (Chicago, Illinois, USA) (RecSys ’11). ACM, New York, NY, USA, 109–116. https://doi.org/10.1145/2043932.2043955
  • Vargas and Castells (2013) Saúl Vargas and Pablo Castells. 2013. Exploiting the diversity of user preferences for recommendation. In Proceedings of the 10th conference on open research areas in information retrieval. 129–136.