This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.


On Three-Layer Data Markets

Alireza Fallah University of California, Berkeley    Michael I. Jordan    Ali Makhdoumi Fuqua School of Business, Duke University    Azarakhsh Malekian Rotman School of Management, University of Toronto.
Abstract

We study a three-layer data market comprising users (data owners), platforms, and a data buyer. Each user benefits from platform services in exchange for data, incurring privacy loss when their data, albeit noisily, is shared with the buyer. The user chooses platforms to share data with, while platforms decide on data noise levels and pricing before selling to the buyer. The buyer selects platforms to purchase data from. We model these interactions via a multi-stage game, focusing on the subgame Nash equilibrium. We find that when the buyer places a high value on user data (and platforms can command high prices), all platforms offer services to the user who joins and shares data with every platform. Conversely, when the buyer’s valuation of user data is low, only large platforms with low service costs can afford to serve users. In this scenario, users exclusively join and share data with these low-cost platforms. Interestingly, increased competition benefits the buyer, not the user: as the number of platforms increases, the user utility does not necessarily improve while the buyer utility improves. However, increasing the competition improves the overall utilitarian welfare. Building on our analysis, we then study regulations to improve the user utility. We discover that banning data sharing maximizes user utility only when all platforms are low-cost. In mixed markets of high- and low-cost platforms, users prefer a minimum noise mandate over a sharing ban. Imposing this mandate on high-cost platforms and banning data sharing for low-cost ones further enhances user utility.

1 Introduction

In the digital era, online platforms have become integral to our daily lives, offering a plethora of services that range from social networking to personalized shopping experiences. While these platforms provide convenience and connectivity, they also engage in extensive data collection practices. Users’ interactions with these platforms generate vast amounts of data, encompassing personal preferences, behaviors, and other sensitive information. This data, a valuable commodity, is often shared with third-party buyers.

With the advent of new functionalities by information technology companies such as Apple or Google that reduce or eliminate tracking via third-party cookies, there has been an increased reliance on first-party data, which is information gathered by a company from its own user base. For example, as elaborated in Cross [2023b], major retailers such as BestBuy and Walmart capitalize on this data through loyalty programs, customer purchase records, and subscriptions. They do this by selling the data to advertisers who aim to place targeted ads on the retailers’ websites or across various online platforms. In a similar vein, Cross [2023a] notes that Mastercard, a leading payment technology firm, trades cardholder transaction information via third-party online data marketplaces and its internal “Data and Services” division. This practice grants numerous entities access to extensive consumer data and insights. For instance, their Intelligent Targeting service allows businesses to leverage “Mastercard 360 data insights” to create and execute advertising campaigns that specifically target potential “high-value” customers.

In this complex digital ecosystem, users face the dilemma of benefiting from the services provided by these platforms while potentially compromising their privacy. Platforms, on the other hand, must balance the profitability derived from data sharing with the need to maintain user trust and satisfaction. This interplay between users’ privacy concerns and the economic incentives of platforms raises important questions about the sustainability of current data market practices. Notable regulatory actions have been taken to improve user experiences. For instance, the General Data Protection Regulation (GDPR) and the California Consumer Privacy Act (CCPA) have been introduced to protect user privacy and to limit the amount of data shared with the platforms. Although, at first sight, this helps users, there have been several studies pointing to its detrimental impact both on platforms and users. This is because these regulations have led to a more concentrated market structure by limiting competition in data markets. This results from the regulation’s restrictions on data sharing, which may slow down data-based innovations and services. For instance, studies have shown that GDPR has led to a significant reduction in the number of available apps in the market, with new app entries falling by half following the regulation’s implementation (Janssen et al. [2022]). This decline in market entries indicates a potential loss of innovation as new and smaller players find it increasingly difficult to compete under these regulations. This effect is particularly detrimental to small companies that rely on these data synergies for innovation and market growth (see also Gal and Aviv [2020]).

In this paper, we aim to understand the user and platform strategies that arise when considering the collection and usage of user data. We also explore regulatory frameworks that could be implemented to enhance user utility while maintaining the benefits provided by online platforms.

1.1 Our model and main results

We adopt an analytical-modeling-based approach and develop a framework to examine the choices of both users and platforms, taking into account users’ privacy concerns. We also explore regulations and policies that could alleviate these concerns. In our model, multiple platforms interact with a user who owns data and seeks the services of these platforms, as well as with a third-party buyer interested in purchasing the user’s data from the platforms. Throughout our study, we consider MasterCard as an example of such a platform. The user represents an individual in a specific population segment with particular preferences, and the buyer is conceptualized as a third-party advertiser aiming to understand and leverage these user preferences for its own benefit.

The utility of each platform is determined by three factors: the gains obtained from using user data to provide services, potential payments received from buyers purchasing this data, and the costs incurred in providing services to the user. Let us highlight the cost of services, as it plays an important role in our analysis. It is particularly important to distinguish between high-cost and low-cost platforms. The low-cost platforms are those that do not necessarily need to sell data to stay in the market, as their gains from the service already cover their costs of operation. In our leading examples, low-cost platforms are exemplified by payment-technology companies or big retailers, where either their operational costs per user are low, or their main source of revenue is not derived from selling data. Conversely, for high-cost platforms, the gains from service provision are not enough to offset their operational costs. Hence, they require a minimum level of data selling to be incentivized to enter the market (i.e., to provide services to the user). Examples of these platforms typically include smaller firms or companies that heavily rely on data collection for advertising to ensure their continued operation.

We model the interaction between the players as a multi-stage game in which, in the first stage, the platforms decide whether they want to provide services to the user and the privacy level they want to deliver to the data buyer (so that the user is incentivized to share). We model the privacy level that the platform provides as a noise level added to the user data. Adding a suitable level of noise is common practice in, for example, the analysis of census data Abowd [2018] and is theoretically justified in the literature on differential privacy Dwork et al. [2014].

In the second stage of the game, the user decides which platform to join. The user’s utility is the service quality of the platform, proportional to the amount of user information revealed to the platform minus the user’s privacy loss. The privacy loss refers to the amount of information all platforms reveal to the data buyer. In the third and fourth stages, the platforms determine the price to charge the data buyer, and the buyer decides from which platforms to purchase. The data buyer’s utility is the user’s leaked information from all platforms minus the payments made to acquire this information. Again, the data buyer can be viewed as a third-party advertiser in our example mentioned above who pays the platforms to obtain user data. They then use this data for potential gain, with their gain being proportional to the accuracy of the user data obtained.

Equilibrium characterization and insights:

We adopt the subgame Nash equilibrium as our equilibrium concept and use backward induction to characterize it. Finding the equilibrium of this game is complex mainly because the platforms’ choice of noise level not only directly impacts the user’s utility but also changes the data buyer’s decision, which indirectly impacts the user’s utility. Also, the competition among platforms further complicates the dynamic of the game. For instance, a platform may still be incentivized to sell data at a noise level higher than needed for the user to be willing to share their data. At first glance, this may seem unreasonable as it reduces the quality of the data that they can sell, and hence, the data buyer would pay less for it. However, increasing the privacy could persuade the user to forgo other platforms that offer less stringent privacy guarantees and only use this platform’s service. As a result, this can give a monopoly to the platform in terms of access to the user’s data, and hence, the data buyer would end up paying more to them.

We start our analysis of equilibrium by focusing on the case of two platforms. This setting allows us to infer insights about equilibrium properties that are applicable to more general cases. Here, we identify up to three primary regimes, depending on the data buyer’s valuation of user data acquisition, denoted by β\beta in our model. When β\beta falls below a certain threshold, neither of the two platforms enters the market due to insufficient payment to cover service costs. This regime ceases to exist when at least one platform is low-cost, meaning it can cover service costs solely through service provision without needing to sell data. Conversely, when β\beta is sufficiently high, both platforms will participate in the market. Interestingly, in this case, and when the user is privacy-conscious and requires a minimum non-zero level of noise for data sharing, the user utility is equal to zero at equilibrium. However, as the data buyer compensates each platform at the marginal value of the data, their utility remains positive. Finally, there exists an intermediate regime of β\beta, where only the lower-cost platform engages in the market. Here, both the user and the data buyer utilities are zero at equilibrium, indicating that having both platforms in the market to compete benefits the data buyers but not the user.

We demonstrate that these findings extend to the general scenario with K>2K>2 platforms. Specifically, we establish that when β\beta exceeds a certain threshold, signifying the data buyer’s willingness to pay more for the user data, all platforms will enter the market and provides services to the user. Conversely, when β\beta is below this threshold, only low-cost platforms will participate, while high-cost ones will abstain. Although competition does not favor the user, it results in positive utility for the data buyer. Moreover, the overall utilitarian welfare increases proportionally with the number of participating platforms.

Regulation and policy design:

Building on our equilibrium characterization, we then study regulations that aim to improve user utility. Here, our analysis provides three main insights. Firstly, we consider a ban on platforms sharing user data with the data buyers. This regulation only maximizes the user utility if all the platforms are low-cost. Specifically, under this regulation, users obtain the best of both worlds: all platforms (that have a low cost for providing services) enter the market, and the user benefits from the services of all of them while incurring zero privacy loss. This observation suggests that if we only had large firms whose costs of providing service per user are small, then the draconian regulation that involves banning data sharing would have been optimal. However, this is no longer the case if the market includes both high and low-cost platforms.

In this case, our second insight reveals that imposing a uniform minimum privacy mandate improves user utility, particularly when the value of the user data to the data buyer is significant enough for platforms to charge a high price for it. This means that all platforms need to perturb the user data by adding to it a noise with a high enough noise level. Here, our analysis identifies specific conditions where GDPR-type policies, which limit user data usage, can benefit users. Finally, our third insight suggests that, despite potential implementation challenges, a non-uniform minimum privacy mandate — banning data sharing for low-cost platforms while imposing a minimum privacy standard for high-cost platforms — further enhances user utility. This regulation entails prohibiting large, low-cost platforms from sharing user data, while permitting smaller, high-cost platforms to do so, albeit in a limited manner through privacy mechanisms. This approach aligns with our earlier discussion that GDPR-type regulations disproportionately harm small businesses. Therefore, a non-uniform regulation favoring small businesses is preferred not only by these businesses but also by the users.

1.2 Related literature

Our paper broadly relates to the emerging literature on data markets and online platform behavior. Earlier works on this topic include Acquisti et al. [2016], Posner and Weyl [2018], Ali et al. [2019], Jones and Tonetti [2020], and Dosis and Sand-Zantman [2022] which study the consequences of protecting and disclosing personal information about individuals. Additionally, Acemoglu et al. [2022], Bergemann et al. [2020], Ichihashi [2020b], and Fainmesser et al. [2022] study the privacy consequences of data externality (whereby a user’s data reveals information about others). Acemoglu et al. [2023a] study user-optimal privacy-preserving mechanisms that platforms can adopt to balance privacy and learning and Acemoglu et al. [2023b] develop an experimentation game to study the implications of the platform’s information advantage in product offering. Finally, Ichihashi [2021] studies data intermediaries, and Ichihashi [2020a] studies consumer privacy choices while online platforms adjust their privacy guarantees dynamically to incentivize more data sharing from consumers.

More closely related to ours are papers that study data sharing among users, platforms, and third-party data buyers and, in particular, the impacts of banning such data sharing. In this regard, Bimpikis et al. [2021] and Argenziano and Bonatti [2023] develop a two-round model of the interaction between a user and two competing platforms that can collect data on user preferences (and potentially use it for price discrimination). Bimpikis et al. [2021] show that banning data sharing can hurt the user when the platforms offer complementary products, and Argenziano and Bonatti [2023] show that banning data sharing can hurt the user if the benefits of knowing user preferences by the seller and personalized services are limited. The above papers consider the interactions between a user and a platform that itself uses the user data for price discrimination and, therefore, can potentially hurt the user. In contrast, we study a different problem that relates to how platforms sell user data to other third-party data buyers. Therefore, in addition to the differences in modeling choices and analysis, we depart from the above papers by considering a three-layer data market that includes the interaction between the platforms and the data buyer and focusing on deriving insights on effective regulations.

Also related to our work is Ravichandran and Korula [2019] who empirically study the effect of the data-sharing ban on platforms’ revenues and Madsen and Vellodi [2023] who study the effects of a complete ban on innovation (we refer to Pino [2022] for a comprehensive survey on the microeconomics of data). We depart from these papers by focusing on the impact of regulations (and in particular data-sharing ban) on user’s utility. Our paper also relates to the literature that studies various forms of monetizing user data by platforms, such as selling cookies Bergemann and Bonatti [2015], efficient pricing for large data sets Abolhassani et al. [2017], dynamic sales Immorlica et al. [2021], Drakopoulos and Makhdoumi [2023], monetization while ensuring the data is replicable Falconer et al. [2023], and the design and price of information where a data buyer faces a decision problem under uncertainty and can augment their initial private information with supplemental data from a data seller Bergemann et al. [2018].

Finally, our paper relates to the growing literature on data acquisition and the design of mechanisms for incentivizing data sharing, including Ghosh and Roth [2011], Ligett and Roth [2012], Nissim et al. [2014], Cummings et al. [2015], Chen et al. [2018], Chen and Zheng [2019], Fallah et al. [2023], Cummings et al. [2023], Fallah et al. [2022], and Karimireddy et al. [2022]. We depart from this line of work as we consider a three-layer market where the privacy cost occurs due to selling data to a third-party buyer.

The rest of the paper proceeds as follows. In Section 2, we describe our model and the interactions among the platforms, the data buyer, and the user. We also present some properties of the revealed information measure in our setting such as its monotonicity and submodularity. In Section 3, we introduce our equilibrium concept and provide preliminary characterizations of it. In Section 4, we focus on a setting with two platforms and provide a full characterization of the equilibrium. The focus on two platforms allows us to describe some of the nuances of the interactions in our model. In Section 5, we establish that our main characterizations continue to hold in the general setting with any number of platforms. We also provide several insights and comparative statics, leading to our discussion of the regulations in Section 6. Finally, Section 7 concludes while the appendix contains all the omitted proofs from the text.

2 Model

We consider KK\in\mathbb{N} platforms that are interacting with a user and a data buyer. The user benefits from using the services of these platforms. In return, the platforms acquire data about the user’s preferences and behavior. The platforms may subsequently sell this data to a third party, such as an advertiser, which we refer to as a data buyer (or, interchangeably, a buyer). The data buyer is interested in learning about the user’s preferences, which can adversely impact the user’s utility. This negative impact can be because of price discrimination, unfairly targeted advertising, manipulation, or intrinsic privacy losses. In the example of MasterCard described in the introduction, MasterCard is one of the platforms, and the data buyer is any other firm that purchases user data from MasterCard.

Formally, we represent the user’s data by 𝜽d\bm{\theta}\in\mathbb{R}^{d}, which follows a Gaussian distribution with zero mean and identity variance, that is, 𝜽𝒩(0,𝕀d)\bm{\theta}\sim\mathcal{N}(0,\mathbb{I}_{d}). This vector 𝜽\bm{\theta} can be interpreted as the user’s feature vector. Each platform i[K]{1,,K}i\in[K]\triangleq\{1,\dots,K\} offers a service whose quality depends on how accurately the platform can learn the match between the user’s feature vector and the platform’s characteristics, i.e.,

𝒙i𝜽.\displaystyle\bm{x}_{i}^{\top}\bm{\theta}.

Here, 𝒙id\bm{x}_{i}\in\mathbb{R}^{d} is a known unit-norm vector symbolizing the characteristic vector of platform ii.

The user decides whether to share their data with each of the KK platforms in exchange for their service. We let ai{0,1}a_{i}\in\{0,1\} denote whether the user has shared data with platform ii and 𝒂{0,1}K\bm{a}\in\{0,1\}^{K} denote the user strategy profile. Specifically, if the user opts to share their data with platform ii, meaning ai=1a_{i}=1, then platform ii receives a signal

si:=𝒙i𝜽+𝒩(0,1).\displaystyle s_{i}:=\bm{x}_{i}^{\top}\bm{\theta}+\mathcal{N}(0,1).

The introduced noise ensures a degree of privacy concerning the user’s data.111We set the noise variance to one to simplify the results. However, the analysis remains valid for any (potentially heterogeneous) level of noise.

The data buyer aims to estimate a linear function of the user’s data, 𝒚𝜽\bm{y}^{\top}\bm{\theta}. Here, 𝒚d\bm{y}\in\mathbb{R}^{d} with 𝒚=1\|\bm{y}\|=1 denotes the data buyer’s known characteristic vector. If platform ii possesses the user’s data (i.e., ai=1a_{i}=1), they present the data buyer with an option to acquire a noisy version of their data at a price pip_{i}. Therefore, the data buyer strategy profile is 𝒃{0,1}K\bm{b}\in\{0,1\}^{K}, with bi=1b_{i}=1 denoting their acceptance of the platform ii’s offer. The data buyer then receives a signal

s~i:=si+𝒩(0,σi2)\displaystyle\tilde{s}_{i}:=s_{i}+\mathcal{N}(0,\sigma_{i}^{2})

for the price pip_{i}, where σi2\sigma_{i}^{2} represents the variance of the noise that platform ii adds to the data they share with the data buyer.

Lastly, each platform i[K]i\in[K] decides on the noise level to add to the user data before offering it to the data buyer, denoted by σi+\sigma_{i}\in\mathbb{R}_{+}. We find it more convenient to work with the standard deviation of the noise σi\sigma_{i} and assume, without loss of generality, it is positive rather than the variance σi2\sigma_{i}^{2}. Each platform also decides whether to enter the market and provide services to the user, denoted by ei{0,1}e_{i}\in\{0,1\}. Therefore, the platforms’ strategy profile is (𝝈,𝒆)+K×{0,1}K(\bm{\sigma},\bm{e})\in\mathbb{R}_{+}^{K}\times\{0,1\}^{K}.

We set s~i=NA\tilde{s}_{i}=\text{NA} when the data buyer does not obtain any information from platform ii. This can occur either because the user has not shared their data with platform ii or because the data buyer declines the offer from platform ii or the platform has not provided the service to the user. In this case, pip_{i} is irrelevant, and we set it to zero.

Throughout the paper, we make the following assumption:

Assumption 1.

Let

γi=𝒙iT𝒚 for all i[K]\displaystyle\gamma_{i}=\bm{x}_{i}^{T}\bm{y}\text{ for all }i\in[K]

denote the correlation between the user ii’s data and the data buyer. Then, we assume

(𝒙iγi𝒚)T(𝒙jγj𝒚)=0 for all ij[K].\displaystyle\left(\bm{x}_{i}-\gamma_{i}\bm{y}\right)^{T}\left(\bm{x}_{j}-\gamma_{j}\bm{y}\right)=0\text{ for all }i\neq j\in[K].

Note that under this assumption, the data offered by different platforms are viewed as substitutes for each other from the data buyer’s perspective. However, if this assumption does not hold, a complementary effect could also be observed. To illustrate this, suppose we have K=2K=2 platforms where x1x_{1} is orthogonal to yy, i.e., γ1=0\gamma_{1}=0, and also x2=γ2y+(1γ2)x1x_{2}=\gamma_{2}y+(1-\gamma_{2})x_{1}. In this scenario, the data from platform 1 alone has no value to the data buyer, as x1θx_{1}^{\top}\theta provides no information about θy\theta^{\top}y. However, if the data buyer decides to purchase data from the second platform, then the data from the first platform could also become valuable. To understand why, observe that the data from the second platform is a convex combination of two components: x1θx_{1}^{\top}\theta and yθy^{\top}\theta. Therefore, if the data buyer acquires the data from the second platform, learning x1θx_{1}^{\top}\theta could assist in better distinguishing these two components. In other words, the data from the first platform is valuable only if accompanied by the data from the second platform. This scenario exemplifies the complementary effect we described earlier. Assumption 1 excludes this possibility, allowing us to primarily focus on the substitution effect. This assumption also implies that the correlation among data for different platforms is given by

𝔼[𝒙i𝒙j]=γiγj.\mathbb{E}[\bm{x}_{i}\bm{x}_{j}]=\gamma_{i}\gamma_{j}.

2.1 Utility functions

To define the utility functions, we need to first quantify how much information is revealed by observing the signals. We do so by using the following definition, which captures the reduction in uncertainty following the observation of a set of signals.

Definition 1 (Revealed information).

Consider a random variable ZZ drawn from the distribution π(.)\pi(.). Let π𝒮(.)\pi_{\mathcal{S}}(.) represent the posterior distribution of ZZ upon observing a set of signals 𝒮\mathcal{S}. The revealed information about ZZ by observing 𝒮\mathcal{S} is quantified as the reduction in the mean-squared error of ZZ, i.e.,

(Z|𝒮)=𝔼[(Z𝔼Zπ[Z])2]𝔼[(Z𝔼Zπ𝒮[Z])2],\mathcal{I}(Z~{}|~{}\mathcal{S})=\mathbb{E}\left[\left(Z-\mathbb{E}_{Z\sim\pi}\left[Z\right]\right)^{2}\right]-\mathbb{E}\left[\left(Z-\mathbb{E}_{Z\sim\pi_{\mathcal{S}}}\left[Z\right]\right)^{2}\right],

where the outer expectations are over the randomness in ZZ and the signals.

Measuring the revealed information, or information gain, through the variance of the posterior distribution has been used widely in economics, statistics, and learning theory (see, for example, Breiman et al. [1984], Bergemann et al. [2020], and Acemoglu et al. [2022]).

Given this definition, the user’s utility function is given by

𝒰user(𝝈,𝒆,𝒂,𝒑,𝒃):=i=1Kaieiiα(𝝈,𝒆,𝒂,𝒃),\mathcal{U}_{\text{user}}(\bm{\sigma},\bm{e},\bm{a},\bm{p},\bm{b}):=\sum_{i=1}^{K}a_{i}e_{i}\mathcal{I}_{i}-\alpha~{}\mathcal{I}(\bm{\sigma},\bm{e},\bm{a},\bm{b}), (1)

where the term

i(𝒙i𝜽|si)\displaystyle\mathcal{I}_{i}\triangleq\mathcal{I}(\bm{x}_{i}^{\top}\bm{\theta}~{}|~{}s_{i})

is the user’s revealed information to platform ii if the user shares her information (i.e., when ai=1a_{i}=1) and the platform is offering service to the user (i.e., when ei=1e_{i}=1). The term

(𝝈,𝒆,𝒂,𝒃)(𝒚𝜽|𝒔~),\displaystyle\mathcal{I}(\bm{\sigma},\bm{e},\bm{a},\bm{b})\triangleq\mathcal{I}(\bm{y}^{\top}\bm{\theta}~{}|~{}\bm{\tilde{s}}),

is the user’s revealed information to the data buyer. Here, 𝒔~\bm{\tilde{s}} denotes the vector (s~i)i=1K(\tilde{s}_{i})_{i=1}^{K}, which is the revealed information to the data buyer. Notice that the user’s utility does not directly depend on 𝒑\bm{p}, but we retain this notation to highlight that through the data buyer’s decision, the revealed information depends on 𝒑\bm{p}. Also, notice that in finding the revealed information to the data buyer, all that matters is the set {σi: s.t. ai=bi=ei=1}\{\sigma_{i}~{}:~{}\text{ s.t. }a_{i}=b_{i}=e_{i}=1\}. Therefore, we also find it convenient to define

(𝝈S) for 𝝈S=(σi:iS) for any S[K],\displaystyle\mathcal{I}(\bm{\sigma}_{S})\text{ for }\bm{\sigma}_{S}=(\sigma_{i}~{}:~{}i\in S)\text{ for any }S\subseteq[K], (2)

where we may drop the subindex SS whenever the indices of the elements of 𝝈\bm{\sigma} are clear from the context.

The first term in (1) captures the user’s gain from the service offered by each platform. Notice that this term is present if the platform has entered the market (i.e., ei=1e_{i}=1) and if the user shares with the platform (i.e., ai=1a_{i}=1). The second term represents the user’s privacy loss from the sharing of their data with the data buyer. Additionally, α\alpha indicates the user’s relative importance on these two opposing terms.

For each i[K]i\in[K], platform ii’s utility is given by

𝒰i(𝝈,𝒆,𝒂,𝒑,𝒃):={aii+bipiciei=10ei=0.\mathcal{U}_{i}(\bm{\sigma},\bm{e},\bm{a},\bm{p},\bm{b}):=\begin{cases}a_{i}\mathcal{I}_{i}+b_{i}p_{i}-c_{i}\quad&e_{i}=1\\ 0\quad&e_{i}=0.\end{cases} (3)

When the platform has not entered the market (i.e., ei=0e_{i}=0), the platform’s utility is zero. When the platform has entered the market (i.e., ei=1e_{i}=1), the second term bipib_{i}p_{i} represents the platform’s revenue from selling the user’s data to the data buyer. The first term aiia_{i}\mathcal{I}_{i}, similar to the user’s utility, reflects the quality of service the platform provides (if the user ends up utilizing their service). This term is concluded to capture the fact that a better service by the platform also enhances the platform’s utility. For example, it could increase ad revenue or expand the user base. Lastly, the cost ci+c_{i}\in\mathbb{R}_{+} reflects the per-user cost incurred by the platform for providing the service.

Finally, the data buyer’s utility is given by

𝒰buyer(𝝈,𝒆,𝒂,𝒑,𝒃):=β(𝝈,𝒆,𝒂,𝒃)i=1Kbipi,\mathcal{U}_{\text{buyer}}(\bm{\sigma},\bm{e},\bm{a},\bm{p},\bm{b}):=\beta~{}\mathcal{I}(\bm{\sigma},\bm{e},\bm{a},\bm{b})-\sum_{i=1}^{K}b_{i}p_{i}, (4)

which is the data buyer’s gain from learning the user’s information minus the amount they pay to all the platforms to purchase their data. Here, β[0,1]\beta\in[0,1] is some constant that captures the importance of the revealed information compared to the payments in the data buyer’s utility.

We refer to the above game among the platforms, user, and buyer of data as a three-layer data market with parameters (α,β,𝒄,𝜸)(\alpha,\beta,\bm{c},\bm{\gamma}). Here is the timing of the game (see Figure 1):

  1. 1.

    In the first stage, all platforms simultaneously choose (𝝈,𝒆)+K×{0,1}K(\bm{\sigma},\bm{e})\in\mathbb{R}_{+}^{K}\times\{0,1\}^{K} that specifies their noise levels and whether they want to enter the market to provide services to the user.

  2. 2.

    In the second stage, the user chooses 𝒂{0,1}K\bm{a}\in\{0,1\}^{K} that specifies which platforms to join and share data with.

  3. 3.

    In the third stage, the platforms simultaneously choose 𝒑K\bm{p}\in\mathbb{R}^{K} that determines their offered prices.

  4. 4.

    In the fourth stage, the data buyer chooses 𝒃{0,1}K\bm{b}\in\{0,1\}^{K} that determines whether they accept or decline each offer made by the platforms.

Refer to caption
Figure 1: The timing of the decisions.

2.2 Revealed Information Properties

Before analyzing this game, we provide some properties of our notion of revealed information. In stating the properties, we recall that the leaked information to the data buyer depends on 𝒆\bm{e}, 𝒂\bm{a}, and 𝒃\bm{b} through the set of platforms that share user data with the data buyer and the noise level of each of them. Therefore, we make use of the shorthand notation for the leaked information given in (2). In what follows, we also consider the lattice (,)(\mathcal{L},\leq), where

:={|{0,1}K} and  if ii for all i[K].\displaystyle\mathcal{L}:=\{\ell~{}|~{}\ell\in\{0,1\}^{K}\}\text{ and }\ell\leq\ell^{\prime}\text{ if }\ell_{i}\leq\ell^{\prime}_{i}\text{ for all }i\in[K]. (5)

The revealed information admits the following closed-form expression.

Lemma 1.

For any i[K]i\in[K], we have i=1/2\mathcal{I}_{i}=1/2. Moreover, for any 𝛔\bm{\sigma} and S[K]S\subseteq[K] we have

(𝝈S)=𝐦TM1𝐦,\displaystyle\mathcal{I}(\bm{\sigma}_{S})=\mathbf{m}^{T}M^{-1}\mathbf{m},

where mi=γim_{i}=\gamma_{i}, [M]ii=2+σi2[M]_{ii}=2+\sigma_{i}^{2}, and [M]ij=γiγj[M]_{ij}=\gamma_{i}\gamma_{j} for all i,jSi,j\in S such that iji\neq j.

Lemma 1 follows by using the properties of multivariate normal distribution. We should highlight that for simplicity, we assumed that the user adds the same noise level to their data before sharing it with any platform, which results in i\mathcal{I}_{i} for all I[K]I\in[K] being equal. Our analysis continues to hold for any heterogeneous level of noise.

Importantly, Lemma 1 enables us to obtain some structural properties of the revealed information, as we now demonstrate.

Lemma 2 (monotonicity).

The revealed information is increasing in the set of platforms that share the user data with the data buyer and is decreasing in 𝛔\bm{\sigma}, i.e., for any 𝛔\bm{\sigma} and S[K]S\subseteq[K], (𝛔S)\mathcal{I}(\bm{\sigma}_{S}) is increasing in SS and decreasing in 𝛔\bm{\sigma}.

Lemma 2 simply states that more data sharing with the data buyer increases its revealed information. Moreover, data with a lower noise level increases the revealed information to the data buyer.

Lemma 3 (submodularity in actions).

For a given vector of noise levels 𝛔\bm{\sigma}, the revealed information is submodular in the set of platforms that share the user data with the data buyer, i.e., for any 𝛔\bm{\sigma}, ST[K]S\subseteq T\subseteq[K], and iTi\not\in T, we have

((σi,𝝈S))(𝝈S)((σi,𝝈T))(𝝈T).\displaystyle\mathcal{I}((\sigma_{i},\bm{\sigma}_{S}))-\mathcal{I}(\bm{\sigma}_{S})\geq\mathcal{I}((\sigma_{i},\bm{\sigma}_{T}))-\mathcal{I}(\bm{\sigma}_{T}).

Lemma 3 states the intuitively appealing result that as more data is shared with the data buyer, the marginal increase in the revealed information from one more unit of shared data becomes smaller.

Lemma 4 (submodularity in noise variance).

For any 𝛔\bm{\sigma}, S[K]S\subseteq[K], iSi\not\in S, and jSj\in S

((σi,𝝈S))(𝝈S)\displaystyle\mathcal{I}((\sigma_{i},\bm{\sigma}_{S}))-\mathcal{I}(\bm{\sigma}_{S})

is decreasing in σi\sigma_{i} and increasing in σj\sigma_{j}.

Lemma 4 states the data of platform ii reveals a relatively higher amount of information for smaller σi\sigma_{i} values and for larger σj\sigma_{j} values. To understand the former, let us compare the case of σi0\sigma_{i}\approx 0 and σi\sigma_{i}\to\infty. When σi0\sigma_{i}\approx 0, the data of platform ii reveals a lot, which helps the data buyer learn much better. However, when σi\sigma_{i}\to\infty, the data of platform ii does not contain any information, and therefore, the relative gain in the revealed information is 0\approx 0. To understand the former, let us again compare the case of σj0\sigma_{j}\approx 0 and σj\sigma_{j}\to\infty. When σj0\sigma_{j}\approx 0, the data of platform jj reveals a lot; therefore, the marginal increase in the revealed information of platform ii is small. However, when σj\sigma_{j}\to\infty, the data of platform jj does not contain any information, and therefore, the relative increase in the revealed information from platform ii’s data is much larger.

3 Preliminary Equilibrium Characterization

We adopt subgame perfect equilibrium as our solution concept, which means that the strategy profile of all players is a Nash equilibrium of every subgame of the game. We next formally define and characterize the subgame perfect equilibrium in our setting, starting from the last stage.

Buyer equilibrium:

For a given (𝝈,𝒆,𝒂,𝒑)(\bm{\sigma},\bm{e},\bm{a},\bm{p}), a data buyer profile 𝒃E\bm{b}^{\mathrm{E}} is an equilibrium if it achieves the data buyer’s highest utility, i.e.,

𝒃Eargmax𝒃{0,1}K𝒰buyer(𝝈,𝒆,𝒂,𝒑,𝒃).\displaystyle\bm{b}^{\mathrm{E}}\in\operatorname*{arg\,max}_{\bm{b}\in\{0,1\}^{K}}\mathcal{U}_{\text{buyer}}(\bm{\sigma},\bm{e},\bm{a},\bm{p},\bm{b}).

Given that there is a finite set of 2n2^{n} possibilities, one of them achieves the maximum data buyer’s utility, and therefore, an equilibrium buyer profile strategy always exists. We let E(𝝈,𝒆,𝒂,𝒑)\mathcal{B}^{\mathrm{E}}(\bm{\sigma},\bm{e},\bm{a},\bm{p}) be the set of such buyer equilibria.

Price equilibrium:

For a given (𝝈,𝒆,𝒂)(\bm{\sigma},\bm{e},\bm{a}), a pricing strategy by the platforms 𝒑E\bm{p}^{\mathrm{E}} is an equilibrium if no platform has a profitable deviation (taking the data buyer’s updated decision into account):

𝒰i(𝝈,𝒆,𝒂,𝒑E,𝒃E)𝒰i(𝝈,𝒆,𝒂,(pi,𝒑iE),𝒃) for all i[K],𝒃EE(𝝈,𝒆,𝒂,𝒑E),pi,𝒃E(𝝈,𝒆,𝒂,(pi,𝒑iE)).\displaystyle\mathcal{U}_{i}(\bm{\sigma},\bm{e},\bm{a},\bm{p}^{\mathrm{E}},\bm{b}^{\mathrm{E}})\geq\mathcal{U}_{i}(\bm{\sigma},\bm{e},\bm{a},(p_{i},\bm{p}^{\mathrm{E}}_{-i}),\bm{b})\text{ for all }i\in[K],\bm{b}^{\mathrm{E}}\in\mathcal{B}^{\mathrm{E}}(\bm{\sigma},\bm{e},\bm{a},\bm{p}^{\mathrm{E}}),p_{i},\bm{b}\in\mathcal{B}^{\mathrm{E}}(\bm{\sigma},\bm{e},\bm{a},(p_{i},\bm{p}^{\mathrm{E}}_{-i})).

Before proceeding with the rest of the game, we establish that the data buyer and price equilibria admit a simple characterization, stated next.

Proposition 1.

For a given (𝛔,𝐞,𝐚)(\bm{\sigma},\bm{e},\bm{a}), there exists a unique price equilibrium given by

pi=β((𝝈,𝒆,𝒂,𝟏)(𝝈,𝒆,𝒂,(bi=0,𝟏i))) for all i[K]\displaystyle p_{i}^{*}=\beta\left(\mathcal{I}(\bm{\sigma},\bm{e},\bm{a},\mathbf{1})-\mathcal{I}(\bm{\sigma},\bm{e},\bm{a},(b_{i}=0,\mathbf{1}_{-i}))\right)\text{ for all }i\in[K]

and the corresponding unique buyer equilibrium is bi=1b_{i}=1 for all i[K]i\in[K], where 𝟏i\mathbf{1}_{-i} is the vector of all ones for all j[K]{i}j\in[K]\setminus\{i\}.

The proof of Proposition 1, given in the appendix, directly follows from the submodularity of the revealed information established in Lemma 3 for the following reason. Consider platform i[K]i\in[K] and suppose the contrary that in equilibrium, the buyer is not purchasing from it. This platform has a profitable deviation by lowering its price so that the data buyer finds it optimal to purchase from it. Also, notice that the submodularity of the revealed information implies that if one more platform’s data is shared with the data buyer, the marginal increase in the revealed information decreases. Therefore, if the data buyer was purchasing from another platform jij\neq i, they still find it optimal to purchase from that platform.

Proposition 1 characterizes the equilibrium of the third and fourth stages of the game. Therefore, it remains to characterize the user sharing profile and the platform’s noise level and entry decisions in equilibrium, taking into account the equilibrium of the subsequent stages that we have already identified.

User equilibrium:

For a given (𝝈,𝒆)(\bm{\sigma},\bm{e}), a user sharing profile 𝐚E\mathbf{a}^{\mathrm{E}} is an equilibrium if it achieves the user’s highest utility, i.e.,

𝒂Eargmax𝒂{0,1}K𝒰user(𝝈,𝒆,𝒂,𝒑,𝟏).\displaystyle\bm{a}^{\mathrm{E}}\in\operatorname*{arg\,max}_{\bm{a}\in\{0,1\}^{K}}\mathcal{U}_{\text{user}}(\bm{\sigma},\bm{e},\bm{a},\bm{p}^{*},\mathbf{1}).

Again, given that there are 2n2^{n} possibilities, one of them achieves the maximum user’s utility, and therefore, an equilibrium sharing profile strategy always exists. We let 𝒜E(𝝈,𝒆)\mathcal{A}^{\mathrm{E}}(\bm{\sigma},\bm{e}) be the set of such user equilibria. This set has the following important lattice structure that we use in the rest of the analysis.

Lemma 5.

For a given (𝛔,𝐞)(\bm{\sigma},\bm{e}), the set of user equilibria 𝒜E(𝛔,𝐞)\mathcal{A}^{\mathrm{E}}(\bm{\sigma},\bm{e}) is a sublattice of the lattice (,)(\mathcal{L},\leq) (defined in (5)). Therefore, it has a maximum and a minimum.

For a platform strategy profile (𝝈,𝒆)(\bm{\sigma},\bm{e}), we select the platforms’ Pareto-optimal one if there are multiple user equilibria. Using Lemma 5, such a user equilibrium exists and is the one that shares with the highest number of platforms.222We can view this as Stackelberg Nash equilibrium in which because the platforms move first, we break the ties in their favor. Similar concepts have appeared in the literature with different terms, such as sender-preferred equilibrium in Bayesian persuasion [Kamenica and Gentzkow, 2011].

Platform equilibrium:

A noise level and entry strategy (𝝈E,𝒆E)(\bm{\sigma}^{\mathrm{E}},\bm{e}^{\mathrm{E}}) is an equilibrium if no platform has a profitable deviation, i.e.,

𝒰i(𝝈E,𝒆E,𝒂E,𝒑,𝟏)𝒰i((σi,𝝈iE),(ei,𝒆iE),𝒂,𝒑,𝟏)\displaystyle\mathcal{U}_{i}(\bm{\sigma}^{\mathrm{E}},\bm{e}^{\mathrm{E}},\bm{a}^{\mathrm{E}},\bm{p}^{*},\mathbf{1})\geq\mathcal{U}_{i}((\sigma_{i},\bm{\sigma}_{-i}^{\mathrm{E}}),(e_{i},\bm{e}_{-i}^{\mathrm{E}}),\bm{a},\bm{p}^{*},\mathbf{1})

for all i[K],𝒂E𝒜E(𝝈E,𝒆E),𝒂𝒜E((σi,𝝈iE),(ei,𝒆iE))i\in[K],\bm{a}^{\mathrm{E}}\in\mathcal{A}^{\mathrm{E}}(\bm{\sigma}^{\mathrm{E}},\bm{e}^{\mathrm{E}}),\bm{a}\in\mathcal{A}^{\mathrm{E}}((\sigma_{i},\bm{\sigma}_{-i}^{\mathrm{E}}),(e_{i},\bm{e}_{-i}^{\mathrm{E}})).

We refer to a platform’s noise level and entry decision as the platform’s decision and to the corresponding equilibrium as the platforms’ equilibrium. The following characterizes the set of possible platforms’ equilibria.

Proposition 2.

For any platforms’ equilibrium strategy (𝛔,𝐞)(\bm{\sigma},\bm{e}), we must have

𝒜E(𝝈,𝒆)={(ai=1: for all i s.t. ei=1))},\displaystyle\mathcal{A}^{\mathrm{E}}(\bm{\sigma},\bm{e})=\{(a_{i}=1~{}:~{}\text{ for all }i\text{ s.t. }e_{i}=1))\},

that is, the user must find it optimal to share with all platforms that have entered the market.

This proposition follows by noting that a platform that has entered the market can always increase its noise level so that the user finds it optimal to share with the platform.

Motivated by Proposition 2 and Lemma 5, we conclude that for any set of platforms that have entered the market (captured by 𝒆{0,1}K\bm{e}\in\{0,1\}^{K}), the set of possible noise levels in equilibrium is Σ𝟏,𝒆\Sigma_{\mathbf{1},\bm{e}}, where we have

Σ𝐚,𝒆:={(σi:i s.t. ei=1):𝒂 is the highest element of 𝒜E(𝝈,𝒆)} for all 𝒂{0,1}K.\displaystyle\Sigma_{\mathbf{\bm{a}},\bm{e}}:=\{(\sigma_{i}~{}:~{}i\text{ s.t. }e_{i}=1)~{}:~{}\bm{a}\text{ is the highest element of }\mathcal{A}^{\mathrm{E}}(\bm{\sigma},\bm{e})\}\text{ for all }\bm{a}\in\{0,1\}^{K}. (6)

We are interested in characterizing Σ𝟏,𝒆\Sigma_{\mathbf{1},\bm{e}} for any 𝒆\bm{e} and finding the equilibrium that belongs to one of these sets for an equilibrium choice of 𝒆\bm{e}.

Throughout the paper, we focus on the case in which the user asks for some level of privacy in order to share their data. This means that if the noise levels are all zero, then the user does not share with the platforms. More formally, we make the following assumption:

Assumption 2.

The weight of privacy on the user utility, i.e., α\alpha, is large enough such that, for any 𝐞{0,1}K\bm{e}\in\{0,1\}^{K}, the point (σi=0:i s.t. ei=1)(\sigma_{i}=0~{}:~{}i\text{ s.t. }e_{i}=1) belongs to the set Σ𝟎,𝐞\Sigma_{\mathbf{0},\bm{e}}.

We will further clarify this assumption’s role in the next section.

4 Equilibrium Characterization with K=2K=2 Platforms

We begin our analysis by considering a setting with K=2K=2 platforms. We then show that the main insights from this analysis extend to the general setting with K>2K>2 platforms.

Depending on which platforms enter the market, we have four cases:

Case 1 where both platforms enter the market:

In this case, the user’s utility if they share their data with both platforms is given by333With a slight abuse of notation, we drop the other dependencies of the user utility here.

𝒰user(σ1,σ2):=1α(σ1,σ2),\mathcal{U}_{\text{user}}(\sigma_{1},\sigma_{2}):=1-\alpha\mathcal{I}(\sigma_{1},\sigma_{2}), (7)

and the user’s utility if they share their data with platform i{1,2}i\in\{1,2\} only is equal to

𝒰user(σi):=12α(σi).\mathcal{U}_{\text{user}}(\sigma_{i}):=\frac{1}{2}-\alpha\mathcal{I}(\sigma_{i}). (8)

Recall the definition of Σ𝒂,𝒆\Sigma_{\bm{a},\bm{e}} with 𝒆=(1,1)\bm{e}=(1,1) in (6). To simplify the notation, we drop the 𝒆\bm{e} and use Σ~𝒂\tilde{\Sigma}_{\bm{a}}. In particular,

Σ~(1,1):={(σ1,σ2):𝒰user(σ1,σ2)=max(𝒰user(σ1,σ2),𝒰user(σ1),𝒰user(σ2),0)}.\displaystyle\tilde{\Sigma}_{(1,1)}:=\{(\sigma_{1},\sigma_{2})~{}:~{}\mathcal{U}_{\text{user}}(\sigma_{1},\sigma_{2})=\max\left(\mathcal{U}_{\text{user}}(\sigma_{1},\sigma_{2}),\mathcal{U}_{\text{user}}(\sigma_{1}),\mathcal{U}_{\text{user}}(\sigma_{2}),0\right)\}.

Similarly, Σ~(1,0)\tilde{\Sigma}_{(1,0)} and Σ~(0,1)\tilde{\Sigma}_{(0,1)} denote the cases where 𝒰user(σ1)\mathcal{U}_{\text{user}}(\sigma_{1}) and 𝒰user(σ2)\mathcal{U}_{\text{user}}(\sigma_{2}), respectively, have the highest value. Finally,

Σ~(0,0):={(σ1,σ2):0=max(𝒰user(σ1,σ2),𝒰user(σ1),𝒰user(σ2),0)}.\displaystyle\tilde{\Sigma}_{(0,0)}:=\{(\sigma_{1},\sigma_{2})~{}:~{}0=\max\left(\mathcal{U}_{\text{user}}(\sigma_{1},\sigma_{2}),\mathcal{U}_{\text{user}}(\sigma_{1}),\mathcal{U}_{\text{user}}(\sigma_{2}),0\right)\}.

Notice that we operate under Assumption 2, which implies that Σ~(0,0)\tilde{\Sigma}_{(0,0)} is non-empty. In particular, for K=2K=2, this assumption translates to

α>4γ12γ222(γ12+γ22γ12γ22).\alpha>\frac{4-\gamma_{1}^{2}\gamma_{2}^{2}}{2\left(\gamma_{1}^{2}+\gamma_{2}^{2}-\gamma_{1}^{2}\gamma_{2}^{2}\right)}. (9)

Figure 2(a) depicts the sets Σ~(1,1)\tilde{\Sigma}_{(1,1)}, Σ~(0,1)\tilde{\Sigma}_{(0,1)}, Σ~(1,0)\tilde{\Sigma}_{(1,0)}, and Σ~(0,0)\tilde{\Sigma}_{(0,0)} as a function of the noise variance that the two platforms add to the data, i.e., (σ12,σ22)(\sigma_{1}^{2},\sigma_{2}^{2}), for parameters γ1=0.8\gamma_{1}=0.8 and γ2=0.7\gamma_{2}=0.7.

Refer to caption
(a)
Refer to caption
(b)
Figure 2: Four possible sets Σ~𝒂\tilde{\Sigma}_{\bm{a}} for 𝒂{0,1}2\bm{a}\in\{0,1\}^{2} for a setting with K=2K=2 platforms, γ1=.8\gamma_{1}=.8, and γ2=.7\gamma_{2}=.7.

Using Proposition 2, the noise level equilibrium is in the set Σ~(1,1)\tilde{\Sigma}_{(1,1)}. In fact, we argue that it must belong to Σ~(1,1)Σ~(0,0)\tilde{\Sigma}_{(1,1)}\cap\tilde{\Sigma}_{(0,0)} because, otherwise, one of the platforms can increase its utility by decreasing its noise level such that the user still shares with it.

Now, for any (σ1,σ2)Σ~(1,1)Σ~(0,0)(\sigma_{1},\sigma_{2})\in\tilde{\Sigma}_{(1,1)}\cap\tilde{\Sigma}_{(0,0)}, the only possible deviation for platform 11 is to increase σ1\sigma_{1} to σ~1\tilde{\sigma}_{1} to move to the point (σ~1,σ2)(\tilde{\sigma}_{1},\sigma_{2}) which resides in Σ~(1,0)\tilde{\Sigma}_{(1,0)}, exactly next to the boundary with Σ~(1,1)\tilde{\Sigma}_{(1,1)} (see Figure 2(a)). At first glance, it might look like this can never be a deviation because it reduces the revealed information to the buyer. However, notice that, by this action, platform 11 pushes platform 22 out of the market because (σ~1,σ2)(\tilde{\sigma}_{1},\sigma_{2}) is in the yellow region Σ~(1,0)\tilde{\Sigma}_{(1,0)} in which the user does not share their data with the second platform. Therefore, it is possible that the first platform’s payment increases, as they are now the only platform with access to the data. Note that this is the only deviation that we need to consider. Because if it ends up not being a deviation, it would imply that a further increase of σ1\sigma_{1} into the yellow region is also not profitable. Also, it is evident that increasing σ1\sigma_{1} but staying in the Σ~(1,1)\tilde{\Sigma}_{(1,1)} region or decreasing it and going into the Σ~(0,0)\tilde{\Sigma}_{(0,0)} region in which the user shares no data are not profitable deviations. Considering these deviations results in the following.

Lemma 6.

Suppose Assumptions 1 and 2 hold. Assuming both platforms have entered the market, there exists an equilibrium noise level (σ1,σ2)Σ~(1,1)Σ~(0,0)(\sigma_{1},\sigma_{2})\in\tilde{\Sigma}_{(1,1)}\cap\tilde{\Sigma}_{(0,0)} if and only if αα¯1.884\alpha\geq\bar{\alpha}\approx 1.884. In particular, for αα¯\alpha\geq\bar{\alpha}, (σ1,σ2)(\sigma_{1},\sigma_{2}) is an equilibrium where

2+σ12γ12=2+σ22γ22=2α1.\frac{2+\sigma_{1}^{2}}{\gamma_{1}^{2}}=\frac{2+\sigma_{2}^{2}}{\gamma_{2}^{2}}=2\alpha-1. (10)

Lemma 6 is under the assumption that both platforms enter the market. The final point to check is whether any platform has an incentive to enter the market.

Proposition 3.

Suppose Assumptions 1 and 2 hold. Also, recall α¯1.88\bar{\alpha}\approx 1.88 from Lemma 6. If an equilibrium exists in which both platforms enter the market, then

αα¯andβ2α(max{c1,c2}12).\alpha\geq\bar{\alpha}\quad\text{and}\quad\beta\geq 2\alpha~{}(\max\{c_{1},c_{2}\}-\frac{1}{2}). (11)

Conversely, for any αα¯\alpha\geq\bar{\alpha} and any ββ¯(α)\beta\geq\bar{\beta}(\alpha) such an equilibrium in which both platforms enter the market exists, where

β¯(α)(2+1α1)α(max{c1,c2}12).\bar{\beta}(\alpha)\triangleq(2+\frac{1}{\alpha-1})~{}\alpha~{}(\max\{c_{1},c_{2}\}-\frac{1}{2}). (12)

Both parts of the result provide a lower bound on β\beta, with the difference that the first part establishes a necessary condition for an equilibrium to exist, while the second part presents a sufficient condition. Note that the ratio between these two lower bounds converges to one as α\alpha grows. However, we next highlight a subtle difference between the two results. Note that, for any value ξ\xi greater than 2(maxc1,c212)2~{}(\max{c_{1},c_{2}}-\frac{1}{2}), we can choose sufficiently large α\alpha and β\beta such that an equilibrium exists for these parameters and the ratio of β\beta over α\alpha is less than or equal to ξ\xi. In other words, we have

inf{βα|α,β s.t. an equilibrium exists}=2(max{c1,c2}12).\inf\left\{\frac{\beta}{\alpha}~{}\Big{|}~{}\alpha,\beta\text{ s.t. an equilibrium exists}\right\}=2\left(\max\{c_{1},c_{2}\}-\frac{1}{2}\right).

This, in fact, is a direct consequence of the second part of Proposition 3 along with the fact that β¯\bar{\beta} converges to 2(max{c1,c2}12)2~{}(\max\{c_{1},c_{2}\}-\frac{1}{2}) as α\alpha grows.

However, the infimum mentioned above does not imply that any α\alpha and β\beta satisfying the condition (11) will necessarily result in an equilibrium. In fact, as our proof establishes, to have an equilibrium for α=α¯\alpha=\bar{\alpha}, we must have ββ¯(α¯)\beta\geq\bar{\beta}(\bar{\alpha}). In other words, we have

supαα¯inf{βα|β s.t. an equilibrium exists with the given α}=β¯(α¯).\displaystyle\sup_{\alpha\geq\bar{\alpha}}\inf\left\{\frac{\beta}{\alpha}~{}\Big{|}~{}\beta\text{ s.t. an equilibrium exists with the given }\alpha\right\}=\bar{\beta}(\bar{\alpha}).

Cases 2 and 3 where only one platform enters the market:

Suppose only platform 11 enters the market at some equilibrium. In this case, the first platform wants to choose the noise level σ1\sigma_{1} as small as possible but also large enough so that user shares their data. Therefore, the platform chooses σ1\sigma_{1} at the value that makes the user indifferent between sharing and not sharing (recall that we assumed the user breaks the ties in favor of the platforms), i.e.,

𝒰user(σ1)=12α(σ1)=0,\mathcal{U}_{\text{user}}(\sigma_{1})=\frac{1}{2}-\alpha\mathcal{I}(\sigma_{1})=0, (13)

which implies σ12=2αγ122\sigma_{1}^{2}=2\alpha\gamma_{1}^{2}-2. Notice that this is equivalent to the σ1\sigma_{1} corresponding to points BB and XX in Figure 2(b).

Now, for this noise level to be an equilibrium, we need to verify two potential deviations: (i) platform one has no incentive to abstain from participating, i.e., their utility should be nonnegative, and (ii) the second platform has no profitable deviation by entering the market. Using (13), the first condition simply implies

12c1+β12α0.\frac{1}{2}-c_{1}+\beta\frac{1}{2\alpha}\geq 0. (14)

For the second deviation check, we argue that it suffices to ensure the second platform’s utility at point BB in Figure 2(b) is nonnegative.

Let us denote the noise levels corresponding to point BB as (σ1B,σ2B)(\sigma_{1}^{B},\sigma_{2}^{B}). To understand why we only need to check the point BB, consider that, given platform one’s chosen noise level (equal to σ1B\sigma_{1}^{B}), if the second platform enters the market and selects a certain noise variance, it will result in a point on the line BXBX. The smallest noise level for the second platform that motivates the user to share their data with them (and thus has the potential to be profitable) is σ2B\sigma_{2}^{B}. One might argue that the second platform could opt for a very large noise level so that the vector of noise levels falls into the green region Σ~(0,1)\tilde{\Sigma}_{(0,1)}, implying that only the second platform receives the data. However, this is not feasible, as the line BXBX stays within the blue region Σ~(1,1)\tilde{\Sigma}_{(1,1)} for any value of σ2\sigma_{2} (and in fact, it becomes tangent to the border of Σ~(0,1)\tilde{\Sigma}_{(0,1)} and Σ~(1,1)\tilde{\Sigma}_{(1,1)} as σ2\sigma_{2} approaches infinity). To confirm this, we need to demonstrate that

𝒰user(σ1B,σ2)=1α(σ1B,σ2)12α(σ2)=𝒰user(σ2),\mathcal{U}_{\text{user}}(\sigma_{1}^{B},\sigma_{2})=1-\alpha\mathcal{I}(\sigma_{1}^{B},\sigma_{2})\geq\frac{1}{2}-\alpha\mathcal{I}(\sigma_{2})=\mathcal{U}_{\text{user}}(\sigma_{2}),

holds for any σ2\sigma_{2}. This is indeed the case, given that due to the submodularity of the revealed information, (σ1B,σ2)(σ2)\mathcal{I}(\sigma_{1}^{B},\sigma_{2})-\mathcal{I}(\sigma_{2}) decreases with increasing σ2\sigma_{2} and equals (σ1B)=12α\mathcal{I}(\sigma_{1}^{B})=\frac{1}{2\alpha} when σ2=\sigma_{2}=\infty.

Now, checking the second platform’s utility at point BB along with (14) yields the following result.

Proposition 4.

Suppose Assumptions 1 and 2 hold. For any i{1,2}i\in\{1,2\}, there exists an equilibrium in which only platform ii enters the market if and only if cicic_{i}\leq c_{-i} and

β[2α(ci12),2α(ci12)].\beta\in\left[2\alpha(c_{i}-\frac{1}{2}),2\alpha(c_{-i}-\frac{1}{2})\right].

Case 4 where no platform enters the market:

Finally, we ask whether there is an equilibrium in which no platform enters the market. In such a scenario, we need to ensure no platform has a profitable deviation by entering the market and choosing some noise level. For instance, if platform 11 wants to deviate, the smallest noise level that they would choose would be σ1B\sigma_{1}^{B} as it makes the user indifferent between sharing and not sharing their data with them. In this case, the platform’s utility, as derived in (14), is given by

12c1+β1α,\frac{1}{2}-c_{1}+\beta\frac{1}{\alpha},

which is nonpositive when β2α(c112)\beta\leq 2\alpha(c_{1}-\frac{1}{2}). Making the same argument for the second platform’s deviation, we obtain the following result.

Proposition 5.

Suppose Assumptions 1 and 2 hold. There exists an equilibrium in which neither of the platforms enters the market if and only if

β2α(min{c1,c2}12).\beta\leq 2\alpha\left(\min\{c_{1},c_{2}\}-\frac{1}{2}\right).

4.1 Insights from the case with K=2K=2

We conclude this section by highlighting a few implications and insights of our results for the case K=2K=2, which, as we will see, also extend to the general case. Note that the term β\beta determines the value the buyer places on the user’s data. The higher β\beta is, the more money the platform can earn by selling the user’s data.

Our results indicate that there are, at most, three main regimes for β\beta, and the market equilibrium depending on the interval β\beta falls into. First, as established in Proposition 5, when β\beta is small, the platform’s monetary gain from selling data is limited, so platforms may decide not to provide any service as they cannot fully compensate for their service costs. This regime would not exist if at least one of the platforms operates at a low cost, meaning its gain from the user’s data already compensates for the cost of service (this corresponds to the case ci1/2c_{i}\leq 1/2 for some ii).

On the other hand, as shown in Proposition 3, when β\beta is large enough and platforms can charge a high price for the user’s data, both platforms would enter the market. Finally, there is an intermediary regime for β\beta where only the platform with the lower cost enters the market, and the platform with a higher service cost opts out.

Interestingly, the user’s utility in all these cases would be equal to zero. However, the buyer’s utility could be positive when both platforms enter the market. As we establish next, these observations continue to hold in the general case with K>2K>2 platforms.

5 Equilibrium Characterization with K>2K>2 Platforms

This section shows how our equilibrium characterization extends to a general setting with any K>2K>2 number of platforms. In what follows, we make use of the following definition.

Definition 2 (High- and low-cost platforms).

We say a platform i[K]i\in[K] is of high cost if ci>12c_{i}>\frac{1}{2} and is of low cost if ci<12c_{i}<\frac{1}{2}.

The comparison of the cost to 1/21/2 in the above definition comes from the fact that the leaked information to any platform is i=1/2\mathcal{I}_{i}=1/2 and therefore, high-cost platforms enter the market only if the payment they receive from the buyer is large enough. Conversely, low-cost platforms enter the market irrespective of the payment they receive from the buyer. We can view low-cost platforms as large platforms with small operation costs that do not need extra payments from the data buyer to enter the market and provide services. High-cost platforms, however, are small platforms with large operation costs that do need extra payments from the buyer to be able to provide service.

Theorem 1.

Suppose Assumptions 1 and 2 hold. There exists α¯\bar{\alpha}, β¯\bar{\beta}, and β¯\underline{\beta} such that for αα¯\alpha\geq\bar{\alpha}, we have the following:

  • If ββ¯\beta\geq\bar{\beta}, an equilibrium of the game exists. Moreover, in any equilibrium, all platforms enter the market, the user utility is zero, and the buyer purchases the data of all platforms and has a strictly positive utility.

  • If ββ¯\beta\leq\underline{\beta}, an equilibirum of the game exists. Moreover, in any equilibrium, only low-cost platforms enter the market, the user utility is zero, and the buyer purchases the data of low-cost platforms and has a strictly positive utility only if there is more than one low-cost platform.

Let us explain the qualifiers of Theorem 1 and then explain its intuition. First, notice that similar to the setting with K=2K=2 platforms, for small enough α\alpha and also for intermediary values of β\beta, a pure-strategy equilibrium may not exist. Therefore, we focus on large enough α\alpha and also avoid intermediary values of β\beta. In the proof of Theorem 1, we have provided explicit expressions for α¯\bar{\alpha}, β¯\bar{\beta}, and β¯\underline{\beta}.

To gain the intuition of the first part of Theorem 1, notice that for large enough β\beta the data buyer gains a lot when the user’s data is revealed to them and therefore is willing to pay a high price for the user data. This, in turn, means that even high-cost platforms find it optimal to enter the market. As established in Proposition 1 and Proposition 2, in equilibrium, the user shares with all platforms, and the data buyer purchases from all platforms. To see that the user utility becomes zero in equilibrium, suppose the contrary, that the user utility is strictly positive. This means that the equilibrium noise level strategy is in the interior of the set Σ𝟏,𝟏\Sigma_{\mathbf{1},\mathbf{1}}. Given that noise level strategy is in the interior of Σ𝟏,𝟏\Sigma_{\mathbf{1},\mathbf{1}}, there exists i[K]i\in[K] such that by decreasing its noise level, we remain in the set Σ𝟏,𝟏\Sigma_{\mathbf{1},\mathbf{1}} (i.e., the user still finds it optimal to share with all platforms), but this deviation increases the payment and therefore the utility of platform ii. Also, Lemma 3 guarantees that the user still finds it optimal to share with all platforms.

To gain the intuition of the second part of Theorem 1, notice that for small enough β\beta, the data buyer’s gain when the user’s data is revealed to them is small and therefore, the buyer is willing to pay only a low price for the user data. This means that high-cost platforms will never find it optimal to enter the market. Nonetheless, the low-cost platforms always enter the market. The fact that the user shares with all the low-cost platforms (that have entered), the buyer purchases from them, and the user utility becomes zero in equilibrium follows from a similar argument to the one discussed above.

Theorem 1 leads to the following insights.

Competition among platforms helps the buyer and not the user:

Let us first characterize the equilibrium when we have only one platform.

Proposition 6.

Suppose Assumption 1 holds. When K=1K=1, we have the following:

  • Suppose α1/γ12\alpha\leq 1/\gamma_{1}^{2}. If c11/2+βγ12/2c_{1}\leq 1/2+\beta\gamma_{1}^{2}/2 (which holds whenever c11/2c_{1}\leq 1/2), then in equilibrium, the platform enters the market and chooses σ1=0\sigma_{1}=0, the user shares with the platform, and the buyer purchases the data. In this case, the user utility is (1αγ12)/2(1-\alpha\gamma_{1}^{2})/2, and the buyer utility is zero. Otherwise, if c1>1/2+βγ12/2c_{1}>1/2+\beta\gamma_{1}^{2}/2, then the platform does not enter the market, and all utilities are zero.

  • Suppose α>1/γ12\alpha>1/\gamma_{1}^{2}. If c11/2+β/2αc_{1}\leq 1/2+\beta/2\alpha, then in equilibrium, the platform enters the market and chooses σ1=2(αγ121)\sigma_{1}=\sqrt{2(\alpha\gamma_{1}^{2}-1)}, the user shares with the platform and the buyer purchases the data. In this case, both the user utility and the buyer utility are zero. Otherwise, if c1>1/2+β/2αc_{1}>1/2+\beta/2\alpha, then the platform does not enter the market, and all utilities are zero.

We are interested in the second regime where α\alpha is large enough so that the user is not incentivized to share with the platform, irrespective of the noise level choice of the platform. In this case, as established in Proposition 6, both the user and the data buyer utilities are zero. Now, let us compare this setting to a setting with K2K\geq 2 platforms that are competing to obtain the user’s data and sell it to the data buyer, established in Theorem 1. Interestingly, we observe that competition does not help the user (as the user obtains zero utility for any KK). However, increasing the number of platforms from K=1K=1 to K2K\geq 2 improves the data buyer’s utility. To gain the intuition for this observation, notice that the user’s utility will always be zero because otherwise, the platforms can always decrease their noise levels by a small margin so that the user still shares and increases their payments. Now, why does the competition help the data buyer? This is because of the data externality among the platforms’ data about the user, similar to that of Acemoglu et al. [2022] and Bergemann et al. [2020]. In particular, the data of a platform (because of the submodularity of the revealed information, proved in Lemma 3) decreases the marginal value of another platform’s data, and therefore, the platforms end up competing with each other, and sell their data at lower marginal prices.

More platforms imply higher utilitarian welfare:

Let us first find the overall utilitarian welfare in equilibrium, where utilitarian welfare is defined as the sum of the utilities of the platforms, the user, and the data buyer.

Corollary 1.

Consider α¯\bar{\alpha}, β¯\bar{\beta}, and β¯\underline{\beta} established in Theorem 1. For αα¯\alpha\geq\bar{\alpha}, we have the following:

  • If ββ¯\beta\geq\bar{\beta}, the utilitarian welfare in equilibrium becomes

    K(12+β2α)i=1Kci.\displaystyle K\left(\frac{1}{2}+\frac{\beta}{2\alpha}\right)-\sum_{i=1}^{K}c_{i}.
  • If ββ¯\beta\leq\underline{\beta}, the utilitarian welfare in equilibrium becomes

    (number of low-cost platforms)(12+β2α)i[K]:ci<12ci.\displaystyle(\text{number of low-cost platforms})\left(\frac{1}{2}+\frac{\beta}{2\alpha}\right)-\sum_{i\in[K]~{}:~{}c_{i}<\frac{1}{2}}c_{i}.

Notably, Corollary 1 shows that, within the ranges of each case, the overall utilitarian welfare is increasing in β\beta and decreasing in α\alpha. This is because the payments always cancel each other out, and as β\beta increases, the data buyer’s gain becomes larger, and as α\alpha increases, the user’s loss becomes larger. Moreover, in the first case, the utilitarian welfare is increasing in the number of platforms, and in the second case, it is increasing in the number of low-cost platforms.

6 Regulations

In this section, we consider regulations of the three-layer data market to improve the user utility. As we have established in Theorem 1, the user utility in equilibrium is zero. We now explore whether we can improve the user utility by imposing a regulation. As discussed in the introduction, one possible regulation is to impose a limit on the amount of information each platform can leak about the user to the data buyer. This means imposing a lower bound on the noise level. Let us formally define this regulation.

Definition 3 (Minimum privacy mandate).

A minimum privacy mandate, represented by 𝛔¯=(σ¯1,,σ¯K)\bar{\bm{\sigma}}=(\bar{\sigma}_{1},\dots,\bar{\sigma}_{K}), is a regulation that mandates that each platform i[K]i\in[K] must choose a noise level above σ¯i+\bar{\sigma}_{i}\in\mathbb{R}_{+}. A special case of this regulation is uniform minimum privacy mandate is such that σ¯i=σ¯\bar{\sigma}_{i}=\bar{\sigma} for all i[K]i\in[K]. Another special case of this regulation is to ban the platforms from sharing with the buyer, i.e., σ¯i=\bar{\sigma}_{i}=\infty for all i[K]i\in[K].

Uniform minimum privacy mandate versus ban:

One may conjecture that banning the platforms from sharing their data with the buyer is user-optimal in the class of all uniform minimum privacy mandate regulations because it reduces the privacy loss of the user to zero. We next prove that this is not necessarily true. The main intuition is that if we ban data sharing, then only low-cost platforms enter the market. This, in turn, implies that the user does not obtain the gain from the services provided by the high-cost platforms, and this loss in the service gain can dominate the gain in privacy loss. We next formalize this intuition.

Proposition 7.

Consider αα¯\alpha\geq\bar{\alpha} for α¯\bar{\alpha} established in Theorem 1. We have:

  • If all platforms are low-cost, then irrespective of the minimum privacy mandate, all platforms enter the market in equilibrium. In this case, banning data sharing achieves a higher user utility than any (uniform or non-uniform) minimum privacy mandate.

  • Otherwise, if there is at least one high-cost platform, then there exists β~\tilde{\beta}, β¯\underline{\beta}, and σ¯\bar{\sigma} such that

    • If ββ~\beta\geq\tilde{\beta}, then under uniform minimum privacy mandate σ¯\bar{\sigma} all platforms enter the market, and the user utility is higher than banning data sharing.

    • If ββ¯\beta\leq\underline{\beta}, then irrespective of the minimum privacy mandate, only low-cost platforms enter the market. In this case, banning data sharing achieves a higher user utility than any uniform minimum privacy mandate.

The first part is rather straightforward: if all platforms are low-cost, then they all enter the market and provide services to the user. Banning data sharing is user-optimal because it decreases the privacy loss of the user to zero while the user obtains the service gains from all platforms. The second part is more nuanced: here, if β\beta is large enough, then all platforms enter the market. Now, there are two opposing forces in place. If we ban data sharing, only low-cost platforms enter the market, and user utility then comprises only the service gained from low-cost platforms, while the privacy loss becomes zero. Therefore, banning data sharing decreases the service gains but also decreases the privacy losses. If we do not ban data sharing, the user utility comprises the service gains from all platforms and the privacy loss incurred by the data sharing of all platforms with the buyer. Therefore, not banning data sharing increases the service gains but also increases the privacy losses. For a large enough minimum uniform privacy mandate, and when there is at least one high-cost platform, the service gains of the user utility dominate the privacy loss, and therefore, not banning data sharing becomes the user-optimal regulation. Finally, to understand the last part, we observe that for small enough β\beta, only low-cost platforms enter the market, and therefore, the situation is effectively the same as the first part of Proposition 7. Notice that the assumption of large enough α\alpha and large/small enough β\beta is adopted for the same reason as that of Theorem 1 so that an equilibrium exists.

Non-uniform versus uniform minimum privacy mandate:

As noted above, when all platforms are low-cost or when β\beta is small enough so that high-cost platforms do not enter the market, then banning data sharing is optimal in the class of all (uniform or non-uniform) minimum privacy mandates. However, corresponding to the second part of Proposition 7, the user prefers a minimum privacy mandate to banning data sharing when there is at least one high-cost platform, and β\beta is large enough. This is because if we ban data sharing, the high-cost platforms do not enter, while with the “right” choice of uniform minimum privacy mandate, they enter the market. In this case, the user benefits from the services, while the privacy loss that they incur is small. However, with a uniform minimum privacy mandate, the low-cost platforms (that enter the market regardless of the minimum privacy mandate in place) also bring about privacy loss to the user. Therefore, a natural idea is to have a non-uniform minimum privacy mandate to help the user further. We next formalize this idea.

Proposition 8.

Suppose there is at least one high-cost platform and consider αα¯\alpha\geq\bar{\alpha} for α¯\bar{\alpha} established in Theorem 1. There exists β^\hat{\beta} and σ¯\bar{\sigma} such that for ββ^\beta\geq\hat{\beta} banning the low-cost platforms from data sharing and imposing minimum privacy mandate equal to σ¯\bar{\sigma} for all high-cost platforms generates a higher user utility than any non-uniform minimum privacy mandate.

With the non-uniform minimum privacy mandate described above, the user obtains the service gain of the low-cost platforms and incurs zero privacy loss from them. Moreover, for large enough β\beta, the high-cost platforms all enter the market. Therefore, the user gains from the high-cost platforms’ services, and when σ¯\bar{\sigma} is large enough, it incurs a small privacy loss. Notice that, as established in Proposition 7, with the optimal uniform minimum privacy mandate, both the low-cost and high-cost platforms enter the market. We prove that if we impose the same minimum privacy mandate but only for high-cost platforms and ban the low-cost platforms, the user obtains the same service gain from all platforms but incurs a smaller privacy loss. Moreover, all platforms still find it optimal to enter the market.

Proposition 7 and Proposition 8 establish that banning the data market is not necessarily user-optimal, especially when we have high-cost platforms (i.e., smaller platforms whose cost of providing the service is high). Moreover, a uniform minimum privacy mandate improves the user utility because it also incentivizes the high-cost platforms to enter the market to provide service to the user. Finally, a non-uniform minimum privacy mandate that bans the low-cost (i.e., large) platforms from data sharing and imposes a minimum privacy mandate on high-cost (i.e., small) platforms further improves the user utility. This is because this regulation not only incentivizes the high-cost platforms to enter the market to provide service to the user but also zeros out the privacy loss of the low-cost platforms.

7 Conclusion

In this paper, we study the intricate interplay between user privacy concerns and the dual incentives of online platforms: to safeguard user privacy and to monetize user data through sales to third-party buyers. To this end, we have constructed a multi-stage game model that captures the interactions between users, platforms, and data buyers. Our model delineates various scenarios where the interplay of platform service costs, the valuation of data, and privacy concerns dictate equilibrium outcomes. Additionally, we examine potential regulations designed to bolster user privacy and enhance market outcomes for users.

There are several exciting avenues for future research. In our framework, we have modeled a user as a representative of a market segment. Extending our analysis to a setting with multiple heterogeneous users is very interesting. This is because data externalities among different users may cause indirect information disclosure, which in turn could alter both user strategies and potential regulations aimed at enhancing user experience. Similarly, our focus has been on a representative data buyer who aims to acquire user data and learn their preferences. Extending our findings beyond this is very valuable. This is because user data is inherently multi-dimensional and can be aggregated from diverse sources, potentially modifying user strategies and calling for tailored regulations. Lastly, we have investigated regulations that restrict the transfer of user data to buyers. The challenge of developing alternative regulatory frameworks that address the intricacies of the data market—frameworks that are both adaptable and resilient enough to keep pace with swift technological advancements—remains an open and important area for exploration.

8 Acknowledgment

Alireza Fallah acknowledges support from the European Research Council Synergy Program, the National Science Foundation under grant number DMS-1928930, and the Alfred P. Sloan Foundation under grant G-2021-16778. The latter two grants correspond to his residency at the Simons Laufer Mathematical Sciences Institute (formerly known as MSRI) in Berkeley, California, during the Fall 2023 semester. Michael Jordan acknowledges support from the Mathematical Data Science program of the Office of Naval Research under grant number N00014-21-1-2840 and the European Research Council Synergy Program.

References

  • Abolhassani et al. [2017] M. Abolhassani, H. Esfandiari, M. Hajiaghayi, B. Lucier, and H. Yami. Market pricing for data streams. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 31, 2017.
  • Abowd [2018] J. M. Abowd. The us census bureau adopts differential privacy. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 2867–2867, 2018.
  • Acemoglu et al. [2022] D. Acemoglu, A. Makhdoumi, A. Malekian, and A. Ozdaglar. Too much data: Prices and inefficiencies in data markets. American Economic Journal: Microeconomics:Micro, 2022.
  • Acemoglu et al. [2023a] D. Acemoglu, A. Fallah, A. Makhdoumi, A. Malekian, and A. Ozdaglar. How good are privacy guarantees? platform architecture and violation of user privacy. Technical report, National Bureau of Economic Research, 2023a.
  • Acemoglu et al. [2023b] D. Acemoglu, A. Makhdoumi, A. Malekian, and A. Ozdaglar. A model of behavioral manipulation. Technical report, National Bureau of Economic Research, 2023b.
  • Acquisti et al. [2016] A. Acquisti, C. Taylor, and L. Wagman. The economics of privacy. Journal of Economic Literature, 54(2):442–492, 2016.
  • Ali et al. [2019] S. N. Ali, G. Lewis, and S. Vasserman. Voluntary disclosure and personalized pricing. NBER Working Paper, (w26592), 2019.
  • Argenziano and Bonatti [2023] R. Argenziano and A. Bonatti. Data markets with privacy-conscious consumers. AEA Papers and Proceedings, 113:191–96, May 2023. doi: 10.1257/pandp.20231083. URL https://www.aeaweb.org/articles?id=10.1257/pandp.20231083.
  • Bergemann and Bonatti [2015] D. Bergemann and A. Bonatti. Selling cookies. American Economic Journal: Microeconomics, 7(3):259–294, 2015.
  • Bergemann et al. [2018] D. Bergemann, A. Bonatti, and A. Smolin. The design and price of information. American Economic Review, 108(1):1–48, 2018.
  • Bergemann et al. [2020] D. Bergemann, A. Bonatti, and T. Gan. The economics of social data. arXiv preprint arXiv:2004.03107, 2020.
  • Bimpikis et al. [2021] K. Bimpikis, I. Morgenstern, and D. Saban. Data tracking under competition. Available at SSRN 3808228, 2021.
  • Breiman et al. [1984] L. Breiman, J. Friedman, R. A. Olshen, and C. J. Stone. Classification and Regression Trees. Routledge, 1st edition, 1984. doi: 10.1201/9781315139470.
  • Chen and Zheng [2019] Y. Chen and S. Zheng. Prior-free data acquisition for accurate statistical estimation. In Proceedings of the 2019 ACM Conference on Economics and Computation, pages 659–677, 2019.
  • Chen et al. [2018] Y. Chen, N. Immorlica, B. Lucier, V. Syrgkanis, and J. Ziani. Optimal data acquisition for statistical estimation. In Proceedings of the 2018 ACM Conference on Economics and Computation, pages 27–44, 2018.
  • Cross [2023a] R. Cross. How mastercard sells its ‘gold mine’ of transaction data, Sep 2023a. URL https://pirg.org/edfund/resources/how-mastercard-sells-data/.
  • Cross [2023b] R. Cross. The new data brokers: retailers, rewards apps & streaming services are selling your data. https://pirg.org/articles/the-new-data-brokers-retailers-rewards-apps-streaming-services-are-selling-your-data/, June 2023b.
  • Cummings et al. [2015] R. Cummings, K. Ligett, A. Roth, Z. S. Wu, and J. Ziani. Accuracy for sale: Aggregating data with a variance constraint. In Proceedings of the 2015 Conference on Innovations in Theoretical Computer Science, pages 317–324, 2015.
  • Cummings et al. [2023] R. Cummings, H. Elzayn, E. Pountourakis, V. Gkatzelis, and J. Ziani. Optimal data acquisition with privacy-aware agents. In 2023 IEEE Conference on Secure and Trustworthy Machine Learning (SaTML), pages 210–224. IEEE, 2023.
  • Dosis and Sand-Zantman [2022] A. Dosis and W. Sand-Zantman. The ownership of data. The Journal of Law, Economics, and Organization, 39(3):615–641, 02 2022.
  • Drakopoulos and Makhdoumi [2023] K. Drakopoulos and A. Makhdoumi. Providing data samples for free. Management Science, 69(6):3536–3560, 2023.
  • Dwork et al. [2014] C. Dwork, A. Roth, et al. The algorithmic foundations of differential privacy. Foundations and Trends® in Theoretical Computer Science, 9(3–4):211–407, 2014.
  • Fainmesser et al. [2022] I. P. Fainmesser, A. Galeotti, and R. Momot. Digital privacy. Management Science, 2022.
  • Falconer et al. [2023] T. Falconer, J. Kazempour, and P. Pinson. Towards replication-robust analytics markets, 2023.
  • Fallah et al. [2022] A. Fallah, A. Makhdoumi, A. Malekian, and A. E. Ozdaglar. Bridging central and local differential privacy in data acquisition mechanisms. Available at SSRN 4311351, 2022.
  • Fallah et al. [2023] A. Fallah, A. Makhdoumi, A. Malekian, and A. Ozdaglar. Optimal and differentially private data acquisition: Central and local mechanisms. Operations Research, 2023.
  • Gal and Aviv [2020] M. S. Gal and O. Aviv. The competitive effects of the GDPR. Journal of Competition Law & Economics, 16(3):349–391, 2020.
  • Ghosh and Roth [2011] A. Ghosh and A. Roth. Selling privacy at auction. In Proceedings of the 12th ACM conference on Electronic commerce, pages 199–208, 2011.
  • Ichihashi [2020a] S. Ichihashi. Dynamic privacy choices. In Proceedings of the 21st ACM Conference on Economics and Computation, pages 539–540, 2020a.
  • Ichihashi [2020b] S. Ichihashi. Online privacy and information disclosure by consumers. American Economic Review, 110(2):569–595, 2020b.
  • Ichihashi [2021] S. Ichihashi. Competing data intermediaries. The RAND Journal of Economics, 52(3):515–537, 2021.
  • Immorlica et al. [2021] N. Immorlica, I. A. Kash, and B. Lucier. Buying data over time: Approximately optimal strategies for dynamic data-driven decisions. In 12th Innovations in Theoretical Computer Science Conference (ITCS 2021). Schloss Dagstuhl-Leibniz-Zentrum für Informatik, 2021.
  • Janssen et al. [2022] R. Janssen, R. Kesler, M. E. Kummer, and J. Waldfogel. GDPR and the lost generation of innovative apps. Technical report, National Bureau of Economic Research, 2022.
  • Jones and Tonetti [2020] C. I. Jones and C. Tonetti. Nonrivalry and the economics of data. American Economic Review, 110(9):2819–2858, 2020.
  • Kamenica and Gentzkow [2011] E. Kamenica and M. Gentzkow. Bayesian persuasion. American Economic Review, 101(6):2590–2615, 2011.
  • Karimireddy et al. [2022] S. P. Karimireddy, W. Guo, and M. I. Jordan. Mechanisms that incentivize data sharing in federated learning. arXiv preprint arXiv:2207.04557, 2022.
  • Ligett and Roth [2012] K. Ligett and A. Roth. Take it or leave it: Running a survey when privacy comes at a cost. In International Workshop on Internet and Network Economics, pages 378–391. Springer, 2012.
  • Madsen and Vellodi [2023] E. Madsen and N. Vellodi. Insider imitation. Available at SSRN 3832712, 2023.
  • Nissim et al. [2014] K. Nissim, S. Vadhan, and D. Xiao. Redrawing the boundaries on purchasing data from privacy-sensitive individuals. In Proceedings of the 5th Conference on Innovations in Theoretical Computer Science, pages 411–422, 2014.
  • Pino [2022] F. Pino. The microeconomics of data–a survey. Journal of Industrial and Business Economics, 49(3):635–665, 2022.
  • Posner and Weyl [2018] E. Posner and E. Weyl. Radical markets: Uprooting Capitalism and Democracy for a Just Society. Princeton University Press, 2018.
  • Ravichandran and Korula [2019] D. Ravichandran and N. Korula. Effect of disabling third-party cookies on publisher revenue. Google Report, 2019.

Appendix A Proofs:

This appendix included the omitted proofs from the text.

Proof of Lemma 1

Without loss of generality, we prove the claim for S=[K]S=[K]. Let us recall a property of normal distributions: If 𝐱\mathbf{x} is a nn-dimensional normal random variable is partitioned as

𝐱=[𝐱1𝐱2] with sizes [k×1(nk)×1]\displaystyle{\mathbf{x}}={\begin{bmatrix}{\mathbf{x}}_{1}\\ {\mathbf{x}}_{2}\end{bmatrix}}{\text{ with sizes }}{\begin{bmatrix}k\times 1\\ (n-k)\times 1\end{bmatrix}}

with with sizes

[k×kk×(nk)(nk)×k(nk)×(nk)]\displaystyle\begin{bmatrix}k\times k&k\times(n-k)\\ (n-k)\times k&(n-k)\times(n-k)\end{bmatrix}

and accordingly 𝝁\bm{\mu} is partitioned as

𝝁=[𝝁1𝝁2]\displaystyle\bm{\mu}=\begin{bmatrix}\bm{\mu}_{1}\\ \bm{\mu}_{2}\end{bmatrix}

with sizes

[k×1(nk)×1]\displaystyle\begin{bmatrix}k\times 1\\ (n-k)\times 1\end{bmatrix}

and 𝚺\bm{\Sigma} is partitioned as

𝚺=[𝚺11𝚺12𝚺21𝚺22]\displaystyle\bm{\Sigma}=\begin{bmatrix}\bm{\Sigma}_{11}&\bm{\Sigma}_{12}\\ \bm{\Sigma}_{21}&\bm{\Sigma}_{22}\end{bmatrix}

with sizes

[k×kk×(nk)(nk)×k(nk)×(nk)]\displaystyle\begin{bmatrix}k\times k&k\times(n-k)\\ (n-k)\times k&(n-k)\times(n-k)\end{bmatrix}

then the distribution of 𝐱1\mathbf{x}_{1} conditional on 𝐱2=𝐚\mathbf{x}_{2}=\mathbf{a} is normal with covariance matrix

𝚺11𝚺12𝚺221𝚺21.\displaystyle{\bm{\Sigma}}_{11}-{\bm{\Sigma}}_{12}{\bm{\Sigma}}_{22}^{-1}{\bm{\Sigma}}_{21}.

The proof of this lemma follows directly from the above fact. \blacksquare

Proof Lemma 2

Both parts follow from the conditional expectation being the MSE estimator. To see the first part, notice that we can always throw away data points. To see the second part, notice that we can always add extra noise to the data points. \blacksquare

Proof of Lemma 3

Using Lemma 1, we first derive the leaked information to the buyer. In particular, if both platforms enter the market and the user’s data is shared with them and then sold to the buyer, then the leaked information is given by

(σ1,σ2)=γ12(2+σ22)+γ22(2+σ12)2γ12γ22(2+σ12)(2+σ22)γ12γ22.\displaystyle\mathcal{I}(\sigma_{1},\sigma_{2})=\frac{\gamma_{1}^{2}(2+\sigma_{2}^{2})+\gamma_{2}^{2}(2+\sigma_{1}^{2})-2\gamma_{1}^{2}\gamma_{2}^{2}}{(2+\sigma_{1}^{2})(2+\sigma_{2}^{2})-\gamma_{1}^{2}\gamma_{2}^{2}}.

On the other hand, if the user’s data is only shared with the buyer through platform ii, then the leaked information to the buyer is given by

(σi)=γi22+σi2.\displaystyle\mathcal{I}(\sigma_{i})=\frac{\gamma_{i}^{2}}{2+\sigma_{i}^{2}}.

To prove this lemma, without loss of generality, we need to establish that if the buyer has the extra data of platform jj, the gain in acquiring the data of platform ii is smaller. By invoking the fact mentioned in the proof of Lemma 1 for the normal variable (𝒚T𝜽,𝒙iT𝜽+𝒩(0,1)+𝒩(0,σi2),𝒙jT𝜽+𝒩(0,1)+𝒩(0,σj2))(\bm{y}^{T}\bm{\theta},\bm{x}_{i}^{T}\bm{\theta}+\mathcal{N}(0,1)+\mathcal{N}(0,\sigma_{i}^{2}),\bm{x}_{j}^{T}\bm{\theta}+\mathcal{N}(0,1)+\mathcal{N}(0,\sigma_{j}^{2})) conditional on the rest of the data points, the inequality stated in the lemma becomes equivalent to

γ~i22+σi2γ~i2(2+σj2)+γ~j2(2+σi2)2γ~i2γ~j2(2+σi2)(2+σj2)γ~i2γ~j2γ~j22+σj2,\displaystyle\frac{\tilde{\gamma}_{i}^{2}}{2+\sigma_{i}^{2}}\geq\frac{\tilde{\gamma}_{i}^{2}(2+\sigma_{j}^{2})+\tilde{\gamma}_{j}^{2}(2+\sigma_{i}^{2})-2\tilde{\gamma}^{2}_{i}\tilde{\gamma}^{2}_{j}}{(2+\sigma_{i}^{2})(2+\sigma_{j}^{2})-\tilde{\gamma}_{i}^{2}\tilde{\gamma}_{j}^{2}}-\frac{\tilde{\gamma}_{j}^{2}}{2+\sigma_{j}^{2}},

where for all [K]\ell\in[K], γ~=γ(var~)1/2\tilde{\gamma}_{\ell}=\gamma_{\ell}(\tilde{\mathrm{var}})^{1/2} where var~\tilde{\mathrm{var}} is the variance of 𝜽T𝜽\bm{\theta}^{T}\bm{\theta} conditional on the rest of the data points. The above inequality holds true as it is equivalent to

γ~i2γ~j2(γ~i22+σi2+γ~j22+σi2)2γ~i2γ~j2\displaystyle\tilde{\gamma}_{i}^{2}\tilde{\gamma}_{j}^{2}\left(\frac{\tilde{\gamma}_{i}^{2}}{2+\sigma_{i}^{2}}+\frac{\tilde{\gamma}_{j}^{2}}{2+\sigma_{i}^{2}}\right)\leq 2\tilde{\gamma}_{i}^{2}\tilde{\gamma}_{j}^{2}

that holds as γ~i,γ~j1\tilde{\gamma}_{i},\tilde{\gamma}_{j}\leq 1. \blacksquare

Proof of Lemma 4

We prove part one and part two follows an identical argument. We use a similar argument to the one presented in the proof of Lemma 3. We establish this inequality by considering a one-by-one change in the noise variance.

If 𝝈j=𝝈j\bm{\sigma}_{-j}=\bm{\sigma}^{\prime}_{-j} and σj2σj2{\sigma^{\prime}_{j}}^{2}\geq\sigma_{j}^{2}, then the inequality becomes equivalent to

γ~i2(2+σj2)+γ~j2(2+σi2)2γ~i2γ~j2(2+σi2)(2+σj2)γ~i2γ~j2γ~j22+σj2\displaystyle\frac{\tilde{\gamma}_{i}^{2}(2+\sigma_{j}^{2})+\tilde{\gamma}_{j}^{2}(2+\sigma_{i}^{2})-2\tilde{\gamma}^{2}_{i}\tilde{\gamma}^{2}_{j}}{(2+\sigma_{i}^{2})(2+\sigma_{j}^{2})-\tilde{\gamma}_{i}^{2}\tilde{\gamma}_{j}^{2}}-\frac{\tilde{\gamma}_{j}^{2}}{2+\sigma_{j}^{2}}
γ~i2(2+σj2)+γ~j2(2+σi2)2γ~i2γ~j2(2+σi2)(2+σj2)γ~i2γ~j2γ~j22+σj2.\displaystyle\leq\frac{\tilde{\gamma}_{i}^{2}(2+{\sigma^{\prime}_{j}}^{2})+\tilde{\gamma}_{j}^{2}(2+\sigma_{i}^{2})-2\tilde{\gamma}^{2}_{i}\tilde{\gamma}^{2}_{j}}{(2+\sigma_{i}^{2})(2+{\sigma^{\prime}_{j}}^{2})-\tilde{\gamma}_{i}^{2}\tilde{\gamma}_{j}^{2}}-\frac{\tilde{\gamma}_{j}^{2}}{2+{\sigma^{\prime}_{j}}^{2}}.

The above inequality holds because the function

xγ~i2(2+x)+γ~j2(2+σi2)2γ~i2γ~j2(2+σi2)(2+x)γ~i2γ~j2γ~j22+x\displaystyle x\mapsto\frac{\tilde{\gamma}_{i}^{2}(2+x)+\tilde{\gamma}_{j}^{2}(2+\sigma_{i}^{2})-2\tilde{\gamma}^{2}_{i}\tilde{\gamma}^{2}_{j}}{(2+\sigma_{i}^{2})(2+x)-\tilde{\gamma}_{i}^{2}\tilde{\gamma}_{j}^{2}}-\frac{\tilde{\gamma}_{j}^{2}}{2+x}

is increasing, which can be seen by evaluating its derivative. \blacksquare

Proof of Proposition 1

We first argue that, for any i[K]i\in[K], the equilibrium price of platform ii must be such that the buyer finds it optimal to purchase from this platform in the corresponding buyer equilibrium. To see this notice that if the buyer does not purchase from platform ii, the platform’s utility from interacting with the buyer becomes 0. Now, if the platform deviates and selects the price

pi=β((𝝈,𝒆,𝒂,𝟏)(𝝈,𝒆,𝒂,(bi=0,𝟏i)))\displaystyle p_{i}=\beta\left(\mathcal{I}(\bm{\sigma},\bm{e},\bm{a},\mathbf{1})-\mathcal{I}(\bm{\sigma},\bm{e},\bm{a},(b_{i}=0,\mathbf{1}_{-i}))\right)

then the buyer always finds it optimal to purchase from this platform, and therefore, the platform’s utility weakly increases from 0. This is because for any 𝒃i\bm{b}_{-i}, we have

𝒰buyer(𝝈,𝒆,𝒂,(pi,𝒑i),(bi=1,𝒃i))𝒰buyer(𝝈,𝒆,𝒂,(pi,𝒑i),(bi=0,𝒃i))\displaystyle\mathcal{U}_{\text{buyer}}(\bm{\sigma},\bm{e},\bm{a},(p_{i},\bm{p}_{-i}),(b_{i}=1,\bm{b}_{-i}))-\mathcal{U}_{\text{buyer}}(\bm{\sigma},\bm{e},\bm{a},(p_{i},\bm{p}_{-i}),(b_{i}=0,\bm{b}_{-i}))
=β[((𝝈,𝒆,𝒂,(bi=1,𝒃i))(𝝈,𝒆,𝒂,(bi=0,𝒃i)))((𝝈,𝒆,𝒂,𝟏)(𝝈,𝒆,𝒂,(bi=0,𝟏i)))]\displaystyle=\beta\left[\left(\mathcal{I}(\bm{\sigma},\bm{e},\bm{a},(b_{i}=1,\bm{b}_{-i}))-\mathcal{I}(\bm{\sigma},\bm{e},\bm{a},(b_{i}=0,\bm{b}_{-i}))\right)-\left(\mathcal{I}(\bm{\sigma},\bm{e},\bm{a},\mathbf{1})-\mathcal{I}(\bm{\sigma},\bm{e},\bm{a},(b_{i}=0,\mathbf{1}_{-i}))\right)\right]
(a)0\displaystyle\overset{(a)}{\geq}0

where (a) follows from Lemma 3. \blacksquare

Proof of Lemma 5

This lemma directly follows the submodularity of the leaked information in actions, established in Lemma 3. \blacksquare

Proof of Proposition 2

Suppose the contrary that there exists a platform equilibrium strategy (𝝈,𝒆)(\bm{\sigma},\bm{e}) for which ei=1e_{i}=1, ai=0a_{i}=0 and 𝒂𝒜E(𝝈,𝒆)\bm{a}\in\mathcal{A}^{\mathrm{E}}(\bm{\sigma},\bm{e}). We argue that platform ii has a profitable deviation. In particular, notice that the current utility of platform ii is ci-c_{i}, as the buyer will not pay any non-zero price for this platform’s data. Now, if platform ii chooses e~i=1\tilde{e}_{i}=1, p~i=0\tilde{p}_{i}=0 and σ~i=\tilde{\sigma}_{i}=\infty, then in any 𝒂𝒜E((σ~i,𝝈i),𝒆)\bm{a}\in\mathcal{A}^{\mathrm{E}}((\tilde{\sigma}_{i},\bm{\sigma}_{-i}),\bm{e}) we must have that ai=1a_{i}=1 and therefore the platform ii’s utility increases to ici\mathcal{I}_{i}-c_{i}. \blacksquare

Proof of Lemma 6

Recall that we argued that the equilibrium, if exists, belong to Σ~(1,1)Σ~(0,0)\tilde{\Sigma}_{(1,1)}\cap\tilde{\Sigma}_{(0,0)}. Furthermore, we stated that we only need to check deviations in Figure 2(a). To make sure such a deviation does not happen for some equilibrium (σ1,σ2)Σ~(1,1)Σ~(0,0)(\sigma_{1},\sigma_{2})\in\tilde{\Sigma}_{(1,1)}\cap\tilde{\Sigma}_{(0,0)}, we must have

𝒰1((σ1,σ2)),𝟏,𝐩,𝟏)=12+β((σ1,σ2)(σ2))12+β(σ~1)=𝒰1((σ~1,σ2)),(1,0)),𝐩,𝟏).\displaystyle\mathcal{U}_{1}((\sigma_{1},\sigma_{2})),\mathbf{1},\mathbf{p}^{*},\mathbf{1})=\frac{1}{2}+\beta\left(\mathcal{I}(\sigma_{1},\sigma_{2})-\mathcal{I}(\sigma_{2})\right)\geq\frac{1}{2}+\beta\mathcal{I}(\tilde{\sigma}_{1})=\mathcal{U}_{1}((\tilde{\sigma}_{1},\sigma_{2})),(1,0)),\mathbf{p}^{*},\mathbf{1}). (15)

Notice that (σ1,σ2)(\sigma_{1},\sigma_{2}) is on the boundary of Σ~(1,1)\tilde{\Sigma}_{(1,1)} and Σ~(0,0)\tilde{\Sigma}_{(0,0)} which implies that

𝒰user(σ1,σ2)=1α(σ1,σ2)=0,\mathcal{U}_{\text{user}}(\sigma_{1},\sigma_{2})=1-\alpha\mathcal{I}(\sigma_{1},\sigma_{2})=0, (16)

and thus, (σ1,σ2)=1/α\mathcal{I}(\sigma_{1},\sigma_{2})=1/\alpha. Using this, we can rewrite the condition (15) as

1α(σ2)+(σ~1).\frac{1}{\alpha}\geq\mathcal{I}(\sigma_{2})+\mathcal{I}(\tilde{\sigma}_{1}). (17)

Notice that we can similarly consider the deviation of the second platform by increasing its noise variance to σ~22\tilde{\sigma}_{2}^{2} in Figure 2(a). To ensure this deviation is also not profitable, we should have

1α(σ1)+(σ~2).\frac{1}{\alpha}\geq\mathcal{I}(\sigma_{1})+\mathcal{I}(\tilde{\sigma}_{2}). (18)

The following lemma establishes the necessary and sufficient condition for (17) and (18) to hold for some (σ1,σ2)Σ~(1,1)Σ~(0,0)(\sigma_{1},\sigma_{2})\in\tilde{\Sigma}_{(1,1)}\cap\tilde{\Sigma}_{(0,0)}.

Notice that if we move up on the boundary Σ~(1,1)Σ~(0,0)\tilde{\Sigma}_{(1,1)}\cap\tilde{\Sigma}_{(0,0)}, σ2\sigma_{2} and σ~1\tilde{\sigma}_{1} both would increase, and hence, (σ2)\mathcal{I}(\sigma_{2}) and (σ~1)\mathcal{I}(\tilde{\sigma}_{1}) both would decrease. Therefore, if (17) holds for some (σ1,σ2)Σ~(1,1)Σ~(0,0)(\sigma_{1},\sigma_{2})\in\tilde{\Sigma}_{(1,1)}\cap\tilde{\Sigma}_{(0,0)}, then it would hold for any upper point on the boundary. Similarly, if (18) holds for some (σ1,σ2)Σ~(1,1)Σ~(0,0)(\sigma_{1},\sigma_{2})\in\tilde{\Sigma}_{(1,1)}\cap\tilde{\Sigma}_{(0,0)}, then it would hold for any lower point on the boundary. Hence, if a point satisfies both conditions, then for any other point on the boundary Σ~(1,1)Σ~(0,0)\tilde{\Sigma}_{(1,1)}\cap\tilde{\Sigma}_{(0,0)}, at least one of the conditions would hold.

Next, we show that, for the point given by (10), both conditions hold if αα¯\alpha\geq\bar{\alpha} and neither hold if α<α¯\alpha<\bar{\alpha}. Given our discussion above, this would show that for α<α¯\alpha<\bar{\alpha}, there is no point for which both conditions hold and hence the proof would be complete.

Now, let us first see why the point given by (10) is actually on the boundary Σ~(1,1)Σ~(0,0)\tilde{\Sigma}_{(1,1)}\cap\tilde{\Sigma}_{(0,0)}. Consider point XX in Figure 3. For this point, we have

γ122+σ12=γ222+σ22=12α.\frac{\gamma_{1}^{2}}{2+\sigma_{1}^{2}}=\frac{\gamma_{2}^{2}}{2+\sigma_{2}^{2}}=\frac{1}{2\alpha}.
Refer to caption
Figure 3: Illustrating the point given by (10).

Hence, the line

2+σ12γ12=2+σ22γ22\frac{2+\sigma_{1}^{2}}{\gamma_{1}^{2}}=\frac{2+\sigma_{2}^{2}}{\gamma_{2}^{2}} (19)

passes through point XX, and therefore, it intersects with the boundary Σ~(1,1)Σ~(0,0)\tilde{\Sigma}_{(1,1)}\cap\tilde{\Sigma}_{(0,0)}. Let us call this intersection AA (see Figure 3) and denote it by (σ1A,σ2A)(\sigma_{1}^{A},\sigma_{2}^{A}). Using (16) along with the fact that AA is on the line (19), we can derive that

2+(σ1A)2γ12=2+(σ2A)2γ22=2α1.\frac{2+(\sigma_{1}^{A})^{2}}{\gamma_{1}^{2}}=\frac{2+(\sigma_{2}^{A})^{2}}{\gamma_{2}^{2}}=2\alpha-1. (20)

Now, let us consider the deviation to (σ1~,σ2A)(\tilde{\sigma_{1}},\sigma_{2}^{A}) which is on the boundary of Σ~(1,1)\tilde{\Sigma}_{(1,1)} and Σ~(1,0)\tilde{\Sigma}_{(1,0)}, i.e.,

1α(σ1~,σ2A)=12α(σ1~).1-\alpha\mathcal{I}(\tilde{\sigma_{1}},\sigma_{2}^{A})=\frac{1}{2}-\alpha\mathcal{I}(\tilde{\sigma_{1}}). (21)

Let

η=γ122+σ1~2.\eta=\frac{\gamma_{1}^{2}}{2+\tilde{\sigma_{1}}^{2}}. (22)

Plugging (20) and (22) into (21) and simplifying the equation yields

η{114α16α216α+14α,114α+16α216α+14α}.\eta\in\left\{1-\frac{1}{4\alpha}-\frac{\sqrt{16\alpha^{2}-16\alpha+1}}{4\alpha},1-\frac{1}{4\alpha}+\frac{\sqrt{16\alpha^{2}-16\alpha+1}}{4\alpha}\right\}. (23)

Next we argue that we should pick the smaller (first) solution. Notice that since σ~1σ1A\tilde{\sigma}_{1}\geq\sigma_{1}^{A}, we have η=(σ~1)(σ1A)=1/(2α1)\eta=\mathcal{I}(\tilde{\sigma}_{1})\leq\mathcal{I}(\sigma_{1}^{A})=1/(2\alpha-1). In addition, Assumption 2, or equivalently (9), implies α3/2\alpha\geq 3/2. In this regime, we have 114α>12α11-\frac{1}{4\alpha}>\frac{1}{2\alpha-1}. Thus, we should pick the smaller solution above.

Now, we check the condition (17). Notice that we need to find α\alpha such that

η+12α11α.\eta+\frac{1}{2\alpha-1}\leq\frac{1}{\alpha}. (24)

We can verify that the right-hand side minus the left-hand side is equal to zero at α¯1.88\bar{\alpha}\approx 1.88 and positive afterward. Notice that η\eta and therefore the condition (24) is symmetric with respect to both platforms, and consequently, considering the other deviation to (σ1A,σ~2)(\sigma_{1}^{A},\tilde{\sigma}_{2}) leads to the exact same condition on α\alpha. This completes the proof of our claim that either both conditions hold if αα¯\alpha\geq\bar{\alpha} and neither hold if α<α¯\alpha<\bar{\alpha}. \blacksquare

Proof of Proposition 3

Suppose an equilibrium (σ1,σ2)(\sigma_{1}^{*},\sigma_{2}^{*}) exists in which both platforms enter the market. As we argued in Section 4, in this case (σ1,σ2)Σ~(1,1)Σ~(0,0)(\sigma_{1}^{*},\sigma_{2}^{*})\in\tilde{\Sigma}_{(1,1)}\cap\tilde{\Sigma}_{(0,0)}. Next, notice that for any i{1,2}i\in\{1,2\}, the platform ii’s utility should be nonnegative. Therefore, we have

12ci+β((σ1,σ2)(σi))0.\frac{1}{2}-c_{i}+\beta(\mathcal{I}(\sigma_{1}^{*},\sigma_{2}^{*})-\mathcal{I}(\sigma_{-i}^{*}))\geq 0. (25)

Notice that, since (σ1,σ2)Σ~(1,1)Σ~(0,0)(\sigma_{1}^{*},\sigma_{2}^{*})\in\tilde{\Sigma}_{(1,1)}\cap\tilde{\Sigma}_{(0,0)}, we have (σ1,σ2)=1/α\mathcal{I}(\sigma_{1}^{*},\sigma_{2}^{*})=1/\alpha. Also, the minimum of (σi)\mathcal{I}(\sigma_{-i}^{*}) corresponds to point BB for i=1i=1 and point CC for i=2i=2 in Figure 2(b) and is equal to 1/(2α)1/(2\alpha). This implies the lower bound on β\beta.

Now, suppose ββ¯\beta\geq\bar{\beta}. We claim the point (σ1A,σ2A)(\sigma_{1}^{A},\sigma_{2}^{A}), defined in the proof of Lemma 6, can be an equilibrium. We just need to check when the platform’s utility is nonnegative. In particular, the first platform’s utility at (σ1A,σ2A)(\sigma_{1}^{A},\sigma_{2}^{A}) is given by

12c1+β((σ1A,σ2A)(σ2A))=12+β(α1)α(2α1),\frac{1}{2}-c_{1}+\beta\left(\mathcal{I}(\sigma_{1}^{A},\sigma_{2}^{A})-\mathcal{I}(\sigma_{2}^{A})\right)=\frac{1}{2}+\frac{\beta(\alpha-1)}{\alpha(2\alpha-1)}, (26)

which is nonnegative for ββ¯\beta\geq\bar{\beta}. We can similarly write the second platform’s utility at the point AA. \blacksquare

Proof of Theorem 1

We make use of the following lemmas in this proof.

Lemma 7.

For any 𝐞\bm{e}, the noise level equilibrium strategy, if it exists, belongs to the boundary of the set Σ𝟏,𝐞\Sigma_{\mathbf{1},\bm{e}}.

Proof of Lemma 7: Using Proposition 2, we know that the equilibrium noise variance belongs to Σ𝟏,𝒆\Sigma_{\mathbf{1},\bm{e}}. We next argue that it cannot belong to the interior of Σ𝟏,𝒆\Sigma_{\mathbf{1},\bm{e}}. To see this, suppose the contrary that 𝝈int(Σ𝟏,𝒆)\bm{\sigma}\in\mathrm{int}(\Sigma_{\mathbf{1}},\bm{e}) is an equilibrium. There exists i[K]i\in[K] such that by locally decreasing σi\sigma_{i}, we remain in the set Σ𝟏,𝒆\Sigma_{\mathbf{1},\bm{e}}. We claim that platform ii has an incentive to decrease its noise variance. First, note that by decreasing its noise variance, we remain in the set Σ𝟏,𝒆\Sigma_{\mathbf{1},\bm{e}} and that the user still finds it optimal to share with platform ii. However, the payment she receives from the buyer, which is given by

β((𝝈,𝒆,𝒂,𝟏)(𝝈,𝒆,𝒂,(bi=0,𝟏i)))\displaystyle\beta\left(\mathcal{I}(\bm{\sigma},\bm{e},\bm{a},\mathbf{1})-\mathcal{I}(\bm{\sigma},\bm{e},\bm{a},(b_{i}=0,\mathbf{1}_{-i}))\right)

increases because of Lemma 4. \blacksquare

Lemma 8.

For any 𝐞\bm{e}, the noise level equilibrium strategy, if it exists, belongs to Σ𝟏,𝐞Σ𝟎,𝐞\Sigma_{\mathbf{1},\bm{e}}\cap\Sigma_{\mathbf{0},\bm{e}}.

Proof of Lemma 8: Using Lemma 7, we know that the equilibrium noise level, if it exists, belongs to the boundary of Σ𝟏,𝒆\Sigma_{\mathbf{1},\bm{e}} and Σ𝒂,𝒆\Sigma_{\bm{a},\bm{e}} for some 𝒂{0,1}K{𝟏}\bm{a}\in\{0,1\}^{K}\setminus\{\mathbf{1}\}. We argue that it cannot belong to Σ𝟏,𝒆Σ𝒂,𝒆\Sigma_{\mathbf{1},\bm{e}}\cap\Sigma_{\bm{a},\bm{e}} when 𝒂𝟎\bm{a}\neq\mathbf{0}. Suppose the contrary that σΣ𝟏,𝒆Σ𝒂,𝒆\sigma\in\Sigma_{\mathbf{1},\bm{e}}\cap\Sigma_{\bm{a},\bm{e}} is an equilibrium where ai=1a_{i}=1 and also aj=0a_{j}=0. We claim that platform ii has an incentive to decrease its noise level. In particular, we argue that by decreasing its noise level, we go to int(Σ𝟏,𝒆)\mathrm{int}(\Sigma_{\mathbf{1},\bm{e}}) and that platform ii’s utility increases. Notice that it suffices to show that by decreasing σi\sigma_{i}, we remain in the set Σ𝟏,𝒆\Sigma_{\mathbf{1},\bm{e}}. This is because by invoking Lemma 4, the payment of platform ii increases, and the users keep sharing with platform ii. For 𝝈\bm{\sigma}, the indifference condition for the user implies

i:ei=1aiiα(𝝈,𝒆,𝒂,𝟏)=i:ei=1iα(𝝈,𝒆,𝟏,𝟏)\displaystyle\sum_{i:~{}e_{i}=1}a_{i}\mathcal{I}_{i}-\alpha\mathcal{I}(\bm{\sigma},\bm{e},\bm{a},\mathbf{1})=\sum_{i:~{}e_{i}=1}\mathcal{I}_{i}-\alpha\mathcal{I}(\bm{\sigma},\bm{e},\mathbf{1},\mathbf{1})

which can be written as

i:ei=1,ai=0i=α((𝝈,𝒆,𝟏,𝟏)(𝝈,𝒆,𝒂,𝟏)).\displaystyle\sum_{i:~{}e_{i}=1,a_{i}=0}\mathcal{I}_{i}=\alpha\left(\mathcal{I}(\bm{\sigma},\bm{e},\mathbf{1},\mathbf{1})-\mathcal{I}(\bm{\sigma},\bm{e},\bm{a},\mathbf{1})\right).

Now if ai=1a_{i}=1, then by decreasing σi\sigma_{i}, the left-hand side does not change, and the right-hand decreases (from Lemma 4). Therefore, by decreasing σi\sigma_{i}, we remain in the set Σ𝟏,𝒆\Sigma_{\mathbf{1},\bm{e}}. \blacksquare

For a given 𝒆\bm{e} with ei=1e_{i}=1, platform ii finds it optimal to enter the market if

12ci+β((𝝈,𝒆,(ai=1,𝒂i),𝟏)(𝝈,𝒆,(ai=0,𝒂i),𝟏))0.\displaystyle\frac{1}{2}-c_{i}+\beta\left(\mathcal{I}(\bm{\sigma},\bm{e},(a_{i}=1,\bm{a}_{-i}),\mathbf{1})-\mathcal{I}(\bm{\sigma},\bm{e},(a_{i}=0,\bm{a}_{-i}),\mathbf{1})\right)\geq 0.

Therefore, by letting

ββ¯:=supi[K],(ei=1,𝒆i),𝝈Σ𝟏,𝒆Σ𝟎,𝒆ci12(𝝈,𝒆,(ai=1,𝒂i),𝟏)(𝝈,𝒆,(ai=0,𝒂i),𝟏),\displaystyle\beta\geq\bar{\beta}:=\sup_{i\in[K],(e_{i}=1,\bm{e}_{-i}),\bm{\sigma}\in\Sigma_{\mathbf{1},\bm{e}}\cap\Sigma_{\mathbf{0},\bm{e}}}\frac{c_{i}-\frac{1}{2}}{\mathcal{I}(\bm{\sigma},\bm{e},(a_{i}=1,\bm{a}_{-i}),\mathbf{1})-\mathcal{I}(\bm{\sigma},\bm{e},(a_{i}=0,\bm{a}_{-i}),\mathbf{1})},

all platforms enter the market. We next argue that this term is finite. Note that, for any given value of 𝒆i\bm{e}_{-i}, the fact that 𝝈Σ𝟏,𝒆Σ𝟎,𝒆\bm{\sigma}\in\Sigma_{\mathbf{1},\bm{e}}\cap\Sigma_{\mathbf{0},\bm{e}} implies that the term (𝝈,𝒆,(ai=1,𝒂i),𝟏)\mathcal{I}(\bm{\sigma},\bm{e},(a_{i}=1,\bm{a}_{-i}),\mathbf{1}) in the denominator is equal to

|{i:ei=1}|2α.\frac{|\{i:e_{i}=1\}|}{2\alpha}.

As a result, we can rewrite β¯\bar{\beta} as

β¯=maxi[K],(ei=1,𝒆i)ci1/2|{i:ei=1}|/(2α)sup𝝈Σ𝟏,𝒆Σ𝟎,𝒆(𝝈,𝒆,(ai=0,𝒂i),𝟏).\bar{\beta}=\max_{i\in[K],(e_{i}=1,\bm{e}_{-i})}\frac{c_{i}-1/2}{{|\{i:e_{i}=1\}|}/{(2\alpha)}-\sup_{\bm{\sigma}\in\Sigma_{\mathbf{1},\bm{e}}\cap\Sigma_{\mathbf{0},\bm{e}}}\mathcal{I}(\bm{\sigma},\bm{e},(a_{i}=0,\bm{a}_{-i}),\mathbf{1})}.

Notice that, for any given βi\beta_{-i}, the denominator is positive because the amount of leaked information strictly decreases when one less user shares information. Finally, β¯\bar{\beta} is finite becasue the outer maximum is taken over finitely many elements.

Similarly, for a given 𝒆\bm{e} with ei=1e_{i}=1, when ci>1/2c_{i}>1/2 platform ii finds it optimal to not enter the market if

12ci+β((𝝈,𝒆,(ai=1,𝒂i),𝟏)(𝝈,𝒆,(ai=0,𝒂i),𝟏))0.\displaystyle\frac{1}{2}-c_{i}+\beta\left(\mathcal{I}(\bm{\sigma},\bm{e},(a_{i}=1,\bm{a}_{-i}),\mathbf{1})-\mathcal{I}(\bm{\sigma},\bm{e},(a_{i}=0,\bm{a}_{-i}),\mathbf{1})\right)\leq 0.

Therefore, by letting

ββ¯:=infi[K],ci>12,(ei=1,𝒆i),𝝈Σ𝟏,𝒆ci12(𝝈,𝒆,(ai=1,𝒂i),𝟏)(𝝈,𝒆,(ai=0,𝒂i),𝟏),\displaystyle\beta\leq\underline{\beta}:=\inf_{i\in[K],c_{i}>\frac{1}{2},(e_{i}=1,\bm{e}_{-i}),\bm{\sigma}\in\Sigma_{\mathbf{1},\bm{e}}}\frac{c_{i}-\frac{1}{2}}{\mathcal{I}(\bm{\sigma},\bm{e},(a_{i}=1,\bm{a}_{-i}),\mathbf{1})-\mathcal{I}(\bm{\sigma},\bm{e},(a_{i}=0,\bm{a}_{-i}),\mathbf{1})},

all platforms enter the market.

We next prove that assuming a set of platforms that have entered the market, for large enough α\alpha, there exists an equilibrium. Without loss of generality, let us assume all platforms have entered the market. 𝝈Σ𝟏,𝟏Σ𝟎,𝟏\bm{\sigma}\in\Sigma_{\mathbf{1},\mathbf{1}}\cap\Sigma_{\mathbf{0},\mathbf{1}} is an equilibrium noise level if no platform has a deviation. Let us consider platform 1. For this platform, the possible deviations are to Σ𝟏,𝟏Σ𝒂,𝟏\Sigma_{\mathbf{1},\mathbf{1}}\cap\Sigma_{\bm{a},\mathbf{1}} where a1=0a_{1}=0 and 𝒂1\bm{a}_{-1} can be any vector in {0,1}K1\{0,1\}^{K-1}. We next argue that 𝝈Σ𝟏,𝟏Σ𝟎,𝟏\bm{\sigma}\in\Sigma_{\mathbf{1},\mathbf{1}}\cap\Sigma_{\mathbf{0},\mathbf{1}} for which

γi2+σi2=t for all i[K]\displaystyle\frac{\gamma_{i}}{2+\sigma_{i}^{2}}=t\text{ for all }i\in[K]

is a noise level equilibrium for large enough α\alpha. Similar to the analysis with K=2K=2 platforms, we let

(σi:iS)\displaystyle\mathcal{I}(\sigma_{i}:i\in S)

for any S[K]S\subseteq[K] denote the leaked information about θ\theta when the data of the user is shared with the buyer through platforms in the set SS and platform iSi\in S is using noise variance σi2\sigma_{i}^{2}.

Before proceeding with the proof, we first find a closed-form expression for the leaked information.

Lemma 9.

  • Suppose γi2+σi2=t\frac{\gamma_{i}}{2+\sigma_{i}^{2}}=t for all i[K]i\in[K]. The leaked information about θ\theta when the data of all platforms is shared with the buyer becomes

    Kt1+(K1)t.\displaystyle\frac{Kt}{1+(K-1)t}.
  • Suppose γi2+σi2=t\frac{\gamma_{i}}{2+\sigma_{i}^{2}}=t for all i[K]{1}i\in[K]\setminus\{1\} and γ12+σ12=t\frac{\gamma_{1}}{2+\sigma_{1}^{2}}=t^{\prime}. The leaked information about θ\theta when the data of all platforms is shared with the buyer becomes

    1+t11+t+(K2)t.\displaystyle 1+\frac{t-1}{1+t^{\prime}+(K-2)t}.

Proof: Using Lemma 1, the leaked information is

(γ1,,γK)(γ12/tγ1γ2γ1γKγ2γ1γ22/tγ2γKγKγ1γKγ2γK2/t)1(γ1,,γK)T\displaystyle(\gamma_{1},\dots,\gamma_{K})\begin{pmatrix}\gamma_{1}^{2}/t&\gamma_{1}\gamma_{2}\cdots&\gamma_{1}\gamma_{K}\\ \gamma_{2}\gamma_{1}&\gamma_{2}^{2}/t\cdots&\gamma_{2}\gamma_{K}\\ \vdots\\ \gamma_{K}\gamma_{1}&\gamma_{K}\gamma_{2}\cdots&\gamma_{K}^{2}/t\end{pmatrix}^{-1}(\gamma_{1},\dots,\gamma_{K})^{T}
=(a)\displaystyle\overset{(a)}{=} 𝜸T(𝜸𝜸T+(1t1)diag(γ12,,γK2))1𝜸\displaystyle\bm{\gamma}^{T}\left(\bm{\gamma}\bm{\gamma}^{T}+(\frac{1}{t}-1)\mathrm{diag}(\gamma_{1}^{2},\dots,\gamma_{K}^{2})\right)^{-1}\bm{\gamma}
=(a)\displaystyle\overset{(a)}{=} 𝜸T(11t1diag(1/γ12,,1/γK2)11t1diag(1/γ12,,1/γK2)𝜸T𝜸11t1diag(1/γ12,,1/γK2)1+𝜸T11t1diag(γ12,,γK2)𝜸)𝜸\displaystyle\bm{\gamma}^{T}\left(\frac{1}{\frac{1}{t}-1}\mathrm{diag}(1/\gamma_{1}^{2},\dots,1/\gamma_{K}^{2})-\frac{\frac{1}{\frac{1}{t}-1}\mathrm{diag}(1/\gamma_{1}^{2},\dots,1/\gamma_{K}^{2})\bm{\gamma}^{T}\bm{\gamma}\frac{1}{\frac{1}{t}-1}\mathrm{diag}(1/\gamma_{1}^{2},\dots,1/\gamma_{K}^{2})}{1+\bm{\gamma}^{T}\frac{1}{\frac{1}{t}-1}\mathrm{diag}(\gamma_{1}^{2},\dots,\gamma_{K}^{2})\bm{\gamma}}\right)\bm{\gamma}
=\displaystyle= Kt1+(K1)t,\displaystyle\frac{Kt}{1+(K-1)t},

where in (a), we used 𝜸=(γ1,,γK)T\bm{\gamma}=(\gamma_{1},\dots,\gamma_{K})^{T} and in (b), we used the Sherman–Morrison formula for the inverse of a rank-one perturbation of a matrix. This completes the proof of the first part. The second part follows from a similar argument. \blacksquare

Let us consider the deviation of platform 1 so that the corresponding user equilibrium is (a1=1,𝒂i)(a_{1}=1,\bm{a}_{-i}) with 𝒂i\bm{a}_{-i} having i1i-1 ones and KiK-i zeros. Notice that if such deviation is not possible by increasing σ1\sigma_{1}, then this is not a profitable deviation for platform 1, and we now verify that even if such deviation is possible, it is not profitable for platform 1. Given that 𝝈Σ𝟏,𝟏Σ𝟎,𝟏\bm{\sigma}\in\Sigma_{\mathbf{1},\mathbf{1}}\cap\Sigma_{\mathbf{0},\mathbf{1}}, we have that the user is indifferent between sharing with all platforms and sharing with none of them. Therefore, we have

i=1Kiα(σ1,,σK)=0\displaystyle\sum_{i=1}^{K}\mathcal{I}_{i}-\alpha\mathcal{I}(\sigma_{1},\dots,\sigma_{K})=0

which gives us

(σ1,,σK)=K2α.\displaystyle\mathcal{I}(\sigma_{1},\dots,\sigma_{K})=\frac{K}{2\alpha}.

Using Lemma 9, the above equality gives us

t=11+2αK.\displaystyle t=\frac{1}{1+2\alpha-K}.

Given the above deviation to (σ~1,,σK)(\tilde{\sigma}_{1},\dots,\sigma_{K}), the user indifference condition gives us

K2α(σ~1,,σK)=i2α(σ~1,,σi).\displaystyle\frac{K}{2}-\alpha\mathcal{I}(\tilde{\sigma}_{1},\dots,\sigma_{K})=\frac{i}{2}-\alpha\mathcal{I}(\tilde{\sigma}_{1},\dots,\sigma_{i}).

Substituting the closed-form expressions of Lemma 9 in the above equality gives us

t=24α+Ki16α2+(Ki)28αK2+4α2K,\displaystyle t^{\prime}=\frac{2-4\alpha+K-i-\sqrt{16\alpha^{2}+(K-i)^{2}-8\alpha K}}{2+4\alpha-2K},

where t=γ122+σ~12t^{\prime}=\frac{\gamma_{1}^{2}}{2+\tilde{\sigma}_{1}^{2}}.

Now, for this deviation to be not profitable, we must have

(σ1,,σK)(σ2,,σK)(σ~1,,σi)(σ2,,σi).\displaystyle\mathcal{I}(\sigma_{1},\dots,\sigma_{K})-\mathcal{I}(\sigma_{2},\dots,\sigma_{K})\geq\mathcal{I}(\tilde{\sigma}_{1},\dots,\sigma_{i})-\mathcal{I}(\sigma_{2},\dots,\sigma_{i}).

Again, using the closed-form expressions of Lemma 9 and writing tt^{\prime} and tt in terms of α\alpha, the above inequality becomes equivalent to

1+1K2α1+i12α1+iK+K2α+2K4αKi+16α2+(Ki)28αK0,\displaystyle-1+\frac{1-K}{2\alpha-1}+\frac{i-1}{2\alpha-1+i-K}+\frac{K}{2\alpha}+\frac{2K-4\alpha}{K-i+\sqrt{16\alpha^{2}+(K-i)^{2}-8\alpha K}}\geq 0,

which we prove next. We can write

3Ki4α+1K2α1+i12α1+iK1+16α2+(Ki)28αK4α\displaystyle\frac{3K-i}{4\alpha}+\frac{1-K}{2\alpha-1}+\frac{i-1}{2\alpha-1+i-K}-1+\frac{\sqrt{16\alpha^{2}+(K-i)^{2}-8\alpha K}}{4\alpha}
(a)3Ki4α+1K2α1+i12α1+iKK2α(1+1K2α)\displaystyle\overset{(a)}{\geq}\frac{3K-i}{4\alpha}+\frac{1-K}{2\alpha-1}+\frac{i-1}{2\alpha-1+i-K}-\frac{K}{2\alpha\left(1+\sqrt{1-\frac{K}{2\alpha}}\right)}
3K14αK12αK2α(1+1K2α)(b)0\displaystyle\geq\frac{3K-1}{4\alpha}-\frac{K-1}{2\alpha}-\frac{K}{2\alpha\left(1+\sqrt{1-\frac{K}{2\alpha}}\right)}\overset{(b)}{\geq}0

where (a) follows by noting that the last term is increasing in ii and (b) holds for

αα¯:=(K+1)28.\displaystyle\alpha\geq\bar{\alpha}:=\frac{(K+1)^{2}}{8}.

Notice that for any other set S[K]S\subseteq[K] of the platforms that have entered the market, the above bound on α\alpha guarantees the existence of platform equilibrium strategy as |S|K|S|\leq K. This completes the proof. \blacksquare

Proof of Proposition 6

The user utility from sharing if the platform enters the market is

1α(σ1)=12αγ122+σ12.\displaystyle\mathcal{I}_{1}-\alpha\mathcal{I}(\sigma_{1})=\frac{1}{2}-\alpha\frac{\gamma_{1}^{2}}{2+\sigma_{1}^{2}}.

The above utility is always positive if α1/γ12\alpha\leq 1/\gamma_{1}^{2}. Therefore, the user always shares their data. This also implies that the platform chooses σ1=0\sigma_{1}=0 to maximize the payment received from the buyer. The platform’s utility then becomes

1c1+β(0)=12c1+βγ122.\displaystyle\mathcal{I}_{1}-c_{1}+\beta\mathcal{I}(0)=\frac{1}{2}-c_{1}+\beta\frac{\gamma_{1}^{2}}{2}.

Therefore, the platform enters the market if and only if the above quantity is positive.

If α>1/γ12\alpha>1/\gamma_{1}^{2}, then the user’s utility is not always positive. In this case, if the platform enters the market, it chooses σ1\sigma_{1} so that the user utility is zero, i.e.,

iα(σ1)=12αγ122+σ12=0.\displaystyle\mathcal{I}_{i}-\alpha\mathcal{I}(\sigma_{1})=\frac{1}{2}-\alpha\frac{\gamma_{1}^{2}}{2+\sigma_{1}^{2}}=0.

The above equality gives us σ1=2(αγ121)\sigma_{1}=\sqrt{2\left(\alpha\gamma_{1}^{2}-1\right)}. The platform’s utility then becomes

1c1+β(σ1)=12c1+β12α.\displaystyle\mathcal{I}_{1}-c_{1}+\beta\mathcal{I}(\sigma_{1})=\frac{1}{2}-c_{1}+\beta\frac{1}{2\alpha}.

Therefore, the platform enters the market if and only if the above quantity is positive. \blacksquare

Proof of Corollary 1

When all platforms enter the market, as we have established in Lemma 8, the equilibrium noise variance is such that the user is indifferent between sharing with all platforms and not sharing at all. This means that

i=1Kiα(σ1,,σK)=0.\displaystyle\sum_{i=1}^{K}\mathcal{I}_{i}-\alpha\mathcal{I}(\sigma_{1},\dots,\sigma_{K})=0. (27)

Moreover, the payments cancel out in the utilitarian welfare and it becomes

2i=1Ki+(βα)(σ1,,σK)i=1Kci=K+(βα)K2αi=1Kci,\displaystyle 2\sum_{i=1}^{K}\mathcal{I}_{i}+\left(\beta-\alpha\right)\mathcal{I}(\sigma_{1},\dots,\sigma_{K})-\sum_{i=1}^{K}c_{i}=K+\left(\beta-\alpha\right)\frac{K}{2\alpha}-\sum_{i=1}^{K}c_{i},

where the above equality follows from i=12\mathcal{I}_{i}=\frac{1}{2} for all i[K]i\in[K] and (27).

The argument is similar to the above one when only low-cost platforms enter the market. \blacksquare

Proof of Proposition 7

If we only have low-cost platforms, then all platforms enter the market as their utility of platform i[K]i\in[K] is given by

ici+pi=12ci+pi0,\displaystyle\mathcal{I}_{i}-c_{i}+p_{i}=\frac{1}{2}-c_{i}+p_{i}\geq 0,

which is always positive. Notice that if we ban data sharing, then the user shares with all platforms (because there is no privacy loss), and its utility becomes

i=1Ki=K2.\displaystyle\sum_{i=1}^{K}\mathcal{I}_{i}=\frac{K}{2}.

Clearly, there is no uniform minimum privacy mandate that can make the user better off because banning data sharing maximizes the user gain from services and minimizes the privacy loss.

Now, let us prove the second part of this proposition. By banning data sharing, only low-cost platforms enter the market, and the user shares data with all of them and obtains the following utility:

12( number of low-cost platforms).\displaystyle\frac{1}{2}(\text{ number of low-cost platforms}).

Now consider a uniform minimum privacy mandate σ¯\bar{\sigma} such that

12Kα(σ¯,,σ¯K many)>12( number of low-cost platforms).\displaystyle\frac{1}{2}K-\alpha\mathcal{I}(\underbrace{\bar{\sigma},\dots,\bar{\sigma}}_{K\text{ many}})>\frac{1}{2}(\text{ number of low-cost platforms}).

For any α\alpha, as σ¯\bar{\sigma} increases, the leaked information goes to zero, and therefore, so long as there exists at least one high-cost platform, a σ¯\bar{\sigma} for which the above inequality holds exist. We next argue that all platforms enter the market for large enough β\beta. To see this, we need to have

ici+β((σ¯,,σ¯K many)(σ¯,,σ¯K1 many))0\displaystyle\mathcal{I}_{i}-c_{i}+\beta\Bigl{(}\mathcal{I}(\underbrace{\bar{\sigma},\dots,\bar{\sigma}}_{K\text{ many}})-\mathcal{I}(\underbrace{\bar{\sigma},\dots,\bar{\sigma}}_{K-1\text{ many}})\Bigr{)}\geq 0

for all i[K]i\in[K]. Therefore, by letting ββ~\beta\geq\tilde{\beta} where

β~=max{β¯,ci12(σ¯,,σ¯K many)(σ¯,,σ¯K1 many)}\displaystyle\tilde{\beta}=\max\Bigl{\{}\bar{\beta},\frac{c_{i}-\frac{1}{2}}{\mathcal{I}(\underbrace{\bar{\sigma},\dots,\bar{\sigma}}_{K\text{ many}})-\mathcal{I}(\underbrace{\bar{\sigma},\dots,\bar{\sigma}}_{K-1\text{ many}})}\Bigr{\}}

and β¯\bar{\beta} is specified in Theorem 1 all platforms enter the market. \blacksquare

Proof Proposition 8

Similar to the proof of Proposition 7 and Theorem 1, we choose α¯\bar{\alpha} and β^\hat{\beta} large enough so that an equilibrium exists and all platforms enter the market. Now suppose σ¯\bar{\sigma} is the optimal uniform minimum privacy mandate so that all platforms enter the market, the user utility is

12Kα(σ¯,,σ¯K many),\displaystyle\frac{1}{2}K-\alpha\mathcal{I}(\underbrace{\bar{\sigma},\dots,\bar{\sigma}}_{K\text{ many}}),

and that the utility of platform ii is

ici+β((σ¯,,σ¯K many)(σ¯,,σ¯K1 many))0.\displaystyle\mathcal{I}_{i}-c_{i}+\beta\Bigl{(}\mathcal{I}(\underbrace{\bar{\sigma},\dots,\bar{\sigma}}_{K\text{ many}})-\mathcal{I}(\underbrace{\bar{\sigma},\dots,\bar{\sigma}}_{K-1\text{ many}})\Bigr{)}\geq 0.

Consider a regulation that bans data sharing for low-cost platforms and uses the same minimum privacy mandate for high-value platforms. We first prove that all platforms still enter the market. First, notice that the low-cost platforms enter the market as they obtain a positive utility. Letting hh be the number of high-cost platforms, the utility of a high-cost platform ii becomes

ici+β((σ¯,,σ¯h many)(σ¯,,σ¯h1 many))ici+β((σ¯,,σ¯K many)(σ¯,,σ¯K1 many))0,\displaystyle\mathcal{I}_{i}-c_{i}+\beta\Bigl{(}\mathcal{I}(\underbrace{\bar{\sigma},\dots,\bar{\sigma}}_{h\text{ many}})-\mathcal{I}(\underbrace{\bar{\sigma},\dots,\bar{\sigma}}_{h-1\text{ many}})\Bigr{)}\geq\mathcal{I}_{i}-c_{i}+\beta\Bigl{(}\mathcal{I}(\underbrace{\bar{\sigma},\dots,\bar{\sigma}}_{K\text{ many}})-\mathcal{I}(\underbrace{\bar{\sigma},\dots,\bar{\sigma}}_{K-1\text{ many}})\Bigr{)}\geq 0,

where the first inequality follows from the submodularity of the leaked information in actions, established in Lemma 3. The utility of the user also increases because we have

12Kα(σ¯,,σ¯h many)12Kα(σ¯,,σ¯K many),\displaystyle\frac{1}{2}K-\alpha\mathcal{I}(\underbrace{\bar{\sigma},\dots,\bar{\sigma}}_{h\text{ many}})\geq\frac{1}{2}K-\alpha\mathcal{I}(\underbrace{\bar{\sigma},\dots,\bar{\sigma}}_{K\text{ many}}),

where the inequality follows from the monotonicity of the leaked information, established in Lemma 2. \blacksquare