This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Receiver-Oriented Cheap Talk Design

Itai Arieli    Ivan Geffner    Moshe Tennenholtz
Abstract

This paper considers the dynamics of cheap talk interactions between a sender and receiver, departing from conventional models by focusing on the receiver’s perspective. We study two models, one with transparent motives and another one in which the receiver can filter the information that is accessible by the sender. We give a geometric characterization of the best receiver equilibrium under transparent motives and prove that the receiver does not benefit from filtering information in this case. However, in general, we show that the receiver can strictly benefit from filtering and provide efficient algorithms for computing optimal equilibria. This innovative analysis aligns with user-based platforms where receivers (users) control information accessible to senders (sellers). Our findings provide insights into communication dynamics, leveling the sender’s inherent advantage, and offering strategic interaction predictions.

1 Introduction

The conventional sender-receiver model focuses on the interaction between an informed sender, who is fully aware of the state of the game, and an oblivious receiver. In these interactions, the sender must disclose some of the information to the receiver in order to influence its behavior, and the receiver must then play an action that affects the utility of both players. Relevant information disclosure models include cheap talk [10, 17], Bayesian persuasion [16], and their extensions to multiple senders and receivers [12, 7, 1, 2, 3, 4]. Traditionally, all these interactions are always viewed from the point of view of the sender. However, in order to get a complete picture of the communication model, it is also critical to consider the receiver’s perspective. Moreover, since the sender, as the informed side, has an inherent advantage in cheap talk models, analyzing the best outcomes for the receiver can level this advantage and provide a more reliable prediction of strategic interactions.

We focus on interactions where the sender and the receiver can communicate via cheap talk and analyze two settings. First, we consider a setting where the sender has transparent motives (i.e., where the sender has a state-independent utility) in a similar fashion as Lipnowski and Ravid [17], and we follow with a setting where there is a pre-play stage before the communication phase in which the receiver can filter the information that the sender receives about the realized state.

Our first result is providing a geometric characterization of the best sequential equilibrium for the receiver when the sender has transparent motives. It follows from this characterization that (a) the maximum utility that the receiver can get in equilibrium can be computed as a function of the prior by taking the maximum of finitely many linear functions, and (b) that the receiver cannot improve her utility by applying filters when the sender has transparent motives. We continue by showing that, in general (i.e., when the sender does not necessarily have transparent motives), the receiver can strictly benefit from filtering information. Moreover, if the receiver can filter, we provide an algorithm that outputs the best sequential equilibrium for the receiver when her action set is binary. Our algorithm runs in O(klogk)O(k\log k) time, where kk is the number of possible states. For our final result, we consider the same setting but with two senders instead, and reduce the problem of finding the best sequential equilibrium for the receiver to a linear programming instance with O(k)O(k) variables and O(k)O(k) constraints. If there are three or more senders, it is easy to check that the receiver does not benefit from filtering information (see Section 4 for more details.)

A novelty in the second part of our analysis is giving the receiver the advantage of determining the information that the sender receives. Intuitively, we can see this setting as the dual of Bayesian persuasion [16]. More precisely, in Bayesian persuasion settings, the sender can freely choose the mechanism by which the information about the realized state is disclosed to the receiver. By contrast, in our setting, the receiver can freely choose the mechanism by which the information about the realized state is disclosed to the sender (note, however, that the sender has to disclose this information to the receiver via cheap talk, which means that the sender can still manipulate the information received if it is in her interests to do so.) This aspect is motivated by several scenarios, as for instance:

  • Interacting with search engines like Google, Yahoo! or Bing. Here, the search engine plays the role of a sender that can access the necessary information about a given query on a large database and the user plays the role of a receiver that can impact the outcome of her interaction with the search engine by using specific keywords that influence how the information in the database is retrieved by the search engine.

  • Scenarios in which the state of the game is a function of both the sender and the receiver’s information and the sender should propose an agreement to the receiver (as, for example, in big economic transactions or in diplomacy and politics). In these cases, the receiver should strategically decide how much information it should disclose to the sender in order to influence her proposal.

  • Online platforms, such as Amazon, store a significant amount of information about their users. They have the ability to determine which information is disclosed to sellers, who, in turn, can utilize this data to target users through advertisements. Our model essentially examines the scenario in which the platform operates for the average user, revealing seller information to maximize the revenue for the average buyer in equilibrium.

Our analysis attempts to model these scenarios more accurately.

1.1 Related Literature

Our work on the setting with transparent motives is closely related to that of Lipnowski and Ravid [17], who studied a cheap talk model where the sender has transparent motives and showed that the best sender equilibrium is the quasi-concave envelope of the sender’s utility function. Our result is complementary since we give the geometric characterization of the best equilibrium for the receiver. In fact, we show that the best equilibrium for the receiver can be computed as the maximum of at most k2k^{2} affine functions, where kk is the number of states in the game.

Restricting the information available to the sender has been increasingly gaining traction in the community. Most notably, Bergemann, Brooks and Morris [6] considered a buyer-seller setting and studied possible ways to limit the seller’s information. In their work, they characterized the set of all pairs of possible utilities achievable in equilibrium. In particular, they provide an algorithm that achieves the optimal way to limit the seller’s information to maximize the expected buyers’ utility. Ichihashi [14] considered a Bayesian persuasion setting and studied how the outcome of the interaction is affected when the sender’s information is restricted. One of their results is that, if the receiver restricts sender information in a pre-play stage, the best utility that the receiver can get in this setting coincides with the one that the receiver would get in the “flipped game”, where the receiver persuades the sender. Other papers have studied similar questions in the context of common-interest coordination games [9, 11] and cheap talk games [15, 13]. The main differences with our work are the fact that (a) we have no restrictions on the utility of the sender or that of the receiver, and (b) we focus on the best equilibrium for the receiver.

The rest of the paper is organized as follows. In Section 2 we introduce the cheap talk setting where the sender has transparent motives. We also provide a geometric characterization of the best equilibrium for the receiver. In Section 3 we introduce filtered information aggregation games, where the receiver can limit the information of the senders. In Section 4 we state our main results regarding the characterization of the best sequential equilibrium for the receiver in filtered information aggregation games, and these are proved in Sections 5, 6 and 7, where we also show how to extend the results of Section 2 to sequential equilibria. We end with a conclusion in Section 8.

2 Cheap Talk with Transparent Motives

2.1 Model

An information transmission game involves a sender ss and a receiver rr, and is defined by a tuple Γ=(A,Ω,p,M,u)\Gamma=(A,\Omega,p,M,u), where A={a1,,a}A=\{a_{1},\ldots,a_{\ell}\} is the set of actions, Ω={ω1,,ωn}\Omega=\{\omega_{1},\ldots,\omega_{n}\} is the set of possible states, pp is a commonly known prior distribution over Ω\Omega that assigns a strictly positive probability to each possible state, MM is a finite set that contains the messages that the sender can send to the receiver (MM is usually assumed to be the set of binary strings of length at most NN, for some positive integer NN), and u:{s,r}×Ω×Au:\{s,r\}\times\Omega\times A\longrightarrow\mathbb{R} is a utility function such that u(i,ω,a)u(i,\omega,a) gives the utility of player ii (where ii is either the sender or the receiver) when action aa is played at state ω\omega. Each information transmission game instance is divided into three phases. In the first phases, a state ω\omega is sampled from Ω\Omega following the prior distribution pp, and this state is disclosed to the sender ss. During the second phase, the sender sends a message mMm\in M to the receiver, and in the third phase the receiver plays an action aAa\in A and each player i{s,r}i\in\{s,r\} receives u(i,ω,a)u(i,\omega,a) utility.

Given an information transmission game Γ=(A,Ω,p,M,u)\Gamma=(A,\Omega,p,M,u), a strategy profile σ\vec{\sigma} for Γ\Gamma consists of a pair of strategies (σs,σr)(\sigma_{s},\sigma_{r}) for the sender and the receiver, where σs\sigma_{s} is a map from Ω\Omega to distributions over MM (which is denoted by Δ(M)\Delta(M)), and σr\sigma_{r} is a map from MM to Δ(A)\Delta(A). We say that σ\vec{\sigma} is a Nash equilibrium if no player can increase its utility by defecting from σ\vec{\sigma}. More precisely, denote by ui(σ)u_{i}(\vec{\sigma}) the expected utility that ii gets when players play σ\vec{\sigma}, then σ\vec{\sigma} is a Nash equilibrium if

ui(σi,τi)ui(σ)u_{i}(\vec{\sigma}_{-i},\tau_{i})\leq u_{i}(\vec{\sigma})

for all players i{s,r}i\in\{s,r\} and all strategies τi\tau_{i} for ii.

For future reference, some of the results in this paper require a stronger notion of equilibrium known as sequential equilibrium. A sequential equilibrium consists of a strategy profile σ\vec{\sigma} and a belief system β\beta that (in this case) assigns a posterior distribution over Ω\Omega to each possible message in MM. Intuitively, β\beta represents the receiver’s beliefs about the posterior distribution over Ω\Omega given the message that it receives from the sender. The pair (σ,β)(\vec{\sigma},\beta) form a sequential equilibrium if

  • β\beta is consistent with σ\vec{\sigma}, which means that there exists a sequence of completely mixed strategies σ1,σ2,\vec{\sigma}^{1},\vec{\sigma}^{2},\ldots that converge to σ\vec{\sigma} such that the sequence of maps from messages to posterior distributions over Ω\Omega obtained by applying Bayes’ rule assuming that the sender plays σs1,σs2,\sigma_{s}^{1},\sigma_{s}^{2},\ldots converges to β\beta.

  • σr\sigma_{r} is, according to β\beta, a best-response for the sender for all messages mMm\in M.

In this section, we are interested in information transmission games in which the sender has transparent motives. This means that the sender’s utility u(s,ω,a)u(s,\omega,a) only depends only on the action aa played by the receiver, and not on the state ω\omega. Thus, for simplicity, we will define functions us:Au_{s}:A\to\mathbb{R} and ur:Ω×Au_{r}:\Omega\times A\to\mathbb{R} that denote the utilities of the sender and the receiver respectively. More precisely, us(a):=u(s,ω,a)u_{s}(a):=u(s,\omega,a) and ur(ω,a):=u(r,ω,a)u_{r}(\omega,a):=u(r,\omega,a) for all ωΩ\omega\in\Omega and aAa\in A.

2.2 Characterizing the Best Equilibrium for the Receiver

In this section, we are interested in finding the best Nash equilibrium for the receiver given an information transmission game with transparent motives. We show later how to extend this equilibrium to a sequential equilibrium in Section 7.

We start by considering the case where the state space Ω\Omega is binary (i.e., Ω={ω0,ω1\Omega=\{\omega_{0},\omega_{1}). We can assume without loss of generality that there exist +1\ell+1 real numbers 0=b1<b2<<b+1=10=b_{1}<b_{2}<\ldots<b_{\ell+1}=1 such that action aia_{i} is optimal for the receiver if and only if the posterior probability qq of state ω1\omega_{1} after her interaction with the sender is in the interval [bi,bi+1][b_{i},b_{i+1}]. To see this, note that the expected utility of the receiver given some action aa is a linear function over qq that goes from (0,ur(ω0,a))(0,u_{r}(\omega_{0},a)) to (1,ur(ω1,a))(1,u_{r}(\omega_{1},a)). Therefore, the set of points in which an action aa gives the maximal utility over all other actions is an interval in [0,1][0,1]. If this interval is empty for some action aa, it means that aa will never be played under any circumstances since there will always be better options for the receiver. This implies that, without loss of generality, we can simply remove aa from AA and keep only those actions that are maximal at some point.

Using a slight abuse of notation, let ur(a,q)u_{r}(a,q) denote the expected utility of the receiver given action aa and the posterior probability qq of ω1\omega_{1}. Note that ur(a,q)u_{r}(a,q) can be computed by the following expression:

ur(a,q):=qur(a,ω1)+(1q)ur(a,ω0).u_{r}(a,q):=q\cdot u_{r}(a,\omega_{1})+(1-q)\cdot u_{r}(a,\omega_{0}).

For each j{1,2,,+1}j\in\{1,2,\ldots,\ell+1\}, let I(j):=[us(aj1),us(aj)]I(j):=[u_{s}(a_{j-1}),u_{s}(a_{j})] (using the convention that a1:=a0a_{-1}:=a_{0} and a+1:=aa_{\ell+1}:=a_{\ell}), and let CC be the subset of pairs (j,k)[l+1]2(j,k)\in[l+1]^{2} such that I(j)I(k)I(j)\cap I(k)\not=\emptyset. At a high level, (j,k)C(j,k)\in C if the optimal actions with prior bib_{i} and the optimal actions with prior bjb_{j} can be combined in such a way that the utility of the sender is equal in both combinations.

ur(a,p)=pur(a,ω1)+(1p)ur(a,ω0).u_{r}(a,p)=pu_{r}(a,\omega_{1})+(1-p)u_{r}(a,\omega_{0}).

Let

u~r(q):=maxbjqbks.t.(j,k)C(αj,kur(aj,bj)+(1αj,k)ur(ak,bk)),\tilde{u}_{r}(q):=\max_{b_{j}\leq q\leq b_{k}\ s.t.\ (j,k)\in C}\left(\alpha_{j,k}u_{r}(a_{j},b_{j})+(1-\alpha_{j,k})u_{r}(a_{k},b_{k})\right),

where q[0,1]q\in[0,1] and αj,k\alpha_{j,k} is a value in [0,1][0,1] that satisfies that αj,kbj+(1αj,k)bk=q\alpha_{j,k}b_{j}+(1-\alpha_{j,k})b_{k}=q. The following theorem states that the best utility for the receiver in an information transmission game with transparent motives is given by u~r\tilde{u}_{r}.

Theorem 1.

Given an information transmission game Γ=(A,Ω,p,M,u)\Gamma=(A,\Omega,p,M,u) with |Ω|=2|\Omega|=2 and transparent motives, the maximum utility that the receiver can get in a Nash equilibrium is given by u~r(q)\tilde{u}_{r}(q), where qq is the prior probability that the realized state is ω1\omega_{1}.

Before proving Theorem 1, we illustrate its use via the following simple example. Consider an information transmission game Γ=(A,Ω,p,M,u)\Gamma=(A,\Omega,p,M,u) with transparent motives such that |Ω|=2|\Omega|=2 and A={a1,a2,a3,a4,a5}A=\{a_{1},a_{2},a_{3},a_{4},a_{5}\}, in which the utilities are defined as follows:

ur(a,ω0)u_{r}(a,\omega_{0}) ur(a,ω1)u_{r}(a,\omega_{1}) us(a)u_{s}(a)
a1a_{1} 12\frac{1}{2} 32-\frac{3}{2} 0
a2a_{2} 13\frac{1}{3} 23-\frac{2}{3} 2
a3a_{3} 0 0 -2
a4a_{4} 23-\frac{2}{3} 13\frac{1}{3} 1
a5a_{5} 12\frac{1}{2} 32-\frac{3}{2} -1

If we plot the utility of the receiver in a bubbling equilibrium in which no information is revealed we get the following figure, where the dotted line represents the utility of the receiver as a function of the prior probability that the realized state is ω1\omega_{1}.

xxu~r(x)\tilde{u}_{r}(x)12\frac{1}{2}16\frac{1}{6}12\frac{1}{2}11a1a_{1}a2a_{2}a3a_{3}a4a_{4}a5a_{5}

If we follow the construction of Theorem 1, we have that b1=0b_{1}=0 b2=16b_{2}=\frac{1}{6}, b3=13b_{3}=\frac{1}{3}, b4=23b_{4}=\frac{2}{3}, b5=56b_{5}=\frac{5}{6} and b6=1b_{6}=1, and the set CC consists of all pairs of the form (j,j+1)(j,j+1) in addition to (1,3)(1,3), (2,5)(2,5), (3,5)(3,5) and (4,6)(4,6). Thus, it follows from Theorem 1 that the function u~r\tilde{u}_{r} that represents the receiver’s utility in a Nash equilibrium is obtained by taking the maximum of the bubbling equilibrium and the following four segments: the segment s1s_{1} that connects (b1,12)(b_{1},\frac{1}{2}) to (b3,0)(b_{3},0), the segment s2s_{2} that connects (b2,16)(b_{2},\frac{1}{6}) to (b5,16)(b_{5},\frac{1}{6}), the segment s3s_{3} that connects (b3,0)(b_{3},0) to (b5,16)(b_{5},\frac{1}{6}), and the segment s4s_{4} that connects (b4,0)(b_{4},0) to (b6,12)(b_{6},\frac{1}{2}), which are represented in the figure below with a solid green line:

xxu~r(x)\tilde{u}_{r}(x)12\frac{1}{2}16\frac{1}{6}12\frac{1}{2}11s1s_{1}s2s_{2}s3s_{3}s4s_{4}

Thus, the maximum utility that the receiver can get in a Nash equilibrium of Γ\Gamma is represented by the solid black line in the figure below:

xxu~r(x)\tilde{u}_{r}(x)12\frac{1}{2}16\frac{1}{6}12\frac{1}{2}11s1s_{1}s2s_{2}s3s_{3}s4s_{4}

Next, we proceed with the proof of Theorem 1.

Proof of Theorem 1.

First we show that the receiver can guarantee at least u~r(q)\tilde{u}_{r}(q) utility in a Nash equilibrium. Let qq be the prior probability that the realized state is ω1\omega_{1} and suppose that u~r(q)=αj,kur(aj,bj)+(1αj,k)ur(ak,bk)\tilde{u}_{r}(q)=\alpha_{j,k}u_{r}(a_{j},b_{j})+(1-\alpha_{j,k})u_{r}(a_{k},b_{k}), for some (j,k)C(j,k)\in C. Since (j,k)C(j,k)\in C, it follows that I(bj)I(bk)I(b_{j})\cap I(b_{k})\neq\emptyset. Therefore, by definition, we can find βj,βk[0,1]\beta_{j},\beta_{k}\in[0,1] such that

βjus(aj1)+(1βj)us(aj)=βkus(ak1)+(1βk)us(ak).\displaystyle\beta_{j}u_{s}(a_{j-1})+(1-\beta_{j})u_{s}(a_{j})=\beta_{k}u_{s}(a_{k-1})+(1-\beta_{k})u_{s}(a_{k}). (1)

We next describe a cheap talk equilibrium that gives the receiver a utility of αj,kur(aj,bj)+(1αj,k)ur(ak,bk)\alpha_{j,k}u_{r}(a_{j},b_{j})+(1-\alpha_{j,k})u_{r}(a_{k},b_{k}). Using the fact that αj,kbj+(1αj,k)bk=q\alpha_{j,k}b_{j}+(1-\alpha_{j,k})b_{k}=q, the splitting lemma of Aumann and Maschler [5] implies that there exists a communication protocol such that

  • The sender sends a message m1m_{1} with probability αj,k\alpha_{j,k} and another message m2m_{2} with probability 1αj,k1-\alpha_{j,k}.

  • The posterior probability of ω1\omega_{1} after the receiver receives message m1m_{1} is bjb_{j} and the posterior probability of ω1\omega_{1} after the receiver receives message m2m_{2} is bkb_{k}.

Consider such a communication protocol in which, in the last phase, the receiver plays action aj1a_{j-1} with probability βj\beta_{j} and action aja_{j} with probability 1βj1-\beta_{j} whenever she receives message m1m_{1}. If she receives message m2m_{2} instead, she plays ak1a_{k-1} with probability βk\beta_{k} and aka_{k} with probability 1βk1-\beta_{k}. If the receiver receives a message that is not m1m_{1} or m2m_{2}, she plays the action that gives the least possible utility for the sender. By construction, since the posterior probability of ω1\omega_{1} after receiving message m1m_{1} is bjb_{j}, both aj1a_{j-1} and aja_{j} are best responses for the receiver. Similarly, since the posterior probability of ω1\omega_{1} after receiving message m2m_{2} is bkb_{k}, so are ak1a_{k-1} and aka_{k} in this scenario. Altogether, this implies that the receiver best-responds to the sender’s strategy. To see that this strategy profile is also incentive-compatible for the sender note that, by Equation 1, the sender is indifferent between sending m1m_{1} or m2m_{2}. Moreover, sending any other message mm would give her less utility than sending m1m_{1} or m2m_{2}. This shows that the receiver can indeed achieve u~r(q)\tilde{u}_{r}(q) utility in equilibrium.

We next show that the receiver’s utility is at most u~r(q)\tilde{u}_{r}(q) in any Nash equilibrium. It follows from standard extreme point considerations that if there exists a Nash equilibrium σ\vec{\sigma} in which the sender sends three or more messages with strictly positive probability, there exists another Nash equilibrium σ\vec{\sigma}^{\prime} in which the sender sends at most two different messages and both the sender and the receiver at least as much utility with σ\vec{\sigma}^{\prime} than with σ\vec{\sigma}. Thus, for the rest of the proof, we’ll restrict our analysis to communication protocols in which the sender sends at most two different messages with positive probability.

Suppose that the sender only sends one message with strictly positive probability. This means that the receiver plays her action with no additional information about the realized state, except for the prior pp. Let qq be the prior probability that the realized state is ω1\omega_{1} and let j[]j\in[\ell] be an index such that q[bj,bj+1]q\in[b_{j},b_{j+1}]. Since, in this case, the posterior distribution is the same as the prior distribution, aja_{j} is the receiver’s best response, and her expected utility is given by (1q)ur(ω0,aj)+qur(ω1,aj)(1-q)\cdot u_{r}(\omega_{0},a_{j})+q\cdot u_{r}(\omega_{1},a_{j}). We can rewrite this as an affine combination of ur(aj,bj)u_{r}(a_{j},b_{j}) and ur(aj,bj+1)u_{r}(a_{j},b_{j+1}) with the following expression:

ur(aj,q)=αur(aj,bj)+(1α)ur(aj,bj+1),u_{r}(a_{j},q)=\alpha u_{r}(a_{j},b_{j})+(1-\alpha)u_{r}(a_{j},b_{j+1}),

where α\alpha is the value in [0,1][0,1] such that αbj+(1α)bj+1=q\alpha b_{j}+(1-\alpha)b_{j+1}=q. This shows that, by definition, u~r(q)ur(aj,q)\tilde{u}_{r}(q)\geq u_{r}(a_{j},q), and therefore the receiver cannot get more utility than u~r(q)\tilde{u}_{r}(q) in Nash equilibria where the sender sends only one message.

Assume instead that the sender sends two messages m1m_{1} and m2m_{2} such that the posterior probability that the realized state is ω1\omega_{1} after the receiver receives m1m_{1} and m2m_{2} are x1x_{1} and x2x_{2}, respectively. We first claim that we can assume without loss of generality that x1=bjx_{1}=b_{j} and x2=bkx_{2}=b_{k} for some j,k[+1]j,k\in[\ell+1]. Suppose that bj<x1<bj+1b_{j}<x_{1}<b_{j+1} for some j[+1]j\in[\ell+1]. Then, it means that the best response for the receiver when receiving message m1m_{1} is to play action aja_{j} with probability 11. Then, we can construct another equilibrium σ\vec{\sigma} in which the posterior probability after receiving message m1m_{1} is bjb_{j} while the posterior probability after receiving message m2m_{2} remains unchanged (note that the conditional probabilities of sending m1m_{1} and m2m_{2} are uniquely determined by Bayes’ plausibility constraints). By definition, playing aja_{j} is still a receiver’s best response to m1m_{1} in σ\vec{\sigma}. Moreover, it is straightforward to check that the sender receives the same utility in the new equilibrium and the original equilibrium. It remains to check that the receiver also (weakly) increases her utility. To see this, note that the posterior distribution of σ\vec{\sigma} is a mean preserving spread of the original one. Therefore the new posterior distribution Blackwell dominates the original distribution and by Blackwell theorem [8] the receiver is (weakly) better. We can therefore assume from now on that x1=bjx_{1}=b_{j} and x2=bkx_{2}=b_{k} for some j,k[+1]j,k\in[\ell+1].

We next show that the posterior probabilities bjb_{j} and bkb_{k} described in the paragraph above satisfy that (j,k)C(j,k)\in C. To see this, note that the sender’s expected utility must be equal when sending m1m_{1} and m2m_{2} (otherwise, she can increase her utility by sending one of these messages by probability 1 and the other with probability 0). By definition, the only best responses for the receiver given posterior probabilities bjb_{j} and bkb_{k} are {aj1,aj}\{a_{j-1},a_{j}\}, and {ak1,ak}\{a_{k-1},a_{k}\}, respectively. Suppose that the sender plays aj1a_{j-1} with probability αj\alpha_{j} when receiving message m1m_{1}, and plays ak1a_{k-1} with probability αk\alpha_{k} when receiving message m2m_{2}. Since the sender is indifferent between m1m_{1} and m2m_{2}, αj\alpha_{j} and αk\alpha_{k} must satisfy that

αjus(aj1)+(1αj)us(aj)=αkus(ak1)+(1αk)us(ak),\alpha_{j}u_{s}(a_{j-1})+(1-\alpha_{j})u_{s}(a_{j})=\alpha_{k}u_{s}(a_{k-1})+(1-\alpha_{k})u_{s}(a_{k}),

which implies that I(j)I(k)I(j)\cap I(k)\not=\emptyset, and therefore that (j,k)C(j,k)\in C. This implies that we can represent the receiver’s utility as an affine combination of ur(aj,bj)u_{r}(a_{j},b_{j}) and ur(ak,bk)u_{r}(a_{k},b_{k}) using the following expression:

ur(σ)=αj,kur(aj,bj)+(1αj,k)ur(ak,bk),u_{r}(\vec{\sigma})=\alpha_{j,k}u_{r}(a_{j},b_{j})+(1-\alpha_{j,k})u_{r}(a_{k},b_{k}),

where αj,k\alpha_{j,k} is the unique value in [0,1][0,1] such that αj,kbj+(1αj,k)bk=q\alpha_{j,k}b_{j}+(1-\alpha_{j,k})b_{k}=q (note that this equality must hold because of Bayes’ plausibility constraints). This completes the proof of Theorem 1. ∎

2.3 General State space

In this section we generalize the proof of Theorem 1 to arbitrary information transmission games with transparent motives. At a high level, both the statement and the proof are quite similar to the case with binary states, although they require additional notation.

Let Γ=(A,Ω,p,M,u)\Gamma=(A,\Omega,p,M,u) be an information transmission game with transparent motives where Ω={ω1,,ωn}\Omega=\{\omega_{1},\ldots,\omega_{n}\} and A={a1,,a}A=\{a_{1},\ldots,a_{\ell}\}. As in the proof of Theorem 1, we identify Δ(Ω)\Delta(\Omega) with Δn1\Delta^{n-1} (the n1n-1-dimensional simplex) and assume without loss of generality that there exists no strictly dominated action. More generally, we assume without loss of generality that all actions are optimal in some posterior distribution, which is that, for all actions aAa\in A, the set

{qΔn1:i=1nqiur(a,ωi)>maxaA,aai=1nqiur(a,ωi)}\left\{q\in\Delta^{n-1}\quad:\quad\sum_{i=1}^{n}q_{i}u_{r}(a,\omega_{i})>\max_{a^{\prime}\in A,\ a^{\prime}\neq a}\sum_{i=1}^{n}q_{i}u_{r}(a^{\prime},\omega_{i})\right\}

is non-empty.

Let D1,D2,,DD_{1},D_{2},\ldots,D_{\ell} be the polytopes contained in Δn1\Delta^{n-1} such that DjD_{j} contains the set of points qΔn1q\in\Delta^{n-1} for which action aja_{j} is a best response given that the posterior distribution is qq. Given a point qΔn1q\in\Delta^{n-1}, let A(q)A(q) be the set of indices j[]j\in[\ell] such that vDjv\in D_{j} and let I(q):=[minjA(v)us(a),maxjA(v)us(a)]I(q):=[\min_{j\in A(v)}u_{s}(a),\max_{j\in A(v)}u_{s}(a)]. Intuitively, I(q)I(q) is the interval ranging from the minimum to the maximum utility that the sender can get with a receiver’s best-response for qq. Let VV be the union of the set of vertices of D1,D2,,DD_{1},D_{2},\ldots,D_{\ell} (i.e., VV is the set of points that are vertices of at least one of the polytopes), and let CC be the set of tuples of points v1,,vkv_{1},\ldots,v_{k} of VV such that knk\leq n and the intersection of all of the intervals I(vi)I(v_{i}) is non-empty. More precisely,

C:={(v1,v2,,vk)Vk:kn,i=1kI(vi)}.C:=\left\{(v_{1},v_{2},\ldots,v_{k})\in V^{k}:k\leq n,\ \bigcap_{i=1}^{k}I(v_{i})\not=\emptyset\right\}.

Moreover, given any vertex v=(v1,,vk)C\vec{v}=(v_{1},\ldots,v_{k})\in C, let H(v)Δn1H(\vec{v})\subseteq\Delta^{n-1} be the convex hull of {v1,,vk}\{v_{1},\ldots,v_{k}\} (i.e., the set of all the affine combinations of points in v\vec{v}). Let

u~r(p):=max{i=1kαiur(a(vi),vi)v=(v1,,vk)C,pH(v)},\tilde{u}_{r}(p):=\max\left\{\sum_{i=1}^{k}\alpha_{i}u_{r}(a(v_{i}),v_{i})\ \mid\ \vec{v}=(v_{1},\ldots,v_{k})\in C,p\in H(\vec{v})\right\},

where a(v)a(v) is the action in A(v)A(v) that has the lowest index and (αi)i=1kΔk1(\alpha_{i})_{i=1}^{k}\in\Delta^{k-1} are some affine parameters such that i=1kαivi=p\sum_{i=1}^{k}\alpha_{i}v_{i}=p. The generalization of Theorem 1 is as follows.

Theorem 2.

Given an information transmission game Γ=(A,Ω,p,M,u)\Gamma=(A,\Omega,p,M,u) with transparent motives, the maximum utility that the receiver can get in a Nash equilibrium is given by u~r(p)\tilde{u}_{r}(p).

Proof.

Showing that u~r(p)\tilde{u}_{r}(p) is achievable by the receiver in a Nash equilibrium is analogous to the case with two states. However, the converse direction requires a slightly different reasoning. As in the case with binary states, for every Nash equilibrium in which the sender sends n+1n+1 or more messages with positive probability (where nn is the number of states), there exists a Nash equilibrium in which the sender sends nn or fewer messages and gives the same utility for the sender and the receiver.

Suppose that σ\vec{\sigma} is a Nash equilibrium where the sender sends knk\leq n different messages m1,,mkm_{1},\ldots,m_{k} with positive probability. For each i[k]i\in[k], denote by viv_{i} the posterior distribution of possible states after the receiver receives message mim_{i} when the sender plays σs\sigma_{s}. As in the case of n=2n=2, we claim that, without loss of generality, we can assume that viVv_{i}\in V for every i[k]i\in[k]. More precisely, we show next that, for every equilibrium σ\vec{\sigma}, there exists possibly a different equilibrium σ\vec{\sigma}^{\prime} in which each message mim_{i} induces a posterior viv^{\prime}_{i} such that (a) viVv^{\prime}_{i}\in V for all i[k]i\in[k], (b) the sender gets the same utility in σ\vec{\sigma} and σ\vec{\sigma}^{\prime}, and (c) the receiver gets a (weakly) higher utility in σ\vec{\sigma}^{\prime} than in σ\vec{\sigma}.

Assume that viVv_{i}\not\in V for some i[k]i\in[k]. We claim that we can find a distribution λΔ(V)\lambda\in\Delta(V) such that (i) vVλ(v)v=vi\sum_{v\in V}\lambda(v)v=v_{i} and (ii) if ajAa_{j}\in A is a best response for the receiver given a posterior distribution viv_{i}, then aja_{j} is also a best response for the receiver in every vVv\in V such that λ(v)>0\lambda(v)>0. To see this, let MiAM_{i}\subseteq A be the set of actions that are a best response for the receiver in viv_{i} and let Ti:=ajMiDjT_{i}:=\cap_{a_{j}\in M_{i}}D_{j}. Since TiT_{i} is a convex polytope that contains viv_{i}, we can define λ\lambda as an affine combination of the vertices of TiT_{i} (where the coefficients represent the probabilities of each vertex). All such vertices satisfy the above (ii) property.

Assume that, in σ\vec{\sigma}, each message mjm_{j} is sent with total probability probability qjq_{j}. We define σ\vec{\sigma}^{\prime} as follows. The strategy σr\sigma^{\prime}_{r} of the receiver is equal to σr\sigma_{r}. The sender does the following. For all jij\neq i, the sender sends message mjm_{j} with total probability qjq_{j} (the same probability as in σs\sigma_{s}). However, instead of sending message mim_{i} with probability qiq_{i}, in σs\sigma^{\prime}_{s} the sender sends each vertex vv of TiT_{i} with probability λ(v)qi\lambda(v)\cdot q_{i}. By property (i), the strategy σs\sigma^{\prime}_{s} is well defined and, by property (ii), strategy σr\sigma^{\prime}_{r} is a best-response for the receiver to σs\sigma^{\prime}_{s}.

By construction, the sender has identical utility in σ\vec{\sigma} and in σ\vec{\sigma}^{\prime}. To see that the receiver (weakly) increases her utility, note that the posterior distribution in σ\vec{\sigma}^{\prime} is a mean preserving spread of σ\vec{\sigma} since we split mim_{i} to the vertices of TiT_{i}. Repeating this process for every miVm_{i}\not\in V gives the desired equilibrium.

To finish the proof, note that we can express the receiver’s expected utility as an affine combination of {ur(a(mi),mi)}i=1k\{u_{r}(a(m_{i}),m_{i})\}_{i=1}^{k}. Moreover, by Bayes’ plausibility assumption, the affine coefficients αi\alpha_{i} of the combination are uniquely determined by the identity i=1kαimi=p\sum_{i=1}^{k}\alpha_{i}m_{i}=p. This shows that ur(σ)u~r(p)u_{r}(\vec{\sigma})\leq\tilde{u}_{r}(p), as desired.

2.4 Efficiently computing u~r(q)\tilde{u}_{r}(q) for a Binary State Sace

In this section we provide an algorithm that, given an information transmission game Γ=(A,Ω,p,M,u)\Gamma=(A,\Omega,p,M,u) with |A|=n|A|=n, computes the best utility of the receiver in a Nash equilibrium and runs in O(nlogn)O(n\log n) time. By Theorem 1, we know that the best utility of the receiver in a Nash equilibrium of Γ\Gamma is a piece-wise linear function. More precisely, there exists a positive integer knk\leq n, a sequence of k+1k+1 real numbers 0=x1<x2<<xk+1=10=x_{1}<x_{2}<\ldots<x_{k+1}=1, and kk linear functions f1,,fkf_{1},\ldots,f_{k} such that, if qq is the prior probability that the realized state is ω1\omega_{1}, then

u~r(q)={f1(q)if q[x1,x2]f2(q)if q[x2,x3]fk(q)if q[xk,xk+1].\tilde{u}_{r}(q)=\left\{\begin{array}[]{ll}f_{1}(q)&\mbox{if }q\in[x_{1},x_{2}]\\ f_{2}(q)&\mbox{if }q\in[x_{2},x_{3}]\\ \vdots&\vdots\\ f_{k}(q)&\mbox{if }q\in[x_{k},x_{k+1}].\\ \end{array}\right.

Before describing the algorithm, we need the following additional notation. Given two pairs of indices (i,j)C(i,j)\in C and (i,j)C(i^{\prime},j^{\prime})\in C, we say that (i,j)(i,j) is dominated by (i,j)(i^{\prime},j^{\prime}) if iii\geq i^{\prime} and jjj\leq j^{\prime}. Intuitively, (i,j)(i,j) is dominated by (i,j)(i^{\prime},j^{\prime}) if they both belong to CC and the interval [i,j][i^{\prime},j^{\prime}] contains the interval [i,j][i,j]. It is easy to check that, if (i,j)(i^{\prime},j^{\prime}) dominates (i,j)(i,j), then, for each point (x,y)(x,y) in the segment si,js_{i,j} that goes from (bi,ur(ai))(b_{i},u_{r}(a_{i})) to (bj,ur(aj))(b_{j},u_{r}(a_{j})), there exists a point (x,y)si,j(x,y^{\prime})\in s_{i^{\prime},j^{\prime}} with yyy^{\prime}\geq y (this follows directly from the fact that the utility of the receiver is a convex function over the prior). This means that, in order to find the maximum of all the segments si,js_{i,j} with (i,j)C(i,j)\in C, we can restrict the search to segments that are not dominated by other segments. In particular, if we denote by Cmax(i)C_{\max}(i) the maximum index jj such that (i,j)C(i,j)\in C, we can restrict our search to all segments of the form si,Cmax(i)s_{i,C_{\max}(i)}.

Suppose that we have an oracle that gives Cmax(i)C_{\max}(i) for each ii in constant time. We show next how to construct an O(nlogn)O(n\log n) algorithm πCmax\pi^{C_{\max}} that outputs kk and the sequences x1,,xk+1x_{1},\ldots,x_{k+1} and f1,,fkf_{1},\ldots,f_{k} using CmaxC_{\max} as a black box. At a high level, the algorithm computes all the segments of the form si,Cmax(i)s_{i,C_{\max}(i)} and runs a variant of the Bentley-Ottmann algorithm in which, (a) we only output the segment intersections that involve the segment with the highest yy-coordinate, and (b) whenever two segments ss and ss^{\prime} intersect, it eliminates the segment with the lowest slope. This guarantees that the number of segment intersections processed by the algorithm is at most nn, and therefore that the total complexity of the algorithm is O(nlogn)O(n\log n). More precisely, πCmax\pi^{C_{\max}} consists of running a vertical sweep-line LL that moves across the xx-axis from left to right while keeping a list of all the segments that LL intersects, sorted by the yy-coordinate of their intersection with LL. In order to do so, we need two data structures: a binary search tree TT whose purpose is to keep track of all the segments that LL intersects, sorted by yy-coordinate, and a priority queue QQ whose purpose is to keep a list of all the events that occur as LL moves from left to right, sorted by xx-coordinate. There are three types of events:

  • A segment ss begins. In this case, we insert ss in TT and query from TT the segments ss^{\prime} and s′′s^{\prime\prime} that are immediately above and below ss, respectively. If ss intersects ss^{\prime} or ss intersects s′′s^{\prime\prime}, we add these intersections to the event queue.

  • A segment ss ends. If ss is still in TT, we remove ss from TT. Let ss^{\prime} and s′′s^{\prime\prime} be the segments in TT that are immediately above and below ss. If ss^{\prime} and s′′s^{\prime\prime} intersect, we add the intersection to the event queue.

  • There is an intersection PP between two segments ss and ss^{\prime}. If either ss or ss^{\prime} has been previously deleted from TT, this event is ignored. Otherwise, we erase from TT the topmost segment between ss and ss^{\prime} and proceed as if the erased segment ended (i.e., we check the segments that are immediately above and below of the erased segment and, if they intersect, we add the intersection). If the segment removed was also the topmost segment of all TT, we append PP to the output.

The full algorithm is as follows. We initialize an empty binary search tree TT and an empty priority queue QQ and push 2n2n events to QQ indicating the beginning and end of segments s1,Cmax(1),s2,Cmax(2),,sn,Cmax(n)s_{1,C_{\max}(1)},s_{2,C_{\max}(2)},\allowbreak\ldots,s_{n,C_{\max}(n)}. We then process the events in QQ as described above until QQ becomes empty and output the sequence of points obtained this way, in addition to (0,ur(a1))(0,u_{r}(a_{1})) and (1,ur(an))(1,u_{r}(a_{n})). The values xix_{i} correspond to the xx-coordinates of the points in the output, and the linear functions fif_{i} are the ones that interpolate each pair of consecutive points.

The proof of correctness is equivalent to that of the Bentley-Ottmann algorithm. Note that the only difference between the algorithms is the way that we process the intersection events in πCmax\pi^{C_{\max}}. The Bentley-Ottmann algorithm outputs all segment intersections and, to do so, it swaps ss and ss^{\prime} inside TT whenever ss and ss^{\prime} intersect (instead of processing them the way that we do). However, since a segment that is surpassed in yy-coordinate by another segment can never be maximal again, we can optimize the algorithm by removing it right after processing its intersection. Since each intersection event erases one of the segments involved, we process at most nn intersection events, which means that we process in total 3n3n events (nn starting points, nn endpoints, and nn intersections). Since each event is processed in O(logn)O(\log n) time, the total time complexity is O(nlogn)O(n\log n). To see that the output of πCmax\pi^{C_{\max}} is correct, note that the vertices that define the piece-wise linear function u~r\tilde{u}_{r} are either (a) intersections of different segments si,js_{i,j} or (b) points in which a segment si,js_{i,j} ends and another segment si,js_{i^{\prime},j^{\prime}} begins. If we use an implementation trick where we always process starting point events before endpoint events, the points defined in (b) would also be considered as intersections. This means that, using this implementation trick, the desired set of vertices is the subset of all segment intersections that occur at the top layer of the segments, which is precisely what πCmax\pi^{C_{\max}} outputs.

As a final note regarding this part of the algorithm, note that, as in Theorem 1, we are assuming that Γ\Gamma is given in a form such that there exist n+1n+1 real numbers 0=b1<b2<<bn+1=10=b_{1}<b_{2}<\ldots<b_{n+1}=1 such that action aia_{i} is the receiver’s optimal action in the interval [bi,bi+1][b_{i},b_{i+1}]. However, even if we can assume that Γ\Gamma has this form without loss of generality, the input of the algorithm may not be reduced to this form yet. However, we can perform the same variant of the Bentley-Ottmann algorithm in order to reduce Γ\Gamma to such a form, and to compute b1,b2,,bn+1b_{1},b_{2},\ldots,b_{n+1}. To see this, note that the linear function that outputs the utility that the receiver gets by playing aiAa_{i}\in A as a function of the prior is a segment going from (0,ur(0,ai))(0,u_{r}(0,a_{i})) to (1,ur(1,ai))(1,u_{r}(1,a_{i})). Reducing Γ\Gamma to the desired form means precisely to compute the upper envelope of this set of segments.

Now, we have an O(nlogn)O(n\log n) algorithm πCmax\pi^{C_{\max}} that, given CmaxC_{\max} as a black box, computes the best utility for the receiver. It remains to show how to compute {Cmax(1),,Cmax(n)}\{C_{\max}(1),\ldots,C_{\max}(n)\} in O(nlogn)O(n\log n) time as well.

Efficient Computation of CmaxC_{\max}

By definition, (j,k)C(j,k)\in C if I(j)I(k)I(j)\cap I(k)\not=\emptyset, where I(j)=[us(aj1),us(aj)]I(j)=[u_{s}(a_{j-1}),u_{s}(a_{j})]. For simplicity, denote by sjs_{j} and eje_{j} the minimum and maximum of I(j)I(j), respectively. Consider the following algorithm ρ\rho:

  1. 1.

    Step 1: Initialize an array CC of nn elements in which C[i]=iC[i]=i for all ini\in n.

  2. 2.

    Step 2: Initialize an empty array AA. For each j[n]j\in[n], append (sj,0,j)(s_{j},0,j) and (ej,1,j)(e_{j},1,j) to AA.

  3. 3.

    Step 3: Sort AA in lexicographical order (i.e., sort by the first coordinate, break ties with the second coordinate and then with the third one).

  4. 4.

    Step 4: Initialize an empty binary search tree TT and two empty arrays Pos1,Pos2Pos_{1},Pos_{2} of nn elements each. Iterate through the elements A[i]:=(a,b,j)A[i]:=(a,b,j) of AA and do the following. If b=0b=0, set Pos1[j]=iPos_{1}[j]=i, insert jj in TT, and set C[a]=TmaxC[a]=T_{\max}, where TmaxT_{\max} is the maximal element of TT. Otherwise, set Pos2[j]=iPos_{2}[j]=i, and remove jj from TT.

  5. 5.

    Step 5: For each i[n]i\in[n], set C[i]C[i] to the maximum between C[i]C[i] and the highest third coordinate between all the elements of AA in the interval [Pos1[i],Pos2[i]][Pos_{1}[i],Pos_{2}[i]].

We claim that ρ\rho outputs CmaxC_{\max} and can be implemented in O(nlogn)O(n\log n) time. To see that it gives the correct output, note that, given k[n]k\in[n], the highest jj such that I(k)I(j)I(k)\cap I(j)\not=\emptyset is either (i) the highest jj such that sjs_{j} or eje_{j} appear between sis_{i} and and eie_{i} (after sorting all the starting points and endpoints of each interval as in Step 3), or (ii) the highest jj such that sjs_{j} appears before sis_{i} and eje_{j} does not appear before sis_{i}. The values described in (i) and (ii) are calculated in step 4 and step 5, respectively. To see that ρ\rho can be implemented in O(nlogn)O(n\log n) time, note that steps 1,2,3 and 4 can be implemented in O(nlogn)O(n\log n) time. A naive implementation of step 5 would be quadratic in nn, but it can be performed in O(nlogn)O(n\log n) using a Segment Tree data structure.

3 Filtered Information Aggregation Games

3.1 Model and Analysis

In this section we analyze the effects of filtering on information aggregation games. At a high level, information aggregation games are information transmission games with multiple senders. For a rigorous definition, an information aggregation game consists of a receiver rr, and a tuple Γ=(S,A,Ω,p,M,u)\Gamma=(S,A,\Omega,p,M,u) S={1,2,,k}S=\{1,2,\ldots,k\} is the set of senders and A={a1,,a}A=\{a_{1},\ldots,a_{\ell}\}, Ω={ω1,,ωk}\Omega=\{\omega_{1},\ldots,\omega_{k}\}, pΔ(Ω)p\in\Delta(\Omega), MM and uu are defined as in Information Transmission games (see Section 2.1), with the only exception that uu is a function from S{r}×Ω×AS\cup\{r\}\times\Omega\times A to \mathbb{R} instead of a function from {s,r}×Ω×A\{s,r\}\times\Omega\times A to \mathbb{R}.

We are interested in a setting where the receiver can filter potential information from the senders. We model this as follows: at the beginning of the game, there is an additional phase (phase 0) where the receiver publicly chooses a function X:ΩΔ({0,1})X:\Omega\to\Delta(\{0,1\}^{*}) that maps each state ω\omega to a distribution over possible signals (we assume that signals are binary strings of arbitrary length). We call this function filter. The rest of the game follows exactly in the same way as in a (standard) information aggregation game, except that, if the realized state is ω\omega, the senders receive a signal xx sampled from X(ω)X(\omega) instead of ω\omega itself. We call this a filtered information aggregation game (FIAG). More formally, a FIAG ΓF=(S,A,Ω,p,M,u)\Gamma^{F}=(S,A,\Omega,p,M,u) runs as follows:

  • Phase 0: The receiver chooses a filter X:ΩΔ({0,1})X:\Omega\to\Delta(\{0,1\}^{*}). The filter is disclosed to the senders.

  • Phase 1: A state ωΩ\omega\in\Omega is sampled according to the distribution pp and is disclosed to the senders.

  • Phase 2: Each sender iSi\in S sends a message miMm_{i}\in M to the receiver rr.

  • Phase 3: The receiver plays an action aAa\in A.

  • Phase 4: Each player iS{r}i\in S\cup\{r\} receives u(i,ω,a)u(i,\omega,a) utility.

For an illustration of how the receiver may benefit from filtering, consider the following example.

Example 1.

Let Γ=(S,A,Ω,p,M,u)\Gamma=(S,A,\Omega,p,M,u) be an information aggregation game with S={s}S=\{s\}, A={0,1}A=\{0,1\}, Ω={ω1,ω2,ω3,ω4}\Omega=\{\omega_{1},\omega_{2},\omega_{3},\omega_{4}\}, pp is the uniform distribution over Ω\Omega and uu is defined by the following table, where the first component of each cell is the utility of the sender and the second component is the utility of the receiver:

u(0,ω)u(0,\omega) u(1,ω)u(1,\omega)
ω1\omega_{1} 0,00,0 1,11,1
ω2\omega_{2} 1,01,0 0,10,1
ω3\omega_{3} 1,11,1 0,00,0
ω4\omega_{4} 0,10,1 1,01,0

We show next that, in Example 1, the receiver can double her maximal utility by filtering the sender’s information.

Proposition 1.

If Γ\Gamma is the game defined in Example 1,

  • (a)

    The utility of the receiver in all sequential equilibria of Γ\Gamma is 12\frac{1}{2}.

  • (b)

    There exists a sequential equilibrium in ΓF\Gamma^{F} where the receiver ends up playing her preferred action in every single state.

Proof.

We claim that in a standard cheap talk equilibrium, the receiver utility is always 12\frac{1}{2} and thus the receiver cannot get more than his utility at the bubbling equilibrium where information is not revealed. To see this consider without loss of generality an equilibrium with only two messages m1,m2m_{1},m_{2} that are sent with positive probability. If, by way of contradiction, the equilibrium induces a utility that is strictly higher than 12\frac{1}{2} to the receiver, then both actions are played with positive probability and the receiver is strictly better off playing one of the actions, say action 11, for at least one of the messages, say m1m_{1}. This means that the sender is strictly better off sending m1m_{1} at states Ω=ω1,ω4\Omega=\omega_{1},\omega_{4} and message m2m_{2} at states ω2,ω3\omega_{2},\omega_{3}. But in this equilibrium, the receiver’s utility is 12\frac{1}{2}. A contradiction.

In contrast, consider the following equilibrium in the filtered game. First, the receiver filters the sender’s information by partitioning Ω\Omega as follows {{ω1,ω2},{ω3,ω4}}\{\{\omega_{1},\omega_{2}\},\{\omega_{3},\omega_{4}\}\} so that the sender only observes the chosen partition element. Then, at the second stage, the sender has two messages that are used to fully reveal his information to the receiver who plays action a=1a=1 if the chosen partition element is {ω1,ω2}\{\omega_{1},\omega_{2}\} and action a=0a=0 if the chosen partition element is {ω3,ω4}\{\omega_{3},\omega_{4}\}. It is easy to see that the sender maximizes his utility given the realized partition element as he is indifferent among the two possible messages. The receiver in contrast receives his maximal utility under this equilibrium. ∎

However, even if filtering can greatly improve the receiver’s utility in in general, the following result complements Theorems 1 and 2 by stating that the receiver cannot improve her utility by filtering information if the sender has transparent motives.

Proposition 2.

Let ΓF=(S,A,Ω,p,M,u)\Gamma^{F}=(S,A,\Omega,p,M,u) be a FIAG where the sender has transparent motives. Then, the receiver cannot increase her utility by filtering information.

Proof.

Given a sequential equilibrium σ\vec{\sigma} for ΓF\Gamma^{F}, consider the following strategy profile σ\vec{\sigma}^{\prime} where the receiver does not filter. In Phase 0, the receiver selects the filter that simply forwards the realized state to the senders (i.e., the information that the senders receive is not filtered). In Phase 2, the sender first simulates what signal would she get with σ\vec{\sigma} and what message mm would she send to the receiver in σ\vec{\sigma}, and then she sends mm to the receiver. In Phase 3, the receiver plays whatever it would have played in σ\vec{\sigma} after receiving message mm.

It is straightforward to check that σ\vec{\sigma}^{\prime} is also sequential equilibrium that gives each player the same utility as in σ\vec{\sigma}. ∎

In the next section we give an efficient protocols that computes the optimal filters for the receiver in arbitrary games with binary actions and one or two senders.

4 Best Sequential Equilibria with Filters and Binary Actions

We next address the question of best receiver equilibrium from a computational aspect. Both of our results restrict attention to the case where the action set of the receiver is binary.

Theorem 3.

Let ΓF=(S,A,Ω,p,u)\Gamma^{F}=(S,A,\Omega,p,u) be a filtered information aggregation game with S={s}S=\{s\} and A={0,1}A=\{0,1\}. Then, there exists an algorithm π\pi that outputs the best sequential equilibrium for the receiver in ΓF\Gamma^{F} that runs in O(klogk)O(k\log k) time, where k=|Ω|k=|\Omega|.

Theorem 4.

Let ΓF=(S,A,Ω,p,u)\Gamma^{F}=(S,A,\Omega,p,u) be a filtered information aggregation game with |Ω|=k|\Omega|=k, S={s1,s2}S=\{s_{1},s_{2}\}, and A={0,1}A=\{0,1\}. Then, finding the best sequential equilibrium for the receiver in ΓF\Gamma^{F} reduces to a linear programming instance with kk variables and 2k+22k+2 constraints.

Note that we only provide results for the cases of one and two senders. If ΓF\Gamma^{F} is a filtered information aggregation game with |S|3|S|\geq 3 there is a trivial strategy profile σMAX\vec{\sigma}_{MAX} that gives the receiver the maximal possible utility of the game. Using this strategy, the receiver puts no filters (more precisely, it selects the identity filter), the senders send the signal they get to the receiver, and then the receiver computes the signal xx that was sent by the majority of the senders and plays the action that gives her the most utility conditioned that the senders got signal xx. It is straightforward to check that this is indeed a Nash equilibrium: the senders do not get any additional utility by defecting since the other senders will still send the true signal (which means that the receiver will be able to compute the true signal as well), and the receiver does not get any additional utility either since she is always playing the optimal action for each possible state. In Section 7 we show that this strategy profile can be extended to a sequential equilibrium (σMAX,b)(\vec{\sigma}_{MAX}^{*},b) by defining an appropriate belief assessment bb. This gives the following result:

Theorem 5.

Let ΓF=(S,A,Ω,p,u)\Gamma^{F}=(S,A,\Omega,p,u) be a filtered information aggregation game with such that |S|3|S|\geq 3. Then, σMAX\vec{\sigma}_{MAX} can be extended to a sequential equilibrium that gives the receiver the maximal possible utility of the game in ΓF\Gamma^{F}.

In particular, we have the following Corollary.

Corollary 1.

If there are three or more senders, the receiver does not get any additional utility by filtering information.

In the following sections, we give algorithms that output the best Nash equilibria for the settings described in Theorems 3, 1 and 4, respectively. Later, in section 7, we show how to extend these Nash equilibria to sequential equilibria. Section 7 also includes the missing details in the proof of Theorem 5.

5 Proof of Theorem 3

In this section we provide an algorithm that outputs the best Nash equilibrium for the receiver in any filtered information game ΓF=(S,A,Ω,p,u)\Gamma^{F}=(S,A,\Omega,p,u) with S={s}S=\{s\} and A={0,1}A=\{0,1\}. In Section 7 we show how to extend the resulting Nash equilibrium to a sequential equilibrium.

Consider a game ΓXF\Gamma^{F}_{X} that is identical to ΓF\Gamma^{F} except for the fact that the filter XX is fixed beforehand (which means that the receiver doesn’t get to choose the filter in Phase 0). The following proposition shows that there exists a Pareto-optimal Nash equilibrium with a relatively simple form.

Proposition 3.

There exists a Pareto-optimal Nash equilibrium σ\vec{\sigma} in ΓXF\Gamma^{F}_{X} such that, in σ\vec{\sigma}, either:

  • The receiver always plays 0.

  • The receiver always plays 1.

  • The receiver always plays the best action for the sender.

Before proving Proposition 3 we need additional notation. First, given a filtered information game ΓXF=(S,A,Ω,p,u)\Gamma^{F}_{X}=(S,A,\Omega,p,u), let ui(x,a)u_{i}(x,a) be the expected utility of player ii on signal xx and action aa. This expected utility can be computed by the following equation:

ui(x,a)=ωΩPr[ωx]ui(ω,a),u_{i}(x,a)=\sum_{\omega\in\Omega}\Pr[\omega\mid x]\cdot u_{i}(\omega,a),

where Pr[ωx]\Pr[\omega\mid x] is the probability that the realized state is ω\omega conditional on the fact that the sender received signal xx. Similarly, Pr[ωx]\Pr[\omega\mid x] has the following expression:

Pr[ωx]=Pr[ωp,xX(ω)]ωΩPr[ωp,xX(ω)].\Pr[\omega\mid x]=\frac{\Pr[\omega\longleftarrow p,\ x\longleftarrow X(\omega)]}{\sum_{\omega\in\Omega}\Pr[\omega\longleftarrow p,\ x\longleftarrow X(\omega)]}.

Let X0X_{0} be the set of signals xx in {X(ω)}ωΩ\{X(\omega)\}_{\omega\in\Omega} such that us(x,0)>us(x,1)u_{s}(x,0)>u_{s}(x,1) (i.e., the set of signals in which the sender prefers 0), let X1X_{1} be the set of signals such that us(x,0)<us(x,1)u_{s}(x,0)<u_{s}(x,1), and X=X_{=} be the set signals such that us(x,0)=us(x,1)u_{s}(x,0)=u_{s}(x,1). Given a strategy profile σ\vec{\sigma} for ΓXF\Gamma^{F}_{X}, denote by σs(x,m)\sigma_{s}(x,m) the probability that the sender sends message mm given signal xx and denote by σr(a,m)\sigma_{r}(a,m) the probability that the receiver plays action aa given message mm. Moreover, let M0σM_{0}^{\vec{\sigma}} denote the set of messages that have a strictly positive probability to be sent by the sender on at least one signal xx in which the sender prefers 0 (i.e., M0σM_{0}^{\vec{\sigma}} is the set of messages mm such that there exists xX0x\in X_{0} such that us(x,m)>0u_{s}(x,m)>0). We define M1σM_{1}^{\vec{\sigma}} and M=σM_{=}^{\vec{\sigma}} analogously.

With this notation, the following lemma describes all strategy profiles in ΓXF\Gamma^{F}_{X} that are incentive-compatible for the sender.

Lemma 1.

A strategy profile σ\vec{\sigma} for ΓXF\Gamma^{F}_{X} is incentive-compatible for the sender if and only if the following is satisfied:

  • (a)

    If mM0σm\in M^{\vec{\sigma}}_{0}, then σr(0,m)σr(0,m)\sigma_{r}(0,m)\geq\sigma_{r}(0,m^{\prime}) for all messages mm^{\prime}.

  • (b)

    If mM1σm\in M^{\vec{\sigma}}_{1}, then σr(1,m)σr(1,m)\sigma_{r}(1,m)\geq\sigma_{r}(1,m^{\prime}) for all messages mm^{\prime}.

Lemma 1 states that, for a strategy profile to be incentive-compatible for the sender, the receiver should play 0 with maximal probability on all messages that could be sent on signals in which the sender prefers 0, and the receiver should play 0 with minimal probability on all messages that could be sent on signals in which the sender prefers 1. In particular, we have the following Corollary.

Corollary 2.

A strategy profile σ\vec{\sigma} for ΓXF\Gamma^{F}_{X} is incentive-compatible for the sender if and only if there exist two real numbers a,b[0,1]a,b\in[0,1] with aba\geq b such that:

  • (a)

    mM0σσr(0,m)=am\in M^{\vec{\sigma}}_{0}\ \Longrightarrow\ \sigma_{r}(0,m)=a.

  • (b)

    mM1σσr(0,m)=bm\in M^{\vec{\sigma}}_{1}\ \Longrightarrow\ \sigma_{r}(0,m)=b.

  • (c)

    mM=σbσr(0,m)am\in M^{\vec{\sigma}}_{=}\ \Longrightarrow\ b\leq\sigma_{r}(0,m)\leq a.

Proof of Lemma 1.

Clearly, if (a) and (b) are satisfied, then σ\vec{\sigma} is incentive-compatible for the sender. Conversely, suppose that σ\vec{\sigma} is incentive-compatible for the sender but it doesn’t satisfy (a). This means that there exists a signal xX0x\in X_{0} and a message mm that satisfies us(x,m)>0u_{s}(x,m)>0 and such that ur(0,m)<ur(0,m)u_{r}(0,m)<u_{r}(0,m^{\prime}) for some other message mm^{\prime}. Therefore, if the sender sends mm^{\prime} instead of mm whenever it receives signal xx, it could increase its expected utility. This contradicts the fact that σ\vec{\sigma} is incentive-compatible for the sender. The proof of the case in which σ\vec{\sigma} doesn’t satisfy (b) is analogous. ∎

Corollary 2 characterizes the necessary and sufficient conditions for a strategy profile σ\vec{\sigma} in ΓXF\Gamma^{F}_{X} to be incentive-compatible for the sender. We show next that Proposition 3 follows from adding the receiver incentive-compatibility constraints into the mix. Denote by urσ(m,0)u_{r}^{\vec{\sigma}}(m,0) the receiver’s expected utility when playing action 0 conditioned on the fact that it received message mm and that the sender plays σs\sigma_{s}. Then, we have the following cases:

Case 1: There exists mM0σm\in M_{0}^{\vec{\sigma}} such that urσ(m,0)<urσ(m,1)u_{r}^{\vec{\sigma}}(m,0)<u_{r}^{\vec{\sigma}}(m,1): Since σ\vec{\sigma} is incentive-compatible for the receiver, it must be the case that a=0a=0 and therefore that σr(0,m)=0\sigma_{r}(0,m)=0 for all messages mm.

Case 2: There exists mM1σm\in M_{1}^{\vec{\sigma}} such that urσ(m,0)>urσ(m,1)u_{r}^{\vec{\sigma}}(m,0)>u_{r}^{\vec{\sigma}}(m,1): This time it must be the case that b=1b=1 and therefore that σr(0,m)=1\sigma_{r}(0,m)=1 for all messages mm.

Case 3: urσ(m,0)urσ(m,1)u_{r}^{\vec{\sigma}}(m,0)\geq u_{r}^{\vec{\sigma}}(m,1) for all mM0σm\in M_{0}^{\vec{\sigma}} and urσ(m,0)urσ(m,1)u_{r}^{\vec{\sigma}}(m,0)\leq u_{r}^{\vec{\sigma}}(m,1) for all mM1σm\in M_{1}^{\vec{\sigma}}: In this case, consider a strategy profile σ\vec{\sigma}^{\prime} that is identical to σ\vec{\sigma} except that, whenever the receiver receives a message mM0σm\in M_{0}^{\vec{\sigma}}, it plays action 0 with probability 1 (as opposed to probability aa), and when the receiver receives a message mM1σm\in M_{1}^{\vec{\sigma}}, it plays action 11 with probability 1 (as opposed to 1b1-b). It is easy to check that, by construction, σ\vec{\sigma}^{\prime} Pareto-dominates σ\vec{\sigma}. Even so, σ\vec{\sigma}^{\prime} can be further improved: consider a strategy profile σs\vec{\sigma}^{s} such that the sender sends message 0 in all subsets CX0C\in X_{0} and in all subsets CX=C\in X_{=} such that ur(C,0)>ur(C,1)u_{r}(C,0)>u_{r}(C,1), and the sender sends message 11 in all other subsets. Additionally, the receiver plays action 0 when receiving message 0 and plays action 11 when receiving message 11. It is easy to check that σs\vec{\sigma}^{s} satisfies that urσs(0,0)urσs(0,1)u_{r}^{\vec{\sigma}^{s}}(0,0)\geq u_{r}^{\vec{\sigma}^{s}}(0,1) and urσs(1,0)urσs(1,1)u_{r}^{\vec{\sigma}^{s}}(1,0)\leq u_{r}^{\vec{\sigma}^{s}}(1,1). Therefore, σs\vec{\sigma}^{s} is a Nash equilibrium of ΓXE\Gamma^{E}_{X}. Moreover, σs\vec{\sigma}^{s} Pareto-dominates σ\vec{\sigma}^{\prime}. To check this, note that both σr\vec{\sigma}^{r} and σ\vec{\sigma}^{\prime} play the best action for the sender at every possible signal xx. The only difference is that, in signals in which the sender is indifferent, the receiver plays the best action for herself in σs\vec{\sigma}^{s} while it doesn’t necessarily do so in σ\vec{\sigma}^{\prime}.

This analysis provides a refinement of Proposition 3. In fact, consider the following two strategy profiles:

  • Strategy σr\vec{\sigma}^{r}: Regardless of the signal received, the sender sends an empty \bot message. The receiver plays the action that gives her the most utility with no information.

  • Strategy σs\vec{\sigma}^{s}: After receiving the signal, the sender sends its preferred action. The receiver plays the action sent by the sender.

We have the following characterization of Pareto-optimal Nash equilibria:

Proposition 4 (Refinement of Proposition 3).

If σs\vec{\sigma}^{s} is incentive-compatible for the receiver, then σs\vec{\sigma}^{s} is a Pareto-optimal Nash equilibrium of ΓXF\Gamma^{F}_{X}. Otherwise, σr\vec{\sigma}^{r} is a Pareto-optimal Nash equilibrium of ΓFX\Gamma_{F}^{X}.

Proposition 4 shows that there is a simple algorithm that outputs a Pareto-optimal Nash equilibrium of ΓXF\Gamma^{F}_{X} once the filter XX has been fixed beforehand. This reduces the problem of finding the best Nash equilibrium of ΓF\Gamma^{F} for the receiver to that of finding the filter XOPTX^{OPT} such that the Pareto-optimal Nash equilibrium of ΓXF\Gamma^{F}_{X} gives the best possible utility for the receiver. In fact, the best Nash equilibrium σOPT\vec{\sigma}^{OPT} goes as follows:

  1. 1.

    The receiver chooses the filter XOPTX^{OPT} such that σs\vec{\sigma}^{s} is incentive-compatible for the receiver and gives the receiver the maximum utility in ΓXF\Gamma^{F}_{X}. If there is no filter such filter XOPTX^{OPT}, the receiver chooses an arbitrary filter.

  2. 2.

    Given the chosen filter XX and the signal xx, the sender computes the Pareto-optimal Nash equilibrium σ\vec{\sigma} of ΓXF\Gamma^{F}_{X} and plays σs\sigma_{s} with signal xx.

  3. 3.

    If the receiver sent an arbitrary filter in step 1, it plays the best action assuming no information. Otherwise, if the receiver receives 0, it plays action 0. If it receives 11, it plays action 11. If it receives something else it plays the best action assuming no information.

The following proposition shows that σOPT\vec{\sigma}^{OPT} satisfies the desired properties.

Proposition 5.

σOPT\vec{\sigma}^{OPT} is the Nash equilibrium of ΓF\Gamma^{F} that gives the most utility to the receiver.

Proof.

Once the sender chooses a filter XX, the receiver and the sender both maximize their utilities if they play the Pareto-optimal strategy profile described in Proposition 4. Therefore, it is optimal for the receiver to choose XOPTX^{OPT}. ∎

5.1 Computation of XOPTX^{OPT}

In this section we show how to compute the optimal filter XOPTX^{OPT} for the receiver. Our aim is to find a filter XOPTX^{OPT} such that σs\vec{\sigma}^{s} is incentive-compatible for the receiver in ΓXOPTF\Gamma^{F}_{X^{OPT}} and such that the expected utility for the receiver with σs\vec{\sigma}^{s} is as large as possible. We begin by showing an algorithm that runs in O(klogk)O(k\log k) time, where kk is the number of states (i.e., the size of Ω\Omega), and then we show the proof of correctness for the algorithm provided.

5.1.1 An O(klogk)O(k\log k) Algorithm

For our algorithm, we restrict our search to binary filters (i.e., filters that only send signals in {0,1}\{0,1\}) and later, in Lemma 2, we show that this is done without loss of generality. With this restriction, we can describe possible candidates XX by a function XX_{*} that maps each state ω\omega to the probability that X(ω)X(\omega) outputs 0. For simplicity, we’ll also say that XX is incentive-compatible for the sender (resp., for the receiver) if the strategy profile in which the sender forwards the signal to the receiver and the receiver plays whatever it gets from the sender is incentive-compatible for the sender (resp., for the receiver). We will also say that the sender (resp., the receiver) gets aa expected utility in XX, if aa is the expected utility it gets when running the previous strategy in ΓXF\Gamma^{F}_{X}. Note that σs\vec{\sigma}^{s} is always incentive-compatible for the sender (for all filters XX), but a binary filter XX might not always be since we are forcing the sender to send the signal received instead of its true preference. However, if σs\vec{\sigma}^{s} is incentive-compatible for the sender (resp., for the receiver) in some game ΓXF\Gamma^{F}_{X} for some binary filter XX, and the sender does not always have the same preference in ΓXF\Gamma^{F}_{X}, then either XX or its complementary (i.e., the filter XX^{\prime} that sends 0 whenever XX sends 11, and sends 11 whenever XX sends 0) is incentive-compatible for the sender (resp., the receiver) and gives the same utility as σs\vec{\sigma}^{s} in ΓXF\Gamma^{F}_{X}. This shows that, without loss of generality, we can simply search for the binary filter that (a) is incentive-compatible for both the sender and the receiver, and (b) that gives the most utility for the receiver.

Before we start, let Ω0\Omega^{0} (resp., Ω1\Omega^{1}) denote the set of states in which both the sender and the receiver prefer 0 and let Ω01\Omega_{0}^{1} (resp., Ω10\Omega_{1}^{0}) denote the set of states in which the sender strictly prefers 0 and the receiver strictly prefers 11 (resp., the sender strictly prefers 11 and the receiver strictly prefers 0). The following algorithm outputs the optimal filter XOPTX^{OPT} for the receiver.

  1. Step 1

    Set XOPT(ω)=1X^{OPT}_{*}(\omega)=1 for all ωΩ0\omega\in\Omega^{0}.

  2. Step 2

    Set XOPT(ω)=0X^{OPT}_{*}(\omega)=0 for all ωΩ1\omega\in\Omega^{1}.

  3. Step 3

    Sort all states ωΩ01Ω10\omega\in\Omega_{0}^{1}\cup\Omega_{1}^{0} according to the value ur(ω,0)ur(ω,1)us(ω,1)us(ω,0)\frac{u_{r}(\omega,0)-u_{r}(\omega,1)}{u_{s}(\omega,1)-u_{s}(\omega,0)}. Let ω1,ω2,,ωk\omega^{1},\omega^{2},\ldots,\omega^{k^{\prime}} be the resulting list.

  4. Step 4

    Set XOPT(ω)=1X^{OPT}_{*}(\omega)=1 for all ωΩ10\omega\in\Omega_{1}^{0} and XOPT(ω)=0X^{OPT}_{*}(\omega)=0 for all ωΩ01\omega\in\Omega_{0}^{1}. If the resulting filter is incentive-compatible for the sender, output XOPTX^{OPT}.

  5. Step 5

    For i=1,2,,ki=1,2,\ldots,k^{\prime} do the following. If XOPT(ωi)X^{OPT}_{*}(\omega^{i}) is set to 11 (resp., to 0), set it to 0 (resp., to 1). If the resulting filter is incentive-compatible for the sender, find the maximum (resp., the minimum) qq such that setting XOPT(ωi)X^{OPT}_{*}(\omega^{i}) to qq is incentive-compatible for the sender (we show below that it can be computed with a linear equation with amortized cost). If the resulting filter is incentive-compatible for the receiver, output XOPTX^{OPT}, otherwise output the constant filter XOPT0X^{OPT}_{*}\equiv 0.

It is important to note that the algorithm always terminates, since if we get to kk^{\prime} in step 5, the resulting filter will always output the signal that the sender prefers. It remains to show how to compute qq in amortized linear time. For this purpose, let disd_{i}^{s} (resp., dird_{i}^{r}) denote the difference in utility for the sender (resp., for the receiver) between playing 0 and playing 11. More precisely,

dis:=us(ωi,0)us(ωi,1)dir:=ur(ωi,0)ur(ωi,1).\begin{array}[]{cc}d_{i}^{s}:=u_{s}(\omega_{i},0)-u_{s}(\omega_{i},1)\\ d_{i}^{r}:=u_{r}(\omega_{i},0)-u_{r}(\omega_{i},1).\\ \end{array}

Then, given a binary filter XX, σs\vec{\sigma}^{s} is incentive-compatible for the sender in ΓXF\Gamma^{F}_{X} if and only if

ip(ωi)disX(ωi)0ip(ωi)dis(1X(ωi))0.\begin{array}[]{ll}\sum_{i}p(\omega_{i})\cdot d_{i}^{s}\cdot X_{*}(\omega_{i})\geq 0\\ \sum_{i}p(\omega_{i})\cdot d_{i}^{s}\cdot(1-X_{*}(\omega_{i}))\leq 0.\end{array} (2)

The first and second equation state that the sender prefers 0 on signal 0 and prefers 11 on signal 11, respectively. Similarly, σs\vec{\sigma}^{s} is incentive-compatible for the receiver in ΓXF\Gamma^{F}_{X} if and only if

ip(ωi)dirX(ωi)0ip(ωi)dir(1X(ωi))0.\begin{array}[]{ll}\sum_{i}p(\omega_{i})\cdot d_{i}^{r}\cdot X_{*}(\omega_{i})\geq 0\\ \sum_{i}p(\omega_{i})\cdot d_{i}^{r}\cdot(1-X_{*}(\omega_{i}))\leq 0.\end{array} (3)

It is straightforward to check that the incentive-compatibility equations for the sender monotonically increase and decrease, respectively, whenever we iterate in the fifth step of the algorithm. Therefore, we can simply pre-compute the sums

Y0s:=ωΩ00p(ω)(us(ω,0)us(ω,1))Y1s:=ωΩ11p(ω)(us(ω,0)us(ω,1))Wi,0(s,s):=j<i,ωjΩ01p(ωj)(us(ωj,0)us(ωj,1))Wi,1(s,s):=j<i,ωjΩ10p(ωj)(us(ωj,0)us(ωj,1))Wi,0(s,r):=j>i,ωjΩ01p(ωj)(us(ωj,0)us(ωj,1))Wi,1(s,r):=j>i,ωjΩ10p(ωj)(us(ωj,0)us(ωj,1))\begin{array}[]{l}Y_{0}^{s}:=\sum_{\omega\in\Omega_{0}^{0}}p(\omega)\cdot(u_{s}(\omega,0)-u_{s}(\omega,1))\\ Y_{1}^{s}:=\sum_{\omega\in\Omega_{1}^{1}}p(\omega)\cdot(u_{s}(\omega,0)-u_{s}(\omega,1))\\ W_{i,0}^{(s,s)}:=\sum_{j<i,\omega^{j}\in\Omega_{0}^{1}}p(\omega^{j})\cdot(u_{s}(\omega^{j},0)-u_{s}(\omega^{j},1))\\ W_{i,1}^{(s,s)}:=\sum_{j<i,\omega^{j}\in\Omega_{1}^{0}}p(\omega^{j})\cdot(u_{s}(\omega^{j},0)-u_{s}(\omega^{j},1))\\ W_{i,0}^{(s,r)}:=\sum_{j>i,\omega^{j}\in\Omega_{0}^{1}}p(\omega^{j})\cdot(u_{s}(\omega^{j},0)-u_{s}(\omega^{j},1))\\ W_{i,1}^{(s,r)}:=\sum_{j>i,\omega^{j}\in\Omega_{1}^{0}}p(\omega^{j})\cdot(u_{s}(\omega^{j},0)-u_{s}(\omega^{j},1))\end{array}

and analogues for the receiver:

Y0r:=ωΩ00p(ω)(ur(ω,0)ur(ω,1))Y1r:=ωΩ11p(ω)(ur(ω,0)ur(ω,1))Wi,0(r,s):=j<i,ωjΩ01p(ωj)(ur(ωj,0)ur(ωj,1))Wi,1(r,s):=j<i,ωjΩ10p(ωj)(ur(ωj,0)ur(ωj,1))Wi,0(r,r):=j>i,ωjΩ01p(ωj)(ur(ωj,0)ur(ωj,1))Wi,1(r,r):=j>i,ωjΩ10p(ωj)(ur(ωj,0)ur(ωj,1)).\begin{array}[]{l}Y_{0}^{r}:=\sum_{\omega\in\Omega_{0}^{0}}p(\omega)\cdot(u_{r}(\omega,0)-u_{r}(\omega,1))\\ Y_{1}^{r}:=\sum_{\omega\in\Omega_{1}^{1}}p(\omega)\cdot(u_{r}(\omega,0)-u_{r}(\omega,1))\\ W_{i,0}^{(r,s)}:=\sum_{j<i,\omega^{j}\in\Omega_{0}^{1}}p(\omega^{j})\cdot(u_{r}(\omega^{j},0)-u_{r}(\omega^{j},1))\\ W_{i,1}^{(r,s)}:=\sum_{j<i,\omega^{j}\in\Omega_{1}^{0}}p(\omega^{j})\cdot(u_{r}(\omega^{j},0)-u_{r}(\omega^{j},1))\\ W_{i,0}^{(r,r)}:=\sum_{j>i,\omega^{j}\in\Omega_{0}^{1}}p(\omega^{j})\cdot(u_{r}(\omega^{j},0)-u_{r}(\omega^{j},1))\\ W_{i,1}^{(r,r)}:=\sum_{j>i,\omega^{j}\in\Omega_{1}^{0}}p(\omega^{j})\cdot(u_{r}(\omega^{j},0)-u_{r}(\omega^{j},1)).\end{array}

Note that these sums can be computed in amortized linear time over the total number of states. Once this sums are computed, if we are in the iith iteration of Step 5, to find qq we can check in constant time if the solution of any of the following linear equations is in [0,1][0,1]:

Y0s+Wi,0(s,s)+p(ωi)qus(ω,0)us(ω,1)+Wi,0(s,r)=0Y1s+Wi,1(s,s)+p(ωi)(1q)us(ω,0)us(ω,1)+Wi,1(s,r)=0.\begin{array}[]{l}Y_{0}^{s}+W_{i,0}^{(s,s)}+p(\omega^{i})\cdot q\cdot{u_{s}(\omega,0)-u_{s}(\omega,1)}+W_{i,0}^{(s,r)}=0\\ Y_{1}^{s}+W_{i,1}^{(s,s)}+p(\omega^{i})\cdot(1-q)\cdot{u_{s}(\omega,0)-u_{s}(\omega,1)}+W_{i,1}^{(s,r)}=0.\\ \end{array}

If there exists such solution qq, to check that it is incentive-compatible for both players, we should simply check if the following inequalities are satisfied:

Y0s+Wi,0(s,s)+p(ωi)qus(ω,0)us(ω,1)+Wi,0(s,r)0Y1s+Wi,1(s,s)+p(ωi)(1q)us(ω,0)us(ω,1)+Wi,1(s,r)0.Y0r+Wi,0(r,s)+p(ωi)qur(ω,0)ur(ω,1)+Wi,0(r,r)0Y1r+Wi,1(r,s)+p(ωi)(1q)ur(ω,0)ur(ω,1)+Wi,1(r,r)0.\begin{array}[]{l}Y_{0}^{s}+W_{i,0}^{(s,s)}+p(\omega^{i})\cdot q\cdot{u_{s}(\omega,0)-u_{s}(\omega,1)}+W_{i,0}^{(s,r)}\geq 0\\ Y_{1}^{s}+W_{i,1}^{(s,s)}+p(\omega^{i})\cdot(1-q)\cdot{u_{s}(\omega,0)-u_{s}(\omega,1)}+W_{i,1}^{(s,r)}\leq 0.\\ Y_{0}^{r}+W_{i,0}^{(r,s)}+p(\omega^{i})\cdot q\cdot{u_{r}(\omega,0)-u_{r}(\omega,1)}+W_{i,0}^{(r,r)}\geq 0\\ Y_{1}^{r}+W_{i,1}^{(r,s)}+p(\omega^{i})\cdot(1-q)\cdot{u_{r}(\omega,0)-u_{r}(\omega,1)}+W_{i,1}^{(r,r)}\leq 0.\\ \end{array}

All of these computations can be performed in constant time if the necessary sums are pre-computed. This completed the description of the algorithm, note that it runs in O(klogk)O(k\log k) time since it takes O(klogk)O(k\log k) operations to sort the elements of Ω01Ω10\Omega_{0}^{1}\cup\Omega_{1}^{0}, O(k)O(k) operations to compute the sums, and O(1)O(1) operations in each iteration of Step 5. In the next section, we show that the algorithm’s output is correct.

5.1.2 Proof of Correctness

In this section, we show that the Algorithm presented in Section 5.1 is correct. We begin by showing that we can indeed restrict our search to binary filters filters.

Lemma 2.

Let XX be a filter such that σs\vec{\sigma}^{s} is incentive-compatible for the receiver in ΓXF\Gamma^{F}_{X}. Then, there exists a binary filter XX^{\prime} such that

  • (a)

    σs\vec{\sigma}^{s} is incentive-compatible for the receiver in ΓXF\Gamma^{F}_{X^{\prime}}

  • (b)

    With strategy profile σs\vec{\sigma}^{s}, the expected utility of the receiver in ΓXF\Gamma^{F}_{X} and ΓXF\Gamma^{F}_{X^{\prime}} is identical.

Proof.

Recall that σs\vec{\sigma}^{s} is a strategy profile in which, for each possible signal xx, the sender sends its preferred action and then the receiver plays whatever is sent by the sender. This means that, if σ\vec{\sigma} is incentive-compatible for the receiver in ΓXF\Gamma^{F}_{X}, then we can merge all signals in which the sender prefers 0, and also merge all signals in which the sender prefers 11. More precisely, given filter XX, let X0X_{0} and X1X_{1} be the sets of signals in which the sender prefers 0 and 11 respectively. Consider a filter XX^{\prime} that sends signal 0 whenever XX would send a signal in X0X_{0}, and sends signal 11 whenever XX would send a signal in X1X_{1}. In ΓXF\Gamma^{F}_{X^{\prime}}, by construction, both the sender and the receiver prefer action 0 on signal 0 and action 11 on signal 11. This means that σs\vec{\sigma}^{s} is a Nash equilibrium of ΓXF\Gamma^{F}_{X^{\prime}}. Moreover, again by construction, the expected utility of the receiver (and the sender) when playing σs\vec{\sigma}^{s} in ΓXF\Gamma^{F}_{X^{\prime}} is identical to the one they’d get in ΓXF\Gamma^{F}_{X}. ∎

Recall that binary filters XX can be described by a function X:Ω[0,1]X_{*}:\Omega\longrightarrow[0,1] that maps each state ωΩ\omega\in\Omega to the probability that X(ω)X(\omega) assigns to 0. Because of Lemma 1, our aim is to find which values in [0,1][0,1] we should assign to each element in Ω\Omega. The following lemmas characterizes these values.

Lemma 3.

There exists a filter XOPTX^{OPT} that maximizes the utility for the receiver such that

  • (a)

    XOPT(ω)=1X^{OPT}_{*}(\omega)=1 for all ωΩ00\omega\in\Omega_{0}^{0}.

  • (b)

    XOPT(ω)=0X^{OPT}_{*}(\omega)=0 for all ωΩ11\omega\in\Omega_{1}^{1}.

  • (c)

    At most one state ωΩ01Ω10\omega\in\Omega_{0}^{1}\cup\Omega_{1}^{0} satisfies that XOPT(ω){0,1}X^{OPT}_{*}(\omega)\not\in\{0,1\}.

Proof.

Given a binary filter XX, if we increase X(ωi)X_{*}(\omega_{i}) by ϵ\epsilon, the expected utility of the receiver increases by ϵp(ωi)dir\epsilon p(\omega_{i})d_{i}^{r}. This means that, if dis0d_{i}^{s}\geq 0 and dir0d_{i}^{r}\geq 0, setting X(ωi)X_{*}(\omega_{i}) to 11 both increases the receiver’s utility and preserves the incentive-compatibility constraints (see Equations 2 and 3). Analogously, the same happens by setting X(ωi)X_{*}(\omega_{i}) to 0 when dis0d_{i}^{s}\leq 0 and dir0d_{i}^{r}\leq 0. This proves (a) and (b).

To prove (c) suppose that there exist two states ωi\omega_{i} and ωj\omega_{j} such that XOPT(ωi),XOPT(ωj){0,1}X_{*}^{OPT}(\omega_{i}),X_{*}^{OPT}(\omega_{j})\not\in\{0,1\}. Then, because of (a) and (b) we can assume without loss of generality that ωi,ωjΩ01Ω10\omega_{i},\omega_{j}\in\Omega_{0}^{1}\cup\Omega_{1}^{0}. Therefore, if we increase X(ωi)X_{*}(\omega_{i}) by ϵdjs\epsilon d_{j}^{s} and decrease X(ωj)X_{*}(\omega_{j}) by ϵdjs\epsilon d_{j}^{s}, we’d have that the sender’s utility and incentive-compatibility constraints remain unchanged, but the receiver’s utility increases by ϵ(djsdirdisdjr)\epsilon(d_{j}^{s}d_{i}^{r}-d_{i}^{s}d_{j}^{r}) (note that this value can be negative). This means that if we choose an ϵ\epsilon that is small enough and is of the same sign as djsdirdisdjrd_{j}^{s}d_{i}^{r}-d_{i}^{s}d_{j}^{r}, not only the expected utility of the receiver increases, but also the values ip(ωi)dirX(ωi)\sum_{i}p(\omega_{i})\cdot d_{i}^{r}\cdot X_{*}(\omega_{i}) and ip(ωi)dir(1X(ωi))\sum_{i}p(\omega_{i})\cdot d_{i}^{r}\cdot(1-X_{*}(\omega_{i})) increase and decrease respectively. This contradicts the fact that the filter XOPTX^{OPT} is optimal for the receiver. ∎

Lemma 3 shows that we can assign XOPT(ω)=1X_{*}^{OPT}(\omega)=1 for all ωΩ00\omega\in\Omega_{0}^{0}, XOPT(ω)=0X_{*}^{OPT}(\omega)=0 for all ωΩ00\omega\in\Omega_{0}^{0}, and we have to assign either probability 11 or probability 0 to all but at most one of the remaining states. Ideally, we would like to assign probability 11 to all states in Ω01\Omega_{0}^{1} and probability 0 to all states in Ω10\Omega_{1}^{0}, since this guarantees that the receiver gets the maximum possible utility. However, this may not always be incentive-compatible for the sender, which means that we may have to assign to some of the states the probability that the sender prefers, as opposed to the probability that the receiver prefers. The following lemma characterizes these states.

Lemma 4.

Given a filtered information game ΓF=(S,A,Ω,p,u)\Gamma^{F}=(S,A,\Omega,p,u), let ω1,ω2,,ω\omega^{1},\omega^{2},\ldots,\omega^{\ell} be the states in Ω01Ω10\Omega_{0}^{1}\cup\Omega^{0}_{1} sorted by ur(ω,0)ur(ω,1)us(ω,1)us(ω,0)\frac{u_{r}(\omega,0)-u_{r}(\omega,1)}{u_{s}(\omega,1)-u_{s}(\omega,0)}. Then, there exists a binary filter XOPTX^{OPT} that is optimal for the receiver and a value \ell^{\prime}\leq\ell such that:

  • (a)

    XOPT(ω)=1X^{OPT}_{*}(\omega)=1 for all ωΩ00\omega\in\Omega_{0}^{0}.

  • (b)

    XOPT(ω)=0X^{OPT}_{*}(\omega)=0 for all ωΩ11\omega\in\Omega_{1}^{1}.

  • (c)

    For i<i<\ell^{\prime}, XOPT(ωi)=0X^{OPT}_{*}(\omega^{i})=0 if ωΩ10\omega\in\Omega_{1}^{0} and XOPT(ωi)=1X^{OPT}_{*}(\omega^{i})=1 if ωΩ01\omega\in\Omega_{0}^{1}.

  • (d)

    For i>i>\ell^{\prime}, XOPT(ωi)=0X^{OPT}_{*}(\omega^{i})=0 if ωΩ01\omega\in\Omega_{0}^{1} and XOPT(ωi)=0X^{OPT}_{*}(\omega^{i})=0 if ωΩ10\omega\in\Omega_{1}^{0}.

Lemma 4 states that, if we sort the states in which the sender and the receiver have different preferences by the ratio between how much the receiver prefers 0 with respect to 11 and how much the sender prefers 11 with respect to 0 (note that these ratios are always positive), we get that we can split these states into two contiguous blocks in which, in the first block, we assign them the probability that is most convenient for the sender, and for the second block we assign them the probability that is most convenient for the receiver. The proof follows the lines of that of Lemma 3 part (c).

Proof of Lemma 4.

Given two states ω,ωΩ01Ω10\omega,\omega^{\prime}\in\Omega_{0}^{1}\cup\Omega^{0}_{1}, we say that ω<ω\omega<\omega^{\prime} if ur(ω,0)ur(ω,1)us(ω,1)us(ω,0)<ur(ω,0)ur(ω,1)us(ω,1)us(ω,0)\frac{u_{r}(\omega,0)-u_{r}(\omega,1)}{u_{s}(\omega,1)-u_{s}(\omega,0)}<\frac{u_{r}(\omega^{\prime},0)-u_{r}(\omega^{\prime},1)}{u_{s}(\omega^{\prime},1)-u_{s}(\omega^{\prime},0)}. Using the same argument as in the Proof of Lemma 3, we can assume w.l.o.g. that (a) and (b) are satisfied. Thus, it only remains to show (c) and (d), which follow from the fact that there exists an there exists an optimal binary filter XOPTX^{OPT} for the receiver such that:

  • (s1)

    If ω,ωΩ10\omega,\omega^{\prime}\in\Omega_{1}^{0}, ω<ω\omega<\omega^{\prime} and XOPT(ω)>0X^{OPT}_{*}(\omega)>0, then XOPT(ω)=1X^{OPT}_{*}(\omega^{\prime})=1.

  • (s2)

    If ω,ωΩ01\omega,\omega^{\prime}\in\Omega_{0}^{1}, ω<ω\omega<\omega^{\prime} and XOPT(ω)<1X^{OPT}_{*}(\omega)<1, then XOPT(ω)=0X^{OPT}_{*}(\omega^{\prime})=0.

  • (s3)

    If ωΩ10\omega\in\Omega_{1}^{0}, ωΩ01\omega^{\prime}\in\Omega_{0}^{1}, ω<ω\omega<\omega^{\prime} and XOPT(ω)>0X^{OPT}_{*}(\omega)>0, then XOPT(ω)=0X^{OPT}_{*}(\omega^{\prime})=0.

  • (s4)

    If ωΩ01\omega\in\Omega_{0}^{1}, ωΩ10\omega^{\prime}\in\Omega_{1}^{0}, ω<ω\omega<\omega^{\prime} and XOPT(ω)<1X^{OPT}_{*}(\omega)<1, then XOPT(ω)=1X^{OPT}_{*}(\omega^{\prime})=1.

We will prove (s1) and (s3), the proofs of (s2) and (s4) are analogous.

Proof of (s1): Suppose that there exist ω,ωΩ10\omega,\omega^{\prime}\in\Omega_{1}^{0} such that ω<ω\omega<\omega^{\prime} and XOPT(ω)>0X^{OPT}_{*}(\omega)>0, but XOPT(ω)<1X^{OPT}_{*}(\omega^{\prime})<1. If this happens, we can set

XOPT(ω)XOPT(ω)+ϵ(us(ω,0)us(ω,1))XOPT(ω)XOPT(ω)+ϵ(us(ω,1)us(ω,0)).\begin{array}[]{l}X^{OPT}_{*}(\omega)\longleftarrow X^{OPT}_{*}(\omega)+\epsilon(u_{s}(\omega^{\prime},0)-u_{s}(\omega^{\prime},1))\\ X^{OPT}_{*}(\omega^{\prime})\longleftarrow X^{OPT}_{*}(\omega)+\epsilon(u_{s}(\omega,1)-u_{s}(\omega,0)).\end{array}

Since us(ω,0)us(ω,1))<0u_{s}(\omega^{\prime},0)-u_{s}(\omega^{\prime},1))<0 and us(ω,0)us(ω,1)<0u_{s}(\omega,0)-u_{s}(\omega,1)<0, there exists some ϵ>0\epsilon>0 that is small enough so that XOPT(ω)X^{OPT}_{*}(\omega) and XOPT(ω)X^{OPT}_{*}(\omega) stay between 0 and 11. Moreover, as in the proof of Lemma 3 part (c), if we perform this change the sender’s utility remains unchanged, but the receiver’s utility increases by

Δ:=ϵ((us(ω,0)us(ω,1))(ur(ω,0)ur(ω,1))+(us(ω,1)us(ω,0))(ur(ω,0)ur(ω,1))).\Delta:=\epsilon((u_{s}(\omega^{\prime},0)-u_{s}(\omega^{\prime},1))(u_{r}(\omega,0)-u_{r}(\omega,1))+(u_{s}(\omega,1)-u_{s}(\omega,0))(u_{r}(\omega^{\prime},0)-u_{r}(\omega^{\prime},1))).

By assumption, we have that

ur(ω,0)ur(ω,1)us(ω,1)us(ω,0)<ur(ω,0)ur(ω,1)us(ω,1)us(ω,0)\frac{u_{r}(\omega,0)-u_{r}(\omega,1)}{u_{s}(\omega,1)-u_{s}(\omega,0)}<\frac{u_{r}(\omega^{\prime},0)-u_{r}(\omega^{\prime},1)}{u_{s}(\omega^{\prime},1)-u_{s}(\omega^{\prime},0)}

which, together with the fact that us(ω,1)us(ω,0)>0u_{s}(\omega,1)-u_{s}(\omega,0)>0 and us(ω,1)us(ω,0)>0u_{s}(\omega^{\prime},1)-u_{s}(\omega^{\prime},0)>0, implies that Δ0\Delta\geq 0 whenever ϵ>0\epsilon>0.

Proof of (s3): The proof is almost identical to that of (s1). Suppose that there exist ωΩ10\omega\in\Omega_{1}^{0}, ωΩ01\omega^{\prime}\in\Omega_{0}^{1} such that ω<ω\omega<\omega^{\prime}, XOPT(ω)>0X^{OPT}_{*}(\omega)>0 and XOPT(ω)>0X^{OPT}_{*}(\omega^{\prime})>0. In this case, we can again set

XOPT(ω)XOPT(ω)+ϵ(us(ω,1)us(ω,0))XOPT(ω)XOPT(ω)+ϵ(us(ω,0)us(ω,1)).\begin{array}[]{l}X^{OPT}_{*}(\omega)\longleftarrow X^{OPT}_{*}(\omega)+\epsilon(u_{s}(\omega^{\prime},1)-u_{s}(\omega^{\prime},0))\\ X^{OPT}_{*}(\omega^{\prime})\longleftarrow X^{OPT}_{*}(\omega)+\epsilon(u_{s}(\omega,0)-u_{s}(\omega,1)).\end{array}

The only difference with the proof of (s1) is that, in this case, us(ω,0)us(ω,1)<0u_{s}(\omega,0)-u_{s}(\omega,1)<0 and us(ω,0)us(ω,1)>0u_{s}(\omega^{\prime},0)-u_{s}(\omega^{\prime},1)>0. Again, there exists a sufficiently small ϵ>0\epsilon>0 such that XOPT(ω)X^{OPT}_{*}(\omega) and XOPT(ω)X^{OPT}_{*}(\omega^{\prime}) remain between 0 and 11. Moreover, a straightforward computation shows that the sender’s utility remains unchanged, but that the receiver’s utility changes by

Δ:=ϵ((us(ω,1)us(ω,0))(ur(ω,0)ur(ω,1))+(us(ω,0)us(ω,1))(ur(ω,0)ur(ω,1))).\Delta:=\epsilon((u_{s}(\omega^{\prime},1)-u_{s}(\omega^{\prime},0))(u_{r}(\omega,0)-u_{r}(\omega,1))+(u_{s}(\omega,0)-u_{s}(\omega,1))(u_{r}(\omega^{\prime},0)-u_{r}(\omega^{\prime},1))).

Using that

ur(ω,0)ur(ω,1)us(ω,1)us(ω,0)<ur(ω,1)ur(ω,0)us(ω,0)us(ω,1),\frac{u_{r}(\omega,0)-u_{r}(\omega,1)}{u_{s}(\omega,1)-u_{s}(\omega,0)}<\frac{u_{r}(\omega^{\prime},1)-u_{r}(\omega^{\prime},0)}{u_{s}(\omega^{\prime},0)-u_{s}(\omega^{\prime},1)},

us(ω,1)us(ω,0)>0u_{s}(\omega,1)-u_{s}(\omega,0)>0, and us(ω,0)us(ω,1)>0u_{s}(\omega^{\prime},0)-u_{s}(\omega^{\prime},1)>0, we get that Δ0\Delta\geq 0 whenever ϵ>0\epsilon>0. ∎

Lemma 4 shows that the algorithm provided in Section 5.1 outputs the correct solution as long as it gets the right value of qq. The next lemma shows that this value is precisely the one that the algorithm finds. This completes the proof of correctness.

Lemma 5.

Let XrX^{r} be the filter that assigns Xr(ω)=1X^{r}_{*}(\omega)=1 for ωΩ00Ω10\omega\in\Omega_{0}^{0}\cup\Omega_{1}^{0} and Xr(ω)=0X^{r}_{*}(\omega)=0 for ωΩ11Ω01\omega\in\Omega_{1}^{1}\cup\Omega_{0}^{1}. If XrX^{r} is incentive-compatible for the sender, XrX^{r} is the optimal filter for the receiver that is incentive-compatible for both players. Otherwise, all filters XX that are incentive-compatible for both players and are optimal for the receiver satisfy at least one of the following equations:

ip(ωi)disX(ωi)=0ip(ωi)dis(1X(ωi))=0.\begin{array}[]{ll}\sum_{i}p(\omega_{i})\cdot d_{i}^{s}\cdot X_{*}(\omega_{i})=0\\ \sum_{i}p(\omega_{i})\cdot d_{i}^{s}\cdot(1-X_{*}(\omega_{i}))=0.\end{array}
Proof.

Filter XrX^{r} gives the maximum utility to the sender of all possible filters. Therefore, if it is incentive-compatible for the sender, it is the optimal for the receiver. Suppose instead that XrX^{r} is not incentive-compatible for the sender but that an optimal filter XOPTX^{OPT} satisfies

ip(ωi)disX(ωi)>0ip(ωi)dis(1X(ωi))<0.\begin{array}[]{ll}\sum_{i}p(\omega_{i})\cdot d_{i}^{s}\cdot X_{*}(\omega_{i})>0\\ \sum_{i}p(\omega_{i})\cdot d_{i}^{s}\cdot(1-X_{*}(\omega_{i}))<0.\end{array} (4)

Since XrX^{r} is not IC for the sender, there exists a state ωΩ0,1\omega\in\Omega_{0,1} such that XOPT>0X^{OPT}_{*}>0 or a state ωΩ1,0\omega\in\Omega_{1,0} such that XOPT<1X^{OPT}_{*}<1. In the first case, we can decrease XOPTX^{OPT}_{*} by a small value ϵ>0\epsilon>0 such that Equation 4 is still satisfied. By doing this, we increase the receiver’s expected utility while obtaining a new filter that is still incentive-compatible for both players. The latter case is analogous except that we increase XOPTX^{OPT}_{*} instead of decreasing it. ∎

6 Proof of Theorem 4

The proof of Theorem 4 follows from the following result of Arieli et al. [3] describing the characterization of the best Nash equilibrium in information aggregation games with two senders, one receiver, binary actions and a mediator.

Theorem 6 ([3]).

Let Γ={S,A,Ω,p,u}\Gamma=\{S,A,\Omega,p,u\} be an information aggregation game (without filters) with S={s1,s2}S=\{s_{1},s_{2}\} and A={0,1}A=\{0,1\} in which the players can communicate with a third-party mediator. Given i,j,k{0,1}i,j,k\in\{0,1\}, let Ωi,jk\Omega_{i,j}^{k} be the set of states in which s1s_{1} prefers action ii, s2s_{2} prefers action jj, and the receiver prefers action kk. Let also Ωi,j:=Ωi,j0Ωi,j1\Omega_{i,j}:=\Omega_{i,j}^{0}\cup\Omega_{i,j}^{1}. Consider the following six maps M1,M2,M3,M4,M5,M6M_{1},M_{2},M_{3},M_{4},M_{5},M_{6} from Ω\Omega to [0,1][0,1]:

  • (a)

    M2(ω)=1M_{2}(\omega)=1 for all ωΩ0,00\omega\in\Omega_{0,0}^{0} and M2(ω)=0M_{2}(\omega)=0 otherwise.

  • (b)

    M1(ω)=0M_{1}(\omega)=0 for all ωΩ1,11\omega\in\Omega_{1,1}^{1} and M1(ω)=1M_{1}(\omega)=1 otherwise.

  • (c)

    M3(ω)=1M_{3}(\omega)=1 for all ωΩ0,0Ω0,1\omega\in\Omega_{0,0}\cup\Omega_{0,1} and M3(ω)=0M_{3}(\omega)=0 otherwise.

  • (d)

    M4(ω)=1M_{4}(\omega)=1 for all ωΩ0,0Ω1,0\omega\in\Omega_{0,0}\cup\Omega_{1,0} and M4(ω)=0M_{4}(\omega)=0 otherwise.

  • (e)

    M5(ω)=1M_{5}(\omega)=1 for all ωΩ\omega\in\Omega.

  • (f)

    M6(ω)=0M_{6}(\omega)=0 for all ωΩ\omega\in\Omega.

Then, there exists an i{1,2,3,4,5,6}i\in\{1,2,3,4,5,6\} and a Nash equilibrium of Γ\Gamma that maximizes the receiver’s utility in which the function that maps each state to the probability that the receiver ends up playing 0 is equal to MiM_{i}.

Intuitively, Theorem 6 states that if there were no filters and the two senders had access to a third-party mediator, there exists a Nash equilibrium that is optimal for the receiver in which the either (a) the receiver only plays 0 if all three players prefer 0, (b) the sender plays 11 if and only if all three players prefer 11, (c) the receiver always plays what the first sender prefers, (d) the receiver always plays what the second sender prefers, (e) the receiver plays always 0, or (f) the receiver plays always 1. In the setting described in Section 3.1, players have no access to a mediator. However, we can easily check if the outcomes described in (a), (b), (c), (d), (e) or (f) are incentive-compatible for the receiver (i.e., if the receiver gets more utility with the outcome than when playing with no information), there exists a strategy in Γ\Gamma (without the mediator) that implements these outcomes.

For instance, suppose that we want to implement the outcome described in (a). Consider the strategy profile σ(0,0)\vec{\sigma}^{(0,0)} in which each sender sends a binary signal m{0,1}m\in\{0,1\} to the receiver that is equal to 0 if and only if the realized state is in Ω0,00\Omega_{0,0}^{0}, and the receiver plays 0 only if both signals are 0. It is easy to check that this is incentive-compatible for the senders: if the realized state is indeed in Ω0,00\Omega_{0,0}^{0}, it is a best response for both to send 0 since it guarantees that the receiver will play 0. Moreover, if the realized state is not in Ω0,00\Omega_{0,0}^{0}, none of the senders gets any additional utility by defecting from the main strategy since the other sender will always send signal 11 (which implies that the receiver will play 11). An analogous reasoning gives a strategy profile σ(1,1)\vec{\sigma}^{(1,1)} that implements (b) and is incentive-compatible for the senders. To implement (c), consider the strategy profile σs1\vec{\sigma}^{s_{1}} in which both senders send a binary signal m{0,1}m\in\{0,1\} that is equal to 0 if and only if the realized state is in Ω0,0Ω0,1\Omega_{0,0}\cup\Omega_{0,1}, and the receiver plays 0 if and only if the signal from the first sender is 0. Again, it is straightforward to check that this is incentive-compatible for the senders: it is a best-response for sender 11 to send its preference, while it doesn’t matter what sender 22 sends since it will be ignored. Implementing (d) with a strategy profile σs2\vec{\sigma}^{s_{2}} that is incentive-compatible for the senders is analogous. To implement (e) (resp., (f)) consider the strategy profile σ0\vec{\sigma}^{0} (resp., σ1\vec{\sigma}^{1}) in which the senders send signal 0 regardless of the realized state and the receiver always plays 0 (resp., 11). It is easy to check that σ0\vec{\sigma}^{0} (resp., σ1\vec{\sigma}^{1}) are incentive-compatible for the senders. Since all outcomes that are implementable without a mediator are also implementable with a mediator, this implies the following proposition:

Proposition 6.

Let Γ={S,A,Ω,p,u}\Gamma=\{S,A,\Omega,p,u\} be an information aggregation game with S={s1,s2}S=\{s_{1},s_{2}\} and A={0,1}A=\{0,1\}. Then, either σ0\vec{\sigma}^{0}, σ1\vec{\sigma}^{1}, σ(0,0)\vec{\sigma}^{(0,0)}, σ(1,1)\vec{\sigma}^{(1,1)}, σs1\vec{\sigma}^{s_{1}} or σs2\vec{\sigma}^{s_{2}} is a Nash equilibrium that is optimal for the receiver.

Proposition 6 implies that we can break the problem of finding the best Nash equilibrium for the receiver in a filtered information aggregation game into four sub-problems:

  • (a)

    Finding a filter X(0,0)OPTX^{OPT}_{(0,0)} such that σ(0,0)\vec{\sigma}^{(0,0)} gives the maximal utility for the receiver in ΓX(0,0)OPTF\Gamma^{F}_{X^{OPT}_{(0,0)}}.

  • (b)

    Finding a filter X(1,1)OPTX^{OPT}_{(1,1)} such that σ(1,1)\vec{\sigma}^{(1,1)} gives the maximal utility for the receiver in ΓX(1,1)OPTF\Gamma^{F}_{X^{OPT}_{(1,1)}}.

  • (c)

    Finding a filter Xs1OPTX^{OPT}_{s_{1}} such that σs1\vec{\sigma}^{s_{1}} gives the maximal utility for the receiver in ΓXs1OPTF\Gamma^{F}_{X^{OPT}_{s_{1}}}.

  • (d)

    Finding a filter Xs2OPTX^{OPT}_{s_{2}} such that σs2\vec{\sigma}^{s_{2}} gives the maximal utility for the receiver in ΓXs2OPTF\Gamma^{F}_{X^{OPT}_{s_{2}}}.

Note that these filters might not always exist (for instance, X(0,0)OPTX^{OPT}_{(0,0)} does not exist when the receiver always prefers 0 and the senders always prefer 11). Moreover, we are not including the optimal filters for σ0\vec{\sigma}^{0} and σ1\vec{\sigma}^{1} since all filters give the same utility with these strategies. Finding Xs1OPTX^{OPT}_{s_{1}} and Xs2OPTX^{OPT}_{s_{2}} (whenever they exist) reduce to the case of one sender (by ignoring s2s_{2} and s1s_{1}, respectively), and thus can be solved with the O(klogk)O(k\log k) algorithm presented in Section 5.1. We next show how to compute X(0,0)OPTX^{OPT}_{(0,0)} using a linear program with kk variables and 2k+22k+2 constraints. Computing X(1,1)OPTX^{OPT}_{(1,1)} is analogous.

Suppose that XX is a filter that maximizes the receiver utility when playing strategy σ(0,0)\vec{\sigma}^{(0,0)}. Denote by M0M_{0} the set of signals of XX in which both senders and the receiver prefer 0 to 11. Consider a filter XX^{\prime} that samples a signal mm according to XX and does the following: if mM0m\in M_{0}, it sends signal 0. Otherwise, it sends signal 11. It is straightforward to check that if σ(0,0)\vec{\sigma}^{(0,0)} is a Nash equilibrium in ΓXF\Gamma^{F}_{X}, it is also a Nash equilibrium in ΓXF\Gamma^{F}_{X^{\prime}}. Moreover, players get the same utilities with XX and XX^{\prime}, which means that XX^{\prime} also maximizes the receiver’s utility. This implies that we can restrict our search to binary filters in which all players prefer action 0 on signal 0 and the receiver prefers action 11 on signal 11.

Our aim then is finding a binary filter XX such that both senders and the receiver prefer action 0 on signal 0, the receiver prefers action 11 on signal 11, and such that the receiver gets the maximal possible utility by playing 0 on signal 0 and 11 on signal 11. Let ω1,ω2,,ωk\omega_{1},\omega_{2},\ldots,\omega_{k} be the elements of Ω\Omega and define Ai:=u(s1,ωi,0)u(s1,ωi,1)A_{i}:=u(s_{1},\omega_{i},0)-u(s_{1},\omega_{i},1), Bi:=u(s2,ωi,0)u(s2,ωi,1)B_{i}:=u(s_{2},\omega_{i},0)-u(s_{2},\omega_{i},1) and Ci:=u(r,ωi,0)u(r,ωi,1)C_{i}:=u(r,\omega_{i},0)-u(r,\omega_{i},1). A binary filter XX over Ω\Omega can be described by a sequence of kk real numbers x1,,xkx_{1},\ldots,x_{k} between 0 and 11 such that xix_{i} denotes the probability that the filter sends signal 0 on state ωi\omega_{i}. The condition that both senders prefer action 0 on signal 0 translates into

i=1kAixi0i=1kBixi0.\begin{array}[]{c}\sum_{i=1}^{k}A_{i}x_{i}\geq 0\\ \sum_{i=1}^{k}B_{i}x_{i}\geq 0.\\ \end{array}

Moreover, the utility of the receiver with filter XX and strategy σ(0,0)\vec{\sigma}^{(0,0)} is given by i=1k(xiu(r,ωi,0)+(1xi)u(r,ωi,1))\sum_{i=1}^{k}(x_{i}u(r,\omega_{i},0)+(1-x_{i})u(r,\omega_{i},1)). This sum can be rearranged into i=1ku(r,ωi,1)+i=1kCixi\sum_{i=1}^{k}u(r,\omega_{i},1)+\sum_{i=1}^{k}C_{i}x_{i}. Therefore, finding a filter X(0,0)OPTX^{OPT}_{(0,0)} such that σ(0,0)\vec{\sigma}^{(0,0)} gives the maximal utility for the receiver in ΓX(0,0)OPTF\Gamma^{F}_{X^{OPT}_{(0,0)}} reduces to solving the following linear programming instance:

maxi=1kCixii=1kAixi0i=1kBixi00xi1i[k].\begin{array}[]{c}\max\sum_{i=1}^{k}C_{i}x_{i}\\ \vspace{1mm}\sum_{i=1}^{k}A_{i}x_{i}\geq 0\\ \vspace{1mm}\sum_{i=1}^{k}B_{i}x_{i}\geq 0\\ \vspace{1mm}0\leq x_{i}\leq 1\quad\forall i\in[k].\end{array}

Note that, even though we are maximizing the receiver’s utility, it may be the case that σ(0,0)\vec{\sigma}^{(0,0)} is not incentive-compatible for the receiver in ΓX(0,0)OPTF\Gamma^{F}_{X^{OPT}_{(0,0)}}. For instance, the receiver might prefer playing 0 whenever the senders send 11. If such a thing happens (which can be tested in linear time), it means that there is no filter satisfying the desired conditions.

7 Extending Nash to sequential equilibria

For simplicity, we showed that the strategies provided in the proofs of Theorems 1, 2, 3 and 4 are Nash equilibria. In this section we show how to extend them to sequential equilibria. This means that, for each of these strategy profiles, we have to provide a compatible belief assessment bb and argue that the behavior of the players defined off the equilibrium path is a best-response according to their beliefs.

Note that, by construction, the only time that a player can be off the equilibrium path is when the receiver receives a message that the sender was not supposed to send under any circumstances (i.e., a message that had probability 0 of being sent). In the proofs of the Theorems stated above, the strategy profiles that we constructed either had the receiver play the best action for her in the bubbling equilibrium (i.e., the best action for the receiver without any information), or to play the action that gives the least utility to the sender between all the relevant actions (i.e., the actions that are not dominated by combinations of other actions). We show next that there exist belief assessments that are compatible with the given strategy profiles in which playing these actions is justified. More generally, we show that, for all relevant actions aa, there exists a belief assessment that justifies playing aa whenever the receiver receives an impossible message.

We prove the claim above for the case of one sender, the extension to Theorem 4 is straightforward. Denote by Ω={ω1,ω2,,ωn}\Omega=\{\omega_{1},\omega_{2},\ldots,\omega_{n}\} the set of states and suppose that, with the proposed strategy σs\sigma_{s}, the sender sends messages in MM following a certain function msg:ΩΔ(M)msg:\Omega\to\Delta(M). Let aa be an action that is optimal for the receiver at some posterior distribution λΔ(Ω)\lambda\in\Delta(\Omega) and let mm be a message such that msg(ω)(m)=0msg(\omega)(m)=0 for all ωΩ\omega\in\Omega. Consider a sequence of strategies σs1,σs2,\sigma_{s}^{1},\sigma_{s}^{2},\ldots for the sender in which, in each strategy σsk\sigma_{s}^{k}, the sender computes the message that she would send to the receiver using function msgmsg. However, before sending the message, the sender tosses a coin that lands tails with probability λ(ω)k\frac{\lambda(\omega)}{k}. If the coin lands heads, she sends the computed message. Otherwise, she sends message mm. Clearly, the sequence σs1,σs2,\sigma_{s}^{1},\sigma_{s}^{2},\ldots converges to σs\sigma_{s}. Moreover, the posterior distribution over states induced by each of these strategies conditioned to the fact that the receiver received message mm is precisely λ\lambda. This means that playing action aa is always a best-response for the receiver when receiving message mm. Therefore, in the belief assessment induced by σs1,σs2,\sigma_{s}^{1},\sigma_{s}^{2},\ldots, aa is also a best response to message mm.

8 Conclusion

This paper diverges from traditional sender-receiver models, emphasizing the receiver’s perspective in cheap talk interactions. We explore two models: one, with transparent motives and state-independent utility, provides a geometric profile of the optimal receiver equilibrium; the second introduces a pre-play stage, empowering the receiver to influence information available to the sender. Our approach aligns with user-centric platforms, where receivers control information, similar to platforms like Amazon. Our findings yield nuanced insights into communication dynamics, enhancing our comprehension of strategic interactions. An intriguing avenue not explored in this paper is the integration of a third intermediary controlling senders’ information to maximize the intermediary’s utility.

References

  • [1] Ricardo Alonso and Odilon Câmara. Persuading voters. American Economic Review, 106(11):3590–3605, 2016.
  • [2] Itai Arieli and Yakov Babichenko. Private bayesian persuasion. Journal of Economic Theory, 182:185–217, 2019.
  • [3] Itai Arieli, Ivan Geffner, and Moshe Tennenholtz. Mediated cheap talk design. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pages 5456–5463, 2023.
  • [4] Itai Arieli, Ivan Geffner, and Moshe Tennenholtz. Resilient information aggregation. In Rineke Verbrugge, editor, Proceedings Nineteenth conference on Theoretical Aspects of Rationality and Knowledge, TARK 2023, Oxford, United Kingdom, 28-30th June 2023, volume 379 of EPTCS, pages 31–45, 2023.
  • [5] Robert J Aumann, Michael Maschler, and Richard E Stearns. Repeated games with incomplete information. MIT press, 1995.
  • [6] Dirk Bergemann, Benjamin Brooks, and Stephen Morris. The limits of price discrimination. American Economic Review, 105(3):921–957, 2015.
  • [7] Sourav Bhattacharya and Arijit Mukherjee. Strategic information revelation when experts compete to influence. The RAND Journal of Economics, 44(3):522–544, 2013.
  • [8] David Blackwell. Equivalent comparisons of experiments. The annals of mathematical statistics, pages 265–272, 1953.
  • [9] Andreas Blume and Oliver Board. Language barriers. Econometrica, 81(2):781–812, 2013.
  • [10] Vincent P Crawford and Joel Sobel. Strategic information transmission. Econometrica: Journal of the Econometric Society, pages 1431–1451, 1982.
  • [11] Kris De Jaegher. A game-theoretic rationale for vagueness. Linguistics and Philosophy, 26(5):637–659, 2003.
  • [12] Matthew Gentzkow and Emir Kamenica. Competition in persuasion. The Review of Economic Studies, 84(1):300–322, 2016.
  • [13] Jeanne Hagenbach and Frédéric Koessler. Cheap talk with coarse understanding. Games and Economic Behavior, 124:105–121, 2020.
  • [14] Shota Ichihashi. Limiting sender’s information in bayesian persuasion. Games and Economic Behavior, 117:276–288, 2019.
  • [15] Gerhard Jäger, Lars P Metzger, and Frank Riedel. Voronoi languages: Equilibria in cheap-talk games with high-dimensional types and few signals. Games and economic behavior, 73(2):517–537, 2011.
  • [16] Emir Kamenica and Matthew Gentzkow. Bayesian persuasion. American Economic Review, 101(6):2590–2615, 2011.
  • [17] Elliot Lipnowski and Doron Ravid. Cheap talk with transparent motives. Econometrica, 88(4):1631–1660, 2020.