This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Observational Learning with Fake Agents

Pawan Poojary and Randall Berry This work was supported in part by the NSF under grants CNS-1701921 and CNS-1908807. ECE Department, Northwestern University, Evanston IL 60208
Email: [email protected], [email protected]
Abstract

It is common in online markets for agents to learn from other’s actions. Such observational learning can lead to herding or information cascades in which agents eventually “follow the crowd”. Models for such cascades have been well studied for Bayes-rational agents that choose pay-off optimal actions. In this paper, we additionally consider the presence of fake agents that seek to influence other agents into taking one particular action. To that end, these agents take a fixed action in order to influence the subsequent agents towards their preferred action. We characterize how the fraction of such fake agents impacts behavior of the remaining agents and show that in certain scenarios, an increase in the fraction of fake agents in fact reduces the chances of their preferred outcome.

I Introduction

Consider a new item that is up for sale in a recommendation-based market where agents arrive sequentially and decide whether to buy the item, with their choice serving as a recommendation for later agents. Every agent makes a pay-off optimal decision based on his own prior knowledge of the item’s quality/utility and by observing the choices of his predecessors. An informational cascade or herding occurs when it is optimal for an agent to ignore his own private signal and follow the actions of the past agents. Subsequent agents follow suit and from the onset of a cascade, the agents’ actions do not reveal any information conveyed to them by their private signals. Thus the phenomenon of cascading could prevent the agents from learning the socially optimal (right) choice, i.e. the agents could herd to a wrong cascade.

Cascades were first studied in [1, 2, 3] as a Bayesian learning model in which a homogeneous sequence of rational agents take pay-off optimal actions. In this paper, in addition to rational agents, we introduce randomly arriving fake agents that always take (or fake) a fixed action, regardless of their pay-off, in order to influence the outcome of a cascade. For example, this could model a setting where a poor quality item is posted for sale on a website that records buyers’ decisions, and fake agents intentionally buy (or appear to buy) this item to make it seem worth buying to future buyers. The model could also represent a setting where rational actions are manipulated or where fake actions are inserted into a recommendation system. The objective is to study the impact of varying the amount of these fake agents on the probability of their preferred cascade. Our main result shows a counter-intuitive phenomenon: the probability with which this preferred cascade occurs is not monotonically increasing in the fraction of fake agents, ϵ\epsilon. In fact, in some cases the presence of fake agents can reduce the chance of their preferred cascade.

As in [1, 2, 3], we consider a setting in which all agents can observe all prior actions (although in our case, each action could be either rational or fake). Many other variations of this basic model have been studied such as [4], where observations depend on an underlying network structure and [5], where agents are allowed to select their observations at a cost.

Closer to our work is the model in [6], which assumes that the recording of actions for subsequent agents is subject to an error that is unbiased towards each of the possible actions. In our setting, an action being either fake or rational depending on the agent-type could equivalently be perceived as an error while recording a rational action (as in [6]); except that in our case, the error is biased only towards a preferred action111This change requires a different analysis approach than that in [6] as the underlying Markov chain now typically has a countably infinite state-space..

There is also a body of work that considers agents similar to our fake agents, who only take a single action regardless of the true state. This includes the crazy agents considered in [7], stubborn agents in [8] and zealots in [9]. While [7] relaxes the assumption of binary signals, it does not consider how changing the fraction of crazy agents affects the cascade probability, which is the main focus of our work. In [8] a non-Bayesian model for learning was considered instead of the Bayesian model considered here. Other related cascade models are investigated in [10, 11, 12, 13, 14, 15, 16, 17].

The paper is organized as follows. We describe our model in Section II. We analyze this model and identify the resulting cascade properties in Section III. In Section IV, we present our Markov process formulation and in Section V, we consider the limiting scenario where the proportion of fake agents approaches unity. Section VI characterizes the asymptotic welfare of rational agents. We conclude in Section VII.

II Model

We consider a model similar to [1] in which there is a countable sequence of agents, indexed i=1,2,i=1,2,\ldots where the index represents both the time and the order of actions. Each agent ii takes an action AiA_{i} of either buying (Y)(Y) or not buying (N)(N) a new item that has a true value (V)(V) which could either be good (G)(G) or bad (B)(B). For simplicity, both possibilities of VV are assumed to be equally likely. The agents are Bayes-rational utility maximizers where the pay-off received by each agent ii, denoted by πi\pi_{i}, depends on its action AiA_{i} and the true value VV as follows. If the agent chooses NN, his payoff is 0. Whereas, if the agent chooses YY, he incurs a cost of C=1/2C=1/2 for buying the item, and gains the amount 0 if V=BV=B and 11 if V=GV=G. The buyer’s net pay-off is then the gain minus the cost of buying the item. Observe that when VV is equiprobable, the ex ante expected pay-off is 0 for either of the actions. Thus, at the start, an agent is indifferent to the two actions.

To incorporate agents’ private beliefs about the new item, every agent ii receives a private signal Si{H(high),L(low)}S_{i}\in\{H\,\text{(high)},L\,\text{(low)}\}. This signal, as shown in Figure 1(a)(a), partially reveals the information about the true value of the item through a binary symmetric channel (BSC) with crossover probability 1p1-p, where 1/2<p<1\nicefrac{{1}}{{2}}<p<1. Each agent ii takes a rational action AiA_{i} that depends on his private signal SiS_{i} and the past observations {O1,O2,,Oi1}\{O_{1},O_{2},\ldots,O_{i-1}\} of actions {A1,A2,,Ai1}\{A_{1},A_{2},\ldots,A_{i-1}\}. Next, we modify the information structure in [1] by assuming that at each time instant, an agent could either be fake with probability (w.p.) ϵ[0,1)\epsilon\in[0,1) or ordinary w.p. 1ϵ1-\epsilon. An ordinary agent ii honestly reports his action, i.e. Oi=AiO_{i}=A_{i}. On the contrary, a fake agent always reports a YY, reflecting his intention of influencing the successors into buying the new item, regardless of its true value. This implies that at any time ii, with probability 1ϵ1-\epsilon, the reported action OiO_{i} satisfies Oi=AiO_{i}=A_{i} and with probability ϵ\epsilon, Oi=YO_{i}=Y. Refer to Figure 1(b)(b).

An equivalent model is where a fake agent always takes an action YY, while Oi=AiO_{i}=A_{i} for all agents. This yields the same information structure as the model above. However we chose the former model mainly to simplify our notation.

HHLLGGBBVVSiS_{i}

pp

pp

1p1-p

1p1-p

(a)
YYNNYYNNAiA_{i}OiO_{i}

11

1ϵ1-\epsilon

ϵ\epsilon

(b)
Figure 1: (a)(a) The BSC through which agents receive private signals. (b)(b) The channel through which agents’ actions are corrupted.

III Optimal decision and cascades

For the nthn^{\text{th}} agent, let the history of past observations be denoted by n1={O1,O2,,On1}\mathcal{H}_{n-1}=\{O_{1},O_{2},\ldots,O_{n-1}\}. As the first agent does not have any observation history, he always follows his private signal i.e., he buys if and only if the signal is HH. For the second agent onwards, the Bayes’ optimal action for every agent nn, AnA_{n} is chosen according to the hypothesis (V=GV=G or BB) that has the higher posterior probability given the information set 𝕀n={Sn,n1}\mathbb{I}_{n}=\{S_{n},\mathcal{H}_{n-1}\}. Let γn(Sn,n1)(G|Sn,n1)\gamma_{n}(S_{n},\mathcal{H}_{n-1})\triangleq\mathbb{P}(G|S_{n},\mathcal{H}_{n-1}) denote the posterior probability for the item being good, V=GV=G . Then the Bayes’ optimal decision rule is:

An={Y,ifγn>1/2,N,ifγn<1/2,followsSn,ifγn=1/2.\displaystyle A_{n}=\begin{cases}Y,\quad&\text{if}\;\;\gamma_{n}>1/2,\\ N,\quad&\text{if}\;\;\gamma_{n}<1/2,\\ \text{follows}\;\;S_{n},\quad&\text{if}\;\;\gamma_{n}=1/2.\end{cases} (1)

Note that when γn=1/2\gamma_{n}=1/2, an agent is indifferent between the actions. Similar to [6], our decision rule in this case follows the private signal SnS_{n}, unlike [1] where they employ random tie-breaking.

Definition 1

An information cascade is said to occur when an agent’s decision becomes independent of his private signal

It follows from (1) that, agent nn cascades to a YY (N)(N) if and only if γn>1/2\gamma_{n}>1/2 (<1/2)(<1/2) for all Sn{H,L}S_{n}\in\{H,L\}. The other case being γn>1/2\gamma_{n}>1/2 for Sn=HS_{n}=H and less than 1/21/2 for Sn=LS_{n}=L; in which case, agent nn follows SnS_{n}. A more intuitive way to express this condition is by using the information contained in the history n1\mathcal{H}_{n-1} observed by agent nn in the form of the likelihood ratio ln1(n1)(n1|B)/(n1|G)l_{n-1}(\mathcal{H}_{n-1})\triangleq\mathbb{P}(\mathcal{H}_{n-1}|B)/\mathbb{P}(\mathcal{H}_{n-1}|G) as follows [7].

Lemma 1

Agent nn cascades to a Y(N)Y\,(N) if and only if ln1<1ppl_{n-1}<\frac{1-p}{p} (ln1>p1p)\big{(}l_{n-1}>\frac{p}{1-p}\big{)} and otherwise follows its private signal SnS_{n}.

This lemma follows by expressing γn\gamma_{n} in terms of ln1l_{n-1} using Bayes’ law, and then using the condition on γn\gamma_{n} for a Y(N)Y\,(N) cascade. If agent nn cascades, then the observation OnO_{n} does not provide any additional information about the true value VV to the successors than what is contained in n1\mathcal{H}_{n-1}. As a result, ln+i=ln1l_{n+i}=l_{n-1} for all i=0,1,2,i=0,1,2,\ldots and hence they remain in the cascade, which leads us to the following property, also exhibited by prior models [1, 2, 3, 6].

Property 1

Once a cascade occurs, it lasts forever.

On the other hand, if agent nn does not cascade, then Property 1 and Lemma 1 imply that all the agents until and including nn follow their own private signals ignoring the observations of their predecessors. Mutual independence between the private signals results in the likelihood ratio lnl_{n} depending only on the number of YY’s (denoted by nYn_{Y}) and NN’s (denoted by nNn_{N}) in the observation history n\mathcal{H}_{n}. Specifically, ln=(1pp)hnl_{n}=\left(\frac{1-p}{p}\right)^{h_{n}} where hn=ηnYnNh_{n}=\eta n_{Y}-n_{N} is the difference between the number of YY’s weighted by η\eta and the number of NN’s. Here, the weight ηlog(a1b)/log(p1p)\eta\triangleq\log(\frac{a}{1-b})/\log(\frac{p}{1-p}), where for all i=1,2,,n1i=1,2,\ldots,n-1,

a=(Oi=Y|V=G)andb=(Oi=N|V=B)\displaystyle a=\mathbb{P}(O_{i}=Y|V=G)\quad\text{and}\quad b=\mathbb{P}(O_{i}=N|V=B)

give the probabilities that the observations follow the true value VV, given that these agents follow their own private signals. It can be shown from Figures 1(a)(a) and 1(b)(b) that in the above case, i.e., when AiA_{i} follows SiS_{i},

a=p+(1p)ϵandb=p(1ϵ).\displaystyle a=p+(1-p)\epsilon\;\;\text{and}\;\;b=p(1-\epsilon).

Thus, agents that have not yet entered a cascade satisfy the following property.

Property 2

Until a cascade occurs, each agent follows its private signal. Moreover, hnh_{n} is a sufficient statistic of the information contained in the past observations.

Note that if ϵ=0\epsilon=0 (no fake agents) then a=b=pa=b=p and η=1\eta=1, in which case hnh_{n} is the same as in [6]. The expression for hnh_{n} shows that, due to the presence of fake agents, the dependence of an agent’s decision on a YY in his observation history reduces by a factor of η\eta, whereas the dependence on a NN remains unaffected. This is to be expected because, unlike a NN which surely comes from an honest agent, a YY incurs the possibility that the agent could be fake. Further, this reduced dependence on YY is exacerbated with an increase in the possibility of fake agents, as η\eta reduces with an increase in ϵ\epsilon.

Using the expression for lnl_{n} in Lemma 1, it follows that until a cascade occurs, 1hn1-1\leq h_{n}\leq 1 for all such times nn, and the update rule for hnh_{n} is given by

hn={hn1+ηifOn=Y,hn11ifOn=N.\displaystyle h_{n}=\begin{cases}h_{n-1}+\eta\quad&\text{if}\;\;O_{n}=Y,\\ h_{n-1}-1\quad&\text{if}\;\;O_{n}=N.\end{cases} (2)

Whereas, once hn>1h_{n}>1 (<1<-1), a Y(N)Y\,(N) cascade begins and hnh_{n} stops updating (Property 1).

IV Markovian Analysis of Cascades

Let us, for the sake of discussion, consider the realization V=BV=B . Similar arguments hold when V=GV=G . It follows from the previous section that the process {hn}\{h_{n}\} is a discrete-time Markov process taking values in [1,1][-1,1] before getting absorbed into the left wall (<1)(<-1) causing a NN cascade or the right wall (>1)(>1) causing a YY cascade. More specifically, equation (2) shows that, until a cascade occurs, {hn}\{h_{n}\} is a random walk that shifts to the right by η\eta w.p. (On=Y|B)=1b\mathbb{P}(O_{n}=Y|B)=1-b or to the left by 11 w.p. (On=N|B)=b\mathbb{P}(O_{n}=N|B)=b. Note that the walk starts at state 0 since the first agent has no observation history.

The random walk is depicted in Figure 2 where pf(On=Y|V)p_{f}\triangleq\mathbb{P}(O_{n}=Y|V) denotes the probability of a YY being observed. Depending on the item’s true value, pf=ap_{f}=a for V=GV=G whereas pf=1bp_{f}=1-b for V=BV=B. Note that in the special case where η\eta satisfies 1/η=r1/\eta=r for some r=1,2,r=1,2,\ldots, the process {hn}\{h_{n}\} is equivalent to a Markov Chain with finite state-space 𝒜={r1,r,,1,0,1,,r,r+1}\mathcal{A}=\{-r-1,-r,\ldots,-1,0,1,\ldots,r,r+1\} , and with r1-r-1 and r+1r+1 being absorption states corresponding to NN and YY cascades respectively. In this case, absorption probabilities can be obtained by solving a system of linear equations. In this paper, our main focus is on the more common case of non-integer values of 1/η1/\eta resulting in {hn}\{h_{n}\} taking possibly uncountable values in [1,1][-1,1]222For example, if η\eta was chosen uniformly at random, then with probability 11 it would fall into this case..

0

pfp_{f}

η\eta

pfp_{f}

2η2\eta

1pf1-p_{f}

1-111
Figure 2: Transition diagram of the random walk.
Refer to caption
Figure 3: Thresholds ϵr\epsilon_{r} for the indicated values of rr versus pp.

IV-A Error thresholds

In the absence of fake agents (ϵ=0)(\epsilon=0) as in [1], η=1\eta=1 and so cascading to a Y(N)Y(N) cascade requires at least two consecutive YY’s (NN’s). However, in the presence of fake agents, even a single NN after a YY could trigger a NN cascade. On the other hand, as ϵ\epsilon increases, a greater number of consecutive YY’s (2)(\geq 2) are required to cause a YY cascade. This is characterized in the following lemma.

Lemma 2

Let α=p/(1p)\alpha=p/(1-p). For r=1,2,r=1,2,\ldots define the increasing sequence of thresholds {ϵr}r=1\{\epsilon_{r}\}_{r=1}^{\infty}, where the rthr^{\text{th}} threshold ϵr\epsilon_{r} is given as:

ϵr=αα1rα1r+11.\displaystyle\epsilon_{r}=\frac{\alpha-\alpha^{\frac{1}{r}}}{\alpha^{\frac{1}{r}+1}-1}. (3)

Define r[ϵr,ϵr+1)\mathcal{I}_{r}\triangleq[\epsilon_{r},\epsilon_{r+1}) as the rthr^{\text{th}} ϵ\epsilon–interval. Then for ϵr\epsilon\in\mathcal{I}_{r}, at least r+1r+1 consecutive YY’s are necessary for a YY cascade to begin.

The proof follows by noting that, given that if a cascade has not begun, when an NN is observed, then the rightmost position that the random walk can be in, is in state 0. From here, at least r+1r+1 consecutive YY’s would be needed only if η1/r\eta\leq 1/r, thereby giving the rthr^{\text{th}} threshold ϵr\epsilon_{r}. It follows that ϵr\epsilon\in\mathcal{I}_{r} implies 1r+1<η1r\frac{1}{r+1}<\eta\leq\frac{1}{r} .

Remark 1

For ϵr\epsilon\in\mathcal{I}_{r}, staring from state 0, r+1r+1 consecutive YY’s start a YY cascade.

Moreover, given ϵ\epsilon, the integer rr such that ϵr\epsilon\in\mathcal{I}_{r} can be readily obtained from Lemma 2 as r=1/ηr=\left\lfloor 1/\eta\right\rfloor. Here, recall that η\eta is a function of ϵ\epsilon.

Figure 3 shows the thresholds ϵr\epsilon_{r} varying with pp, for different values of rr. For a fixed ϵ\epsilon, we see that as the signal quality (pp) improves, more consecutive YY’s are required for a YY cascade to begin. This is because, an increase in pp increases the information contained in a YY, but not as much as the corresponding increase in the information contained in a NN. Further, note that as ϵ1\epsilon\rightarrow 1, rr\rightarrow\infty which implies that infinitely many consecutive YY’s are required for a YY cascade to begin. Equivalently, the information contained in a YY observation becomes negligible. We further investigate learning in this asymptotic scenario in Section V.

IV-B YY cascade probability

In this subsection, we will compute the probability of absorption of {hn}\{h_{n}\} to a YY cascade (right wall). The following iterative process depicted in Figure 4 describes all possible sequences that can lead to a YY cascade. For the rest of the paper, we assume ϵr{ϵr}\epsilon\in\mathcal{I}_{r}\smallsetminus\{\epsilon_{r}\} for some r=1,2,r=1,2,\ldots. We initialize Stage 11 with r1=r+1r_{1}=r+1. Now, starting from state 0, consider the sequences shown in Stage 11 in Figure 4. The first sequence of r1r_{1} consecutive YY’s, denoted by Yr1Y^{r_{1}}, clearly hits the right wall (Remark 1). The rest of the sequences, each of length r1+1r_{1}+1, are simply permutations of each other that contain only a single NN. Two NN’s or more are not possible without hitting the left wall. So each of these r1r_{1} distinct sequences results in the same net right shift, which ends somewhere in the region [0,η][0,\eta]. From here, it would take either rr or r+1r+1 consecutive YY’s to hit the right wall. Let this value be denoted by r2r_{2}. The sequences that would then be possible from this point onwards could again be enumerated exactly as in the first stage, except that r2r_{2} now replaces r1r_{1}. This forms Stage 22. Now, unless there are r2r_{2} consecutive YY’s, the second stage would again end in the region [0,η][0,\eta], and then the process continues to the next stage. Here, rnr_{n} denotes the number of consecutive YY’s required to hit the right wall in the nthn^{\text{th}} stage. Note that by considering ϵr{ϵr}\epsilon\in\mathcal{I}_{r}\smallsetminus\{\epsilon_{r}\}, we avoid the case of integer values of 1/η1/\eta which could result in certain pathological sequences with non-zero probability, that are not enumerated in Figure 4.

YY cascade

Yr1Y^{r_{1}}NYr1N\;Y^{r_{1}}YNYr11Y\;N\;Y^{r_{1}-1}Y2NYr12Y^{2}\;N\;Y^{r_{1}-2}Yr11NYY^{r_{1}-1}\;N\;YStage (1)(1)

YY cascade

Yr2Y^{r_{2}}NYr2N\;Y^{r_{2}}YNYr21Y\;N\;Y^{r_{2}-1}Y2NYr22Y^{2}\;N\;Y^{r_{2}-2}Yr21NYY^{r_{2}-1}\;N\;YStage (2)(2)\ldots
Figure 4: An enumeration of all possible sequences that would lead to a YY cascade.

Let SnS_{n} denote the probability of hitting the right wall given that the sequence has not terminated before the nthn^{\text{th}} stage. Thus, the following recursion holds:

Sn=pfrn[1+rn(1pf)Sn+1],forn=1,2,\displaystyle S_{n}=p_{f}^{r_{n}}\big{[}1+r_{n}(1-p_{f})S_{n+1}\big{]},\;\;\text{for}\;n=1,2,\ldots (4)

and the probability of a YY cascade, denoted by Y-cas\mathbb{P}_{Y\text{-cas}} is:

Y-cas(ϵ)=S1,forϵr{ϵr}.\displaystyle\mathbb{P}_{Y\text{-cas}}(\epsilon)=S_{1},\quad\text{for}\;\;\epsilon\in\mathcal{I}_{r}\smallsetminus\{\epsilon_{r}\}. (5)

Here, while r1=r+1r_{1}=r+1 for ϵr\epsilon\in\mathcal{I}_{r}, successive values of rir_{i} for i=2,3,i=2,3,\ldots can be obtained from r1r_{1} using the updates:

rn={r,ifi=1n1(riη1)+rη>1,r+1,o.w.\displaystyle r_{n}=\begin{cases}\text{\scalebox{0.9}{$r$}},\quad&\text{if}\;\;\text{\scalebox{0.9}{$\sum_{i=1}^{n-1}(r_{i}\eta-1)+r\eta>1$}},\\ \text{\scalebox{0.9}{$r+1$}},\quad&\text{o.w.}\end{cases} (6)

Since (4) is an infinite recursion, to compute Y-cas\mathbb{P}_{Y\text{-cas}} in practice, we truncate the process to a finite number of iterations MM. To this end, we first assume that SM+1=1S_{M+1}=1. Next, we use (4) to successively compute SkS_{k} while kk counts down from MM to 11, performing a total of MM iterations. We denote the obtained value as Y-casM\mathbb{P}_{Y\text{-cas}}^{M} . We now state in the following theorem that Y-casM\mathbb{P}_{Y\text{-cas}}^{M} is in fact a tight upper bound to Y-cas\mathbb{P}_{Y\text{-cas}} as MM\rightarrow\infty. Moreover, the difference Y-casMY-cas\mathbb{P}_{Y\text{-cas}}^{M}-\mathbb{P}_{Y\text{-cas}} decays to zero at least as fast as {0.5M}\{0.5^{M}\}, in the number of iterations MM.

Refer to caption
Figure 5: Probability of YY cascade as a function of ϵ\epsilon for V=BV=B and p=0.7p=0.7.
Theorem 1

Let ϵr\epsilon\in\mathcal{I}_{r} for some r=1,2,r=1,2,\ldots, with pfp_{f} denoting the probability of a YY. Then, for any M=1,2,M=1,2,\ldots,

0Y-casM(ϵ)Y-cas(ϵ)kM,\displaystyle 0\,\leq\;\mathbb{P}_{Y\text{-cas}}^{M}(\epsilon)-\mathbb{P}_{Y\text{-cas}}(\epsilon)\;\leq\,k^{M},

where k(r+1)(1pf)pfrk\triangleq(r+1)(1-p_{f})p_{f}^{r} and for any p(0.5,1)p\in(0.5,1) and ϵ[0,1)\epsilon\in[0,1), kk satisfies 0<k1/20<k\leq 1/2.

Proof:

Recall that SM+1=1S_{M+1}=1 while computing Y-casM\mathbb{P}_{Y\text{-cas}}^{M} , which implies that Y-casM\mathbb{P}_{Y\text{-cas}}^{M} is the net probability of: (a)(a) sequences that terminate in a YY cascade by the MthM^{\text{th}} stage and (b)(b) sequences that do not terminate by the MthM^{\text{th}} stage. Thus, the difference: Y-casMY-cas\mathbb{P}_{Y\text{-cas}}^{M}-\mathbb{P}_{Y\text{-cas}} is upper-bounded by the probability of (b)(b). Accounting for all possible sequence combinations that persist through stages 1,2,,M1,2,\ldots,M yields the probability kk. Next, for a fixed pp and for ϵr\epsilon\in\mathcal{I}_{r}, kk is maximized only if ϵ\epsilon is such that pf=r1+rp_{f}=\frac{r}{1+r}, which may not be satisfied for any ϵ\epsilon in r\mathcal{I}_{r}. Assuming kk is maximized, the maximal value of kk would be (r1+r)r\big{(}\frac{r}{1+r}\big{)}^{r}. As this maximal value decreases in rr, evaluating it at r=1r=1 yields k1/2k\leq 1/2 for any rr. ∎

Figure 5 shows a plot of Y-cas\mathbb{P}_{Y\text{-cas}} with respect to ϵ\epsilon, for the case V=BV=B with p=0.7p=0.7. The plot uses M=10M=10 which gives an error of less than 10310^{-3}. It can be seen that in the rthr^{\text{th}} ϵ\epsilon–interval r\mathcal{I}_{r}, Y-cas\mathbb{P}_{Y\text{-cas}} increases with ϵ\epsilon, but with infinitely many discontinuities (where Y-cas(ϵ)\mathbb{P}_{Y\text{-cas}}(\epsilon) decreases). Despite the discontinuities, Y-cas\mathbb{P}_{Y\text{-cas}} achieves the minimum and maximum values at the edge points of r\mathcal{I}_{r}, i.e., ϵr+\epsilon_{r}^{+} and ϵr+1\epsilon_{r+1}^{-}, respectively. Further, note that the relatively larger drops in Y-cas(ϵ)\mathbb{P}_{Y\text{-cas}}(\epsilon) observed in Figure 5 (marked by ×\times and \circ) occur exactly at the threshold points {ϵr}r=1\{\epsilon_{r}\}_{r=1}^{\infty}. Here, counter to expectation, a slight increase in ϵ\epsilon beyond ϵr\epsilon_{r} causes a significant decrease in the probability of a YY cascade.

We now quantify Y-cas\mathbb{P}_{Y\text{-cas}} as ϵ\epsilon tends to each threshold point ϵr\epsilon_{r}. For this, we state the following lemma:

Lemma 3

For all i2i\geq 2, rirr_{i}\rightarrow r as ϵϵr\epsilon\rightarrow\epsilon_{r}.

Proof:

We consider two cases: ϵϵr+\epsilon\rightarrow\epsilon_{r}^{+} and ϵϵr\epsilon\rightarrow\epsilon_{r}^{-}. The proof outline for the first case is as follows: assume rj=rr_{j}=r for all 2ji12\leq j\leq i-1 and note that r1=r+1r_{1}=r+1 as ϵr+r\epsilon_{r}^{+}\in\mathcal{I}_{r}. Then it follows from (6) that ri=rr_{i}=r only if η>(r+1i)1\eta>\big{(}r+\frac{1}{i}\big{)}^{-1}. Now as ϵϵr+\epsilon\rightarrow\epsilon_{r}^{+}, η1/r\eta\rightarrow 1/r and hence the condition for ri=rr_{i}=r is satisfied. Using this argument inductively shows that ri=rr_{i}=r for all i2i\geq 2. The second case is proved similarly. ∎

It follows from Lemma 3 that as ϵϵr\epsilon\rightarrow\epsilon_{r}, the recursion in (4) results in the same infinite computation to obtain SiS_{i} as for Si+1S_{i+1}, for all i2i\geq 2. Thus, all SiS_{i} for i2i\geq 2 have the same value which satisfies: Si=pfr[1+r(1pf)Si]S_{i}=p_{f}^{r}\big{[}1+r(1-p_{f})S_{i}\big{]} . Solving this equation for i=2i=2 gives the value for S2S_{2} which when used in equation (4) for n=1n=1 yields S1S_{1}, i.e., Y-cas\mathbb{P}_{Y\text{-cas}}. However, note that while solving equation (4), r1=r+1r_{1}=r+1 for ϵ=ϵr+\epsilon=\epsilon_{r}^{+} whereas r1=rr_{1}=r for ϵ=ϵr\epsilon=\epsilon_{r}^{-}. This corresponds to the following two different values of Y-cas\mathbb{P}_{Y\text{-cas}} as ϵϵr\epsilon\rightarrow\epsilon_{r}:

Y-cas(ϵr+)\displaystyle\mathbb{P}_{Y\text{-cas}}(\epsilon_{r}^{+}) =pfr+11+(1pf)pfr1r(1pf)pfr,\displaystyle=\text{\scalebox{0.9}{$\displaystyle p_{f}^{r+1}\frac{1+(1-p_{f})p_{f}^{r}}{1-r(1-p_{f})p_{f}^{r}}$}}, (7)
Y-cas(ϵr)\displaystyle\mathbb{P}_{Y\text{-cas}}(\epsilon_{r}^{-}) =pfr11r(1pf)pfr.\displaystyle=\text{\scalebox{0.9}{$\displaystyle p_{f}^{r}\frac{1}{1-r(1-p_{f})p_{f}^{r}}$}}. (8)

Hence, the fractional decrease in Y-cas\mathbb{P}_{Y\text{-cas}} that occurs abruptly at ϵr\epsilon_{r}, defined as δr=[Y-cas(ϵr)Y-cas(ϵr+)]/Y-cas(ϵr)\delta_{r}=\big{[}\mathbb{P}_{Y\text{-cas}}(\epsilon_{r}^{-})-\mathbb{P}_{Y\text{-cas}}(\epsilon_{r}^{+})\big{]}/\mathbb{P}_{Y\text{-cas}}(\epsilon_{r}^{-}) is:

δr=(1pf)(1pfr+1).\displaystyle\delta_{r}=(1-p_{f})\,(1-p_{f}^{r+1}). (9)

It can be verified that as rr\rightarrow\infty, δr0\delta_{r}\rightarrow 0. This is depicted in Figure 5 where the sequences {Y-cas(ϵr)}\{\mathbb{P}_{Y\text{-cas}}(\epsilon_{r}^{-})\} and {Y-cas(ϵr+)}\{\mathbb{P}_{Y\text{-cas}}(\epsilon_{r}^{+})\} , marked by ×\times and \circ respectively, converge to a limiting value as rr\rightarrow\infty.

Property 3

If the possibility of fake agents in the history equals the rthr^{th} ϵ\epsilon-threshold, r=2,3,r=2,3,\ldots, then a further increase in fake agents reduces the chances of a YY cascade by a factor of δr\delta_{r}, rather than increasing it.

Next, we consider the cascade behaviour for low values of ϵ\epsilon and state the following theorem.

Theorem 2

Given the private signal quality p(0.5,1)p\in(0.5,1), and the item’s true value V{G,B}V\in\{G,B\}, there exists some ϵ¯=f(V,p)>0\underline{\epsilon}=f(V,p)>0 such that

Y-cas(ϵ)<Y-cas(0),ϵ(0,ϵ¯).\displaystyle\mathbb{P}_{Y\text{-cas}}(\epsilon)<\mathbb{P}_{Y\text{-cas}}(0),\quad\forall\;\;\epsilon\in(0,\underline{\epsilon}). (10)

From the point of view of the fake agents, the above theorem implies that if they occur with a probability of less than ϵ¯\underline{\epsilon}, then the effect that their presence has on the honest buyers is opposite to what they had intended. That is, they reduce the chances of a YY cascade instead of increasing it. On the other hand, for honest buyers, if V=BV=B, then they benefit from the presence of fake agents. Otherwise if V=GV=G, then they are harmed by the presence of fake agents.

The proof for Theorem 2 follows by noting that as ϵ0\epsilon\rightarrow 0, the limiting value of Y-cas\mathbb{P}_{Y\text{-cas}} can be obtained from (7) with r=1r=1 and pf1pp_{f}\rightarrow 1-p for V=BV=B, whereas pfpp_{f}\rightarrow p for V=GV=G, which gives

limϵ0Y-cas(ϵ)={(1p)21+p(1p)1p(1p),forV=B,p21+p(1p)1p(1p),forV=G.\displaystyle\text{\scalebox{0.95}{$\lim_{\epsilon\rightarrow 0}\mathbb{P}_{Y\text{-cas}}(\epsilon)=$}}\begin{cases}\text{\scalebox{0.9}{$\displaystyle(1-p)^{2}\frac{1+p(1-p)}{1-p(1-p)}$}},\quad&\text{for}\;\;V=B,\vspace{2mm}\\ \text{\scalebox{0.9}{$\displaystyle p^{2}\frac{1+p(1-p)}{1-p(1-p)}$}},\quad&\text{for}\;\;V=G.\end{cases} (11)

However, in the absence of fake agents (ϵ=0)(\epsilon=0) as in [1], η=1\eta=1. It then follows from Figure 2 that {hn}\{h_{n}\} has a finite state-space {2,1,0,1,2}\{-2,-1,0,1,2\}, and a YY (N)(N) cascade starts when hn=2h_{n}=2 (2)(-2). Here, solving for Y-cas\mathbb{P}_{Y\text{-cas}} gives

Y-cas(0)={p2/[p2+(1p)2],forV=G,(1p)2/[p2+(1p)2],forV=B.\displaystyle\mathbb{P}_{Y\text{-cas}}(0)=\begin{cases}p^{2}/\big{[}p^{2}+(1-p)^{2}\big{]},&\text{for}\;V=G,\\ (1-p)^{2}/\big{[}p^{2}+(1-p)^{2}\big{]},\;&\text{for}\;V=B.\end{cases} (12)

By comparing the above expression with the one in (11) for the corresponding values of V,V, it follows that

limϵ0Y-cas(ϵ)<Y-cas(0),p(0.5,1),V{G,B}.\displaystyle\lim_{\epsilon\rightarrow 0}\mathbb{P}_{Y\text{-cas}}(\epsilon)<\mathbb{P}_{Y\text{-cas}}(0),\quad\forall\;p\in(0.5,1),V\in\{G,B\}.

Thus, there exists some ϵ¯>0\underline{\epsilon}>0 such that (10) holds true.

V Learning in the limit as ϵ1\epsilon\rightarrow 1

As the probability of an agent being fake tends to 11, the information contained in a YY observation becomes negligible. As a result, an agent would need to observe infinitely many consecutive YY’s in his history for him to be convinced of starting a YY cascade. Hence, one would expect that if V=GV=G, learning would never occur, whereas if V=BV=B, then learning would always occur. However, recall that as ϵ1\epsilon\rightarrow 1, the occurrence of YY’s becomes increasingly frequent, i.e. pf1p_{f}\rightarrow 1; for both V=BV=B and GG. This motivates studying Y-cas\mathbb{P}_{Y\text{-cas}} in this limiting scenario. First, recall that in the process of enumerating all sequences leading to a YY cascade, for ϵr\epsilon\in\mathcal{I}_{r}, in each stage i2i\geq 2, rir_{i} is either rr or r+1r+1. However, ϵ1\epsilon\rightarrow 1 implies rr\rightarrow\infty, in which case rr+1r\approx r+1. As a result, the expressions obtained in (7) and (8) yield the same limiting value for Y-cas\mathbb{P}_{Y\text{-cas}} as ϵr1\epsilon_{r}\rightarrow 1, i.e., rr\rightarrow\infty. In particular,

limϵ1Y-cas(ϵ)=limrY-cas(ϵr+)=(a)limrY-cas(ϵr),\displaystyle\lim_{\epsilon\rightarrow 1}\mathbb{P}_{Y\text{-cas}}(\epsilon)=\lim_{r\rightarrow\infty}\mathbb{P}_{Y\text{-cas}}(\epsilon_{r}^{+})\overset{(a)}{=}\lim_{r\rightarrow\infty}\mathbb{P}_{Y\text{-cas}}(\epsilon_{r}^{-}),

where Step (a)(a) can be proved by recalling that δr0\delta_{r}\rightarrow 0 as rr\rightarrow\infty. Further, the limiting probability of a YY cascade, in terms of α=p/(1p)\alpha=p/(1-p), is given by (refer to References for proof):

limϵ1Y-cas(ϵ)\displaystyle\lim_{\epsilon\rightarrow 1}\mathbb{P}_{Y\text{-cas}}(\epsilon) =limrpfr11r(1pf)pfr=1ett;\displaystyle=\lim_{r\rightarrow\infty}\text{\scalebox{0.9}{$\displaystyle p_{f}^{r}\frac{1}{1-r(1-p_{f})p_{f}^{r}}=\frac{1}{e^{t}-t}$}}; (13)

where t=1α1logαt=\frac{1}{\alpha-1}\log\alpha for V=GV=G , and t=αα1logαt=\frac{\alpha}{\alpha-1}\log\alpha for V=BV=B .

Refer to caption
(a) V=BV=B
Refer to caption
(b) V=GV=G
Figure 6: Probability of YY cascade versus private signal quality for the indicated values of ϵ\epsilon.

Figure 6 illustrates (13) and the corresponding probability when ϵ=0.9\epsilon=0.9 as a function of the signal quality. For both V=BV=B and V=GV=G, a better signal quality leads to improved learning even when fake agents have overwhelmed the ordinary agents. Also note that for V=BV=B, for a weak signal quality, the incorrect cascade is more likely than the correct one (while for V=GV=G this is never true).

VI Asymptotic welfare for ordinary agents

In this section, we analyze the expected pay-off or welfare of an ordinary (rational) agent nn, 𝔼[πn]\mathbb{E}[\pi_{n}] as nn\rightarrow\infty. Recall from Section II that πn=0\pi_{n}=0 if An=NA_{n}=N whereas πn=1/2\pi_{n}=1/2 or 1/2-1/2 if An=YA_{n}=Y depending on whether V=GV=G or BB, respectively. Thus, agent nn’s welfare is given by:

𝔼[πn]\displaystyle\mathbb{E}[\pi_{n}] =(An=Y|V=G)14(An=Y|V=B)14.\displaystyle=\mathbb{P}(A_{n}=Y|V=G)\frac{1}{4}-\mathbb{P}(A_{n}=Y|V=B)\frac{1}{4}. (14)

Next, by taking the limit as nn\rightarrow\infty of (14), the asymptotic welfare limnE[πn(ϵ)]\underset{n\rightarrow\infty}{\lim}E[\pi_{n}(\epsilon)] denoted by Π(ϵ)\Pi(\epsilon) is given by:

Π(ϵ)\displaystyle\Pi(\epsilon) =14[Y-casG(ϵ)Y-casB(ϵ)]=(b)14(12wrong-cas(ϵ)).\displaystyle=\frac{1}{4}\left[\mathbb{P}_{Y\text{-cas}}^{G}(\epsilon)-\mathbb{P}_{Y\text{-cas}}^{B}(\epsilon)\right]\overset{(b)}{=}\frac{1}{4}\big{(}1-2\mathbb{P}_{\text{wrong-cas}}(\epsilon)\big{)}. (15)

In Step (b)(b), wrong-cas:=[N-casG+Y-casB]/2\mathbb{P}_{\text{wrong-cas}}:=\left[\mathbb{P}_{N\text{-cas}}^{G}+\mathbb{P}_{Y\text{-cas}}^{B}\right]/2 refers to the unconditional probability of a wrong cascade, i.e., the probability that a NN cascade occurs and V=GV=G or a YY cascade occurs and V=BV=B. The relation between wrong-cas(ϵ)\mathbb{P}_{\text{wrong-cas}}(\epsilon) and Π(ϵ)\Pi(\epsilon) presented by equation (15) substantiates the notion that a lower probability of wrong cascade results in higher asymptotic welfare. The probability PY-casV(ϵ)P_{Y\text{-cas}}^{V}(\epsilon) for V{G,B}V\in\{G,B\} can be computed using the recursive method described in Section IV-B by equations (4) and (5). Then, substituting these obtained values in equation (15) yields the value for Π(ϵ)\Pi(\epsilon). Figure 7 shows a plot of Π(ϵ)\Pi(\epsilon) with respect to ϵ(0,1)\epsilon\in(0,1), for p=0.7p=0.7 and compares it with the constant level of Π(0)\Pi(0) which refers to the asymptotic welfare in the absence of fake agents. We substitute (12) in (15) to obtain Π(0)\Pi(0), which gives

Π(0)=(1/4)(2p1)/[p2+(1p)2].\displaystyle\Pi(0)=(1/4)(2p-1)/\big{[}\mspace{2.0mu}p^{2}+(1-p)^{2}\mspace{2.0mu}\big{]}. (16)

It can be observed in Figure 7 that for p=0.7p=0.7, Π(ϵ)<Π(0)\Pi(\epsilon)<\Pi(0) for all ϵ(0,1)\epsilon\in(0,1). We find this ordering to be consistent over all values in the range of pp.

Remark 2

For any private signal quality pp, we observe that the presence of fake agents in any amount worsens the asymptotic welfare.

Further, note that the relatively larger jumps in Π(ϵ)\Pi(\epsilon) observed in Figure 7 (marked by ×\times and \circ) occur exactly at the threshold points {ϵr}r=2\{\epsilon_{r}\}_{r=2}^{\infty}. Here, counter to expectation, a slight increase in ϵ\epsilon beyond ϵr\epsilon_{r} causes an abrupt and significant increase in the asymptotic welfare. Therefore increasing ϵ\epsilon over the rthr^{th} ϵ\epsilon-threshold is not only counter-productive for the fake agents but it also leads to a higher asymptotic welfare for the ordinary agents.

Refer to caption
Figure 7: Asymptotic welfare as a function of ϵ\epsilon for p=0.7p=0.7.

VII Conclusions and future work

We studied the effect of randomly arriving fake agents that seek to influence the outcome of an information cascade. We focussed on the impact of varying the amount of fake agents on the probability of their preferred cascade and identified scenarios where the presence of fake agents reduces the chances of their preferred cascade. In particular, the chances of the preferred cascade decrease abruptly as the fraction of agents increase beyond specific thresholds. We quantified the correct cascade probability as a function of the private signal quality in the asymptotic regime where the fraction of fake agents approaches unity. We observed that the presence of fake agents in any amount worsens the asymptotic welfare. Studying the effects of multiple types of fake agents and non-Bayesian rationality are possible directions for future work.

References

  • [1] S. Bikhchandani, D. Hirshleifer, and I. Welch, “A theory of fads, fashion, custom, and cultural change as informational cascades,” Journal of political Economy, vol. 100, no. 5, pp. 992–1026, 1992.
  • [2] A. V. Banerjee, “A simple model of herd behavior,” The quarterly journal of economics, vol. 107, no. 3, pp. 797–817, 1992.
  • [3] I. Welch, “Sequential sales, learning, and cascades,” The Journal of finance, vol. 47, no. 2, pp. 695–732, 1992.
  • [4] D. Acemoglu, M. A. Dahleh, I. Lobel, and A. Ozdaglar, “Bayesian learning in social networks,” The Review of Economic Studies, vol. 78, no. 4, pp. 1201–1236, 2011.
  • [5] Y. Song, “Social learning with endogenous network formation,” arXiv preprint arXiv:1504.05222, 2015.
  • [6] T. N. Le, V. G. Subramanian, and R. A. Berry, “Information cascades with noise,” IEEE Transactions on Signal and Information Processing over Networks, vol. 3, no. 2, pp. 239–251, 2017.
  • [7] L. Smith and P. Sørensen, “Pathological outcomes of observational learning,” Econometrica, vol. 68, no. 2, pp. 371–398, 2000.
  • [8] D. Acemoğlu, G. Como, F. Fagnani, and A. Ozdaglar, “Opinion fluctuations and disagreement in social networks,” Mathematics of Operations Research, vol. 38, no. 1, pp. 1–27, 2013.
  • [9] M. Mobilia, “Does a single zealot affect an infinite group of voters?” Physical review letters, vol. 91, no. 2, p. 028701, 2003.
  • [10] I. H. Lee, “On the convergence of informational cascades,” Journal of Economic theory, vol. 61, no. 2, pp. 395–411, 1993.
  • [11] Y. Wang and P. M. Djurić, “Social learning with bayesian agents and random decision making,” IEEE Transactions on Signal Processing, vol. 63, no. 12, pp. 3241–3250, 2015.
  • [12] T. Le, V. Subramanian, and R. Berry, “Quantifying the utility of noisy reviews in stopping information cascades,” in IEEE CDC Proceeding, 2016.
  • [13] Y. Peres, M. Z. Rácz, A. Sly, and I. Stuhl, “How fragile are information cascades?” arXiv preprint arXiv:1711.04024, 2017.
  • [14] T. N. Le, V. G. Subramanian, and R. A. Berry, “Bayesian learning with random arrivals,” in 2018 IEEE International Symposium on Information Theory (ISIT).   IEEE, 2018, pp. 926–930.
  • [15] I. Bistritz and A. Anastasopoulos, “Characterizing non-myopic information cascades in bayesian learning,” in 2018 IEEE Conference on Decision and Control (CDC).   IEEE, 2018, pp. 2716–2721.
  • [16] Y. Cheng, W. Hann-Caruthers, and O. Tamuz, “A deterministic protocol for sequential asymptotic learning,” in 2018 IEEE International Symposium on Information Theory (ISIT).   IEEE, 2018, pp. 1735–1738.
  • [17] G. Schoenebeck, S. Su, and V. G. Subramanian, “Social learning with questions,” CoRR, vol. abs/1811.00226, 2018. [Online]. Available: http://arxiv.org/abs/1811.00226

Proof for equation (13)

Proof:

First, we first obtain the limitng value for the term pfrp_{f}^{r} as rr\rightarrow\infty. For this, recall that as r=1/ηr=\left\lfloor 1/\eta\right\rfloor, we have pf1/ηpfr<pf1/η1p_{f}^{1/\eta}\leq p_{f}^{r}<p_{f}^{1/\eta-1}. Since the upper and lower bounds converge to the same limit, it implies that

limrpfr=limϵ1pf1/η.\displaystyle\underset{r\rightarrow\infty}{\lim}p_{f}^{r}=\underset{\epsilon\rightarrow 1}{\lim}p_{f}^{1/\eta}. (17)

Further, given that ηlog(a1b)/log(p1p)\eta\triangleq\log(\frac{a}{1-b})/\log(\frac{p}{1-p}), the numerator of η\eta can be bounded using the inequality: 1x1log(x)x11-x^{-1}\leq\log(x)\leq x-1, which holds for all x>0x>0. As a result, η\eta can be bounded as:

(2p1)(1ϵ)(p+(1p)ϵ)logαη(2p1)(1ϵ)(1p(1ϵ))logα.\displaystyle\frac{(2p-1)(1-\epsilon)}{\big{(}p+(1-p)\epsilon\big{)}\log\alpha}\leq\eta\leq\frac{(2p-1)(1-\epsilon)}{\big{(}1-p(1-\epsilon)\big{)}\log\alpha}. (18)

This in turn extends to the following bounds on pf1/ηp_{f}^{1/\eta},

pf11pfatpf1/ηpf11pf(1b)t,\displaystyle p_{f}^{\frac{1}{1-p_{f}}at}\leq p_{f}^{1/\eta}\leq p_{f}^{\frac{1}{1-p_{f}}(1-b)t}, (19)

where t=1α1logαt=\frac{1}{\alpha-1}\log\alpha for V=GV=G , and t=αα1logαt=\frac{\alpha}{\alpha-1}\log\alpha for V=BV=B . Now, as both aa and 1b1-b tend to 11 as ϵ1\epsilon\rightarrow 1; both the upper and the lower bounds presented in (19) achieve the same limit. Hence,

limrpfr=(17)limϵ1pf1/η=(limϵ1pf11pf)t=(a)et,\displaystyle\underset{r\rightarrow\infty}{\lim}p_{f}^{r}\overset{\eqref{p_f^r_limit}}{=}\underset{\epsilon\rightarrow 1}{\lim}p_{f}^{1/\eta}=\left(\underset{\epsilon\rightarrow 1}{\lim}p_{f}^{\frac{1}{1-p_{f}}}\right)^{t}\overset{(a)}{=}e^{-t}, (20)

where Step (a)(a) follows by reducing the limit term such that it maps to the identity: limx0+(1x)1x=1/e\underset{x\rightarrow 0^{+}}{\lim}(1-x)^{\frac{1}{x}}=1/e.

Secondly, we obtain the limitng value for the term r(1pf)r(1-p_{f}) which equals the limit for (1pf)/η(1-p_{f})/\eta because r=1/ηr=\left\lfloor 1/\eta\right\rfloor. Now, inequality (18) leads to the following bounds on (1pf)/η(1-p_{f})/\eta:

(1b)(1pf)(2p1)(1ϵ)logα1pfηa(1pf)(2p1)(1ϵ)logα.\displaystyle\frac{(1-b)(1-p_{f})}{(2p-1)(1-\epsilon)}\log\alpha\leq\frac{1-p_{f}}{\eta}\leq\frac{a(1-p_{f})}{(2p-1)(1-\epsilon)}\log\alpha.

But as both aa and 1b1-b tend to 11 as ϵ1\epsilon\rightarrow 1; both the upper and the lower bounds presented above achieve the same limit. Hence,

limϵ1r(1pf)=limϵ11pfη=limϵ1(1pf)logα(2p1)(1ϵ)=t,\displaystyle\underset{\epsilon\rightarrow 1}{\lim}\;r(1-p_{f})=\underset{\epsilon\rightarrow 1}{\lim}\frac{1-p_{f}}{\eta}=\underset{\epsilon\rightarrow 1}{\lim}\frac{(1-p_{f})\log\alpha}{(2p-1)(1-\epsilon)}=t, (21)

where tt has been defined for V{G,B}V\in\{G,B\} in (19).

Finally, we use the limits obtained in (20) and (21) for evaluating the limit stated in (13), which gives the desired result. ∎