This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

The Power of Menus in Contract Design

Guru Guruganesh
Google Research
[email protected]
   Jon Schneider
Google Research
[email protected]
   Joshua Wang
Google Research
[email protected]
   Junyao Zhao
Stanford University
[email protected]
Abstract

We study the power of menus of contracts in principal-agent problems with adverse selection (agents can be one of several types) and moral hazard (we cannot observe agent actions directly). For principal-agent problems with TT types and nn actions, we show that the best menu of contracts can obtain a factor Ω(max(n,logT))\Omega(\max(n,\log T)) more utility for the principal than the best individual contract, partially resolving an open question of Guruganesh et al. (2021). We then turn our attention to randomized menus of linear contracts, where we likewise show that randomized linear menus can be Ω(T)\Omega(T) better than the best single linear contract. As a corollary, we show this implies an analogous gap between deterministic menus of (general) contracts and randomized menus of contracts (as introduced by Castiglioni et al. (2022)).

1 Introduction

The principal-agent problem studies a setting where one party (the principal) wishes to incentivize a second party (the agent) to exert effort on behalf of the principal. The agent has different potential actions they could take (e.g., levels of effort they exert), each of which stochastically results in one of several outcomes. The principal cannot directly observe the agent’s action (“moral hazard”) but they have preferences over the different outcomes and would like to incentivize the agent to take a favorable action for the principal. To accomplish this, the principal can offer the agent a contract – a mechanism describing how the principal will reward the agent contingent on the ultimate realized outcome. Principal-agent problems arise in a wide variety of different disciplines (e.g., law, insurance, employment) and the development of contract theory has proven to be an invaluable tool in the economics literature for the analysis of such problems.

Over the past few years, there has been a surge of activity in the application of computational methods to contract theory (much in the same way that computational methods have been successfully applied to mechanism design and auction theory). Many of these papers take the perspective of understanding the power of one class of contracting mechanisms as it compares to another. For example, in Dütting et al. (2019), the authors show that in principal-agent settings with nn actions, the optimal general contract obtains at most Ω(n)\Omega(n) times as much utility as the optimal linear contract (a much simpler subclass of contracts where the principal promises to share a fixed proportion of their utility with the agent), and that there are cases where this is tight. This methodology has been applied to many variants of the principal-agent problem (e.g. variants with combinatorially structured actions and outcomes).

One particularly interesting variant of the principal-agent problem introduces private information (or “adverse selection”) to the problem instance. In this setting an agent may be one of several types (unknown to the principal), and their relevant properties (e.g., the probability they induce a certain outcome by playing a specific action) may differ from type to type, and this allows us to model the uncertainty the principal may have over the agent they are contracting (and capture a much more realistic set of problems).

In the typed principal-agent problem, there are (at least two) natural classes of mechanisms to compare. One option is to, as before, offer a fixed single contract to the agent. However, with the presence of types, the principal also has the option to offer the agent a menu of different contracts (from which the agent selects their favorite). Is it ever in the principal’s interest to do this? This question was originally proposed by Guruganesh et al. (2021), who showed there exist instances where it is strictly beneficial to offer a menu over a single contract. However, Guruganesh et al. (2021) were only able to show that the gap in power between menus and single contracts lies somewhere between Ω(nlogT)\Omega(n\log T) (for a setting with nn actions and TT types) and 22. This gap (and even the question of whether it is super-constant) remains open to this day.

Making things even more interesting, Castiglioni et al. (2022) recently showed that there is a third, even more powerful class of mechanisms where the principal offers the agent a menu of randomized contracts (i.e., the agent selects a distribution over contracts and receives a contract randomly drawn from this distribution). Although much is known about the computational properties of such mechanisms111Intriguingly, while single contracts and ordinary menus are computationally hard to optimize in the typed setting, there exist efficient algorithms to find the best menu of randomized contracts., essentially nothing is currently known about the power of such mechanisms – either in comparison to ordinary menus or to individual contracts.

1.1 Our results

In this paper we pin down more closely the power of menus in contract design. We begin by significantly closing the gap present between single contracts and menus of (deterministic) contracts present in Guruganesh et al. (2021). In particular, Guruganesh et al. (2021) prove that in any principal-agent problem with nn actions and TT types, the profit of the best menu of contracts is at most O(nlogT)O(n\log T) larger than the profit of the best individual contract. However, the only lower bound they provide (and the only lower bound known to date) is that there are menus of contracts that outperform individual contracts by a constant factor. In our first two theorems, we show that both the dependence on nn and TT in the upper bound are asymptotically tight.

Theorem 1.1 (Restatement of Theorem 3.1).

There exists a principal-agent problem with 33 types and nn actions where the profit of the optimal menu of (deterministic) contracts is at least Ω(n)\Omega(n) larger than the profit of the optimal single contract.

Theorem 1.2 (Restatement of Theorem 3.2).

There exists a principal-agent problem with TT types and 22 actions where the profit of the optimal menu of (deterministic) contracts is at least Ω(logT)\Omega(\log T) larger than the profit of the optimal single contract.

We remark that both the optimal menu of deterministic contracts and the optimal single contract are known to be computationally intractable and thus can be tricky to work with. Previous approaches to show such gaps (such as Guruganesh et al. (2021)) worked via adding gadgets to equate the value of one class to the value of welfare and the value of the other side to the value of linear contracts. Our proof introduces some novel elements that help bar single contracts from doing as well as menus, and does not attempt to reduce to welfare versus linear contracts.

We then switch our attention to the setting of linear contracts. Linear contracts are an important subclass of contracts in which the principal gives the agent a fixed percentage of the principal’s total reward. Linear contracts are commonly used in practice, and much of the recent algorithmic work in contract theory has been focused on better understanding linear contracts in a variety of settings.

At first, it may not seem that menus of linear contracts are particularly useful – given a menu of different fixed percentages, it is always in an agent’s interest to choose the highest fixed percentage (and thus the principal could simply have offered that single linear contract). However, this assumes that we are offering the agent a menu of deterministic linear contracts. This is not the case for menus of randomized linear contracts (in the same sense as the menus of randomized general contracts of Castiglioni et al. (2022)) which do have the potential to outperform a single linear contract. The interaction between the principal and the agent for a menu of randomized linear contracts proceeds as follows:

  1. 1.

    The principal begins by constructing a menu of randomized linear contracts. Each menu item is a distribution over transfer coefficients α0\alpha\geq 0, where a specific α\alpha represents the linear contract where the principal offers an α\alpha fraction of their reward to the agent.

  2. 2.

    The agent (with a randomly drawn type) selects a single distribution 𝒟\mathcal{D} from this menu.

  3. 3.

    The principal then samples a linear contract α\alpha from 𝒟\mathcal{D} and reveals it to the agent.

  4. 4.

    The agent observes the linear contract α\alpha and then takes an action in response. An outcome occurs, the principal receives some reward RR, gives αR\alpha R to the agent, and keeps (1α)R(1-\alpha)R.

It turns out that randomized menus of linear contracts are in general more powerful than individual linear contracts. In particular, we show that there exist principal-agent problems where the principal can obtain a strictly larger profit by offering a menu of randomized linear contracts than by offering a single linear contract (see Example 4.1).

We then demonstrate how to find the optimal randomized menu of linear contracts by solving a single linear program with poly(n,T)\mathrm{poly}(n,T) variables and constraints (Theorem 4.4). In the course of doing so, we also prove the following facts about the structure of randomized menus of linear contracts, which may be of independent interest. First, we prove that without loss of generality, all randomized linear contracts in the optimal menu have identical support, and this support has size at most O(nT)O(nT) (see Lemma 4.2). Secondly (and much more counterintuitively), this optimal support sometimes contains contracts where the principal is guaranteed to lose money; i.e., sometimes it is necessary to include linear contracts with α>1\alpha>1 in the menu to achieve the optimal profit for the principal. We give one example of this in Lemma 4.3.

Finally, we attempt to characterize the gap in power between menus of randomized linear contracts and individual linear contracts. An upper bound immediately results from the gap of Guruganesh et al. (2021) between a single linear contract and the maximum achievable welfare in a principal agent problem. We prove a tight lower bound for one specific regime of nn and TT (specifically, where nn and TT have roughly the same magnitude).

Theorem 1.3 (Restatement of Theorem 4.7).

There exists a principal-agent problem with TT types and O(T)O(T) actions where the profit of the optimal menu of randomized linear contracts is at least Ω(T)\Omega(T) as large as the profit of the optimal individual linear contract.

Combining Theorem 1.3 with the facts that (i) there are some principal-agent problems where any contract can be equivalently written as a linear contract and (ii) deterministic menus of linear contracts are equal in power to individual linear contracts, we immediately obtain an analogous gap between menus of (general) deterministic contracts and menus of (general) randomized contracts.

Corollary 1.4 (Restatement of Corollary 4.8).

There exists a principal-agent problem with TT types and O(T)O(T) actions where the profit of the optimal menu of randomized contracts is at least Ω(T)\Omega(T) as large as the profit of the optimal menu of deterministic contracts.

It is natural to ask whether it is possible to strengthen Theorem 1.3 to understand how this gap scales individually with nn and TT. That is, how large is the gap between these two types of mechanisms in the regime where there are a constant number of types but the number of actions for each type grows large (analogous to Theorem 1.1) or in the regime where there is a fixed number of actions per type, but the number of different types of agents grows large (analogous to Theorem 1.2)? We do not resolve this question in this paper – however, we make the following related observation: if every type has exactly one non-null action (e.g., every type decides between working and not working), the ratio in the revenue achieved between menus of randomized linear contracts and a single linear contract is at most a (universal) constant.

Theorem 1.5 (Restatement of Theorem 4.6).

In any principal-agent problem where each type of agent has one null action (guaranteeing the principal zero utility) and one non-null action, the profit of the best single linear contract is at least a constant fraction of the profit of the best menu of randomized linear contracts.

1.2 Related work

The principal-agent problem has a long history in economics, stemming back to the foundational papers of Holmström (1979) and Grossman and Hart (1983). Our paper fits into a recent line of work applying techniques from algorithmic mechanism design to contract theory (Alon et al., 2021, Castiglioni et al., 2021, 2022, Dütting et al., 2019, 2020, 2021, Guruganesh et al., 2021, Cohen et al., 2022).

Of these, the most closely relevant are a handful of recent works that consider the intersection of contract theory with private information (“contracts with types”). We build most directly off of Guruganesh et al. (2021), which introduced the problem of determining the gap in power between single contracts and menus of contracts, and Castiglioni et al. (2022), which introduced the concept of menus of randomized contracts and demonstrated they can be computed efficiently. Other works in this space include Alon et al. (2021), which studies a one-dimensional subclass of the typed principal-agent problem where the type simply scales the cost, and Castiglioni et al. (2021), which studies the computational hardness and power of optimal single contracts in the typed principal-agent setting. Also of note is the recent paper of Gan et al. (2022), which provides a revelation principle for general principal-agent problems which implies (among other things) that menus of randomized contracts are the most general possible mechanisms for our setting.

We briefly mention that the theme of more powerful mechanisms (menus / randomized menus in particular) outperforming simpler mechanisms is a common theme throughout many other parts of algorithmic game theory (for example, the classic work of Briest et al. (2010), which proves that randomization is essential to get approximately revenue-optimal mechanisms for selling multiple goods). These settings generally lack the property of moral hazard (i.e., hidden actions) inherent to the contract design setting, and hence these existing results and techniques do not seem to transfer over to the principal-agent problem.

2 Preliminaries

We introduce the typed principal-agent problem following the notation of Guruganesh et al. (2021). In this problem, there are TT different types of agent (drawn from a known prior distribution), each of which has nn actions which result in one of mm possible outcomes. When the agent of type t[T]t\in[T] takes action i[n]i\in[n] they incur a cost ci0c_{i}\geq 0 for the agent (regardless of their type) and cause outcome j[m]j\in[m] to occur with probability Fi,j(t)F^{(t)}_{i,j}. When outcome jj occurs, the principal receives a reward rjr_{j}. Notably, the principal can only observe the eventual outcome jj and cannot observe the type tt of the agent or the action ii taken by the agent (the hidden-action model with private information). We write 𝒄\bm{c}, 𝑭\bm{F}, and 𝒓\bm{r} to denote the collections of costs, outcome probabilities, and rewards respectively. Together, the tuple (𝒄,𝑭,𝒓)\left(\bm{c},\bm{F},\bm{r}\right) completely specifies a typed principal-agent problem instance.

We further always make the assumption that there exists a null action with zero cost (c0=0c_{0}=0) which deterministically results in a null outcome with zero reward for the principal (r0=0r_{0}=0; other actions may also lead to the null outcome with some probability). In this way we model the possibility for an agent to opt out of participating entirely.

We consider three classes of contracting mechanisms for the principal (and their linear variants, which we define later): (i) single contracts, (ii) menus of deterministic contracts, and (iii) menus of randomized contracts. We discuss these in order.

Single contracts

A contract is simply an mm-dimensional vector 𝒙\bm{x} with non-negative entries. For each j[m]j\in[m], xjx_{j} specifies the amount the principal promises to transfer to the agent in the case that outcome jj occurs. In response to a contract 𝒙\bm{x}, the agent of type tt chooses the action i(t)(𝒙)i_{*}^{(t)}(\bm{x}) that optimizes their utility, namely

i(t)(𝒙)argmaxi(j=1mFi,j(t)xj)ci.\displaystyle i_{*}^{(t)}(\bm{x})\in\operatorname*{argmax}_{i}\left(\sum_{j=1}^{m}F^{(t)}_{i,j}x_{j}\right)-c_{i}\text{.}

We call the principal’s net expected utility from offering contract 𝒙\bm{x} their profit from offering contract 𝒙\bm{x}. The principal’s profit from an agent of type tt is given by

Profit(t)(𝒄,𝑭,𝒓,𝒙)j=1mFi(t)(𝒙),j(t)(rjxj)\displaystyle\textsc{Profit}^{(t)}(\bm{c},\bm{F},\bm{r},\bm{x})\triangleq\sum_{j=1}^{m}F^{(t)}_{i_{*}^{(t)}(\bm{x}),j}(r_{j}-x_{j})

and their overall expected profit is given by

Profit(𝒄,𝑭,𝒓,𝒙)𝔼t[Profit(t)(𝒄,𝑭,𝒓,𝒙)].\displaystyle\textsc{Profit}(\bm{c},\bm{F},\bm{r},\bm{x})\triangleq\mathbb{E}_{t}\left[\textsc{Profit}^{(t)}(\bm{c},\bm{F},\bm{r},\bm{x})\right]\text{.}

We are interested in the largest profit obtainable for the principal via a single contract. We write Opt-Single(𝒄,𝑭,𝒓)\textsc{Opt-Single}\left(\bm{c},\bm{F},\bm{r}\right) to denote this maximal profit for the principal agent problem (𝒄,𝑭,𝒓)\left(\bm{c},\bm{F},\bm{r}\right).

Menus of deterministic contracts

In the presence of types, the principal can sometimes obtain additional profit by offering the agent a menu of individual contracts (from which the agent picks the contract that maximizes their expected utility). By the revelation principle, it suffices to specify a single contract for each type, and we can write a general menu of deterministic contracts in the form 𝑿=(𝒙(1),,𝒙(T))\bm{X}=\left(\bm{x}^{(1)},\ldots,\bm{x}^{(T)}\right), where an agent who reports their type to be tt receives the contract 𝒙(t)\bm{x}^{(t)}. Such a menu is “incentive-compatible” (IC) if no type tt has an incentive to misreport their type as a different type ttt^{\prime}\neq t:

t,t[T]maxi(j=1mFi,j(t)xj(t))cimaxi(j=1mFi,j(t)xj(t))ci.\displaystyle\forall t,t^{\prime}\in[T]\quad\max_{i}\left(\sum_{j=1}^{m}F^{(t)}_{i,j}x^{(t)}_{j}\right)-c_{i}\geq\max_{i}\left(\sum_{j=1}^{m}F^{(t)}_{i,j}x^{(t^{\prime})}_{j}\right)-c_{i}\text{.} (1)

Unless otherwise stated, we restrict our attention to incentive-compatible menus of deterministic contracts. As with single contracts, we write Profit(𝒄,𝑭,𝒓,𝑿)\textsc{Profit}(\bm{c},\bm{F},\bm{r},\bm{X}) to denote the expected profit for the principal from offering the menu of deterministic contracts 𝑿\bm{X}, and we write Opt-DetMenu(𝒄,𝑭,𝒓)\textsc{Opt-DetMenu}{}\left(\bm{c},\bm{F},\bm{r}\right) to be the profit of the best menu of deterministic contracts.

Menus of randomized contracts

Deterministic menus can be generalized further by offering distributions over contracts instead of individual contracts. This class of mechanisms was proposed by Castiglioni et al. (2022), and works as follows. The principal offers the agent a menu of distributions over contracts (from which the agent picks the distribution that maximizes their expected utility, with randomness in both which contract gets chosen and which outcome occurs). By the revelation principle, it suffices to specify a distribution for each type, and we can write a general menu of randomized contracts in the form 𝒳=(𝒟(1),𝒟(2),,𝒟(T))\mathcal{X}=\left(\mathcal{D}^{(1)},\mathcal{D}^{(2)},\ldots,\mathcal{D}^{(T)}\right), where an agent who reports their type to be tt receives the distribution 𝒟(t)\mathcal{D}^{(t)}. The principal then draws a contract from 𝒟(t)\mathcal{D}^{(t)} and relays it to the agent, who takes an action in response. Such a menu is “incentive-compatible” (IC) if no type tt has an incentive to misreport their type as a different type ttt^{\prime}\neq t:

t,t[T]𝔼x𝒟(t)[maxi(j=1mFi,j(t)xj)ci]𝔼x𝒟(t)[maxi(j=1mFi,j(t)xj)ci].\displaystyle\forall t,t^{\prime}\in[T]\quad\mathbb{E}_{x\sim\mathcal{D}^{(t)}}\left[\max_{i}\left(\sum_{j=1}^{m}F^{(t)}_{i,j}x_{j}\right)-c_{i}\right]\geq\mathbb{E}_{x\sim\mathcal{D}^{(t^{\prime})}}\left[\max_{i}\left(\sum_{j=1}^{m}F^{(t)}_{i,j}x_{j}\right)-c_{i}\right]\text{.}

We restrict our attention to incentive-compatible menus of randomized contracts. As before, we write Profit(𝒄,𝑭,𝒓,𝒳)\textsc{Profit}(\bm{c},\bm{F},\bm{r},\mathcal{X}) to denote the expected profit for the principal from offering the menu of randomized contracts 𝒳\mathcal{X}, and we write Opt-RndMenu(𝒄,𝑭,𝒓)\textsc{Opt-RndMenu}{}\left(\bm{c},\bm{F},\bm{r}\right) to be the profit of the best menu of randomized contracts.

Linear contracts

We now discuss how to restrict the previous three classes to linear contracts. Contracts with the property that xj=αrjx_{j}=\alpha\cdot r_{j} for all j[m]j\in[m] and some α0\alpha\geq 0 are called linear contracts.

The most straightforward class to restrict is that of single contracts, where all we do is require that the chosen contract 𝒙\bm{x} is linear. We write Opt-Linear(𝒄,𝑭,𝒓)\textsc{Opt-Linear}{}\left(\bm{c},\bm{F},\bm{r}\right) to be the profit of the best (single) linear contract.

The next class to restrict is menus of deterministic contracts, where we require that each individual contract offered by the menu is linear. Specifically, for 𝑿=(𝒙(1),𝒙(2),,𝒙(T))\bm{X}=\left(\bm{x}^{(1)},\bm{x}^{(2)},\ldots,\bm{x}^{(T)}\right), we would like each 𝒙(t)\bm{x}^{(t)} to be linear. Interestingly, this turns out to be no more powerful than (single) linear contracts, because agents might as well pick the offered linear contract with the highest α\alpha. As a result, Opt-Linear(𝒄,𝑭,𝒓)\textsc{Opt-Linear}{}\left(\bm{c},\bm{F},\bm{r}\right) also designates the profit of the best menu of deterministic linear contracts.

The final class to restrict is menus of randomized contracts, where we require that each distribution offered by the menu only contains linear contracts in its support. Specifically, for Γ=(γ(1),γ(2),,γ(T))\Gamma=\left(\mathcal{\gamma}^{(1)},\mathcal{\gamma}^{(2)},\ldots,\mathcal{\gamma}^{(T)}\right) we would like the support of γ(1)\mathcal{\gamma}^{(1)} to be linear contracts. We will show this is more powerful than (single) linear contracts. We write Opt-RndMenuLinear(𝒄,𝑭,𝒓)\textsc{Opt-RndMenuLinear}{}\left(\bm{c},\bm{F},\bm{r}\right) to be the profit of the best menu of randomized linear contracts. These menus will be our main object of study in Section 4; as we will see, unlike menus of deterministic linear contracts, these are occasionally more powerful than single linear contracts.

Menus ofRandomized ContractsMenus ofDeterministic ContractsMenus of RandomizedLinear ContractsSingle ContractsLinear Contracts
Figure 1: Hierarchy of contracting mechanisms for typed contract problems. Edges link more general, powerful classes (higher) to more restricted, weaker classes (lower). The left (menus of deterministic contracts and single contracts) and right (menus of randomized linear contracts) sides of the diagram are incomparable; for some instances a left class is more powerful and for other instances a right class is more powerful.

This completes our overview of five classes of contracting mechanisms for the principal. Our five classes of interest are depicted in Figure 1, which also illustrates which classes contain each other.

2.1 Known upper bounds

Guruganesh et al. (2021) proved an upper bound of O(nlogT)O(n\log T) between a single linear contract and the maximum achievable welfare in a typed principal agent problem. Every class mentioned above is at least as powerful as single linear contracts and at most as powerful as extracting all welfare, so this O(nlogT)O(n\log T) bound applies to any pair of classes as well. However, this upper bound hinges on two key assumptions: (i) all types have the same set of costs and (ii) the distribution over types is uniform. Removing either of these assumptions breaks the proof, leaving only a O(nT)O(nT) bound. In this paper, we will work in the more general setting where the costs may be vary with types and the distribution is nonuniform. However, all proofs can be adapted to work in the more restricted setting above.

3 Gaps between menus and single contracts

In this section, we show that deterministic menus of contracts can be much more powerful than a single contract:

  • When the number of types TT is a constant, there is a principal-agent problem instance for which optimal deterministic menu of contracts can extract Ω(n)\Omega(n) times as much profit as the optimal single contract.

  • When the number of actions nn is a constant, there is an instance where optimal deterministic menu of contracts can extract Ω(logT)\Omega(\log T) times as much profit as the optimal single contract.

3.1 Ω(n)\Omega(n)-gap with three types

Theorem 3.1.

There is a principal-agent problem (𝐜,𝐅,𝐫)\left(\bm{c},\bm{F},\bm{r}\right) with T=3T=3 agent types and nn actions, for which optimal deterministic menu can extract Ω(n)\Omega(n) times as much profit as optimal single contract can:

Opt-DetMenu(𝒄,𝑭,𝒓)Ω(n)Opt-Single(𝒄,𝑭,𝒓).\textsc{Opt-DetMenu}\left(\bm{c},\bm{F},\bm{r}\right)\geq\Omega(n)\cdot\textsc{Opt-Single}\left(\bm{c},\bm{F},\bm{r}\right).

We provide here the Ω(n)\Omega(n)-gap instance and explain the high-level ideas behind its construction and analysis. We defer the full proof of Theorem 3.1 (which is rather technical) to Section A.1.

Construction of Ω(n)\Omega(n)-gap instance

We begin by providing a formal description of the instance, which is also recorded in Table 1. There are three agent types. There are four outcomes, and their rewards are r1=1r_{1}=1, r2=0r_{2}=0, r3=1r_{3}=1 and r4=0r_{4}=0. We will assume we have 2n+12n+1 non-null actions (assuming n12n\geq 12 without loss of generality), with costs given by ci=niinn+1c_{i}=\frac{n^{i}-i}{n^{n+1}} for i[n]i\in[n], cn+i=4(nii)nn+1c_{n+i}=\frac{4(n^{i}-i)}{n^{n+1}} for i[n]i\in[n], and c2n+1=0c_{2n+1}=0.

We now construct the forecast tensor 𝑭\bm{F}. First, we let F2n+1,j(1)=0F^{(1)}_{2n+1,j}=0, Fi+n,j(1)=0F^{(1)}_{i+n,j}=0 and Fi,j(2)=0F^{(2)}_{i,j}=0 for all i[n]i\in[n] and j{1,2,3}j\in\{1,2,3\}, and we let Fi,j(3)=0F^{(3)}_{i,j}=0 for all i[2n+1]i\in[2n+1] and j{1,2,3}j\in\{1,2,3\}. That is, outcome 44 always occurs when the type 11 agent plays the last n+1n+1 actions, when the type 22 agent plays the first nn actions, and when the type 33 agent plays any action.

For the type 11 agent, we let Fi,1(1)=ni1nF^{(1)}_{i,1}=n^{i-1-n} for all i[n]i\in[n], and we let Fi,2(1)=niinnnFn,2(1)F^{(1)}_{i,2}=\frac{n^{i}-i}{n^{n}-n}\cdot F^{(1)}_{n,2} for all i[n]i\in[n], and moreover, we let Fi,3(1)=0F^{(1)}_{i,3}=0 for all i[n]i\in[n]. For the type 22 agent, we let Fn+i,3(2)=4ni1nF^{(2)}_{n+i,3}=4n^{i-1-n} for all i[n]i\in[n], let Fn+i,2(2)=4nn11Fn,2(1)F^{(2)}_{n+i,2}=\frac{4}{n^{n-1}-1}\cdot F^{(1)}_{n,2} for all i[n]i\in[n], and let Fn+i,1(2)=0F^{(2)}_{n+i,1}=0 for all i[n]i\in[n]. Furthermore, we let F2n+1,1(2)=F2n+1,3(2)=0F^{(2)}_{2n+1,1}=F^{(2)}_{2n+1,3}=0, and F2n+1,2(2)=4nn11Fn,2(1)F^{(2)}_{2n+1,2}=\frac{4}{n^{n-1}-1}\cdot F^{(1)}_{n,2}. The specific value of Fn,2(1)F^{(1)}_{n,2} will not play any role in the analysis, and we can for example let Fn,2(1)=13F^{(1)}_{n,2}=\frac{1}{3} for completeness (in Table 1 we denote this value by γ\gamma).

When n12n\geq 12, it is easy to check that Fi,j(t)13F^{(t)}_{i,j}\leq\frac{1}{3} for all i[2n+1]i\in[2n+1], j{1,2,3}j\in\{1,2,3\} and t{1,2,3}t\in\{1,2,3\}. Therefore, we can let Fi,4(t)=1j{1,2,3}Fi,j(t)F^{(t)}_{i,4}=1-\sum_{j\in\{1,2,3\}}F^{(t)}_{i,j} for all i[2n+1]i\in[2n+1] and t{1,2,3,4}t\in\{1,2,3,4\} such that the forecast tensor 𝑭\bm{F} is well-defined. Finally, we specify the probability of agent types: Pr[agent type=1]=Pr[agent type=2]=12n2n\Pr[\textrm{agent type}=1]=\Pr[\textrm{agent type}=2]=\frac{1}{2n^{2n}} and Pr[agent type=3]=11n2n\Pr[\textrm{agent type}=3]=1-\frac{1}{n^{2n}}.

Type 11 Outcome 11 Outcome 22 Outcome 33 Outcome 44
Reward 11 Reward 0 Reward 11 Reward 0
Action i[n]i\in[n] 1nn(i1)\frac{1}{n^{n-(i-1)}} niinnnγ\frac{n^{i}-i}{n^{n}-n}\gamma 0 1j=13Fi,j(1)1-\sum_{j=1}^{3}F^{(1)}_{i,j}
Cost (nii)/nn+1(n^{i}-i)/n^{n+1}
Action n+i,i[n]n+i,i\in[n] 0 0 0 11
Cost 4(nii)/nn+14(n^{i}-i)/n^{n+1}
Action 2n+12n+1 0 0 0 11
Cost 0
Type 22 Outcome 11 Outcome 22 Outcome 33 Outcome 44
Reward 11 Reward 0 Reward 11 Reward 0
Action i[n]i\in[n] 0 0 0 11
Cost (nii)/nn+1(n^{i}-i)/n^{n+1}
Action n+i,i[n]n+i,i\in[n] 0 4nn11γ\frac{4}{n^{n-1}-1}\gamma 4nn(i1)\frac{4}{n^{n-(i-1)}} 1j=13Fn+i,j(2)1-\sum_{j=1}^{3}F^{(2)}_{n+i,j}
Cost 4(nii)/nn+14(n^{i}-i)/n^{n+1}
Action 2n+12n+1 0 4nn11γ\frac{4}{n^{n-1}-1}\gamma 0 1j=13F2n+1,j(2)1-\sum_{j=1}^{3}F^{(2)}_{2n+1,j}
Cost 0
Type 33 Outcome 11 Outcome 22 Outcome 33 Outcome 44
Reward 11 Reward 0 Reward 11 Reward 0
Action i[2n+1]i\in[2n+1] 0 0 0 11
Cost cic_{i}
Table 1: The Ω(n)\Omega(n) gap instance of Theorem 3.1. Here we can choose γ\gamma to be any (sufficiently small) positive constant; setting γ=1/3\gamma=1/3 ensures that this forms a valid typed principal-agent problem for all n12n\geq 12. Agent types occur with probabilities 1/(2n)2n1/(2n)^{2n}, 1/(2n)2n1/(2n)^{2n}, and 12/(2n)2n1-2/(2n)^{2n} respectively. Outcome 4 can be thought of as a null action, and type 3 exists solely to incentivize the principal to transfer no reward on Outcome 4 (Lemma A.1). The best single contract achieves a profit of O(n(n+1))O(n^{-(n+1)}), but the optimal menu (which contains two contracts, one transferring only on outcome 2, and one transferring only on outcome 3) achieves a profit of Ω(nn)\Omega(n^{-n}).

High-level idea

First of all, we include the zero-reward outcome 44 merely to satisfy the technical constraint that the probabilities over outcomes sum up to 11 for any type and action. Ideally, we would like to ignore the existence of outcome 44, but in principle, a contract can specify a strictly positive payment for outcome 44. Therefore, we introduce type 33, which has much higher probability than the other two types and always results in outcome 44 to make sure the payment for outcome 44 is negligible (Lemma A.1) and hence does not interfere with the analysis. This is the only role of type 33, and thus we will ignore outcome 44 and type 33 for the remainder of this discussion.

Essentially, the above Ω(n)\Omega(n)-gap instance is constructed such that the principal can only get low profit from type 22 agent regardless of the contract (Lemma A.2), and moreover, the only way to get high profit from type 11 agent is using a contract that associates a sufficiently low payment with outcome 11 and a sufficiently high payment with outcome 22 (Lemma A.3 and Lemma A.4), i.e., the contract must incentivize type 11 agent to play a high-profit action through a high payment for outcome 22. However, notice that in the Ω(n)\Omega(n)-gap instance, the type 22 agent also has significant probability for outcome 22 for every non-null action. Thus, if the contract specifies a high payment for outcome 22, it makes a high payment to the type 22 agent for almost nothing in return and hence achieves negative profit from the type 22 agent. Overall, this lets us show that the combined profit from these two types of agents is low (the soundness part of the proof of Theorem 3.1). This shows that any single contract has low profit.

On the other hand, notice that in this instance, outcome 3 has a strictly positive reward, and only type 22 agent can possibly cause outcome 3 to happen. Therefore, if we use a menu of two contracts, we can use the first contract to incentivize type 11 agent through a high payment for outcome 2 and use the second contract to lure type 22 agent away from the first contract through a high payment for outcome 3. We can show that the second contract can simultaneously (i) be as attractive as the first contract for type 22 agent and (ii) make zero profit (which is non-negative) from type 22 agent (the completeness part of the proof of Theorem 3.1). Thus, overall the menu achieves high profit (in particular, at least Ω(n)\Omega(n) times the profit of the best single contract).

3.2 Ω(logT)\Omega(\log T)-gap with two non-null actions

Theorem 3.2.

There is a principal-agent problem (𝐜,𝐅,𝐫)\left(\bm{c},\bm{F},\bm{r}\right) with TT agent types and n=2n=2 non-null actions, for which optimal deterministic menu can extract Ω(logT)\Omega(\log T) times as much profit as optimal single contract can:

Opt-DetMenu(𝒄,𝑭,𝒓)Ω(logT)Opt-Single(𝒄,𝑭,𝒓).\textsc{Opt-DetMenu}\left(\bm{c},\bm{F},\bm{r}\right)\geq\Omega(\log T)\cdot\textsc{Opt-Single}\left(\bm{c},\bm{F},\bm{r}\right).

We first give the construction of the Ω(logT)\Omega(\log T)-gap instances and explain the high-level idea. Then, we formally implement our idea through multiple lemmas and prove Theorem 3.2.

Type (k1)2N+(k-1)2^{N}+\ell Outcome 11 Outcome j{2,3,,T1}j\in\{2,3,\ldots,T-1\} Outcome TT
(k[N],[2N]k\in[N],\ell\in[2^{N}]) Reward 11 Reward 0 Reward 0
Action 11 12T(12k)\frac{1}{2T(1-2^{-k})} 12T(12k) if j=(k1)2N+\frac{1}{2T(1-2^{-k})}\text{ if }j=(k-1)2^{N}+\ell 11T(12k)1-\frac{1}{T(1-2^{-k})}
Cost 12T\frac{1}{2T}
Action 22 0 12T(12k)2k+1 if j(k1)2N+\frac{1}{2T(1-2^{-k})2^{k+1}}\text{ if }j\neq(k-1)2^{N}+\ell 1T32T(12k)2k+11-\frac{T-3}{2T(1-2^{-k})2^{k+1}}
Cost 0 0 otherwise
Type N2N+1N\cdot 2^{N}+1 Outcome 11 Outcome j{2,3,,T1}j\in\{2,3,\ldots,T-1\} Outcome TT
Reward 11 Reward 0 Reward 0
Action 11 0 0 11
Cost 12T\frac{1}{2T}
Action 22 0 0 11
Cost 0
Table 2: The Ω(logT)\Omega(\log T) gap instance of Theorem 3.2. Agent type (k1)2N+(k-1)2^{N}+\ell occurs with probability 2k1/16N2^{k-1}/16^{N} for k[N]k\in[N] and [2N]\ell\in[2^{N}], and the final agent type N2N+1N\cdot 2^{N}+1 occurs with the remaining probability, which is the vast majority. This final agent type serves to tax the final outcome so it cannot be used to a meaningful extent. A single contract has difficulty using Outcome 1 alone to incentivize action 1, because of the varied breakpoints, and if the single contract uses too many other outcomes instead to incentivize action 1, most types of agents would prefer playing action 2.

Construction of Ω(logT)\Omega(\log T)-gap instances

We begin by providing a formal description of the instance, which is also recorded in Table 2. For convenience we let T=N2N+1T=N\cdot 2^{N}+1 for N3N\geq 3 (and we will prove an Ω(N)\Omega(N)-gap). There are two non-null actions and TT outcomes. Action 11 has cost c1=12Tc_{1}=\frac{1}{2T}, and action 22 has cost c2=0c_{2}=0. Outcome 11 has reward r1=1r_{1}=1, and any other outcome j{2,3,,T+1}j\in\{2,3,\dots,T+1\} has reward rj=0r_{j}=0.

Now we construct the forecast tensor 𝑭\bm{F}. We start with defining 𝑭\bm{F} for outcomes [T1][T-1]. For all k[N]k\in[N] and [2N]\ell\in[2^{N}], first let F1,1((k1)2N+)=12T(12k)F^{((k-1)2^{N}+\ell)}_{1,1}=\frac{1}{2T(1-2^{-k})}, and moreover for j[T1]j\in[T-1] such that j(k1)2N+j\neq(k-1)2^{N}+\ell or 11, let F2,j((k1)2N+)=F1,1((k1)2N+)2k+1F^{((k-1)2^{N}+\ell)}_{2,j}=\frac{F^{((k-1)2^{N}+\ell)}_{1,1}}{2^{k+1}}. Then, let F1,t(t)=F1,1(t)F^{(t)}_{1,t}=F^{(t)}_{1,1} for all t[T1]t\in[T-1]. All the other Fi,j(t)F^{(t)}_{i,j} for t[T]t\in[T], i{1,2}i\in\{1,2\} and j[T1]j\in[T-1] that have not been defined are zero.

It is easy to check that Fi,j(t)1TF^{(t)}_{i,j}\leq\frac{1}{T} for all t[T]t\in[T], i{1,2}i\in\{1,2\} and j[T1]j\in[T-1]. Thus, we can let Fi,T(t)=1j[T1]Fi,j(t)F^{(t)}_{i,T}=1-\sum_{j\in[T-1]}F^{(t)}_{i,j} for all t[T]t\in[T] and i{1,2}i\in\{1,2\}, such that the forecast tensor 𝑭\bm{F} is well-defined.

Finally, we specify the probability of agent types: Pr[agent type=(k1)2N+]=2k116N\Pr[\textrm{agent type}=(k-1)2^{N}+\ell]=\frac{2^{k-1}}{16^{N}} for all k[N]k\in[N] and [2N]\ell\in[2^{N}], and Pr[agent type=T]=12N18N\Pr[\textrm{agent type}=T]=1-\frac{2^{N}-1}{8^{N}}.

High-level idea

Similar to the Ω(n)\Omega(n)-gap instance, here we also have an extra zero-reward outcome TT to make the forecast tensor well-defined, and we introduce type TT, which has much larger probability than other types and always results in outcome TT, in order to make sure that a contract with non-negative profit in expectation over type distribution can only make a negligible payment xTx_{T} for outcome TT (Lemma A.5), such that xTx_{T} does not interfere with the analysis. Thus, let us ignore outcome TT and type TT in the following discussion.

In the above Ω(logT)\Omega(\log T)-gap instance, only outcome 11 generates strictly positive reward, and hence, only action 11 generates strictly positive welfare for any type (because only action 11 has strictly positive probability for outcome 11). Moreover, the expected welfare generated by action 11 times the probability of agent type is roughly the same (up to a constant factor) for all types: That is, for all k[N]k\in[N] and [2N]\ell\in[2^{N}],

Pr[agent type=(k1)2N+](F1,1((k1)2N+)1c1)=14T16N(12k)[14T16N,12T16N].\Pr[\textrm{agent type}=(k-1)2^{N}+\ell]\cdot\left(F^{((k-1)2^{N}+\ell)}_{1,1}\cdot 1-c_{1}\right)=\frac{1}{4T\cdot 16^{N}(1-2^{-k})}\in\left[\frac{1}{4T\cdot 16^{N}},\frac{1}{2T\cdot 16^{N}}\right]. (2)

Therefore, to extract a constant fraction of the expected welfare of a random-type agent, the contract must incentivize many agent types tt to play action 11. For the type tt agent, the only two outcomes that action 11 can result into are outcomes 11 and tt.

Because the probability that action 11 results in outcome 11 is different between different types, the contract cannot incentivize different types using only payment for outcome 11. Specifically, a small payment for outcome 11 is not enough to cover the expected cost of action 11 for many types, and a large payment for outcome 11 would overpay many types and hence give up a lot of profit (Lemma A.6).

If we use a menu of contracts, we can have one contract 𝒙t\bm{x}_{t} for each agent type tt. Each contract has the same base payment for outcome 11, which by itself is not enough to incentivize action 11. Then, contract 𝒙t\bm{x}_{t} can incentivize the type tt agent to play action 11 by an additional payment through outcome tt. Therefore, such menu can extract a constant fraction of the expected welfare of the agent (the completeness part of the proof of Theorem 3.2).

However, if we adopt the above idea for the single contract rather than the menu, then, we need to make additional payments for many different outcomes in the single contract. In the above Ω(logT)\Omega(\log T)-gap instance, for the type tt agent, action 22 has zero cost and has strictly positive probabilities for all but outcomes 11 and tt. Therefore, when faced with a single contract with additional payments for many different outcomes, the type tt agent would prefer playing action 22, which generates zero profit for the principal. As a consequence, when we use a single contract, there cannot be many types of agents simultaneously preferring action 11 (Lemma A.7), and hence, overall the expected profit is low (the soundness part of the proof of Theorem 3.2).

Remark 3.3.

Under the assumption that the probability of every agent type with non-zero welfare is the same in the prior distribution of the agent type, Guruganesh et al. (2021) proved that linear contracts can achieve a Ω(1/(nlogT))\Omega(1/(n\log T)) fraction of the profit of the optimal deterministic menu. Thus, our Ω(logT)\Omega(\log T)-gap instance (where n=3n=3) is tight for this setting.

Proof.

Note that in the Ω(logT)\Omega(\log T)-gap instance, only agent types [T1][T-1] has non-zero welfare, but their probabilities in the prior distribution are different. However, we can construct a new instance with equal probabilities of agent types from the Ω(logT)\Omega(\log T)-gap instance by making copies of each agent type such that the number of copies is proportional to the probability of that agent type. That is, in the new instance, for each t[T1]t\in[T-1], we make 16NPr[agent type=t]16^{N}\cdot\Pr[\textrm{agent type}=t] (it is easy to check 16NPr[agent type=t]16^{N}\cdot\Pr[\textrm{agent type}=t] is an integer) copies of agent type tt, and we make 16N16^{N} copies of agent type TT. In the new instance, all the types have the same probability in the new prior distribution. Clearly, the analysis of Theorem 3.2 still works by replacing the probability of each agent type with the number of copies of that agent type. ∎

4 Menus of randomized linear contracts

We now switch our attention to menus involving linear contracts. As we have already mentioned in Section 2, simple menus of (deterministic) linear contracts are no more powerful than a single linear contract. Instead, our main object of study are menus of randomized linear contracts, formed by specializing the menus of randomized contracts in Castiglioni et al. (2022) to the special case of linear contracts.

Recall that a randomized linear contract is a probability distribution γ\gamma over 0\mathbb{R}_{\geq 0} representing all the possible linear contracts (including linear contracts with α>1\alpha>1, where the principal will pay out more than their entire reward). A menu of randomized linear contracts is defined by a tuple Γ=(γ(1),,γ(T))\Gamma=\left(\gamma^{(1)},\dots,\gamma^{(T)}\right) where each γ(i)\gamma^{(i)} is a distribution over the transfer coefficients. The interaction between the principal and an agent of type tt proceeds as follows:

  1. 1.

    The principal publicly commits to a menu Γ=(γ(1),,γ(T)).\Gamma=\left(\gamma^{(1)},\dots,\gamma^{(T)}\right).

  2. 2.

    The agent reports a type t^\hat{t} to the principal, possibly different from the true type t.t.

  3. 3.

    The principal draws a transfer coefficient αγ(t^)\alpha\sim\gamma^{(\hat{t})} and communicates it to the agent.

  4. 4.

    The agent plays the action that maximizes his utility when offered α\alpha of the expected reward.

The very first question we might ask ourselves is: are randomized linear contracts actually more powerful than individual linear contracts. The answer (in contrast to the situation with menus of deterministic linear contracts) is yes.

Before we present the example, we will establish some common notation for the rest of the section. Let Ut(α)U_{t}(\alpha) be the utility obtained by type tt when offered a linear contract with coefficient α\alpha. If action ii has cost cic_{i} and expected reward Ri(t)R^{(t)}_{i} for the principal, then in fact Ut(α)=maxi(Ri(t)αci)U_{t}(\alpha)=\max_{i}(R^{(t)}_{i}\alpha-c_{i}). From this we can observe that each UtU_{t} is a convex, piece-wise linear function with at most nn pieces. Since we are only dealing with menus of randomized linear contracts, it suffices to specify the slopes UtU^{\prime}_{t} of these piecewise linear functions to completely specify the problem instance.

Example 4.1 (Menus of Randomized Linear Contracts Differ From Linear Contracts).

Consider the following typed principal-agent problem with two types (and the underlying distribution is uniform), where ϵ>0\epsilon>0 is a small constant. Let the slopes of their utility functions be defined as follows:

U1(α)={0 if α[0,1ϵ)12ϵ1 if α[1ϵ,1]\displaystyle U_{1}^{\prime}(\alpha)=\begin{cases}0&\text{ if }\alpha\in[0,1-\epsilon)\\ \frac{1}{2\epsilon}-1&\text{ if }\alpha\in[1-\epsilon,1]\end{cases}\qquad U2(α)={12 if α[0,12)1 if α[12,1]\displaystyle U_{2}^{\prime}(\alpha)=\begin{cases}\frac{1}{2}&\text{ if }\alpha\in[0,\frac{1}{2})\\ 1&\text{ if }\alpha\in[\frac{1}{2},1]\end{cases}

This problem is designed so that any linear contract where α\alpha is a breakpoint (one of {0,1/2,1ϵ}\{0,1/2,1-\epsilon\}) yields a total utility of 1/21/2; linear contracts at non-breakpoints only get less utility. The following menu achieves slightly more than 1/21/2 utility:

γ(1)(α)={2/3 if α=01/3 if α=1ϵ0 otherwiseγ(2)(α)={1 if α=120 otherwise\displaystyle\gamma^{(1)}(\alpha)=\begin{cases}2/3&\text{ if }\alpha=0\\ 1/3&\text{ if }\alpha=1-\epsilon\\ 0&\text{ otherwise}\end{cases}\qquad\gamma^{(2)}(\alpha)=\begin{cases}1&\text{ if }\alpha=\frac{1}{2}\\ 0&\text{ otherwise}\end{cases}

Type one is indifferent between the two menu items because it has zero utility at all breakpoints. From type two’s point of view, its menu item has utility 1/41/4 and the other menu item has utility 13(14+12ϵ)<1/4\frac{1}{3}(\frac{1}{4}+\frac{1}{2}-\epsilon)<1/4. The first menu item extracts a profit of 1613ϵ\frac{1}{6}-\frac{1}{3}\epsilon from the first type. The second menu item extracts a profit of 12\frac{1}{2} from the second type. The total profit is 2313ϵ\frac{2}{3}-\frac{1}{3}\epsilon, which is better for ϵ<12\epsilon<\frac{1}{2}. This completes the example.

In the remainder of this section, we study the following three aspects of randomized linear contracts. First, in Section 4.1, we explore various structure of randomized linear contracts: e.g., How large must the support of γ(t)\gamma^{(t)} be in the optimal menu of randomized linear contracts? Is it necessary for the support to contain α>1\alpha>1? Second, in Section 4.2, we present an efficient algorithm for computing the optimal menu of randomized linear contracts. Finally, in Sections 4.4 and 4.3, we explore (and bound) the gap in power between menus of randomized linear contracts and linear contracts.

4.1 The structure of randomized linear contracts

We begin by establishing some basic structural properties of the randomized linear contracts in belonging to the optimal menu. In particular, we will show that all such contracts are supported on breakpoints α\alpha where an agent of one type would switch from playing one action to another.

Let pp be the total number of distinct breakpoints over all TT functions Ut(α)U_{t}(\alpha) (in particular, pnTp\leq nT). We will label these breakpoints as α1αp\alpha_{1}\dots\leq\alpha_{p}, and for convenience of notation write α0=0\alpha_{0}=0. The first claim we prove is that, without loss of generality, every optimal menu is supported on this set of breakpoints (including α0\alpha_{0}), and possibly one other point larger than αp\alpha_{p}.

Lemma 4.2.

For any principal-agent problem with TT types and nn actions, there exists an optimal menu Γ=(γ(1),,γ(T))\Gamma=(\gamma^{(1)},\dots,\gamma^{(T)}) of randomized linear contracts such that, for each t[T]t\in[T],

supp(γ(t)){α0,α1,,αp,αp+1t}\mathrm{supp}\left(\gamma^{(t)}\right)\subseteq\{\alpha_{0},\alpha_{1},\dots,\alpha_{p},\alpha^{t}_{p+1}\}

for some choice of αp+1t>αp\alpha^{t}_{p+1}>\alpha_{p} for each type tt. In other words, all randomized contracts in this menu only place weight on points of the form αi\alpha_{i} or one other (undetermined) point.

We defer the full proof to Appendix A.4 but present a sketch below.

Proof Sketch.

Note that any mass put on a break point α[αj,αj+1)\alpha\in[\alpha_{j},\alpha_{j+1}) can be moved to both neighboring breakpoints without affecting the constraints and only increasing the objective. For the probability mass put on the points above the last breakpoint, we can create a new point based on its conditional expectation in a way that maintains the constraints and improves the objective. ∎

One obvious peculiarity of the characterization in Lemma 4.2 is that it leaves open the option of placing mass above the largest breakpoint αp\alpha_{p}, without any bound on the size of αp+1t\alpha^{t}_{p+1}. In fact, a priori, there is nothing preventing the optimal menu from containing contracts supported on coefficients α\alpha greater than 11, even though by offering such a contract, the principal is guaranteed to lose utility. Even barring placing mass beyond αp\alpha_{p}, it could still be the case that some breakpoints αi\alpha_{i} themselves are greater than 11; these would normally correspond to actions that are not reasonable to incentivize with a single contract (they have negative net welfare for the principal and agent combined). Is it ever useful to incentivize these actions when offering a menu of randomized linear contracts?

Interestingly, the answer is (again) yes. We say that a randomized linear contract is bounded if always chooses a transfer coefficients bounded above by 11. Otherwise we say a randomized linear contract is unbounded. In the example below, we show that a menu of unbounded randomized linear contracts can extract strictly larger revenue for the principal than a menu of bounded randomized linear contracts.

Lemma 4.3.

There exists a principal-agent problem with T=2T=2 types and n=3n=3 actions such that the profit obtained by the optimal menu of bounded randomized linear contracts is strictly smaller than the revenue achieved by the optimal menu of unbounded randomized linear contracts.

We provide the instance below but defer the full proof to the appendix.

Proof Sketch.

Since we are only concerned with linear contracts, it suffices to specify the utility functions U1(α),U2(α):[0,)[0,)U_{1}(\alpha),U_{2}(\alpha):[0,\infty)\to[0,\infty). In fact it suffices to specify the derivatives Ui(α)U_{i}^{\prime}(\alpha) (since we always have that Ut(α)=0αUt(x)𝑑xU_{t}(\alpha)=\int_{0}^{\alpha}U^{\prime}_{t}(x)dx).

We specify the slopes below:

U1(α)={15 if α[0,23)14 if α[23,43)1 if α[43,)\displaystyle U_{1}^{\prime}(\alpha)=\begin{cases}\frac{1}{5}&\text{ if }\alpha\in[0,\frac{2}{3})\\ \frac{1}{4}&\text{ if }\alpha\in[\frac{2}{3},\frac{4}{3})\\ 1&\text{ if }\alpha\in[\frac{4}{3},\infty)\end{cases}\qquad U2(α)={120 if α[0,13)14 if α[13,23)720 if α[23,)\displaystyle U_{2}^{\prime}(\alpha)=\begin{cases}\frac{1}{20}&\text{ if }\alpha\in[0,\frac{1}{3})\\ \frac{1}{4}&\text{ if }\alpha\in[\frac{1}{3},\frac{2}{3})\\ \frac{7}{20}&\text{ if }\alpha\in[\frac{2}{3},\infty)\end{cases}

We show (computationally) that in the above instance, there exists a menu of unbounded randomized linear contracts which extracts a revenue strictly greater than 31/10031/100. On the other hand, the optimal menu of bounded randomized linear contracts extracts a revenue less than 31/10031/100. ∎

4.2 Computing the optimal menu of randomized linear contracts

We now switch our focus to the computational problem of efficiently computing the optimal menu of randomized linear contracts. Ultimately, we will show that we can find this optimal menu by solving a concise (polynomially-sized) linear program.

We begin by writing out a much larger program in terms of the variables γ(i)\gamma^{(i)} (where recall that each γ(i):[0,)[0,)\gamma^{(i)}:[0,\infty)\rightarrow[0,\infty) encodes a distribution over the positive reals via its pdf). The optimal menu will satisfy the following mathematical program:

supγ(1),,γ(T)i=1T\displaystyle\sup_{\gamma^{(1)},\dots,\gamma^{(T)}}\sum_{i=1}^{T} b=0Ui(b)(1b)γ(i)(b)𝑑b\displaystyle\int_{b=0}^{\infty}U^{\prime}_{i}(b)(1-b)\gamma^{(i)}(b)db (Rev)
b=0γ(i)(b)𝑑b=1i\displaystyle\int_{b=0}^{\infty}\gamma^{(i)}(b)db=1\qquad\forall i (Prob)
b=0Ui(b)(γ(i)(b)γ(i)(b))𝑑b0i,i\displaystyle\int_{b=0}^{\infty}U_{i}(b)(\gamma^{(i)}(b)-\gamma^{(i^{\prime})}(b))db\geq 0\qquad\forall i,i^{\prime} (IC)
γ(i)(b)0b0\displaystyle\gamma^{(i)}(b)\geq 0\qquad\forall b\in\mathbb{R}_{\geq 0}

The first constraint simply states that each γ(i)\gamma^{(i)} is a probability distribution. Equation IC is the incentive compatibility constraint which enforces type that tt will choose the contract γ(t)\gamma^{(t)}. Finally, the objective (Equation Rev) maximizes the expected revenue for the principal over all possible menus of randomized linear contracts.

In the following theorem, we show how to rewrite this program as a concise linear program (relying heavily on our characterization in Lemma 4.2 of the support of these distributions).

Theorem 4.4.

Given an instance principal agent problem with TT types and nn actions per type with break points α0=0αp\alpha_{0}=0\leq\dots\alpha_{p}, we can compute a menu Γ=(γ(1),,γ(T))\Gamma=(\gamma^{(1)},\dots,\gamma^{(T)}) which achieves the optimal revenue in time that is polynomial in poly(T,n,|F|)\mathrm{poly}(T,n,|F|) where |F||F| denotes the bit-complexity of the probability matrix.

Proof.

We can write a linear program which computes the optimal menu by essentially solving the mathematical program described above. Using Lemma 4.2, we know that each randomized linear contract will only put mass on the breakpoints and one additional point αp+1t\alpha^{t}_{p+1} above the last break point. One difficulty is that we don’t know the value (or even have a bound on) the final break point αp+1i.\alpha^{i}_{p+1}. To get around this, we instead track the variables zt:=αp+1iγp+1(i)z_{t}:=\alpha^{i}_{p+1}\cdot\gamma^{(i)}_{p+1} and γp+1(i)\gamma^{(i)}_{p+1} (i.e., the total mass above αp\alpha_{p} and the average coefficient).

We can therefore represent each menu as a vector (γ0(t),,γp(t),γp+1(t))(\gamma^{(t)}_{0},\dots,\gamma^{(t)}_{p},\gamma^{(t)}_{p+1}) where γi(t)\gamma^{(t)}_{i} denotes the probability of selecting a contract with transfer coefficient αi\alpha_{i} when ipi\leq p. γp+1(t)\gamma^{(t)}_{p+1} denotes the probability of choosing a point higher than the breakpoints. Rewriting the original mathematical program using the above simplification, we get the following linear program:

max\displaystyle\max ti=0pγi(t)Ut(αi)(1αi)+Ut(αp)(γp+1(t)zt)\displaystyle\sum_{t}\sum_{i=0}^{p}\gamma^{(t)}_{i}\cdot U^{\prime}_{t}(\alpha_{i})(1-\alpha_{i})+U^{\prime}_{t}(\alpha_{p})\left(\gamma^{(t)}_{p+1}-z^{t}\right) (Primal)
s.t.\displaystyle s.t. i=0pUt(αi)(γi(t)γi(t))+Ut(αp)(ztztαpγp+1(t)+αpγp+1(t))\displaystyle\sum_{i=0}^{p}U^{t}(\alpha_{i})\cdot(\gamma^{(t)}_{i}-\gamma^{(t^{\prime})}_{i})+U^{\prime}_{t}(\alpha_{p})\left(z_{t}-z_{t^{\prime}}-\alpha_{p}\cdot\gamma^{(t)}_{p+1}+\alpha_{p}\cdot\gamma^{(t^{\prime})}_{p+1}\right)
+Ut(αp)(γp+1(t)γp+1(t))\displaystyle{}+U_{t}(\alpha_{p})(\gamma^{(t)}_{p+1}-\gamma^{(t)}_{p+1}) 0\displaystyle\geq 0
αpγp+1(t)\displaystyle\alpha_{p}\cdot\gamma^{(t)}_{p+1} zt\displaystyle\leq z^{t}
i=0p+1γi(t)\displaystyle\sum_{i=0}^{p+1}\gamma^{(t)}_{i} =1\displaystyle=1

Furthermore, any solution the above linear program can be converted to an optimal menu Γ=(γ(1),,γ(T))\Gamma=(\gamma^{(1)},\dots,\gamma^{(T)}) with the same objective value. Simply let αp+1t=zt/γp+1(t)\alpha^{t}_{p+1}=z_{t}/\gamma^{(t)}_{p+1} and γp+1(t)\gamma^{(t)}_{p+1} denotes the probability that the contract for type tt plays the transfer coefficient αp+1t.\alpha^{t}_{p+1}. The resulting menu satisfies the original mathematical program and has the same revenue as the objective. ∎

One corollary of this proof is that it implies an upper bound on the maximum transfer coefficient (although not a particularly strong one).

Corollary 4.5.

There is an upper bound BB which is polynomial in the size of the (𝐜,𝐅,𝐫)\left(\bm{c},\bm{F},\bm{r}\right) such that the optimal menu will use a transfer coefficient whose size is at most BB.

Proof.

Since ztz_{t} and γp+1(t)\gamma^{(t)}_{p+1} are both variables in the linear program, we know that their size is bounded by poly(n,T,|F|).\mathrm{poly}(n,T,|F|). Finally, the breakpoint αp+1t\alpha^{t}_{p+1} is a ratio of these quantities and therefore is also bounded by poly(n,T,|F|).\mathrm{poly}(n,T,|F|).

Although we know that the largest transfer is bounded in bit complexity and can be computed in polynomial time, it remains an open question how powerful these menus of bounded randomized linear contracts really are.

Open Question: What is the gap in revenue between a menu of bounded randomized linear contracts when compared to menus of unbounded randomized linear contracts? What is the best upper bound on the value of the largest transfer coefficient in the support of any contract?

4.3 Upper bounds for menus of randomized linear contracts

While Theorem 4.7 demonstrates a super-constant gap between linear contracts and menus of randomized linear contracts, the instance in the proof require both the number of actions and the number of types to grow without bound. It is natural to ask how this gap depends on each of these parameters – the number of actions and the number of types – individually. For example:

Open Question: What is the dependence on the gap between randomized linear contracts and single linear contracts as a function of the number of actions nn (keeping the number of types fixed)? As a function of the number of types TT (keeping the number of actions fixed)?

We do not answer this question in generality – however, interestingly, we show that in the special case where the number of actions is fixed to 22 (one of the simplest non-trivial cases of the above question), the gap is bounded by a (universal) constant.

Theorem 4.6.

Suppose we have a principal-agent problem (𝐜,𝐅,𝐫)\left(\bm{c},\bm{F},\bm{r}\right){} with n=2n=2 actions (a null action and a non-null action). Then linear contracts can extract a constant fraction of the profit that randomized menus of linear contracts can:

Opt-Linear(𝒄,𝑭,𝒓)Ω(1)Opt-RndMenuLinear(𝒄,𝑭,𝒓).\displaystyle\textsc{Opt-Linear}\left(\bm{c},\bm{F},\bm{r}\right)\geq\Omega(1)\cdot\textsc{Opt-RndMenuLinear}\left(\bm{c},\bm{F},\bm{r}\right).
Proof Sketch.

We bucket the types by the breakpoint of α\alpha that shifts the agent to taking its non-null action. If a linear contract does not do well, then these breakpoints must be somewhat spread out. We then argue about the performance of a menu of randomized linear contracts; if it extracts a lot of profit from a type with high breakpoint then that menu item gives a lot of utility to a type with low breakpoint. If it repeatedly extracts a lot of profit over varied breakpoints, then eventually it must be playing breakpoints with greater than one probability, which is a contradiction.

We defer the full proof of Theorem 4.6 to Appendix A.7. ∎

4.4 Lower bound for menus of randomized linear contracts

We will now establish a lower bound on the gap in power between single linear contracts and menus of randomized linear contracts. Specifically, we will exhibit a family of instances with TT types and O(T)O(T) actions per type where this gap is at least Ω(T)\Omega(T). Note that since this gap is bounded above by O(T)O(T) (by simply choosing the best linear contract for one of the TT agents; see Lemma A.8), this lower bound is asymptotically tight in TT. We provide this construction in the following theorem.

Theorem 4.7.

There is a principal-agent problem (𝐜,𝐅,𝐫)\left(\bm{c},\bm{F},\bm{r}\right) with TT agent types and 2T+32T+3 actions per agent, for which the optimal menu of randomized linear contracts can extract Ω(T)\Omega(T) times as much profit as the optimal single linear contract:

Opt-RndMenuLinear(𝒄,𝑭,𝒓)Ω(T)Opt-Linear(𝒄,𝑭,𝒓).\textsc{Opt-RndMenuLinear}\left(\bm{c},\bm{F},\bm{r}\right)\geq\Omega(T)\cdot\textsc{Opt-Linear}\left(\bm{c},\bm{F},\bm{r}\right).
Proof.

Since we are only concerned with benchmarks that pertain to linear menus, to specify a typed principal-agent problem it suffices to specify the utility function Ut(α):[0,1]0U_{t}(\alpha):[0,1]\rightarrow\mathbb{R}_{\geq 0} for each agent t[T]t\in[T] and the distribution over types (which we will take to be uniform). Since we want each agent to have at most n=2T+3n=2T+3 actions, each function Ut(α)U_{t}(\alpha) must be a convex, increasing piecewise linear function with at most nn pieces. In our construction, all TT such functions will share the same breakpoints, which we label α1,α2,,αn1\alpha_{1},\alpha_{2},\dots,\alpha_{n-1} in increasing order. We will set αi=1ηi\alpha_{i}=1-\eta^{i}, where η=1/2T\eta=1/2T. For notational convenience, we’ll let α0=0\alpha_{0}=0 and αn=1\alpha_{n}=1.

For simplicity, it will be easier to work with the derivatives Ut(α)U^{\prime}_{t}(\alpha) of Ut(α)U_{t}(\alpha); note that since Ut(0)=0U_{t}(0)=0 for each type tt, each UtU_{t} is completely specified by UtU^{\prime}_{t} via the relation Ut(α)=0αUt(x)𝑑xU_{t}(\alpha)=\int_{0}^{\alpha}U^{\prime}_{t}(x)dx. Each Ut(α)U^{\prime}_{t}(\alpha) is an increasing piecewise constant function with the same breakpoints αi\alpha_{i} as the UtU_{t} functions. We will construct them from the following pieces: for each 0i<n0\leq i<n, define ui(α)u_{i}(\alpha) to be the piecewise constant function defined to equal

ui(α)=110n(αi+1αi)=ηi10n(1η),u_{i}(\alpha)=\frac{1}{10n}\cdot(\alpha_{i+1}-\alpha_{i})=\frac{\eta^{-i}}{10n(1-\eta)},

for α[αi,αi+1)\alpha\in[\alpha_{i},\alpha_{i+1}) and 0 otherwise. (Note: the second equality above is slightly inaccurate in the case where i=n1i=n-1, where the RHS should instead equal η(n1)/10n\eta^{-(n-1)}/10n; this will not affect any of the calculations that follow).

We now construct the functions UtU^{\prime}_{t} as follows. Each UtU^{\prime}_{t} will be a slight “perturbation” of the sum of all the functions uiu_{i}; in particular, we have that

Ut(α)=(i=0n1ui(α))+(Tut1(α)+T2ut(α))+(Tunt2(α)+T2unt1(α)).U^{\prime}_{t}(\alpha)=\left(\sum_{i=0}^{n-1}u_{i}(\alpha)\right)+\left(T\cdot u_{t-1}(\alpha)+\frac{T}{2}\cdot u_{t}(\alpha)\right)+\left(T\cdot u_{n-t-2}(\alpha)+\frac{T}{2}\cdot u_{n-t-1}(\alpha)\right). (3)

Note that each such function is indeed increasing (since η=1/2T\eta=1/2T, Tηi<η(i+1)T\eta^{-i}<\eta^{-(i+1)}) so these specify valid utility functions. One useful way of thinking about this example is that we divide our actions into two blocks of approximately TT actions each. In Ut(α)U_{t}(\alpha), each action corresponds to a linear piece with a fixed height of 1/10n1/10n (since αiαi+1ui(x)𝑑x=1/(10n)\int_{\alpha_{i}}^{\alpha_{i+1}}u_{i}(x)dx=1/(10n) is independent of ii), except for four of these actions, which have heights that are Θ(T)\Theta(T) times larger. Moreover, these four actions are chosen symmetrically; two consecutive actions from the first block of TT actions, and their “mirror image” from the second block of TT actions.

To prove a lower bound on the gap provided by this instance, we will begin by proposing a menu of randomized linear contracts, and show that each type is incentivized to choose a specific contract from this menu. Recall that γ(t)\gamma^{(t)} denotes the randomized linear contract that type tt will choose from the menu. We will let γ(t)\gamma^{(t)} be the uniform distribution of support size 22 that places 50%50\% of its weight on the point αt\alpha_{t} and the other 50%50\% of its weight on the point αnt1\alpha_{n-t-1}.

We now compute the utility achieved by this type (and other types) who receive this contract. To begin, note that

0αi(i=0n1ui(x))𝑑x=i10n.\int_{0}^{\alpha_{i}}\left(\sum_{i=0}^{n-1}u_{i}(x)\right)dx=\frac{i}{10n}.

In particular, this means that

𝔼αγ(t)[0α(i=0n1ui(x))𝑑x]=120n,\mathbb{E}_{\alpha\sim\gamma^{(t)}}\left[\int_{0}^{\alpha}\left(\sum_{i=0}^{n-1}u_{i}(x)\right)dx\right]=\frac{1}{20n},

or in other words, the total contribution to each agent’s utility from the ui(α)\sum u_{i}(\alpha) term in (3) is independent of the contract γ(t)\gamma^{(t)} they select. It thus suffices to only consider the latter four terms of (3).

Now, if the type tt agent selects contract γ(t)\gamma^{(t)}, they receive an extra 1.75T/10n1.75T/10n utility in expectation: when they draw the αt\alpha_{t} contract, they receive T/10nT/10n extra utility, and when they draw the αnt1\alpha_{n-t-1} contract, they receive (T+T/2+T)/10n(T+T/2+T)/10n extra utility. On the other hand, if the type tt agent selects contract γ(t)\gamma^{(t^{\prime})} for any ttt^{\prime}\neq t, we claim they only receive an extra 1.5T/10n1.5T/10n utility in expectation. Specifically, if t>tt^{\prime}>t, then they are guaranteed to receive (1.5T/10n)(1.5T/10n) utility from the first two terms of (3) but never receive any utility from the second two terms of (3) (regardless of which contract they draw). Similarly, if t<tt^{\prime}<t, then they receive no extra utility when they draw the αt\alpha_{t^{\prime}} contract, but receive 3T/10n3T/10n utility (from all four terms of (3)) when they draw the αnt1\alpha_{n-t^{\prime}-1} contract. This shows it is strictly dominant for an agent of type tt to select the contract γ(t)\gamma^{(t)}.

We now wish to examine the profit received by both i. this menu of randomized linear contracts and ii. the optimal single linear contract and show that the gap between these two values is indeed Ω(T)\Omega(T). We begin with the profit obtained by this menu. Recall that the profit Profit(t)(α)\textsc{Profit}^{(t)}(\alpha) obtained by offering a linear contract α\alpha to an agent of type tt is equal to (1α)Ut(α)(1-\alpha)U^{\prime}_{t}(\alpha). Now, note that (for any 0i<n10\leq i<n-1) that

(1αi)ui(αi)=ηiui(αi)=110n(1η).(1-\alpha_{i})u_{i}(\alpha_{i})=\eta^{i}u_{i}(\alpha_{i})=\frac{1}{10n(1-\eta)}. (4)

So, the expected profit obtained by offering contract γ(t)\gamma^{(t)} to type tt is given by

Profit(t)(γ(t))\displaystyle\textsc{Profit}^{(t)}(\gamma^{(t)}) =\displaystyle= 12(1αt)Ut(αt)+12(1αnt1)Ut(αnt1)\displaystyle\frac{1}{2}(1-\alpha_{t})U^{\prime}_{t}(\alpha_{t})+\frac{1}{2}(1-\alpha_{n-t-1})U^{\prime}_{t}(\alpha_{n-t-1})
=\displaystyle= (1+T2)110n(1η)\displaystyle\left(1+\frac{T}{2}\right)\cdot\frac{1}{10n(1-\eta)}
=\displaystyle= Ω(1).\displaystyle\Omega(1).

The total profit from this menu is therefore 1Tt=1TProfit(t)(γ(t))=Ω(1)\frac{1}{T}\sum_{t=1}^{T}\textsc{Profit}^{(t)}(\gamma^{(t)})=\Omega(1). On the other hand, if we offer a single linear contract α\alpha to all tt types, we receive profit

Profit(α)=(1α)1Tt=1TUt(α).\textsc{Profit}(\alpha)=(1-\alpha)\frac{1}{T}\sum_{t=1}^{T}U^{\prime}_{t}(\alpha).

But now, by (3), note that

1Tt=1TUt(α)52(i=0n1ui(α))\frac{1}{T}\sum_{t=1}^{T}U^{\prime}_{t}(\alpha)\leq\frac{5}{2}\left(\sum_{i=0}^{n-1}u_{i}(\alpha)\right)

But now, (1α)(i=0n1ui(α))(1-\alpha)\left(\sum_{i=0}^{n-1}u_{i}(\alpha)\right) is maximized at one of the αi\alpha_{i}, where by (4) it is at most 110n(1η)\frac{1}{10n(1-\eta)}. It follows that Profit(α)(5/2)/(10n(1η))=O(1/T)\textsc{Profit}(\alpha)\leq(5/2)/(10n(1-\eta))=O(1/T), and therefore that Opt-RndMenuLinear(𝒄,𝑭,𝒓)Ω(T)Opt-Single(𝒄,𝑭,𝒓)\textsc{Opt-RndMenuLinear}\left(\bm{c},\bm{F},\bm{r}\right)\geq\Omega(T)\cdot\textsc{Opt-Single}\left(\bm{c},\bm{F},\bm{r}\right).

As a corollary of Theorem 4.7, we can show the same gap between menus of deterministic general contracts (Opt-DetMenu) and menus of randomized general contracts (Opt-RndMenu).

Corollary 4.8.

There is a principal-agent problem (𝐜,𝐅,𝐫)\left(\bm{c},\bm{F},\bm{r}\right) with TT agent types and 2T+32T+3 actions per agent, for which the optimal menu of randomized (general) contracts can extract Ω(T)\Omega(T) times as much profit as the optimal single (general) contract:

Opt-RndMenu(𝒄,𝑭,𝒓)Ω(T)Opt-DetMenu(𝒄,𝑭,𝒓).\textsc{Opt-RndMenu}\left(\bm{c},\bm{F},\bm{r}\right)\geq\Omega(T)\cdot\textsc{Opt-DetMenu}\left(\bm{c},\bm{F},\bm{r}\right).

The proof of Corollary 4.8 is based on the observation that in principal-agent problems with m=2m=2 outcomes, all general contracts are very close to being linear contracts (in fact, they are affine contracts). We defer the proof to Appendix A.6.

References

  • Alon et al. (2021) T. Alon, P. Dütting, and I. Talgam-Cohen. Contracts with private cost per unit-of-effort. In Proceedings of the 22nd ACM Conference on Economics and Computation, pages 52–69, 2021.
  • Briest et al. (2010) P. Briest, S. Chawla, R. Kleinberg, and S. M. Weinberg. Pricing randomized allocations. In Proceedings of the twenty-first annual ACM-SIAM symposium on Discrete Algorithms, pages 585–597. SIAM, 2010.
  • Castiglioni et al. (2021) M. Castiglioni, A. Marchesi, and N. Gatti. Bayesian agency: Linear versus tractable contracts. In Proceedings of the 22nd ACM Conference on Economics and Computation, pages 285–286, 2021.
  • Castiglioni et al. (2022) M. Castiglioni, A. Marchesi, and N. Gatti. Designing menus of contracts efficiently: The power of randomization. In Proceedings of the 23rd ACM Conference on Economics and Computation, pages 705–735, 2022.
  • Cohen et al. (2022) A. Cohen, A. Deligkas, and M. Koren. Learning approximately optimal contracts. In Algorithmic Game Theory: 15th International Symposium, SAGT 2022, Colchester, UK, September 12–15, 2022, Proceedings, pages 331–346. Springer, 2022.
  • Dütting et al. (2019) P. Dütting, T. Roughgarden, and I. Talgam-Cohen. Simple versus optimal contracts. In Proceedings of the 20th ACM Conference on Economics and Computation, pages 369–387, 2019.
  • Dütting et al. (2020) P. Dütting, T. Roughgarden, and I. Talgam-Cohen. The complexity of contracts. In SODA, pages 2688–2707, 2020.
  • Dütting et al. (2021) P. Dütting, T. Ezra, M. Feldman, and T. Kesselheim. Combinatorial contracts. In FOCS, pages 815–826, 2021.
  • Gan et al. (2022) J. Gan, M. Han, J. Wu, and H. Xu. Optimal coordination in generalized principal-agent problems: A revisit and extensions. arXiv preprint arXiv:2209.01146, 2022.
  • Grossman and Hart (1983) S. J. Grossman and O. D. Hart. An analysis of the principal-agent problem. Econometrica, 51:7–45, 1983.
  • Guruganesh et al. (2021) G. Guruganesh, J. Schneider, and J. R. Wang. Contracts under moral hazard and adverse selection. In Proceedings of the 22nd ACM Conference on Economics and Computation, pages 563–582, 2021.
  • Holmström (1979) B. Holmström. Moral hazard and observability. Bell J. Econ., 10:74–91, 1979.

Appendix A Omitted proofs

A.1 Proof of Theorem 3.1

Here we provide the detailed proof of Theorem 3.1 (that there exists an Ω(n)\Omega(n) gap between single contracts and menus of deterministic contracts). We begin by establishing some lemmas.

Lemma A.1.

Consider the principal-agent problem (𝐜,𝐅,𝐫)\left(\bm{c},\bm{F},\bm{r}\right) defined in the above Ω(n)\Omega(n)-gap instance. For any contract 𝐱=(x1,x2,x3,x4)04\bm{x}=(x_{1},x_{2},x_{3},x_{4})\in\mathbb{R}_{\geq 0}^{4} that has non-negative profit, it must hold that x45n3nx_{4}\leq\frac{5}{n^{3n}}.

Proof.

First, note that for type 11 agent, only first nn actions can generate strict positive welfare, and the welfare of action i[n]i\in[n] is Fi,2(1)ci=inn+1F^{(1)}_{i,2}-c_{i}=\frac{i}{n^{n+1}}, and hence the maximum welfare of type 11 agent is 1nn\frac{1}{n^{n}} generated by action nn. For type 22 agent, only actions {n+1,,2n}\{n+1,\dots,2n\} can generated strict positive welfare, and the welfare of action n+in+i for i[n]i\in[n] is Fn+i,3(2)cn+i=4inn+1F^{(2)}_{n+i,3}-c_{n+i}=\frac{4i}{n^{n+1}}, and hence the maximum welfare of type 22 agent is 4nn\frac{4}{n^{n}} which is generated from action 2n2n. Therefore, the maximum possible profits a contract can get from type 11 agent and type 22 agent are 1nn\frac{1}{n^{n}} and 4nn\frac{4}{n^{n}} respectively.

Notice that regardless of which action type 33 agent plays, the result is always outcome 44 which has zero reward, and thus, the contract 𝒙=(x1,x2,x3,x4)\bm{x}=(x_{1},x_{2},x_{3},x_{4}) pays type 33 agent x4x_{4} for nothing in return.

Therefore, the profit in expectation over agent types of the contract 𝒙\bm{x} is at most Pr[agent type=1]1nn+Pr[agent type=2]4nnPr[agent type=3]x4\Pr[\textrm{agent type}=1]\cdot\frac{1}{n^{n}}+\Pr[\textrm{agent type}=2]\cdot\frac{4}{n^{n}}-\Pr[\textrm{agent type}=3]\cdot x_{4}, and for this to be non-negative, it must hold that

x4Pr[agent type=1]1nn+Pr[agent type=2]4nnPr[agent type=3]=12n2n1nn+12n2n4nn11n2n5n3n.x_{4}\leq\frac{\Pr[\textrm{agent type}=1]\cdot\frac{1}{n^{n}}+\Pr[\textrm{agent type}=2]\cdot\frac{4}{n^{n}}}{\Pr[\textrm{agent type}=3]}=\frac{\frac{1}{2n^{2n}}\cdot\frac{1}{n^{n}}+\frac{1}{2n^{2n}}\cdot\frac{4}{n^{n}}}{1-\frac{1}{n^{2n}}}\leq\frac{5}{n^{3n}}.

Because of Lemma A.1, we henceforth only consider contract 𝒙\bm{x} with x45n3nx_{4}\leq\frac{5}{n^{3n}} for the Ω(n)\Omega(n)-gap instance. As a consequence, if an action only results in outcome 44 for a type of agent, then the resulting payment to that agent is at most x45n3nx_{4}\leq\frac{5}{n^{3n}}, which is strictly less than the cost of any action i[2n]i\in[2n]. In the construction of Ω(n)\Omega(n)-gap instance, we mentioned that outcome 44 always occurs when type 11 agent plays the last n+1n+1 actions and when type 22 agent plays the first nn actions and when type 33 agent plays any action. Thus, we can henceforth assume that type 11 agent never plays actions {n+1,,2n}\{n+1,\dots,2n\}, and type 22 agent never plays actions [n][n], and type 33 agent never plays actions [2n][2n].

Lemma A.2.

Consider the principal-agent problem (𝐜,𝐅,𝐫)\left(\bm{c},\bm{F},\bm{r}\right) defined in the above Ω(n)\Omega(n)-gap instance. For any contract 𝐱\bm{x} with non-negative profit, we have

Profit(2)(𝒄,𝑭,𝒓,𝒙)9nn+1x2F2n+1,2(2).\textsc{Profit}^{(2)}(\bm{c},\bm{F},\bm{r},\bm{x})\leq\frac{9}{n^{n+1}}-x_{2}\cdot F^{(2)}_{2n+1,2}. (5)
Proof.

Consider any contract 𝒙=(x1,x2,x3,x4)04\bm{x}=(x_{1},x_{2},x_{3},x_{4})\in\mathbb{R}_{\geq 0}^{4}. First, since Fn+i,1(2)=0F^{(2)}_{n+i,1}=0 for all i[n+1]i\in[n+1], type 22 agent’s utility and the principal’s profit from type 22 agent are both independent of x1x_{1}, and thus, we can ignore outcome 1 and assume x1=0x_{1}=0 without loss of generality.

Moreover, since Fn+i,2(2)F^{(2)}_{n+i,2} is the same for all i[n+1]i\in[n+1], the expected payment type 22 agent gets from outcome 2 is x2x_{2} times a constant regardless of which non-null action type 22 agent plays. Hence, among all non-null actions (also note that action 2n+12n+1 is always no worse than the null action for type 22 agent), type 22 agent prefers the non-null action n+n+\ell that maximizes the expected payment for outcome 3 minus the cost of the action, i.e., =argmaxi[n+1]x3Fn+i,3(2)+x4Fn+i,4(2)ci\ell=\operatorname*{argmax}_{i\in[n+1]}x_{3}\cdot F^{(2)}_{n+i,3}+x_{4}\cdot F^{(2)}_{n+i,4}-c_{i}.

Note that \ell does not depend on x2x_{2} at all. Thus, type 22 agent will still prefers the same action when faced with contract (0,0,x3,x4)(0,0,x_{3},x_{4}) instead of (0,x2,x3,x4)(0,x_{2},x_{3},x_{4}). Therefore, it remains to prove that the profit which contract (0,0,x3,x4)(0,0,x_{3},x_{4}) gets from type 22 agent is at most 9nn+1\frac{9}{n^{n+1}} (which implies that the profit of contract (0,x2,x3,x4)(0,x_{2},x_{3},x_{4}) is at most 9nn+1x2F2n+1,2(2)\frac{9}{n^{n+1}}-x_{2}\cdot F^{(2)}_{2n+1,2}).

Henceforth, we further assume that

=argmaxi[n]x3Fn+i,3(2)+x4Fn+i,4(2)ci,\ell=\operatorname*{argmax}_{i\in[n]}x_{3}\cdot F^{(2)}_{n+i,3}+x_{4}\cdot F^{(2)}_{n+i,4}-c_{i}, (6)

i.e., we ignore action 2n+12n+1 without loss of generality, because if type 22 agent prefers action 2n+12n+1, then it generates zero reward and hence zero profit for the principal. Next, we show that for any x3,x4x_{3},x_{4}, the principal’s profit from type 22 agent is at most 9nn+1\frac{9}{n^{n+1}} when type 22 agent plays n+n+\ell.

We first prove a necessary condition for type 22 agent to play action n+n+\ell for {2,3,,n}\ell\in\{2,3,\dots,n\}: If type 22 agent prefers playing action n+n+\ell, then it must hold that x311nn154n2nx_{3}\geq 1-\frac{1}{n^{\ell}-n^{\ell-1}}-\frac{5}{4n^{2n}}.

To this end, note that if type 22 agent prefers playing action n+n+\ell, then in particular, type 22 agent prefers action n+n+\ell over n+1n+\ell-1, which by Eq. (6) implies that

x3Fn+,3(2)+x4Fn+,4(2)cx3Fn+1,3(2)+x4Fn+1,4(2)c1.x_{3}\cdot F^{(2)}_{n+\ell,3}+x_{4}\cdot F^{(2)}_{n+\ell,4}-c_{\ell}\geq x_{3}\cdot F^{(2)}_{n+\ell-1,3}+x_{4}\cdot F^{(2)}_{n+\ell-1,4}-c_{\ell-1}.

After rearrangement, this is equivalent to

x3\displaystyle x_{3} cc1+x4(Fn+1,4(2)Fn+,4(2))Fn+,3(2)Fn+1,3(2)\displaystyle\geq\frac{c_{\ell}-c_{\ell-1}+x_{4}\cdot(F^{(2)}_{n+\ell-1,4}-F^{(2)}_{n+\ell,4})}{F^{(2)}_{n+\ell,3}-F^{(2)}_{n+\ell-1,3}}
cc1x4Fn+,3(2)Fn+1,3(2)\displaystyle\geq\frac{c_{\ell}-c_{\ell-1}-x_{4}}{F^{(2)}_{n+\ell,3}-F^{(2)}_{n+\ell-1,3}} (Fn+1,4(2),Fn+,4(2)[0,1]F^{(2)}_{n+\ell-1,4},F^{(2)}_{n+\ell,4}\in[0,1])
cc15n3nFn+,3(2)Fn+1,3(2)\displaystyle\geq\frac{c_{\ell}-c_{\ell-1}-\frac{5}{n^{3n}}}{F^{(2)}_{n+\ell,3}-F^{(2)}_{n+\ell-1,3}} (By Lemma A.1)
=4(nn1+1)nn+15n3n4(nn1)nn+1\displaystyle=\frac{\frac{4(n^{\ell}-n^{\ell-1}+1)}{n^{n+1}}-\frac{5}{n^{3n}}}{\frac{4(n^{\ell}-n^{\ell-1})}{n^{n+1}}} (By definition of Fn+,3(2),Fn+1,3(2),c,c1F^{(2)}_{n+\ell,3},F^{(2)}_{n+\ell-1,3},c_{\ell},c_{\ell-1})
11nn154n2n\displaystyle\geq 1-\frac{1}{n^{\ell}-n^{\ell-1}}-\frac{5}{4n^{2n}} (nn1n for any {2,3,,n}).\displaystyle\text{($n^{\ell}-n^{\ell-1}\geq n$ for any $\ell\in\{2,3,\dots,n\}$)}.

If =1\ell=1, type 22 agent plays action n+1n+1, which generates welfare Fn+1,3(2)1cn+1=4nn+1F^{(2)}_{n+1,3}\cdot 1-c_{n+1}=\frac{4}{n^{n+1}}, and obviously, the principal’s profit from type 22 agent cannot exceed this. Hence, we consider the case when type 22 agent plays action n+n+\ell for {2,3,,n}\ell\in\{2,3,\dots,n\}. In this case, as we have shown, the necessary condition for type 22 agent to prefer action n+n+\ell is x311nn154n2nx_{3}\geq 1-\frac{1}{n^{\ell}-n^{\ell-1}}-\frac{5}{4n^{2n}}. The principal’s profit from type 22 agent is at most (1x3)Fn+,3(2)=(1x3)4nn+1(1-x_{3})\cdot F^{(2)}_{n+\ell,3}=(1-x_{3})\cdot\frac{4}{n^{n-\ell+1}}, which is at most (1nn1+54n2n)4nn+1=4nn+1nn+5n3n+18nn+1+1nn+1=9nn+1(\frac{1}{n^{\ell}-n^{\ell-1}}+\frac{5}{4n^{2n}})\cdot\frac{4}{n^{n-\ell+1}}=\frac{4}{n^{n+1}-n^{n}}+\frac{5}{n^{3n-\ell+1}}\leq\frac{8}{n^{n+1}}+\frac{1}{n^{n+1}}=\frac{9}{n^{n+1}}. ∎

Lemma A.3.

Consider the principal-agent problem (𝐜,𝐅,𝐫)\left(\bm{c},\bm{F},\bm{r}\right) defined in the above Ω(n)\Omega(n)-gap instance. For any contract 𝐱=(x1,x2,x3,x4)04\bm{x}=(x_{1},x_{2},x_{3},x_{4})\in\mathbb{R}_{\geq 0}^{4} with non-negative profit, we have

Profit(1)(𝒄,𝑭,𝒓,𝒙)3nn+1+2(1x1)nn.\textsc{Profit}^{(1)}(\bm{c},\bm{F},\bm{r},\bm{x})\leq\frac{3}{n^{n+1}}+\frac{2(1-x_{1})}{n^{n}}. (7)
Proof.

Consider any contract 𝒙=(x1,x2,x3,x4)04\bm{x}=(x_{1},x_{2},x_{3},x_{4})\in\mathbb{R}_{\geq 0}^{4}. Since Fi,3(1)=0F^{(1)}_{i,3}=0 for all i[n]i\in[n], type 11 agent’s utility and the principal’s profit from type 11 agent are both independent of x3x_{3}, and thus, we can ignore outcome 3 and assume x3=0x_{3}=0 without loss of generality. Also, notice that outcome 22 and 44 have zero reward and hence do not contribute to the principal’s profit.

Given contract 𝒙\bm{x}, let \ell be the action that type 11 agent prefers. Without loss of generality, we assume that \ell is not the null action or action 2n+12n+1 (because otherwise Eq. (7) holds trivially). Notice that the welfare generated by action \ell is F,1(1)1cF^{(1)}_{\ell,1}\cdot 1-c_{\ell}, which is equal to nn+1\frac{\ell}{n^{n+1}} by definition of F,1(1),cF^{(1)}_{\ell,1},c_{\ell}, and we define δ[0,1]\delta\in[0,1] such that Profit(1)(𝒄,𝑭,𝒓,𝒙)=δnn+1\textsc{Profit}^{(1)}(\bm{c},\bm{F},\bm{r},\bm{x})=\frac{\delta\ell}{n^{n+1}}. Hence, our goal is to upper bound δnn+1\frac{\delta\ell}{n^{n+1}}. Without loss of generality, we also assume that 2\ell\geq 2 (because the welfare generated by action 11 is F1,1(1)1c1=1nn+1F^{(1)}_{1,1}\cdot 1-c_{1}=\frac{1}{n^{n+1}}, and the principal’s profit from type 11 agent cannot exceed this).

Since type 11 agent plays action \ell given contract 𝒙\bm{x}, we have that Profit(1)(𝒄,𝑭,𝒓,𝒙)=(1x1)F,1(1)x2F,2(1)x4F,4(1)\textsc{Profit}^{(1)}(\bm{c},\bm{F},\bm{r},\bm{x})=(1-x_{1})\cdot F^{(1)}_{\ell,1}-x_{2}\cdot F^{(1)}_{\ell,2}-x_{4}\cdot F^{(1)}_{\ell,4}. Since Profit(1)(𝒄,𝑭,𝒓,𝒙)=δnn+1\textsc{Profit}^{(1)}(\bm{c},\bm{F},\bm{r},\bm{x})=\frac{\delta\ell}{n^{n+1}} by definition of δ\delta, it follows that

x2F,2(1)=(1x1)F,1(1)x4F,4(1)δnn+1.x_{2}\cdot F^{(1)}_{\ell,2}=(1-x_{1})\cdot F^{(1)}_{\ell,1}-x_{4}\cdot F^{(1)}_{\ell,4}-\frac{\delta\ell}{n^{n+1}}. (8)

Now we derive the utility generated by any action i[n]i\in[n] for type 1 agent (denoted by ui(1)u_{i}^{(1)}):

ui(1)=\displaystyle u_{i}^{(1)}= x1Fi,1(1)+x2Fi,2(1)+x4Fi,4(1)ci\displaystyle x_{1}\cdot F^{(1)}_{i,1}+x_{2}\cdot F^{(1)}_{i,2}+x_{4}\cdot F^{(1)}_{i,4}-c_{i}
=\displaystyle= Fi,1(1)ci(1x1)Fi,1(1)+x2Fi,2(1)+x4Fi,4(1)\displaystyle F^{(1)}_{i,1}-c_{i}-(1-x_{1})\cdot F^{(1)}_{i,1}+x_{2}\cdot F^{(1)}_{i,2}+x_{4}\cdot F^{(1)}_{i,4}
=\displaystyle= inn+11x1nn+1i+x2Fi,2(1)+x4Fi,4(1)\displaystyle\frac{i}{n^{n+1}}-\frac{1-x_{1}}{n^{n+1-i}}+x_{2}\cdot F^{(1)}_{i,2}+x_{4}\cdot F^{(1)}_{i,4} (By definition of Fi,1(1),ciF^{(1)}_{i,1},c_{i})
=\displaystyle= inn+11x1nn+1i+x2F,2(1)niin+x4Fi,4(1)\displaystyle\frac{i}{n^{n+1}}-\frac{1-x_{1}}{n^{n+1-i}}+x_{2}\cdot F^{(1)}_{\ell,2}\cdot\frac{n^{i}-i}{n^{\ell}-\ell}+x_{4}\cdot F^{(1)}_{i,4} (By definition of Fi,2(1),F,2(1)F^{(1)}_{i,2},F^{(1)}_{\ell,2})
=\displaystyle= inn+11x1nn+1i+niin((1x1)F,1(1)δnn+1)\displaystyle\frac{i}{n^{n+1}}-\frac{1-x_{1}}{n^{n+1-i}}+\frac{n^{i}-i}{n^{\ell}-\ell}\cdot\left((1-x_{1})\cdot F^{(1)}_{\ell,1}-\frac{\delta\ell}{n^{n+1}}\right)
+x4(Fi,4(1)niinF,4(1))\displaystyle\qquad\qquad\qquad\qquad+x_{4}\cdot(F^{(1)}_{i,4}-\frac{n^{i}-i}{n^{\ell}-\ell}\cdot F^{(1)}_{\ell,4}) (By Eq. (8)).\displaystyle\text{(By Eq.\leavevmode\nobreak\ \eqref{eq:x_2_for_delta_Omega_n_gap})}. (9)

Using Eq. (9), we calculate the the utility generated by action \ell for type 1 agent:

u(1)=\displaystyle u_{\ell}^{(1)}= nn+11x1nn+1+((1x1)F,1(1)δnn+1)\displaystyle\frac{\ell}{n^{n+1}}-\frac{1-x_{1}}{n^{n+1-\ell}}+\left((1-x_{1})\cdot F^{(1)}_{\ell,1}-\frac{\delta\ell}{n^{n+1}}\right) (By Eq. (9))
=\displaystyle= (1δ)nn+1\displaystyle(1-\delta)\cdot\frac{\ell}{n^{n+1}} (By definition of F,1(1)),\displaystyle\text{(By definition of $F^{(1)}_{\ell,1}$)},

and we lower bound the the utility generated by action 1\ell-1 for type 1 agent:

u1(1)x4(F1,4(1)n1+1nF,4(1))\displaystyle u_{\ell-1}^{(1)}-x_{4}\cdot(F^{(1)}_{\ell-1,4}-\frac{n^{\ell-1}-\ell+1}{n^{\ell}-\ell}\cdot F^{(1)}_{\ell,4})
=\displaystyle= 1nn+11x1nn+2+n1+1n((1x1)F,1(1)δnn+1)\displaystyle\frac{\ell-1}{n^{n+1}}-\frac{1-x_{1}}{n^{n+2-\ell}}+\frac{n^{\ell-1}-\ell+1}{n^{\ell}-\ell}\cdot\left((1-x_{1})\cdot F^{(1)}_{\ell,1}-\frac{\delta\ell}{n^{n+1}}\right) (By Eq. (9))
=\displaystyle= 1nn+11x1nn+2+1n(1n(1)n)((1x1)F,1(1)δnn+1)\displaystyle\frac{\ell-1}{n^{n+1}}-\frac{1-x_{1}}{n^{n+2-\ell}}+\frac{1}{n}\cdot\left(1-\frac{n(\ell-1)-\ell}{n^{\ell}-\ell}\right)\cdot\left((1-x_{1})\cdot F^{(1)}_{\ell,1}-\frac{\delta\ell}{n^{n+1}}\right)
=\displaystyle= 1nn+1n(1)n1x1nn+21n(1n(1)n)δnn+1\displaystyle\frac{\ell-1}{n^{n+1}}-\frac{n(\ell-1)-\ell}{n^{\ell}-\ell}\cdot\frac{1-x_{1}}{n^{n+2-\ell}}-\frac{1}{n}\cdot\left(1-\frac{n(\ell-1)-\ell}{n^{\ell}-\ell}\right)\cdot\frac{\delta\ell}{n^{n+1}} (By definition of F,1(1)F^{(1)}_{\ell,1})
\displaystyle\geq 1nn+1n(1)n1x1nn+21nδnn+1\displaystyle\frac{\ell-1}{n^{n+1}}-\frac{n(\ell-1)-\ell}{n^{\ell}-\ell}\cdot\frac{1-x_{1}}{n^{n+2-\ell}}-\frac{1}{n}\cdot\frac{\delta\ell}{n^{n+1}} (n(1)n1\frac{n(\ell-1)-\ell}{n^{\ell}-\ell}\leq 1 as 2\ell\geq 2)
\displaystyle\geq 1nn+1n(1)n1x1nn+21nn+1\displaystyle\frac{\ell-1}{n^{n+1}}-\frac{n(\ell-1)-\ell}{n^{\ell}-\ell}\cdot\frac{1-x_{1}}{n^{n+2-\ell}}-\frac{1}{n^{n+1}} (n\ell\leq n)
=\displaystyle= 2nn+1(1+n)(1n)1x1nn+1\displaystyle\frac{\ell-2}{n^{n+1}}-\left(1+\frac{\ell}{n^{\ell}-\ell}\right)\cdot\left(\ell-1-\frac{\ell}{n}\right)\cdot\frac{1-x_{1}}{n^{n+1}}
\displaystyle\geq 2nn+1(1+n)1x1nn\displaystyle\frac{\ell-2}{n^{n+1}}-\left(1+\frac{\ell}{n^{\ell}-\ell}\right)\cdot\frac{1-x_{1}}{n^{n}} (1nn\ell-1-\frac{\ell}{n}\leq n)
\displaystyle\geq 2nn+121x1nn\displaystyle\frac{\ell-2}{n^{n+1}}-2\cdot\frac{1-x_{1}}{n^{n}} (2n).\displaystyle\text{($2\ell\leq n^{\ell}$)}.

Recall that \ell is type 11 agent’s favorite action, and hence, for type 11 agent,

u1(1)u(1),u_{\ell-1}^{(1)}\leq u_{\ell}^{(1)},

which implies that 2nn+121x1nn+x4(F1,4(1)n1+1nF,4(1))(1δ)nn+1\frac{\ell-2}{n^{n+1}}-2\cdot\frac{1-x_{1}}{n^{n}}+x_{4}\cdot(F^{(1)}_{\ell-1,4}-\frac{n^{\ell-1}-\ell+1}{n^{\ell}-\ell}\cdot F^{(1)}_{\ell,4})\leq(1-\delta)\cdot\frac{\ell}{n^{n+1}}. After rearrangement, this is equivalent to

δnn+1\displaystyle\frac{\delta\ell}{n^{n+1}} 2nn+1+2(1x1)nnx4(F1,4(1)n1+1nF,4(1))\displaystyle\leq\frac{2}{n^{n+1}}+\frac{2(1-x_{1})}{n^{n}}-x_{4}\cdot\left(F^{(1)}_{\ell-1,4}-\frac{n^{\ell-1}-\ell+1}{n^{\ell}-\ell}\cdot F^{(1)}_{\ell,4}\right)
2nn+1+2(1x1)nn+x4n1+1n\displaystyle\leq\frac{2}{n^{n+1}}+\frac{2(1-x_{1})}{n^{n}}+x_{4}\cdot\frac{n^{\ell-1}-\ell+1}{n^{\ell}-\ell} (F1,4(1),F,4(1)[0,1]F^{(1)}_{\ell-1,4},F^{(1)}_{\ell,4}\in[0,1])
2nn+1+2(1x1)nn+x4\displaystyle\leq\frac{2}{n^{n+1}}+\frac{2(1-x_{1})}{n^{n}}+x_{4}
3nn+1+2(1x1)nn\displaystyle\leq\frac{3}{n^{n+1}}+\frac{2(1-x_{1})}{n^{n}} (By Lemma A.1).\displaystyle\text{(By Lemma\leavevmode\nobreak\ \ref{lemma:x4_upper_bound_Omega_n_gap})}.

Lemma A.4.

Consider the principal-agent problem (𝐜,𝐅,𝐫)\left(\bm{c},\bm{F},\bm{r}\right) defined in the above Ω(n)\Omega(n)-gap instance. For any contract 𝐱=(x1,x2,x3,x4)04\bm{x}=(x_{1},x_{2},x_{3},x_{4})\in\mathbb{R}_{\geq 0}^{4} with x1=1nαx_{1}=1-n^{-\alpha} for any α1\alpha\leq 1, if there is an action {2,3,,n}\ell\in\{2,3,\dots,n\} that generates non-negative utility for type 11 agent, then it must hold that

x2nn11Fn,2(1)(12nn+α5n3n).x_{2}\geq\frac{n^{n-1}-1}{F^{(1)}_{n,2}}\cdot\left(\frac{1}{2n^{n+\alpha}}-\frac{5}{n^{3n}}\right). (10)
Proof.

Consider any contract 𝒙=(x1,x2,x3,x4)04\bm{x}=(x_{1},x_{2},x_{3},x_{4})\in\mathbb{R}_{\geq 0}^{4} with x1=1nαx_{1}=1-n^{-\alpha} with α1\alpha\leq 1. The utility of action {2,3,,n}\ell\in\{2,3,\dots,n\} for type 11 agent is F,1(1)x1+F,2(1)x2+F,4(1)x4cF^{(1)}_{\ell,1}\cdot x_{1}+F^{(1)}_{\ell,2}\cdot x_{2}+F^{(1)}_{\ell,4}\cdot x_{4}-c_{\ell}. For this to be non-negative, it must hold that

x2\displaystyle x_{2} cF,1(1)+F,1(1)(1x1)F,4(1)x4F,2(1)\displaystyle\geq\frac{c_{\ell}-F^{(1)}_{\ell,1}+F^{(1)}_{\ell,1}\cdot(1-x_{1})-F^{(1)}_{\ell,4}\cdot x_{4}}{F^{(1)}_{\ell,2}}
=+nαF,2(1)1nn+1F,4(1)x4F,2(1)\displaystyle=\frac{-\ell+n^{\ell-\alpha}}{F^{(1)}_{\ell,2}}\cdot\frac{1}{n^{n+1}}-\frac{F^{(1)}_{\ell,4}\cdot x_{4}}{F^{(1)}_{\ell,2}} (By definition of c,F,1(1),x1c_{\ell},F^{(1)}_{\ell,1},x_{1})
=nn11Fn,2(1)(nαn1nnF,4(1)x4n(n))\displaystyle=\frac{n^{n-1}-1}{F^{(1)}_{n,2}}\cdot\left(\frac{n^{\ell-\alpha}-\ell}{n^{\ell}-\ell}\cdot\frac{1}{n^{n}}-\frac{F^{(1)}_{\ell,4}\cdot x_{4}\cdot n}{(n^{\ell}-\ell)}\right) (By definition of F,2(1)F^{(1)}_{\ell,2})
nn11Fn,2(1)(nαn1nnx4)\displaystyle\geq\frac{n^{n-1}-1}{F^{(1)}_{n,2}}\cdot\left(\frac{n^{\ell-\alpha}-\ell}{n^{\ell}-\ell}\cdot\frac{1}{n^{n}}-x_{4}\right) (F,4(1)1F^{(1)}_{\ell,4}\leq 1 and nnn^{\ell}-\ell\geq n for 2\ell\geq 2)
nn11Fn,2(1)(nαn1nn5n3n)\displaystyle\geq\frac{n^{n-1}-1}{F^{(1)}_{n,2}}\cdot\left(\frac{n^{\ell-\alpha}-\ell}{n^{\ell}-\ell}\cdot\frac{1}{n^{n}}-\frac{5}{n^{3n}}\right) (By Lemma A.1).\displaystyle\text{(By Lemma\leavevmode\nobreak\ \ref{lemma:x4_upper_bound_Omega_n_gap})}.

It remains to prove that nαnnα2\frac{n^{\ell-\alpha}-\ell}{n^{\ell}-\ell}\geq\frac{n^{-\alpha}}{2}. To this end, note that n12n^{\ell-1}\geq 2\ell for 2\ell\geq 2 and n12n\geq 12, and we derive that

nαn\displaystyle\frac{n^{\ell-\alpha}-\ell}{n^{\ell}-\ell} =nα/n1/n\displaystyle=\frac{n^{-\alpha}-\ell/n^{\ell}}{1-\ell/n^{\ell}}
nαn1/21/n\displaystyle\geq\frac{n^{-\alpha}-n^{-1}/2}{1-\ell/n^{\ell}} (By n12n^{\ell-1}\geq 2\ell)
nαn12\displaystyle\geq n^{-\alpha}-\frac{n^{-1}}{2}
nα2\displaystyle\geq\frac{n^{-\alpha}}{2} (By α1).\displaystyle\text{(By $\alpha\leq 1$)}.

We can now proceed to prove the main theorem.

Proof of Theorem 3.1.

Soundness. We prove that any single contract 𝒙=(x1,x2,x3,x4)04\bm{x}=(x_{1},x_{2},x_{3},x_{4})\in\mathbb{R}_{\geq 0}^{4} can only get total profit at most 14nn+1\frac{14}{n^{n+1}} from first two types of agents (recall type 33 agent generates zero reward). By Lemma A.2, contract 𝒙\bm{x} can only get profit at most 9nn+1\frac{9}{n^{n+1}} from type 22 agent. By Lemma A.3, if x111nx_{1}\geq 1-\frac{1}{n}, then contract 𝒙\bm{x} can only get profit at most 5nn+1\frac{5}{n^{n+1}} from type 11 agent, and thus, the total profit from two types of agents is at most 14nn+1\frac{14}{n^{n+1}}.

Therefore, it remains to analyze the case x1=1nαx_{1}=1-n^{-\alpha} for α1\alpha\leq 1. In this case, we can without loss of generality assume that type 11 agent prefers an action {2,,n}\ell\in\{2,\dots,n\}, because otherwise type 11 agent prefers either the null action (which generates zero profit) or action 11 (which generates F1,1(1)1c1=1nn+1F^{(1)}_{1,1}\cdot 1-c_{1}=\frac{1}{n^{n+1}} welfare and hence at most 1nn+1\frac{1}{n^{n+1}} profit). Since type 11 agent prefers action {2,,n}\ell\in\{2,\dots,n\} over the null action, action \ell must have non-negative utility, and it follows by Lemma 10 that Ineq. (10) holds. Combining Ineq. (5) in Lemma A.2 with Ineq. (10), we have that

Profit(2)(𝒄,𝑭,𝒓,𝒙)\displaystyle\textsc{Profit}^{(2)}(\bm{c},\bm{F},\bm{r},\bm{x}) 9nn+1F2n+1,2(2)(nn11)Fn,2(1)(12nn+α5n3n)\displaystyle\leq\frac{9}{n^{n+1}}-\frac{F^{(2)}_{2n+1,2}\cdot(n^{n-1}-1)}{F^{(1)}_{n,2}}\cdot\left(\frac{1}{2n^{n+\alpha}}-\frac{5}{n^{3n}}\right)
=9nn+1(2nn+α20n3n)\displaystyle=\frac{9}{n^{n+1}}-\left(\frac{2}{n^{n+\alpha}}-\frac{20}{n^{3n}}\right) (By definition of F2n+1,2(2)F^{(2)}_{2n+1,2})
10nn+12nn+α.\displaystyle\leq\frac{10}{n^{n+1}}-\frac{2}{n^{n+\alpha}}.

On the other hand, by Lemma A.3, we have that

Profit(1)(𝒄,𝑭,𝒓,𝒙)\displaystyle\textsc{Profit}^{(1)}(\bm{c},\bm{F},\bm{r},\bm{x}) 3nn+1+2(1x1)nn\displaystyle\leq\frac{3}{n^{n+1}}+\frac{2(1-x_{1})}{n^{n}}
=3nn+1+2nn+α.\displaystyle=\frac{3}{n^{n+1}}+\frac{2}{n^{n+\alpha}}.

Hence, the total profit of contract 𝒙\bm{x} from first two types of agents is at most 13nn+1\frac{13}{n^{n+1}} in this case.

Completeness. We show a menu of two contracts that gets total profit 1nn\frac{1}{n^{n}} from first two types of agents and zero profit from type 33 agent. The two contracts are 𝒙1=(0,nn11nnFn,2(1),0,0)\bm{x}_{1}=(0,\frac{n^{n-1}-1}{n^{n}\cdot F^{(1)}_{n,2}},0,0) and 𝒙2=(0,0,1,0)\bm{x}_{2}=(0,0,1,0). First, it is easy to check that given contract 𝒙1\bm{x}_{1}, every action among [n][n] has zero utility for type 11 agent (i.e., Fi,2(1)nn11nnFn,2(1)=ciF^{(1)}_{i,2}\cdot\frac{n^{n-1}-1}{n^{n}\cdot F^{(1)}_{n,2}}=c_{i} for all i[n]i\in[n]), and given contract 𝒙2\bm{x}_{2}, none of the actions can make strictly positive utility for type 11 agent. Thus, we can assume that type 11 agent chooses contract 𝒙1\bm{x}_{1} and plays action nn, and then the principal’s profit from type 11 agent is Fn,1(1)1Fn,2(1)nn11nnFn,2(1)=1nnF^{(1)}_{n,1}\cdot 1-F^{(1)}_{n,2}\cdot\frac{n^{n-1}-1}{n^{n}\cdot F^{(1)}_{n,2}}=\frac{1}{n^{n}}.

On the other hand, type 22 agent will get utility F2n,3(2)1c2n=4nnF^{(2)}_{2n,3}\cdot 1-c_{2n}=\frac{4}{n^{n}} by choosing contract 𝒙2\bm{x}_{2} and playing action 2n2n. Moreover, if choosing contract 𝒙1\bm{x}_{1}, type 22 agent will get utility F,2(2)nn11nnFn,2(1)c=4nncF^{(2)}_{\ell,2}\cdot\frac{n^{n-1}-1}{n^{n}\cdot F^{(1)}_{n,2}}-c_{\ell}=\frac{4}{n^{n}}-c_{\ell} for action {n+1,n+2,,2n+1}\ell\in\{n+1,n+2,\dots,2n+1\}, which is at most 4nn\frac{4}{n^{n}}. Thus, we can assume that type 22 agent chooses contract 𝒙2\bm{x}_{2}, and then the principal’s profit from type 22 agent is non-negative, because for any outcome, contract 𝒙2\bm{x}_{2} does not pay more than its reward. It follows that the total profit from two types of agents is at least 1nn\frac{1}{n^{n}}.

Finally, recall that regardless of which action type 33 agent plays, the result is always outcome 44. Since the payment for outcome 44 is zero for both contracts 𝒙1,𝒙2\bm{x}_{1},\bm{x}_{2}, the profit from type 33 agent is zero. ∎

A.2 Proof of Theorem 3.2

Here we provide the detailed proof of Theorem 3.2 (that there exists an Ω(logT)\Omega(\log T) gap between single contracts and menus of deterministic contracts).

Lemma A.5.

Consider the principal-agent problem (𝐜,𝐅,𝐫)\left(\bm{c},\bm{F},\bm{r}\right) defined in the above Ω(logT)\Omega(\log T)-gap instance. For any contract 𝐱=(x1,,xT)0T\bm{x}=(x_{1},\dots,x_{T})\in\mathbb{R}_{\geq 0}^{T} that has non-negative profit, it must hold that xT116Nx_{T}\leq\frac{1}{16^{N}}.

Proof.

Note that only outcome 1 has strictly positive reward, and only action 11 can result in outcome 1. Hence, the welfare of type ((k1)2N+)((k-1)2^{N}+\ell) agent is F1,1((k1)2N+)r1c1=2k2T(12k)F^{((k-1)2^{N}+\ell)}_{1,1}\cdot r_{1}-c_{1}=\frac{2^{-k}}{2T(1-2^{-k})} for all k[N]k\in[N] and [2N]\ell\in[2^{N}], and the principal’s profit from this type of agent is at most this amount.

Moreover, observe that type TT agent can only result in outcome TT regardless of which action is played. Thus, contract (x1,,xT)(x_{1},\dots,x_{T}) pays type TT agent xTx_{T} for nothing in return.

Altogether, the profit of contract (x1,,xT)(x_{1},\dots,x_{T}) in expectation over agent types is at most

k[N],[2N]2k116N2k2T(12k)(12N18N)xT\displaystyle\sum_{k\in[N],\ell\in[2^{N}]}\frac{2^{k-1}}{16^{N}}\cdot\frac{2^{-k}}{2T(1-2^{-k})}-(1-\frac{2^{N}-1}{8^{N}})\cdot x_{T} k[N],[2N]2k16N2k2T12xT\displaystyle\leq\sum_{k\in[N],\ell\in[2^{N}]}\frac{2^{k}}{16^{N}}\cdot\frac{2^{-k}}{2T}-\frac{1}{2}\cdot x_{T}
=N8N2T12xT,\displaystyle=\frac{N}{8^{N}\cdot 2T}-\frac{1}{2}\cdot x_{T},

and for this to be non-negative, it must hold that xTN8NT116Nx_{T}\leq\frac{N}{8^{N}\cdot T}\leq\frac{1}{16^{N}}. ∎

Lemma A.6.

Consider the principal-agent problem (𝐜,𝐅,𝐫)\left(\bm{c},\bm{F},\bm{r}\right) defined in the above Ω(logT)\Omega(\log T)-gap instance. For any d[N]d\in[N], for any contract 𝐱=(x1,,xT)0T\bm{x}=(x_{1},\dots,x_{T})\in\mathbb{R}_{\geq 0}^{T} with x112d+1x_{1}\geq 1-2^{-d+1}, we have

k[d],[2N]Pr[agent type=(k1)2N+]Profit((k1)2N+)(𝒄,𝑭,𝒓,𝒙)28NT.\sum_{k\in[d],\ell\in[2^{N}]}\Pr[\textrm{agent type}=(k-1)2^{N}+\ell]\cdot\textsc{Profit}^{((k-1)2^{N}+\ell)}(\bm{c},\bm{F},\bm{r},\bm{x})\leq\frac{2}{8^{N}\cdot T}. (11)
Proof.

Note that only outcome 1 has strictly positive reward, and only action 11 can result in outcome 1. Moreover, since F1,1(t)1TF^{(t)}_{1,1}\leq\frac{1}{T} for all t[T]t\in[T], outcome 1 happens with probability at most 1T\frac{1}{T} regardless of the contract and the agent type. Thus, contract (x1,,xT)(x_{1},\dots,x_{T}) gets profit at most 1x1T\frac{1-x_{1}}{T} (which in turn is at most 2d+1T\frac{2^{-d+1}}{T} by our assumption) regardless of the agent type. It follows that

k[d],[2N]Pr[agent type=(k1)2N+]Profit((k1)2N+)(𝒄,𝑭,𝒓,𝒙)\displaystyle\sum_{k\in[d],\ell\in[2^{N}]}\Pr[\textrm{agent type}=(k-1)2^{N}+\ell]\cdot\textsc{Profit}^{((k-1)2^{N}+\ell)}(\bm{c},\bm{F},\bm{r},\bm{x})
\displaystyle\leq k[d],[2N]Pr[agent type=(k1)2N+]2d+1T\displaystyle\sum_{k\in[d],\ell\in[2^{N}]}\Pr[\textrm{agent type}=(k-1)2^{N}+\ell]\cdot\frac{2^{-d+1}}{T}
=\displaystyle= k[d],[2N]2k116N2d+1T=2(12d)8NT28NT.\displaystyle\sum_{k\in[d],\ell\in[2^{N}]}\frac{2^{k-1}}{16^{N}}\cdot\frac{2^{-d+1}}{T}=\frac{2(1-2^{-d})}{8^{N}\cdot T}\leq\frac{2}{8^{N}\cdot T}.

Lemma A.7.

Consider the principal-agent problem (𝐜,𝐅,𝐫)\left(\bm{c},\bm{F},\bm{r}\right) defined in the above Ω(logT)\Omega(\log T)-gap instance. For any d[N]d\in[N], given any contract 𝐱=(x1,,xT)0T\bm{x}=(x_{1},\dots,x_{T})\in\mathbb{R}_{\geq 0}^{T} with x1[12d+1,12d]x_{1}\in[1-2^{-d+1},1-2^{-d}] that has non-negative profit, we have

|{t{d2N+1,,T1} s.t. type t agent prefers playing action 1}|2N+2+1.|\{t\in\{d\cdot 2^{N}+1,\dots,T-1\}\textrm{ s.t. type $t$ agent prefers playing action $1$}\}|\leq 2^{N+2}+1.
Proof.

Consider any t{d2N+1,,T1}t\in\{d\cdot 2^{N}+1,\dots,T-1\} such that type tt agent prefers playing action 11 given contract (x1,,xT)(x_{1},\dots,x_{T}). First, since type tt agent prefers action 11 over the null action, action 11 must have non-negative utility. Because for type tt agent, action 1 only has strictly positive probability for outcomes 11, tt and TT, it follows that F1,1(t)x1+F1,t(t)xt+F1,T(T)xTc1F^{(t)}_{1,1}\cdot x_{1}+F^{(t)}_{1,t}\cdot x_{t}+F^{(T)}_{1,T}\cdot x_{T}\geq c_{1}. After rearrangement, this is equivalent to

xt\displaystyle x_{t} c1F1,1(t)x1F1,T(T)xTF1,t(t)\displaystyle\geq\frac{c_{1}-F^{(t)}_{1,1}\cdot x_{1}-F^{(T)}_{1,T}\cdot x_{T}}{F^{(t)}_{1,t}}
c1F1,1(t)x11/16NF1,t(t)\displaystyle\geq\frac{c_{1}-F^{(t)}_{1,1}\cdot x_{1}-1/16^{N}}{F^{(t)}_{1,t}} (By Lemma A.5 and F1,T(T)1F^{(T)}_{1,T}\leq 1)
=1/(2T)F1,1(t)x11/16NF1,1(t)\displaystyle=\frac{1/(2T)}{F^{(t)}_{1,1}}-x_{1}-\frac{1/16^{N}}{F^{(t)}_{1,1}} (By definition of F1,t(t)F^{(t)}_{1,t} and c1c_{1})
12d1x12T16N\displaystyle\geq 1-2^{-d-1}-x_{1}-\frac{2T}{16^{N}} (12TF1,1(t)12T(12d1)\frac{1}{2T}\leq F^{(t)}_{1,1}\leq\frac{1}{2T(1-2^{-d-1})} for td2N+1t\geq d\cdot 2^{N}+1)
2d12T16N\displaystyle\geq 2^{-d-1}-\frac{2T}{16^{N}} (By assumption x112dx_{1}\leq 1-2^{-d})
12N+12N8N12N+2\displaystyle\geq\frac{1}{2^{N+1}}-\frac{2N}{8^{N}}\geq\frac{1}{2^{N+2}} (By dN and TN2N and N3).\displaystyle\text{(By $d\leq N$ and $T\geq N\cdot 2^{N}$ and $N\geq 3$)}. (12)

Moreover, since type tt agent prefers action 11 over action 22, action 11 must have higher utility than action 22. Namely, F1,1(t)x1+F1,t(t)xt+F1,T(T)xTc1s[T1]st,1F2,s(t)xs+F2,T(t)xTF^{(t)}_{1,1}\cdot x_{1}+F^{(t)}_{1,t}\cdot x_{t}+F^{(T)}_{1,T}\cdot x_{T}-c_{1}\geq\sum_{\begin{subarray}{c}s\in[T-1]\\ s\neq t,1\end{subarray}}F^{(t)}_{2,s}\cdot x_{s}+F^{(t)}_{2,T}\cdot x_{T} (recall F2,1(t),F2,t(t)=0F^{(t)}_{2,1},F^{(t)}_{2,t}=0). After rearrangement, this is equivalent to

xt\displaystyle x_{t} c1F1,1(t)x1+(F2,T(t)F1,T(t))xTF1,t(t)+s[T1]st,1F2,s(t)F1,t(t)xs\displaystyle\geq\frac{c_{1}-F^{(t)}_{1,1}\cdot x_{1}+(F^{(t)}_{2,T}-F^{(t)}_{1,T})\cdot x_{T}}{F^{(t)}_{1,t}}+\sum_{\begin{subarray}{c}s\in[T-1]\\ s\neq t,1\end{subarray}}\frac{F^{(t)}_{2,s}}{F^{(t)}_{1,t}}\cdot x_{s}
c1F1,1(t)x11/16NF1,t(t)+s[T1]st,1F2,s(t)F1,t(t)xs\displaystyle\geq\frac{c_{1}-F^{(t)}_{1,1}\cdot x_{1}-1/16^{N}}{F^{(t)}_{1,t}}+\sum_{\begin{subarray}{c}s\in[T-1]\\ s\neq t,1\end{subarray}}\frac{F^{(t)}_{2,s}}{F^{(t)}_{1,t}}\cdot x_{s} (By Lemma A.5 and F2,T(t)F1,T(t)1F^{(t)}_{2,T}-F^{(t)}_{1,T}\geq-1)
c1F1,1(t)x11/16NF1,t(t)+s[T1]st,112N+1xs\displaystyle\geq\frac{c_{1}-F^{(t)}_{1,1}\cdot x_{1}-1/16^{N}}{F^{(t)}_{1,t}}+\sum_{\begin{subarray}{c}s\in[T-1]\\ s\neq t,1\end{subarray}}\frac{1}{2^{N+1}}\cdot x_{s} (By definition of F2,s(t)F^{(t)}_{2,s})
12N+2+s[T1]st,112N+1xs,\displaystyle\geq\frac{1}{2^{N+2}}+\sum_{\begin{subarray}{c}s\in[T-1]\\ s\neq t,1\end{subarray}}\frac{1}{2^{N+1}}\cdot x_{s}, (13)

where the final step follows from the second to the last steps of the derivation of Ineq. (A.2).

Now let S1={t{d2N+1,,T1} s.t. type t agent prefers playing action 1}S_{1}=\{t\in\{d\cdot 2^{N}+1,\dots,T-1\}\textrm{ s.t. type $t$ agent prefers playing action $1$}\}, and assume for contradiction |S1|2N+2+2|S_{1}|\geq 2^{N+2}+2. By Ineq. (A.2), we have that for all tS1t\in S_{1}

xt\displaystyle x_{t} 12N+2+s[T1]st,112N+1xs\displaystyle\geq\frac{1}{2^{N+2}}+\sum_{\begin{subarray}{c}s\in[T-1]\\ s\neq t,1\end{subarray}}\frac{1}{2^{N+1}}\cdot x_{s}
sS1st,112N+1xs\displaystyle\geq\sum_{\begin{subarray}{c}s\in S_{1}\\ s\neq t,1\end{subarray}}\frac{1}{2^{N+1}}\cdot x_{s}
(|S1|2)122N+3\displaystyle\geq(|S_{1}|-2)\cdot\frac{1}{2^{2N+3}} (By Ineq. (A.2))
12N+1\displaystyle\geq\frac{1}{2^{N+1}} (By assumption |S1|2N+2+2).\displaystyle\text{(By assumption $|S_{1}|\geq 2^{N+2}+2$)}. (14)

Notice that Ineq. (A.2) gives a new lower bound of xtx_{t} for all tS1t\in S_{1}, which is twice as large as the lower bound given by Ineq. (A.2). Using this larger lower bound, we can repeat the above derivation of Ineq. (A.2) (i.e., we use the larger lower bound instead in the third inequality of the derivation) and double the lower bound again. By repeating such derivation arbitrarily many times, we have that xtx_{t} for all tS1t\in S_{1} is arbitrarily large, and thus, contract 𝒙\bm{x} makes arbitrarily large payment to each type of agent in S1S_{1}. Therefore, contract 𝒙\bm{x} must have strictly negative profit (because the reward generated by any type of agent is always bounded), which contradicts our assumption in the lemma statement. ∎

Proof of Theorem 3.2.

Soundness. We show that any contract 𝒙=(x1,,xT)0T\bm{x}=(x_{1},\dots,x_{T})\in\mathbb{R}_{\geq 0}^{T} with non-negative profit at most has expected profit 6N16N\frac{6}{N\cdot 16^{N}} (the expectation is over the distribution of agent type). Let d[N]d\in[N] be such that x1[12d+1,12d]x_{1}\in[1-2^{-d+1},1-2^{-d}]. Now we upper bound the expected profit of contract 𝒙\bm{x} as follows:

expected profit =k[N],[2N]Pr[agent type=(k1)2N+]Profit((k1)2N+)(𝒄,𝑭,𝒓,𝒙)\displaystyle=\sum_{k\in[N],\ell\in[2^{N}]}\Pr[\textrm{agent type}=(k-1)2^{N}+\ell]\cdot\textsc{Profit}^{((k-1)2^{N}+\ell)}(\bm{c},\bm{F},\bm{r},\bm{x})
=k[d],[2N]Pr[agent type=(k1)2N+]Profit((k1)2N+)(𝒄,𝑭,𝒓,𝒙)\displaystyle=\sum_{k\in[d],\ell\in[2^{N}]}\Pr[\textrm{agent type}=(k-1)2^{N}+\ell]\cdot\textsc{Profit}^{((k-1)2^{N}+\ell)}(\bm{c},\bm{F},\bm{r},\bm{x})
+t{d2N+1,N2N}Pr[agent type=t]Profit(t)(𝒄,𝑭,𝒓,𝒙)\displaystyle\qquad+\sum_{t\in\{d\cdot 2^{N}+1,\dots N\cdot 2^{N}\}}\Pr[\textrm{agent type}=t]\cdot\textsc{Profit}^{(t)}(\bm{c},\bm{F},\bm{r},\bm{x})
28NT+t{d2N+1,N2N}Pr[agent type=t]Profit(t)(𝒄,𝑭,𝒓,𝒙)(By Lemma A.6).\displaystyle\leq\frac{2}{8^{N}\cdot T}+\sum_{t\in\{d\cdot 2^{N}+1,\dots N\cdot 2^{N}\}}\Pr[\textrm{agent type}=t]\cdot\textsc{Profit}^{(t)}(\bm{c},\bm{F},\bm{r},\bm{x})\,\,\,\,\text{(By Lemma\leavevmode\nobreak\ \ref{lemma:bound_lower_part_agents_Omega_log_T_gap})}. (15)

Moreover, because only action 11 can generate strictly positive reward, we have that

t{d2N+1,N2N}Pr[agent type=t]Profit(t)(𝒄,𝑭,𝒓,𝒙)\displaystyle\sum_{t\in\{d\cdot 2^{N}+1,\dots N\cdot 2^{N}\}}\Pr[\textrm{agent type}=t]\cdot\textsc{Profit}^{(t)}(\bm{c},\bm{F},\bm{r},\bm{x})
=t{d2N+1,N2N} s.t. type t agent prefers action 1Pr[agent type=t]Profit(t)(𝒄,𝑭,𝒓,𝒙)\displaystyle=\sum_{\begin{subarray}{c}t\in\{d\cdot 2^{N}+1,\dots N\cdot 2^{N}\}\\ \textrm{ s.t. type $t$ agent prefers action $1$}\end{subarray}}\Pr[\textrm{agent type}=t]\cdot\textsc{Profit}^{(t)}(\bm{c},\bm{F},\bm{r},\bm{x})
t{d2N+1,N2N} s.t. type t agent prefers action 1Pr[agent type=t]expected welfare of action 1 for type t agent\displaystyle\leq\sum_{\begin{subarray}{c}t\in\{d\cdot 2^{N}+1,\dots N\cdot 2^{N}\}\\ \textrm{ s.t. type $t$ agent prefers action $1$}\end{subarray}}\Pr[\textrm{agent type}=t]\cdot\textrm{expected welfare of action $1$ for type $t$ agent}
=t{d2N+1,N2N} s.t. type t agent prefers action 1Pr[agent type=t](F1,1(t)1c1)\displaystyle=\sum_{\begin{subarray}{c}t\in\{d\cdot 2^{N}+1,\dots N\cdot 2^{N}\}\\ \textrm{ s.t. type $t$ agent prefers action $1$}\end{subarray}}\Pr[\textrm{agent type}=t]\cdot(F^{(t)}_{1,1}\cdot 1-c_{1})
t{d2N+1,N2N} s.t. type t agent prefers action 112T16N(By Ineq. (2))\displaystyle\leq\sum_{\begin{subarray}{c}t\in\{d\cdot 2^{N}+1,\dots N\cdot 2^{N}\}\\ \textrm{ s.t. type $t$ agent prefers action $1$}\end{subarray}}\frac{1}{2T\cdot 16^{N}}\qquad\text{(By Ineq.\leavevmode\nobreak\ \eqref{eq:equal_revenue_Omega_log_T_gap})}
2N+2+12T16N4T8N(By Lemma A.7).\displaystyle\leq\frac{2^{N+2}+1}{2T\cdot 16^{N}}\leq\frac{4}{T\cdot 8^{N}}\qquad\qquad\qquad\qquad\quad\,\text{(By Lemma\leavevmode\nobreak\ \ref{lemma:bound_upper_part_agents_Omega_log_T_gap})}. (16)

Combining Ineq. (A.2) and Ineq. (A.2), we get that the expected profit of contract 𝒙\bm{x} is at most 6T8N6N16N\frac{6}{T\cdot 8^{N}}\leq\frac{6}{N\cdot 16^{N}}.

Completeness. We show that there is a menu of T1T-1 contracts {𝒙1,,𝒙T1}\{\bm{x}_{1},\dots,\bm{x}_{T-1}\} that achieve expected profit 116N+1\frac{1}{16^{N+1}}. Specifically, for each k[N]k\in[N] and [2N]\ell\in[2^{N}], 𝒙(k1)2N+:=(x1(k,),x2(k,),,xT(k,))\bm{x}_{(k-1)2^{N}+\ell}:=(x_{1}^{(k,\ell)},x_{2}^{(k,\ell)},\dots,x_{T}^{(k,\ell)}) is defined as follows: let x1(k,)=12x_{1}^{(k,\ell)}=\frac{1}{2}, and let x(k1)2N+(k,)=1212k+1x_{(k-1)2^{N}+\ell}^{(k,\ell)}=\frac{1}{2}-\frac{1}{2^{k+1}} if (k1)2N+1(k-1)2^{N}+\ell\neq 1, and let the other entries be zero.

Because all the contracts in the menu make zero payment for outcome TT, and type TT agent can only result in outcome TT which has zero reward, we can ignore type TT agent.

Now we show that given the above menu, for all k[N]k\in[N] and [2N]\ell\in[2^{N}], type ((k1)2N+)((k-1)2^{N}+\ell) agent wants to pick contract 𝒙(k1)2N+\bm{x}_{(k-1)2^{N}+\ell} and play action 11. First, it is easy to check that for type ((k1)2N+)((k-1)2^{N}+\ell) agent, given contract 𝒙(k1)2N+\bm{x}_{(k-1)2^{N}+\ell}, action 22 generates zero utility, and action 11 generates utility

12F1,1((k1)2N+)+(1212k+1)F1,(k1)2N+((k1)2N+)c1=(112k+1)F1,1((k1)2N+)c1=2k2T(12k),\frac{1}{2}\cdot F^{((k-1)2^{N}+\ell)}_{1,1}+\left(\frac{1}{2}-\frac{1}{2^{k+1}}\right)\cdot F^{((k-1)2^{N}+\ell)}_{1,(k-1)2^{N}+\ell}-c_{1}=\left(1-\frac{1}{2^{k+1}}\right)\cdot F^{((k-1)2^{N}+\ell)}_{1,1}-c_{1}=\frac{2^{-k-2}}{T(1-2^{-k})}, (17)

which is strictly positive. Thus, it remains to show that by picking any other contract 𝒙(k1)2N+\bm{x}_{(k^{\prime}-1)2^{N}+\ell^{\prime}}, type ((k1)2N+)((k-1)2^{N}+\ell) agent cannot get more than the utility in Eq. (17). To this end, we notice that given contract 𝒙(k1)2N+\bm{x}_{(k^{\prime}-1)2^{N}+\ell^{\prime}}, type ((k1)2N+)((k-1)2^{N}+\ell) agent would get the following utilities by playing action 11 and action 22 respectively:

utility of action 11 =12F1,1((k1)2N+)c10,\displaystyle=\frac{1}{2}\cdot F^{((k-1)2^{N}+\ell)}_{1,1}-c_{1}\leq 0,
utility of action 22 =(1212k+1)F2,(k1)2N+((k1)2N+)12F2,(k1)2N+((k1)2N+)=2k3T(12k).\displaystyle=(\frac{1}{2}-\frac{1}{2^{k^{\prime}+1}})\cdot F^{((k-1)2^{N}+\ell)}_{2,(k^{\prime}-1)2^{N}+\ell}\leq\frac{1}{2}\cdot F^{((k-1)2^{N}+\ell)}_{2,(k^{\prime}-1)2^{N}+\ell}=\frac{2^{-k-3}}{T(1-2^{-k})}.

Finally, we derive the expected profit of the above menu (given that we have shown type ((k1)2N+)((k-1)2^{N}+\ell) agent picks contract 𝒙(k1)2N+\bm{x}_{(k-1)2^{N}+\ell} and plays action 11) as follows:

expected profit =k[N],[2N]Pr[agent type=(k1)2N+]\displaystyle=\sum_{k\in[N],\ell\in[2^{N}]}\Pr[\textrm{agent type}=(k-1)2^{N}+\ell]
×(F1,1((k1)2N+)(1x1(k,))F1,(k1)2N+((k1)2N+)x(k1)2N+(k,))\displaystyle\qquad\qquad\qquad\times\left(F^{((k-1)2^{N}+\ell)}_{1,1}\cdot\left(1-x_{1}^{(k,\ell)}\right)-F^{((k-1)2^{N}+\ell)}_{1,(k-1)2^{N}+\ell}\cdot x_{(k-1)2^{N}+\ell}^{(k,\ell)}\right)
=k[N],[2N]Pr[agent type=(k1)2N+]F1,1((k1)2N+)(1x1(k,)x(k1)2N+(k,))\displaystyle=\sum_{k\in[N],\ell\in[2^{N}]}\Pr[\textrm{agent type}=(k-1)2^{N}+\ell]\cdot F^{((k-1)2^{N}+\ell)}_{1,1}\cdot\left(1-x_{1}^{(k,\ell)}-x_{(k-1)2^{N}+\ell}^{(k,\ell)}\right)
=k[N],[2N]2k12T(12k)16N12k+1=k[N],[2N]18T16N(12k)\displaystyle=\sum_{k\in[N],\ell\in[2^{N}]}\frac{2^{k-1}}{2T(1-2^{-k})\cdot 16^{N}}\cdot\frac{1}{2^{k+1}}=\sum_{k\in[N],\ell\in[2^{N}]}\frac{1}{8T\cdot 16^{N}(1-2^{-k})}
=T18T16N(12k)116N+1.\displaystyle=\frac{T-1}{8T\cdot 16^{N}(1-2^{-k})}\geq\frac{1}{16^{N+1}}.

A.3 Bounding the gap between linear contracts and menus of randomized linear contracts

Lemma A.8.

The optimal linear contract achieves a O(T)O(T) approximation with respect to the optimal menu of randomized linear contracts. In particular,

Opt-Linear(𝒄,𝑭,𝒓)Ω(1/T)Opt-RndMenuLinear(𝒄,𝑭,𝒓).\displaystyle\textsc{Opt-Linear}\left(\bm{c},\bm{F},\bm{r}\right)\geq\Omega(1/T)\cdot\textsc{Opt-RndMenuLinear}\left(\bm{c},\bm{F},\bm{r}\right).
Proof.

For each type, we can compute the best linear contract. This contract generates at least as much revenue as any component of the randomized linear contract. Hence, we can extract at least an Ω(1/T)\Omega(1/T) fraction of the optimal revenue. ∎

A.4 Proof of Lemma 4.2

Proof of Lemma 4.2.

Given any optimal menu Γ~=(γ~(1),,γ~(T))\tilde{\Gamma}=\left(\tilde{\gamma}^{(1)},\dots,\tilde{\gamma}^{(T)}\right), we will construct a new menu Γ=(γ(1),,γ(T))\Gamma=\left(\gamma^{(1)},\dots,\gamma^{(T)}\right) which satisfies the above constraints and whose revenue is at least as large as Γ~\tilde{\Gamma}. In particular, we will demonstrate how to redistribute probability mass not supported on one of these points to these points while leaving the performance of the menu unchanged.

For any point ααp\alpha\leq\alpha_{p} that is not a break point, we can write α=rαj+(1r)αj+1\alpha=r\cdot\alpha_{j}+(1-r)\cdot\alpha_{j+1} as a convex combination of two adjacent breakpoints. Now we simply increment γ(t)(αj)\gamma^{(t)}(\alpha_{j}) by rγ~(t)(α)r\cdot\tilde{\gamma}^{(t)}(\alpha) and γ(t+1)(αj+1)\gamma^{(t+1)}(\alpha_{j+1}) by(1r)γ~(t)(αj+1)(1-r)\cdot\tilde{\gamma}^{(t)}(\alpha_{j+1}). Since we are simply moving the mass at a given point to its two neighbors, and all utility functions UtU_{t} are linear in between any pair of consecutive breakpoints, the utility of each agent for each randomized linear contract in the menu remains the same (and hence each agent still selects the same contract from the menu and plays the same action).

The profit obtained by the principal by offering linear contract α\alpha to the agent of type tt is given by (1α)Ut(α)(1-\alpha)U^{\prime}_{t}(\alpha) (when taking the derivative, we arbitrarily break ties in the favor of the principal). Since Ui(α)=Ui(αj)Ui(αj+1)U_{i}^{\prime}(\alpha)=U_{i}^{\prime}(\alpha_{j})\leq U_{i}^{\prime}(\alpha_{j+1}), we note that the revenue achieved by the new menu is at least what is achieved by Γ~\tilde{\Gamma}.

It remains to deal with the probability mass above αp\alpha_{p}. Define the point

αp+1i:=b=αpbγ~(i)(b)𝑑bb=αpγ~(i)(b)𝑑b\alpha^{i}_{p+1}:=\frac{\int_{b=\alpha_{p}}^{\infty}b\cdot\tilde{\gamma}^{(i)}(b)db}{\int_{b=\alpha_{p}}^{\infty}\tilde{\gamma}^{(i)}(b)db}

whenever the b=αpγ(i)(b)𝑑b>0\int_{b=\alpha_{p}}^{\infty}\gamma^{(i)}(b)db>0. By definition, αp+1iαp\alpha^{i}_{p+1}\geq\alpha_{p}. We will set γ(i)(αp+1i)=b=αpγ(i)(b)𝑑b\gamma^{(i)}(\alpha^{i}_{p+1})=\int_{b=\alpha_{p}}^{\infty}\gamma^{(i)}(b)db. Since we are simply contracting the mass in this last interval (αp,)(\alpha_{p},\infty), the resulting distribution is still a probability distribution. Furthermore,

b=αpf(b)γ~(i)(b)𝑑b=f(αp+1)γ(i)(αp+1)\int_{b=\alpha_{p}}^{\infty}f(b)\tilde{\gamma}^{(i)}(b)db=f(\alpha_{p+1})\gamma^{(i)}(\alpha_{p+1})

for any linear function ff. Again, since (for this range of α\alpha) all utilities for each agent (and the principal’s profit) are linear functions of γ~\tilde{\gamma}, they are preserved under the new contract γ\gamma.

A.5 Proof of Lemma 4.3

Proof of Lemma 4.3.

Recall that the breakpoints in the instance are multiples of 1/31/3, i.e. αi=i/3\alpha_{i}=i/3 for i6.i\leq 6. First, we exhibit a menu whose revenue is large 0.316250.31625. Consider the menu Γ=(γ(1),γ(2))\Gamma=(\gamma^{(1)},\gamma^{(2)})

γ(1)(α)={35 if α=079200 if α=231200 if α=530 otherwiseγ(2)(α)={1 if α=130 otherwise\displaystyle\gamma^{(1)}(\alpha)=\begin{cases}\frac{3}{5}&\text{ if }\alpha=0\\ \frac{79}{200}&\text{ if }\alpha=\frac{2}{3}\\ \frac{1}{200}&\text{ if }\alpha=\frac{5}{3}\\ 0&\text{ otherwise}\end{cases}\qquad\gamma^{(2)}(\alpha)=\begin{cases}1&\text{ if }\alpha=\frac{1}{3}\\ 0&\text{ otherwise}\end{cases}

The first menu item extracts a revenue of 3592400\frac{359}{2400} and the second extracts a revenue of 16\frac{1}{6} to get a total revenue of 253800=0.31625\frac{253}{800}=0.31625.

To show that no bounded randomized linear contract cannot get a fractional value of greater than 0.316250.31625, we exhibit a dual solution to the dual of linear program presented in Section 4.

min\displaystyle\min s1+s2\displaystyle s_{1}+s_{2}
s.t.\displaystyle s.t. s1\displaystyle s_{1} (U1(αj)λ1,2U2(αj)λ2,1)+U1(αj)(1αj)0j3\displaystyle\geq\left(U_{1}(\alpha_{j})\lambda_{1,2}-U_{2}(\alpha_{j})\lambda_{2,1}\right)+U_{1}^{\prime}(\alpha_{j})(1-\alpha_{j})\qquad\forall 0\leq j\leq 3
s2\displaystyle s_{2} (U2(αj)λ2,1U1(αj)λ1,2)+U2(αj)(1αj)0j3\displaystyle\geq\left(U_{2}(\alpha_{j})\lambda_{2,1}-U_{1}(\alpha_{j})\lambda_{1,2}\right)+U_{2}^{\prime}(\alpha_{j})(1-\alpha_{j})\qquad\forall 0\leq j\leq 3
λ,s\displaystyle\lambda,s 0\displaystyle\geq 0

We derived this dual by restricting the breakpoints to just α0=0,α1=1/3,α2=2/3,α3=1\alpha_{0}=0,\alpha_{1}=1/3,\alpha_{2}=2/3,\alpha_{3}=1. Hence any feasible dual solution is an upper bound on menus of bounded randomized linear contracts. In particular, assigning values λ1,2=1.857154080498256\lambda_{1,2}=1.857154080498256 , λ2,1=0.8095306797975786\lambda_{2,1}=0.8095306797975786, s1=0.25s_{1}=0.25 and s2=0.056345s_{2}=0.056345 achieves an objective of approximately 0.30635<0.310.30635<0.31. ∎

A.6 Proof of Corollary 4.8

Proof of Corollary 4.8.

Note that the example in Theorem 4.7 only has one non-null outcome. We first claim that for such instances, every general contract is equivalent to an affine contract: a contract that transfers α\alpha times the principal’s reward plus an additive β\beta (in particular, a transfer of x0x_{0} on the null outcome and x1x_{1} on the non-null outcome is equivalent to the affine contract with β=x0\beta=x_{0} and α=(x1x0)/r1\alpha=(x_{1}-x_{0})/r_{1}).

Because of this, the best menu of randomized (general) contracts performs equally as well as the best menu of randomized affine contracts. Since all linear contracts are affine, this menu performs at least as well as the best menu of randomized linear contracts. We thus have that

Opt-RndMenu(𝒄,𝑭,𝒓)Opt-RndMenuLinear(𝒄,𝑭,𝒓).\textsc{Opt-RndMenu}\left(\bm{c},\bm{F},\bm{r}\right)\geq\textsc{Opt-RndMenuLinear}\left(\bm{c},\bm{F},\bm{r}\right).

On the other hand, Lemma 7 in Guruganesh et al. (2021) states that for typed principal-agent problems with two outcomes, Opt-DetMenu(𝒄,𝑭,𝒓)=Opt-Linear(𝒄,𝑭,𝒓)\textsc{Opt-DetMenu}\left(\bm{c},\bm{F},\bm{r}\right)=\textsc{Opt-Linear}\left(\bm{c},\bm{F},\bm{r}\right). Combining these two facts with Theorem 4.7 we arrive at the theorem statement. ∎

A.7 Proof of Theorem 4.6

Proof of Theorem 4.6.

Consider any (𝒄,𝑭,𝒓)\left(\bm{c},\bm{F},\bm{r}\right){} with n=2n=2 actions (a null action 0 and a non-null action 11), TT agent types, and mm outcomes. For each agent type tt, we let Ut(α):00U_{t}(\alpha):\mathbb{R}_{\geq 0}\rightarrow\mathbb{R}_{\geq 0} denote the utility function for type-tt agent given linear contract α\alpha (i.e., pay αrj\alpha\cdot r_{j} for each outcome jj) as input variable. Notice that Ut(α)=max{0,j[m]αF1,j(t)rjc1}U_{t}(\alpha)=\max\{0,\,\sum_{j\in[m]}\alpha\cdot F_{1,j}^{(t)}r_{j}-c_{1}\} (i.e., the best of the utilities of actions 0 and 11). We let R(t):=j[m]F1,j(t)rjR^{(t)}:=\sum_{j\in[m]}F_{1,j}^{(t)}r_{j}. Then, the utility function can be simplified as Ut(α)=max{0,R(t)αc1}U_{t}(\alpha)=\max\{0,\,R^{(t)}\cdot\alpha-c_{1}\}. For each agent type tt, we let αt\alpha_{t} denote the minimum linear contract for which type-tt agent would like to play action 11, i.e., αt=c1R(t)\alpha_{t}=\frac{c_{1}}{R^{(t)}}.

Bucketizing the agent types

Let w=10w=10 (any constant larger than 11 would work) and αmin:=mint[T]αt\alpha_{\min}:=\min_{t\in[T]}\alpha_{t} and αmax:=maxt[T]αt\alpha_{\max}:=\max_{t\in[T]}\alpha_{t}. Now we bucketize the agent types into bb buckets according to their αt\alpha_{t}, and bb is chosen such that 1αmin[wb1(1αmax),wb(1αmax))1-\alpha_{\min}\in[w^{b-1}(1-\alpha_{\max}),w^{b}(1-\alpha_{\max})). The ii-th bucket BiB_{i} contains every agent type tt such that 1αt[wi1(1αmax),wi(1αmax))1-\alpha_{t}\in[w^{i-1}(1-\alpha_{\max}),w^{i}(1-\alpha_{\max})). Now consider two principal-agent problem instances by modifying the instance (𝒄,𝑭,𝒓)\left(\bm{c},\bm{F},\bm{r}\right){}:

  • Odd instance (𝒄1,𝑭1,𝒓1)(\bm{c}_{1},\bm{F}_{1},\bm{r}_{1}): Remove every agent type that belongs to a bucket with even index. Add a dummy agent type (which always results into the dummy outcome with zero reward) such that its probability is the sum of the probabilities of the removed agent types.

  • Even instance (𝒄2,𝑭2,𝒓2)(\bm{c}_{2},\bm{F}_{2},\bm{r}_{2}): Same as above except that we now remove the agent types in the buckets with odd indices.

Obviously Opt-Linear(𝒄,𝑭,𝒓)Opt-Linear(𝒄1,𝑭1,𝒓1),Opt-Linear(𝒄2,𝑭2,𝒓2)\textsc{Opt-Linear}(\bm{c},\bm{F},\bm{r})\geq\textsc{Opt-Linear}(\bm{c}_{1},\bm{F}_{1},\bm{r}_{1}),\textsc{Opt-Linear}(\bm{c}_{2},\bm{F}_{2},\bm{r}_{2}), and

Opt-RndMenuLinear(𝒄,𝑭,𝒓)\displaystyle\textsc{Opt-RndMenuLinear}(\bm{c},\bm{F},\bm{r})\leq Opt-RndMenuLinear(𝒄1,𝑭1,𝒓1)\displaystyle\textsc{Opt-RndMenuLinear}(\bm{c}_{1},\bm{F}_{1},\bm{r}_{1})
+Opt-RndMenuLinear(𝒄2,𝑭2,𝒓2).\displaystyle+\textsc{Opt-RndMenuLinear}(\bm{c}_{2},\bm{F}_{2},\bm{r}_{2}).

Assume that Opt-RndMenuLinear(𝒄1,𝑭1,𝒓1)Opt-RndMenuLinear(𝒄2,𝑭2,𝒓2)\textsc{Opt-RndMenuLinear}(\bm{c}_{1},\bm{F}_{1},\bm{r}_{1})\geq\textsc{Opt-RndMenuLinear}(\bm{c}_{2},\bm{F}_{2},\bm{r}_{2}) without loss of generality. Then, we have that

Opt-Linear(𝒄,𝑭,𝒓)Opt-RndMenuLinear(𝒄,𝑭,𝒓)Opt-Linear(𝒄1,𝑭1,𝒓1)2Opt-RndMenuLinear(𝒄1,𝑭1,𝒓1).\frac{\textsc{Opt-Linear}(\bm{c},\bm{F},\bm{r})}{\textsc{Opt-RndMenuLinear}(\bm{c},\bm{F},\bm{r})}\geq\frac{\textsc{Opt-Linear}(\bm{c}_{1},\bm{F}_{1},\bm{r}_{1})}{2\cdot\textsc{Opt-RndMenuLinear}(\bm{c}_{1},\bm{F}_{1},\bm{r}_{1})}.

Henceforth, it suffices to show Opt-Linear(𝒄1,𝑭1,𝒓1)Ω(1)Opt-RndMenuLinear(𝒄1,𝑭1,𝒓1)\textsc{Opt-Linear}(\bm{c}_{1},\bm{F}_{1},\bm{r}_{1})\geq\Omega(1)\cdot\textsc{Opt-RndMenuLinear}(\bm{c}_{1},\bm{F}_{1},\bm{r}_{1}). In the principal-agent problem (𝒄1,𝑭1,𝒓1)(\bm{c}_{1},\bm{F}_{1},\bm{r}_{1}), if two agent types t,tt,t^{\prime} belong to two different buckets (say tBi,tBit\in B_{i},t^{\prime}\in B_{i^{\prime}} for i<ii<i^{\prime}), then i+1<ii+1<i^{\prime} because i,ii,i^{\prime} must be odd. It follows that w(1αt)1αtw(1-\alpha_{t})\leq 1-\alpha_{t^{\prime}} by definition of the buckets.

Choosing one agent type from each bucket

Notice that for each bucket BiB_{i}, linear contract βi:=1wi1(1αmax)\beta_{i}:=1-w^{i-1}(1-\alpha_{\max}) can extract 1w\frac{1}{w}-fraction of the welfare of any agent type tBit\in B_{i}. Indeed, for any tBit\in B_{i}, we have βiαt\beta_{i}\geq\alpha_{t}, and hence type-tt agent would play action 11 and generates a profit (1βi)R(t)(1-\beta_{i})R^{(t)} for the principal, which is 1βi1αt1βiwi(1αmax)=1w\frac{1-\beta_{i}}{1-\alpha_{t}}\geq\frac{1-\beta_{i}}{w^{i}(1-\alpha_{\max})}=\frac{1}{w} fraction of type-tt agent’s welfare.

Suppose the optimal linear contract only achieves expected profit pp_{\ell} for (𝒄1,𝑭1,𝒓1)(\bm{c}_{1},\bm{F}_{1},\bm{r}_{1}). Then the expected welfare of bucket BiB_{i} (i.e., tBiPr[agent type=t]type-t agent’s welfare\sum_{t\in B_{i}}\Pr[\textrm{agent type}=t]\cdot\textrm{type-$t$ agent's welfare}) for any odd ii is at most wpw\cdot p_{\ell}, because we have shown there is a linear contract that extracts 1w\frac{1}{w}-fraction of the welfare of any agent type tBit\in B_{i}. Moreover, suppose the optimal menu of randomized linear contracts Γ\Gamma extracts ηi\eta_{i} fraction of the expected welfare of bucket BiB_{i}. Then the expected profit of Γ\Gamma is at most odd i[b]ηiwp\sum_{\textrm{odd }i\in[b]}\eta_{i}w\cdot p_{\ell}. Since ww is a constant, it suffices to prove odd i[b]ηi=O(1)\sum_{\textrm{odd }i\in[b]}\eta_{i}=O(1).

Moreover, since Γ\Gamma extracts ηi\eta_{i} fraction of the expected welfare of bucket BiB_{i}, there should be at least one agent type in BiB_{i} from which Γ\Gamma extracts at least ηi\eta_{i} fraction of the welfare. We choose one such agent type for each bucket BiB_{i} with odd i[b]i\in[b].

Summarizing what we have done so far, we have chosen k=b2k=\left\lceil\frac{b}{2}\right\rceil agent types, which we denote by t1,t2,,tkt_{1},t_{2},\dots,t_{k} (ordered s.t. αti\alpha_{t_{i}} is non-decreasing in ii), such that

  • w(1αti+1)1αtiw(1-\alpha_{t_{i+1}})\leq 1-\alpha_{t_{i}} for each i[k1]i\in[k-1] (recall this follows from picking the odd instance),

  • and menu Γ\Gamma extracts certain ρi\rho_{i} fraction of the welfare of the type-tit_{i} agent for each tit_{i}, such that i[k]ρiodd i[b]ηi\sum_{i\in[k]}\rho_{i}\geq\sum_{\textrm{odd }i\in[b]}\eta_{i} (this is by our choice of the agent types),

and our final step is to show i[k]ρi=O(1)\sum_{i\in[k]}\rho_{i}=O(1), which implies odd i[b]ηi=O(1)\sum_{\textrm{odd }i\in[b]}\eta_{i}=O(1).

Proving that i[k]ρi=O(1)\sum_{i\in[k]}\rho_{i}=O(1)

Let Di(α)D_{i}(\alpha) denote the CDF of type-tit_{i} agent’s favorite randomized linear contract for each i[k]i\in[k] (we will also use DiD_{i} to refer to type-tit_{i} agent’s favorite randomized linear contract). Then, we have that

profit of Di from type-ti agenttype-ti agent’s welfare=ρi=αti1α1αti𝑑Di(α),\displaystyle\underbrace{\frac{\textrm{profit of $D_{i}$ from type-$t_{i}$ agent}}{\textrm{type-$t_{i}$ agent's welfare}}}_{=\rho_{i}}=\int_{\alpha_{t_{i}}}^{\infty}\frac{1-\alpha}{1-\alpha_{t_{i}}}dD_{i}(\alpha),
type-ti agent’s utility from Ditype-ti agent’s welfare:=ui=αtiααti1αti𝑑Di(α),\displaystyle\underbrace{\frac{\textrm{type-$t_{i}$ agent's utility from $D_{i}$}}{\textrm{type-$t_{i}$ agent's welfare}}}_{:=u_{i}}=\int_{\alpha_{t_{i}}}^{\infty}\frac{\alpha-\alpha_{t_{i}}}{1-\alpha_{t_{i}}}dD_{i}(\alpha),

where the integrals start from αti\alpha_{t_{i}} because αti\alpha_{t_{i}} is the minimum α\alpha for which type-tit_{i} agent would like to play action 11. Note that the LHS of the first equation is just ρi\rho_{i} by definition, and we let uiu_{i} denote the LHS of the second equation (observe that ρi+ui=1\rho_{i}+u_{i}=1). Without loss of generality, we assume that ui1u_{i}\leq 1 for all i[k]i\in[k], because otherwise we can first remove all the agent types tit_{i} with ui>1u_{i}>1 and then prove i[k] s.t. ui1ρi=O(1)\sum_{i\in[k]\textrm{ s.t. }u_{i}\leq 1}\rho_{i}=O(1) for the remaining agent types (which obviously implies i[k]ρi=O(1)\sum_{i\in[k]}\rho_{i}=O(1)).

Now we establish a lower bound for ui1uiu_{i-1}-u_{i}. Because type-ti1t_{i-1} agent prefers Di1D_{i-1} over DiD_{i}, we have that

type-ti1 agent’s utility from Di1type-ti1 agent’s welfare=ui1type-ti1 agent’s utility from Ditype-ti1 agent’s welfare:=ui1,\underbrace{\frac{\textrm{type-$t_{i-1}$ agent's utility from $D_{i-1}$}}{\textrm{type-$t_{i-1}$ agent's welfare}}}_{=u_{i-1}}\geq\underbrace{\frac{\textrm{type-$t_{i-1}$ agent's utility from $D_{i}$}}{\textrm{type-$t_{i-1}$ agent's welfare}}}_{:=u_{i-1}^{\prime}},

where we denote the RHS by ui1u_{i-1}^{\prime}. Thus we can prove a lower bound for ui1uiu_{i-1}^{\prime}-u_{i} instead. To this end, we derive that

ui1ui\displaystyle u_{i-1}^{\prime}-u_{i} =αti1ααti11αti1𝑑Di(α)αtiααti1αti𝑑Di(α)\displaystyle=\int_{\alpha_{t_{i-1}}}^{\infty}\frac{\alpha-\alpha_{t_{i-1}}}{1-\alpha_{t_{i-1}}}dD_{i}(\alpha)-\int_{\alpha_{t_{i}}}^{\infty}\frac{\alpha-\alpha_{t_{i}}}{1-\alpha_{t_{i}}}dD_{i}(\alpha)
αtiααti11αti1𝑑Di(α)αtiααti1αti𝑑Di(α)\displaystyle\geq\int_{\alpha_{t_{i}}}^{\infty}\frac{\alpha-\alpha_{t_{i-1}}}{1-\alpha_{t_{i-1}}}dD_{i}(\alpha)-\int_{\alpha_{t_{i}}}^{\infty}\frac{\alpha-\alpha_{t_{i}}}{1-\alpha_{t_{i}}}dD_{i}(\alpha) (By αtiαti1\alpha_{t_{i}}\geq\alpha_{t_{i-1}})
=αti(1αti)(ααti1)(1αti1)(ααti)(1αti)(1αti1)𝑑Di(α)\displaystyle=\int_{\alpha_{t_{i}}}^{\infty}\frac{(1-\alpha_{t_{i}})(\alpha-\alpha_{t_{i-1}})-(1-\alpha_{t_{i-1}})(\alpha-\alpha_{t_{i}})}{(1-\alpha_{t_{i}})(1-\alpha_{t_{i-1}})}dD_{i}(\alpha)
=αtiαti1αtiα+αti+αti1α(1αti)(1αti1)𝑑Di(α)\displaystyle=\int_{\alpha_{t_{i}}}^{\infty}\frac{-\alpha_{t_{i-1}}-\alpha_{t_{i}}\alpha+\alpha_{t_{i}}+\alpha_{t_{i-1}}\alpha}{(1-\alpha_{t_{i}})(1-\alpha_{t_{i-1}})}dD_{i}(\alpha)
=αti1α1αtiαtiαti11αti1𝑑Di(α)\displaystyle=\int_{\alpha_{t_{i}}}^{\infty}\frac{1-\alpha}{1-\alpha_{t_{i}}}\cdot\frac{\alpha_{t_{i}}-\alpha_{t_{i-1}}}{1-\alpha_{t_{i-1}}}dD_{i}(\alpha)
αti1α1αti(11w)𝑑Di(α)\displaystyle\geq\int_{\alpha_{t_{i}}}^{\infty}\frac{1-\alpha}{1-\alpha_{t_{i}}}\cdot\left(1-\frac{1}{w}\right)dD_{i}(\alpha) (By w(1αti)(1αti1)w(1-\alpha_{t_{i}})\leq(1-\alpha_{t_{i-1}}))
=(11w)ρi\displaystyle=\left(1-\frac{1}{w}\right)\rho_{i} (By definition of ρi).\displaystyle\text{(By definition of $\rho_{i}$)}.

Therefore, we have shown that ui1ui(11w)ρiu_{i-1}-u_{i}\geq\left(1-\frac{1}{w}\right)\rho_{i}. Taking a telescoping sum, we have that u1uk=i{2,,k}ui1ui(11w)i{2,,k}ρiu_{1}-u_{k}=\sum_{i\in\{2,\dots,k\}}u_{i-1}-u_{i}\geq\left(1-\frac{1}{w}\right)\sum_{i\in\{2,\dots,k\}}\rho_{i}. It follows that i[k]ρi=O(1)\sum_{i\in[k]}\rho_{i}=O(1) because ρ1,u11\rho_{1},u_{1}\leq 1 and uk0u_{k}\geq 0, and ww is a constant. ∎