This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

11institutetext: Peking University 22institutetext: Columbia University
22email: allenyzj@stu.pku.edu.cn, {ll3530, christian.kroer}@columbia.edu

Greedy-Based Online Fair Allocation with Adversarial Input: Enabling Best-of-Many-Worlds Guarantees

Zongjun Yang 11    Luofeng Liao 22    Christian Kroer 22
Abstract

We study an online allocation problem with sequentially arriving items and adversarially chosen agent values, with the goal of balancing fairness and efficiency. Our goal is to study the performance of algorithms that achieve strong guarantees under other input models such as stochastic inputs, in order to achieve robust guarantees against a variety of inputs. To that end, we study the PACE (Pacing According to Current Estimated utility) algorithm, an existing algorithm designed for stochastic input. We show that in the equal-budgets case, PACE is equivalent to the integral greedy algorithm. We go on to show that with natural restrictions on the adversarial input model, both integral greedy allocation and PACE have asymptotically bounded multiplicative envy as well as competitive ratio for Nash welfare, with the multiplicative factors either constant or with optimal order dependence on the number of agents. This completes a “best-of-many-worlds” guarantee for PACE, since past work showed that PACE achieves guarantees for stationary and stochastic-but-non-stationary input models.

Keywords:
online fair allocation envy analysis Nash Welfare optimization.

1 Introduction

We study online fair allocation, where items arrive sequentially in TT rounds, and we need to distribute them among a set of nn agents with heterogeneous preferences, with the goal of achieving both fairness and efficiency properties. In each round, we observe each agent’s value for the item and make an irrevocable allocation. Frequently in the literature on this problem, the items are assumed to be divisible, and agents’ utilities are linear and additive. We also consider linear and additive utilities, but indivisible items.

Fair allocation problem in the offline setting has been well-studied. A classical objective to optimize is the Nash welfare (NW), defined as the geometric mean of the agents’ utilities. Maximizing Nash welfare provides a balance between efficiency and fairness due to the multiplicative nature of the objective. For divisible items, an offline optimal allocation can be computed via solving the Eisenberg-Gale (EG) convex program  (Eisenberg and Gale, 1959). The solution enjoys both envy-freeness and proportionality, which are important measures of fairness. For indivisible items, finding an allocation that maximizes NW is an APX-hard problem even in the offline setting (Moulin, 2004; Lee, 2017), though constant-factor approximation algorithms are known (Cole and Gkatzelis, 2018; Cole et al., 2017; McGlaughlin and Garg, 2020).

In the online setting, Gao et al. (2021) provides a simple allocation algorithm called PACE (Pace According to Current Estimated utility), which generates asymptotically fair and efficient allocations when items are drawn in an i.i.d. manner. PACE gives each agent a per-round budget of faux currency, and simulates a first-price auction in each round. The fair allocation is achieved by having each agent shade their bid with a pacing multiplier, which is a projection of their current estimated inverse bang-per-buck to a fixed interval. Liao et al. (2022) extend these results to non-stationary inputs, where the distribution of items may change over time. They show that in this case, PACE still achieves asymptotic fairness and efficiency guarantees, up to linear error terms from the amount of non-stationarity.

Yet in many real-world scenarios, we cannot expect items to be drawn in a stochastic manner, even if from non-stationary distributions. This motivates the investigation of algorithms with competitive ratio guarantee for adversarial settings. To fit arbitrary inputs, including extreme ones, some algorithms adopt “conservative” designs for fairness, such as allocating half of each item purely equally  (Banerjee et al., 2022). Although this helps to provide worst-case guarantees, it damages the efficiency in the average case, which may not be acceptable in some practical applications. Moreover, this conservative allocation requires each item to be divisible, or at least for random allocation to be acceptable.

This motivates us to move in another direction: Instead of developing algorithms to fit extreme adversarial inputs, we seek to find worst-case guarantees for existing algorithms that are designed for stochastic inputs. In particular, we focus on the performance of the algorithm PACE  (Gao et al., 2021; Liao et al., 2022), and explore the question:

How does PACE perform under adversarial input?

Our first contribution is to show that, in the case where all agents have the same weight (or budget in market equilibrium terminology), PACE is equivalent to the integral greedy algorithm, assuming no projection of the pacing multiplier. Due to this equivalence, we start by studying the integral greedy allocation method. Our results for that method are of independent interest, as it is a natural allocation method.

Although both integral greedy allocation and PACE have infinite envy and Ω(T)\Omega(T) competitive ratio when inputs are completely arbitrary, we notice that such pessimistic results only occur under extreme inputs where the ratio between the largest and smallest non-zero values for an agent differ by an exponential factor. We show that, once we rule out such extreme instances by introducing mild assumptions (such as a constant ratio between nonzero values observed by an agent) both algorithms converge with approximate envy-freeness and bounded competitive ratio as TT increases to infinity. The upper bounds are either constant, or in near-optimal order of nn, see Table 1. Combined with existing results under stationary  (Gao et al., 2021) and non-stationary  (Liao et al., 2022) input models, this establishes a “best-of-many-worlds” guarantee for PACE: it is the first online algorithm that simultaneously guarantees asymptotic fairness and efficiency guarantees under stochastic, stochastic but nonstationary, and adversarial inputs. As such, we believe our results show that PACE is a natural and robust algorithm for online fair allocation in real-world settings, since it achieves strong guarantees under many different utility models, and is thus likely to perform well on a variety of real-world inputs.

Algorithm Assumptions Measure Upper-Bound (TT\rightarrow\infty) Theorem
Integral Greedy Assumption 3.1 multiplicative envy 1+2log1ε1+2\log\frac{1}{\varepsilon} Theorem 3.1
competitive ratio w.r.t. NW λ(n!)1/n+α,α>0{\lambda\cdot(n!)^{1/n+\alpha},\forall\alpha>0}^{*} Theorem 3.3
seed utility δ\delta utility ratio with seeds O(logT)O(\log T) Theorem 3.5
PACE Assumption 4.1 multiplicative envy 1+2log1ε1+2\log\frac{1}{\varepsilon} Theorem 4.1
Assumption 4.1 competitive ratio w.r.t. NW (1+2log1ε)1c\left(1+2\log\frac{1}{\varepsilon}\right)\frac{1}{c} Theorem 4.2

The lower bound for online algorithms is at least (n!)1/n(n!)^{1/n} when nn\rightarrow\infty.

Table 1: Summary of Results

1.1 Related Work

In this subsection, we review previous works that are most closely related to ours. An extensive review of other related works is provided in Appendix 0.A.

The PACE Algorithm

Our work is a direct generalization of the PACE algorithm (Pace According to Current Estimated utility) (Gao et al., 2021; Liao et al., 2022) to adversarial inputs. We will review PACE in detail in Section 2.

Online Fair Division

For maximizing Nash welfare in the online setting, the competitive ratio with respect to the offline optimal allocation is trivially Ω(n)\Omega(n) and Ω(T)\Omega(T), when the input is completely adversarial (Banerjee et al., 2022). These pessimistic results under arbitrary input motivate the introduction of assumptions. Azar et al. (2016) adopted an assumption that we also make: that the minimum nonzero valuation of each agent is at least ε\varepsilon times the largest. However, their analysis does not remove the dependence on TT in their upper bound of the competitive ratio O(lognTε)O\left(\log\frac{nT}{\varepsilon}\right), meaning that the ratio is unbounded in the online setting with a fixed number of agents and an unbounded horizon. Banerjee et al. (2022) assumes extra prior knowledge of the monopolistic utility of each agent, and gives an O(logn)O(\log n) and O(logT)O(\log T)-competitive algorithm; their algorithm involves allocating half of each item uniformly across agents; since we are interested in algorithms with asymptotic convergence guarantees on non-adversarial inputs, such an approach cannot be used. Huang et al. (2022) assumes the input to be λ\lambda-balanced or μ\mu-impartial, where λ\lambda and μ\mu characterize the desired properties of the input; their competitive ratio upper bounds are logarithmic in the parameter and nn. However, the parameters still implicitly depend on the horizon length TT.

The above works on online NW maximization consider greedy-style algorithms as we do. In contrast to them, our work focuses on an asymptotic bound that does not depend on TT as TT\rightarrow\infty. Moreover, while their algorithms can only deal with divisible items, our paper adopts integral allocation, and show that integral decision is enough for the convergence of multiplicative envy and competitive ratio w.r.t. the optimal continuous allocation, given our assumptions.

Online Allocation with Resource Constraints

We also briefly discuss how this paper differs from existing works in online resource allocation, where a sequence of requests arrive over time, with each request consisting of a reward and cost function, and at each time step the algorithm must select a decision with the goal of maximizing the sum of rewards while satisfying long-term cost constraints on each resource. In that setting, strong best-of-many-worlds guarantees are known (Balseiro et al., 2023; Celli et al., 2022; Castiglioni et al., 2022). In online resource allocation, the optimization objective is separable across timesteps, e.g., in the form of s=1Tfs(x)\sum_{s=1}^{T}f_{s}(x) (Balseiro et al., 2023). Time-separability is crucial for the regret bounds in these works, as it enables translating dual regret to primal regret through weak duality. However, in online fair allocation with the Nash welfare objective, time-separability no longer holds, since we take the logarithm of the utility over time. Therefore, our results cannot be derived with similar techniques to those papers. Indeed, the types of competitive-ratio guarantees achieved e.g. by Balseiro et al. (2023) are impossible in the online fair allocation setting, where hard input sequences are known (Banerjee et al., 2022; Gao et al., 2021).

2 Setup

2.1 Online Fair Allocation

Consider a problem instance with nn agents and TT items. For i[n]i\in[n] and t[T]t\in[T], let vit0v_{i}^{t}\geq 0 be agent ii’s value for a unit of item tt. The input of our problem is a sequence of agent valuations 𝒗=(vit)n×T\bm{v}=(v_{i}^{t})^{n\times T}. We assume that for each item jj at least one agent values it, i.e. there exists an agent ii such that vit>0v_{i}^{t}>0. Each agent i[n]i\in[n] has a non-negative weight BiB_{i}, which can also be interpreted as a budget of faux currency in a Fisher market (Varian, 1974). An allocation 𝒙=(xit)n×T\bm{x}=(x_{i}^{t})^{n\times T} distributes each item to an agent, where xitx_{i}^{t} is the amount of item tt that is allocated to agent ii gets item tt. We assume each item has a unit supply. An allocation is feasible if i[n]xit1\sum_{i\in[n]}x_{i}^{t}\leq 1 for each tt. An allocation is integral if xit{0,1}x_{i}^{t}\in\{0,1\}. For a feasible, integral allocation xx, let Ai={t:xit=1}A_{i}=\{t:x_{i}^{t}=1\} be the set of items allocated to agent ii.

We assume additive, linear utility for all agents. That is to say, Uit=s=1txisvisU_{i}^{t}=\sum_{s=1}^{t}x_{i}^{s}v_{i}^{s}, where UitU_{i}^{t} is the utility agent ii derives from the first tt items. For a subset of items A[T]A\subseteq[T], agent ii’s total value on the bundle AA is denoted as Ui(A)=sAvisU_{i}(A)=\sum_{s\in A}v_{i}^{s}. Agent ii’s monopolistic utility ViV_{i} is defined as his total value for all items Vi=Ui([T])=t=1Tvit.V_{i}=U_{i}([T])=\sum_{t=1}^{T}v_{i}^{t}.

We focus on the online setting where items arrive sequentially, while the set of agents is fixed. An online allocation algorithm is one that makes an irrevocable choice to distribute the item in each round based only on the information of past rounds. Concretely, it maps the history t={(vis)s=1t,(xis)s=1t1}i=1n\mathcal{H}^{t}=\left\{(v_{i}^{s})_{s=1}^{t},(x_{i}^{s})_{s=1}^{t-1}\right\}_{i=1}^{n} to a decision (x1t,,xnt)(x_{1}^{t},\cdots,x_{n}^{t}) such that i=1nxit=1\sum_{i=1}^{n}x_{i}^{t}=1.

In this paper, we are interested in envy and Nash welfare (NW) as measures of fairness. The multiplicative envy of agent ii to agent jj is defined as the ratio between the utility that agent ii would get from the allocation xjx_{j} of agent jj to the utility of their own allocation xix_{i}, adjusted by their respective budget. As a criterion for fairness, it measures the extent to which an agent prefers someone else’s bundle to his own.

Envyij=BiBjt=1Txjtvitt=1Txitvit.\mathrm{Envy}_{ij}=\frac{B_{i}}{B_{j}}\frac{\sum_{t=1}^{T}x_{j}^{t}v_{i}^{t}}{\sum_{t=1}^{T}x_{i}^{t}v_{i}^{t}}.

Notice that when the allocation is integral, the above definition becomes Ui(Aj)/Ui(Ai){U_{i}(A_{j})}/{U_{i}(A_{i})}.

The Nash welfare of an allocation is defined as the weighted geometric mean of all agents’ utilities:

NW=i=1n(UiT)Bi/Bj.\mathrm{NW}=\prod_{i=1}^{n}\left(U_{i}^{T}\right)^{B_{i}/\sum B_{j}}.

Maximizing the geometric mean is a well-studied proxy for balancing fairness and efficiency in fair allocation problems. It is also equivalent to maximizing the objective of the Eisenberg-Gale convex program in a Fisher market. For a Nash welfare maximizing allocation 𝒙={xi,t}\bm{x}^{\star}=\{x^{\star,t}_{i}\}, let Ui=t=1Txi,tvitU_{i}^{\star}=\sum_{t=1}^{T}x^{\star,t}_{i}v_{i}^{t} be the utility of agent i[n]i\in[n]. Notice that the optimal allocation might not be unique. In this paper, we measure the performance in terms of Nash welfare maximization based on the competitive ratio of our allocation, which is defined as the supremum of the ratio between the online allocation and an optimal offline NW-maximizing allocation, over the space of all possible inputs.

While we allow the input sequence to be adversarial, we restrict our attention to a subset of adversarial input sequences, where we use 𝒱T\mathcal{V}^{T} to denote the space of valid length TT input sequences. Concrete assumptions on 𝒱T\mathcal{V}^{T} will be specified in Section 3 and Section 4 before analyzing specific algorithms.

We consider the asymptotic worst-case envy and competitive ratio (w.r.t. Nash welfare) over all possible inputs 𝒱T\mathcal{V}^{T} when TT\rightarrow\infty. Concretely, we will investigate

limTsupv𝒱Tmaxi,jEnvyij,limTsupv𝒱Ti=1n(UiTUi)Bi/Bj.\lim_{T\rightarrow\infty}\sup_{v\in\mathcal{V}^{T}}\max_{i,j}\mathrm{Envy}_{ij},\ \lim_{T\rightarrow\infty}\sup_{v\in\mathcal{V}^{T}}\prod_{i=1}^{n}\left(\frac{U_{i}^{T}}{U_{i}^{\star}}\right)^{B_{i}/\sum B_{j}}.

In our analysis, we will show that with proper assumptions, both measures converge with an asymptotic upper bound which is independent of TT. Our upper bounds will be either constant or with a near-optimal order dependence on nn.

We emphasize that both measures adopted in this paper, multiplicative envy, and competitive ratio, are defined as ratios, not differences. This is mainly because of the scale invariant property of fair division, which is an important and desirable property for allocation algorithms: if an agent’s values for all items are multiplied by a constant factor, the resulting allocation stays the same. The algorithms that we are interested in, together with NW-maximizing allocations, are all scale invariant. Hence, it is more useful to consider multiplicative performance measures, which are also invariant to valuation scaling.

2.2 Algorithms

We introduce the two major algorithms that we study in this paper: the PACE (Pace According to Current Estimated Utility) algorithm (Gao et al., 2021; Liao et al., 2022), and the integral greedy algorithm. Moreover, we will discuss the “greedy-based” nature of PACE by showing its equivalence to the integral greedy algorithm under certain conditions.

Pseudocode for PACE is shown in Algorithm 1. In each round tt, the agent utilities are revealed. Each agent then places a bid for the item, which is equal to their value for the item multiplied by the current pacing multiplier βit\beta_{i}^{t}. The whole item is allocated to the highest bidder, preferring the bidder with the smallest index when a tie occurs. Each agent then observes their realized utility at this round, and updates their current estimated utility. The pacing multiplier is updated to be the weight BiB_{i} divided by the estimated utility, then projected to an interval [a,b][a,b].

Input: number of agents nn, time horizon TT, truncation parameters aa and bb.
Initialization: Ui0=0,β1=1nU_{i}^{0}=0,\beta^{1}=1^{n}.
for t=1,,Tt=1,\cdots,T, do
       Agent ii bids βitvit\beta_{i}^{t}v_{i}^{t}.
      The whole item tt is allocated to the highest bidder, with arbitrary tie-breaking:
it:=minargmaxi[n]βitvit,xit=𝟏(i=it).i^{t}:=\min\arg\max_{i\in[n]}\beta_{i}^{t}v_{i}^{t},\ \ x_{i}^{t}=\bm{1}(i=i^{t}).
Agent ii updates his estimated utility
u¯it=1txitvit+t1tu¯it1.\bar{u}_{i}^{t}=\frac{1}{t}\cdot x_{i}^{t}v_{i}^{t}+\frac{t-1}{t}\bar{u}_{i}^{t-1}.
Agent ii updates the pacing multiplier
βit+1=[a,b](Biu¯it)\beta_{i}^{t+1}=\prod_{[a,b]}\left(\frac{B_{i}}{\bar{u}_{i}^{t}}\right)
Algorithm 1 PACE

2.2.1 Performance Under Stochastic Input

As shown by Gao et al. (2021), PACE is an instantiation of stochastic unregularized dual averaging  (Xiao, 2009) applied to the dual of the underlying allocation program where the supplies are given by the density of each item. With i.i.d. input, PACE converges to the equilibrium of a potentially infinite-dimensional Fisher market (Gao and Kroer, 2023), which is also closely related to the game-theoretic solution concept of pacing equilibrium (Conitzer et al., 2022). The agent utilities under PACE converge to those associated with the offline NW-maximizing allocation in the mean-square sense.

Theorem 2.1.

(Theorem 4. in Gao et al. (2021)) Let ui:=Ui/Tu_{i}^{\star}:=U_{i}^{\star}/T be agent ii’s time-averaged utility under the Nash-welfare-maximizing allocation with supplies given by some underlying distribution. When the values in different rounds are i.i.d. chosen from the same distribution, it holds that

𝑬[t=1T(u¯itui)]ClogTT,\bm{E}\left[\sum_{t=1}^{T}(\bar{u}_{i}^{t}-u_{i}^{\star})\right]\leq C\cdot\frac{\log T}{T}, (1)

where CC is a constant independent of TT.

Liao et al. (2022) generalizes Theorem 2.1 to non-stationary inputs, which have a stochastic component yet change over time. Particularly, they consider three types of inputs: independent yet adversarially corrupted input, ergodic input, and periodic input, and show that for all three cases 𝑬uTu0\bm{E}\|u^{T}-u^{\star}\|\rightarrow 0 is still preserved, up to errors due to non-stationarity.

In the stationary and non-stationary cases, mean-square convergence of time-averaged utility implies an asymptotic competitive ratio of 11 w.r.t Nash welfare. For both cases, Gao et al. (2021); Liao et al. (2022) also provide a theoretical guarantee that PACE is asymptotically envy-free (again up to a non-stationarity error in the non-stationary case). In this paper, we provide bounds on PACE’s performance with adversarial inputs with assumptions. Combined with the results from Gao et al. (2021); Liao et al. (2022), this is the first “best-of-many-worlds” guarantee for online fair allocation under stationary, non-stationary, and adversarial inputs.

While we make an attempt to take an algorithm for stochastic inputs and show its performance on adversarial input, the other direction seems to be difficult. It is hard for some algorithms that are designed for adversarial inputs to achieve optimality in stochastic scenarios, due to “conservative” routines that they adopt to deal with extremely bad inputs. For instance, Banerjee et al. (2022) divides half of the resources equally, which can be undesirable with stochastic input, see Example 1

Example 1.

Consider an online scenario with nn agents and nn types of items {θj}j=1n\{\theta_{j}\}_{j=1}^{n}. Agent nn’s value for a unit of type-jj item vi(θj)=1v_{i}(\theta_{j})=1 if i=ji=j, and vi(θj)=0.01v_{i}(\theta_{j})=0.01 otherwise. In each round, the item type is drawn i.i.d. from a uniform distribution. In this scenario,

  • The equilibrium of the underlying Fisher market allocates all type-ii items to agent ii. PACE converges to this equilibrium.

  • The proposed algorithm proposed by (Banerjee et al., 2022) allocates at most (12+1n)\left(\frac{1}{2}+\frac{1}{n}\right) fraction of type-ii item to agent ii, which is clearly not optimal.

2.2.2 Greedy Interpretation of PACE.

In the standard configuration of online fair allocation, where agents have equal weight B1==Bn=1B_{1}=\cdots=B_{n}=1, we now show that PACE, when the projection of multipliers is disregarded, can be interpreted as greedily maximizing Nash welfare with integral decisions.

To show this, the following optimization program (2) maximizes NW up to round tt greedily, given the history of previous t1t-1 rounds. Its decision is integral.

maxxit{0,1}\displaystyle\max_{x_{i}^{t}\in\{0,1\}} i=1nBilogUit\displaystyle\ \ \sum_{i=1}^{n}B_{i}\log U_{i}^{t} (2)
s.t. Uit=s=1t1xisvis+xitvit,i[n]\displaystyle\ \ U_{i}^{t}=\sum_{s=1}^{t-1}x_{i}^{s}v_{i}^{s}+x_{i}^{t}v_{i}^{t},\ \forall i\in[n]
i=1nxit=1\displaystyle\ \ \sum_{i=1}^{n}x_{i}^{t}=1

The above program allocates the item to the agent iti^{t} that gives the maximum increment to the objective. This is equivalent to the following decision rule:

itargmaxi[n]Bilog(1+vitUit1).i^{t}\in\arg\max_{i\in[n]}B_{i}\log\left(1+\frac{v_{i}^{t}}{U_{i}^{t-1}}\right).

When agent weights are equal, iti^{t} is the agent ii that maximizes vit/Uit1{v_{i}^{t}}/{U_{i}^{t-1}}; this coincides exactly with the decision of PACE with no projections.

This interpretation motivates us to first consider the behavior of a greedy algorithm (without projecting the multiplier), and then study PACE based on the insights derived for integral greedy. We focus on weight-adapted integral greedy algorithm, or simply integral greedy algorithm, as shown in Algorithm 2.

Input: number of agents nn, time horizon TT
Initialization: Ui0=0U^{0}_{i}=0 for all ii.
for t=1,,Tt=1,\cdots,T, do
       Observe agent values for item tt, and allocates the whole item to agent iti^{t}:
it:=min(argmaxi[n]BivitUit1),xit=𝟏(i=it).i^{t}:=\min\bigg{(}\mathrm{arg}\max_{i\in[n]}\frac{B_{i}v_{i}^{t}}{U_{i}^{t-1}}\bigg{)},\ \ x_{i}^{t}=\bm{1}(i=i^{t}).
Agent ii updates his current utility
Uit=Uit1+xitvit.U_{i}^{t}=U_{i}^{t-1}+x_{i}^{t}v_{i}^{t}.
Algorithm 2 Integral Greedy Algorithm (weight-adapted)

Algorithm 2 is equivalent to greedily maximizing the NW objective with integral decisions when agent weights are equal. However, with unequal weights the equivalence breaks down, so does its equivalence between PACE. For this consideration, in the following discussions, we assume equal weights B1==BnB_{1}=\cdots=B_{n}, which is a standard setting in the Nash welfare maximization and online fair allocation literature. We will also remark on part of our results that can be generalized to unequal weights. Note that the PACE algorithm itself extends to unequal weights, and the stationary and nonstationary results extend as well.

With equal weight assumption, PACE can be regarded as greedy-based decision using U^it\hat{U}_{i}^{t} as estimated utility. U^it\hat{U}_{i}^{t} is a projected utility, where \ell and rr are reinterpreted projection bounds:

it:=minargmaxi[n]vitU^it1,U^it=[t,rt]Uit.i^{t}:=\min\mathrm{arg}\max_{i\in[n]}\frac{v_{i}^{t}}{\hat{U}_{i}^{t-1}},\ \hat{U}_{i}^{t}=\prod_{[\ell t,rt]}U_{i}^{t}.

3 Analysis of the Integral Greedy Algorithm

This section is devoted to show that integral greedy algorithm achieves convergence of multiplicative envy and competitive ratio, under reasonable assumptions. The missing proofs in this section are deferred to Appendix 0.B.

3.1 Assumptions on the Input

We begin with introducing the assumptions on the input space, as well as the necessity of doing so. We focus on input space 𝒱εT\mathcal{V}_{\varepsilon}^{T}, which is parametrized by ε(0,1]\varepsilon\in(0,1] and defined as follows:

{assumption}

For each integer TT, 𝒱εT\mathcal{V}_{\varepsilon}^{T} is the set of inputs which satisfy the following requirements:

  • The monopolistic utility of each agent is infinite: Vi=(T).V_{i}=\infty\ (T\rightarrow\infty).

  • For each i[n]i\in[n] and t[T]t\in[T], vit{0}[ε,1]v_{i}^{t}\in\{0\}\cup[\varepsilon,1].

The first requirement helps to avoid allocating nothing to some agent, since the decision is integral. We further require the number of nonzero valuations of each agent is not upper-bounded by a constant, so it is meaningful to consider the asymptotic sense in TT. For the second requirement, we note that it is equivalent to assuming a constant bound on the ratio of the minimum nonzero item value and the maximum:

mint{vit:vit>0}maxt{vit}ε,i[n].\frac{\min_{t}\{v_{i}^{t}:v_{i}^{t}>0\}}{\max_{t}\{v_{i}^{t}\}}\geq\varepsilon,\ \forall i\in[n].

This equivalence is due to the scale-invariant property of integral greedy allocation, as well as the NW maximizing allocation.

If we do not include any extra assumptions on the input space, the worst-case envy and the competitive ratio is infinite, which makes the analysis of the integral greedy algorithm trivial and uninteresting.

Lemma 1.

When the values are arbitrarily chosen from [0,1][0,1], it holds that

  1. 1.

    (Banerjee et al., 2022) Any online allocation algorithm has Ω(T)\Omega(T) competitive ratio with respective to NW.

  2. 2.

    The integral greedy algorithm has Ω(T)\Omega(T) worst-case multiplicative envy.

even when the first requirement in Section 3.1 is satisfied.

Proof.

We construct a sequence of instances where Envy12=Ω(T)\mathrm{Envy}_{12}=\Omega(T). Given horizon length TT, fix agent 11’s valuation to 11 in all rounds. For agent 22, we set v2t=atT1v_{2}^{t}=a^{t-T}\leq 1 for some a>2a>2. In the integral greedy allocation, agent 11 receives only one item and has total utility 11. This proves the lower bound of multiplicative envy. ∎

The above hard instance features exponential growth in agent 22’s valuation. Hard instances with similar spirit for the continuous problem have previously been given by Banerjee et al. (2022). The vulnerability of allocation algorithms under such instances can be explained by their non-anticipating nature: it does not know the future, so it fails to recognize that agent 22’s valuation up to current rounds is only a negligible fraction of the entire horizon. More generally, it is difficult for online algorithms to distinguish agents that are hard to satisfy in the future, with those who are easily satisfied.

However, we note that such adversarial instances are arguably not “natural.” For allocation problem in a real-world market, items are usually similar in nature, e.g., they are all food or ad slots. It is unlikely for an agent to have exponentially diverging nonzero values on these items. By requiring the ratio of minimal and maximal nonzero value to be bounded by ε\varepsilon, section 3.1 rules out extreme cases where values from a single agent diverge drastically. In the rest of this section, we will show that once the above assumptions are introduced, both multiplicative envy and competitive ratio of the greedy algorithm are independent of TT asymptotically, i.e., converge to a constant (which depends only on nn and ε\varepsilon).

3.2 Envy Analysis for Greedy

In this subsection we analyze the worst-case envy of agents in integral greedy allocation. We observe that, the envy between any pair of agents can be reduced to the case n=2n=2 by the following lemma, which characterized an “induction” structure of the integral greedy allocation.

Lemma 2 (Inductive structure of greedy allocation.).

For any nn-agent instance 𝐯\bm{v} and agent subset I[n]I\subseteq[n], define a new instance 𝐯|I\bm{v}|_{I} obtained by transforming 𝐯\bm{v} as:

  • Remove all agents that are not in II (by setting their values to 0 on all items).

  • Remove all items that are not in iIAi\bigcup_{i\in I}A_{i} (by setting all agents’ value to 0 on them).

Then, the resulting allocation is the same when the integral greedy algorithm is run on 𝐯\bm{v} and 𝐯|I\bm{v}|_{I}.

We show that the multiplicative envy of the integral greedy algorithm is upper bounded by 11 plus a logarithmic term in 1/ε1/\varepsilon, which also generalizes to unequal weights.

Theorem 3.1 (Upper Bound for Multiplicative Envy).

Even with unequal agent weights, for inputs 𝐯\bm{v} satisfying Assumption 3.1 and any i,ji,j,

supv𝒱εTEnvyij1+2log1ε+O(1T).\sup_{v\in\mathcal{V}_{\varepsilon}^{T}}\mathrm{Envy}_{ij}\leq 1+2\log\frac{1}{\varepsilon}+O\left(\frac{1}{T}\right).
Proof Sketch.

Due to Lemma 2 and symmetry, it suffices to consider Envy21\mathrm{Envy}_{21} in 22-agent inputs. In the proof sketch, we assume B1=B2=1B_{1}=B_{2}=1 for simplicity. We transform the input by:

  1. 1.

    Set agent 11’s valuation for all items in A2A_{2} to 0.

  2. 2.

    Move all items in A2A_{2} to the beginning of the input sequence, and all items in A1A_{1} to the end of the input sequence (preserving order).

One can show that the agent’s utilities under the budget-adapted greedy algorithm are invariant to this transformation. Therefore, it suffices to consider only the transformed inputs, where agent 22 receives his entire share only in beginning RR rounds.

To find the worst-case envy for transformed inputs, a question from an adversarial point of view will be: given (U1R,U2R)=(0,U)(U_{1}^{R},U_{2}^{R})=(0,U), how can we design a value sequence for the coming rounds, such that agent 22’s total valuation on items over SS rounds is maximized, while ensuring that nothing is allocated to agent 22? This can be characterized by an optimization program:

maxv1t,v2t\displaystyle\max_{v_{1}^{t},v_{2}^{t}} 1Ut=1v2t\displaystyle\ \ \frac{1}{U}\sum_{t=1}^{\infty}v_{2}^{t} (3)
s.t. v1t,v2t{0}[ε,1],\displaystyle\ \ v_{1}^{t},v_{2}^{t}\in\{0\}\cup[\varepsilon,1],\ \ t1\displaystyle\forall t\geq 1
v2tv1tUU1t1,\displaystyle\ \ \frac{v_{2}^{t}}{v_{1}^{t}}\leq\frac{U}{U_{1}^{t-1}},\ \ t1\displaystyle\forall t\geq 1
U1ts=1tv1s,\displaystyle\ \ U_{1}^{t}\geq\sum_{s=1}^{t}v_{1}^{s},\ \ t1\displaystyle\forall t\geq 1

Notice that in (3) we re-index the rounds by starting with index 11 at round R+1R+1. We call (3) a canonical optimization program for multiplicative envy maximization, parametrized by UU. We observe that v2t/v1tv_{2}^{t}/v_{1}^{t} can be upper bounded by q(U1t1)q(U_{1}^{t-1}), defined as

q(U1t1)={min{UU1t1,1ε},0U1t1Uε0,U1t1>Uε.q(U_{1}^{t-1})=\begin{cases}\min\left\{\frac{U}{U_{1}^{t-1}},\frac{1}{\varepsilon}\right\},&0\leq U_{1}^{t-1}\leq\frac{U}{\varepsilon}\\ 0,&U_{1}^{t-1}>\frac{U}{\varepsilon}\end{cases}.

We can then give an upper-bound of the objective of the canonical optimization program (3),

1Ut=1v2t=1Ut=1v1tv2tv1t1Ut=1v1tq(U1t1).\frac{1}{U}\sum_{t=1}^{\infty}v_{2}^{t}=\frac{1}{U}\sum_{t=1}^{\infty}v_{1}^{t}\cdot\frac{v_{2}^{t}}{v_{1}^{t}}\leq\frac{1}{U}\sum_{t=1}^{\infty}v_{1}^{t}\cdot q(U_{1}^{t-1}). (4)

Each increment v1t1v_{1}^{t}\leq 1 is small when UU\rightarrow\infty. One can show that the right hand side of (4) converges to a definite integral asymptotically as UU\rightarrow\infty:

1Ut=1v1tq(U1t1)1U0U/εq(U1)dU1=1+2log1ε.\frac{1}{U}\sum_{t=1}^{\infty}v_{1}^{t}\cdot q(U_{1}^{t-1})\rightarrow\frac{1}{U}\int_{0}^{U/\varepsilon}q(U_{1})\mathrm{d}U_{1}=1+2\log\frac{1}{\varepsilon}.

The convergence rate on the order of O(1/T)O(1/T) is then achieved by more carefully calculating the above upper bound. \hfill\square

The analysis of the canonical optimization program (3) also gives an upper bound on the number of items that agent 22 “envies of” agent 11, concretely, items in A1A_{1} with nonzero value from agent 22. Lemma 3 will be useful in Section 4.

Lemma 3.

Let Ci:={t:vit>0}C_{i}:=\{t:v_{i}^{t}>0\} be the set of items on which agent ii has non-zero value. When agents have equal weights, for 𝐯𝒱εT\bm{v}\in\mathcal{V}_{\varepsilon}^{T} and any i,j[n]i,j\in[n],

limTsupvVεT|CiAj|Ui(Ai)1ε(1+log1ε).\lim_{T\rightarrow\infty}\sup_{v\in V_{\varepsilon}^{T}}\frac{|C_{i}\cap A_{j}|}{U_{i}(A_{i})}\leq\frac{1}{\varepsilon}\left(1+\log\frac{1}{\varepsilon}\right).

Next, we complement Theorem 3.1 with a lower bound showing that the bound 1+2log1/ε1+2\log 1/\varepsilon on envy is tight for the integral greedy allocation.

Theorem 3.2 (Lower Bound for Multiplicative Envy).

For input 𝐯𝒱εT\bm{v}\in\mathcal{V}_{\varepsilon}^{T}, we have for any i,j[n]i,j\in[n],

limTsupv𝒱εTBiBjUi(Aj)Ui(Ai)1+2log1ε.\lim_{T\rightarrow\infty}\sup_{v\in\mathcal{V}_{\varepsilon}^{T}}\frac{B_{i}}{B_{j}}\cdot\frac{U_{i}(A_{j})}{U_{i}(A_{i})}\geq 1+2\log\frac{1}{\varepsilon}.

3.3 Nash Welfare Analysis for Greedy

We give an upper bound on the asymptotic competitive ratio w.r.t. Nash welfare for the integral greedy algorithm.

Theorem 3.3 (Upper Bound for Competitive Ratio).

For input space 𝒱εT\mathcal{V}_{\varepsilon}^{T} and any given α>0\alpha>0 there exists a constant λ>0\lambda>0 (independent of nn and TT), such that

limTsupv𝒱εT(i=1nUi(Ai)i=1nUi(Ai))1/nλ(n!)1+αn.\lim_{T\rightarrow\infty}\sup_{v\in\mathcal{V}^{T}_{\varepsilon}}\left(\frac{\prod_{i=1}^{n}U_{i}(A_{i})}{\prod_{i=1}^{n}U_{i}(A_{i}^{\star})}\right)^{1/n}\leq\lambda\cdot\left(n!\right)^{\frac{1+\alpha}{n}}.

For ε=1\varepsilon=1, the above holds with λ=1,α=0\lambda=1,\alpha=0.

Proof Sketch

Assume without loss of generality that U1(A1)Un(An)U_{1}(A_{1})\leq\cdots\leq U_{n}(A_{n}). The main idea of the proof is to show that Ui/UiU_{i}^{\star}/U_{i} is bounded by (ni+1)iα(n-i+1)\cdot i^{\alpha} asymptotically. Suppose this is not true, we show that xix_{i}^{\star} will include large proportion of xjx_{j} (j<ij<i), which will lead to contradiction with the optimality of 𝒙\bm{x}^{\star}.

Although Azar et al. (2016) gives an O(log(nT/ε))O(\log(nT/\varepsilon)) algorithm, we show that (n!)1/n(n!)^{1/n} factor is inevitable if one aims to remove the dependence on TT. Hence, integral greedy algorithm is near-optimal in terms of nn.

Theorem 3.4 (Lower Bound for Competitive Ratio).

Even in the case ε=1\varepsilon=1, for any feasible, deterministic online algorithm we have

(i=1nUi(Ai)i=1nUi)1/n(n!)1/n.\left(\frac{\prod_{i=1}^{n}U_{i}(A_{i})}{\prod_{i=1}^{n}U_{i}^{\star}}\right)^{1/n}\geq(n!)^{1/n}.
Proof Sketch.

We sketch the construction of an adaptive adversary, who attempts to make low-utility agents hard to satisfy in the future. Divide the horizon into nn phases, each with length TiT_{i}, satisfying Ti/Ti1.T_{i}/T_{i-1}\rightarrow\infty. The adversary maintains a set of “active agents”, initially containing all agents. In each round, only currently active agents see nonzero values. At the end of each phase, the agent in the active set who has lowest utility is eliminated. This results in a competitive ratio of (n!)1/n(n!)^{1/n}, see the detailed proof in Section 0.B.8.

3.4 Nash Welfare Analysis without Section 3.1

As an extension for our analysis on the integral greedy algorithm, we show how it can be adapted when Section 3.1 does not hold. Despite the Ω(T)\Omega(T) lower bound, we show that, when each agent begins with a seed utility δ\delta, the competitive ratio of integral greedy algorithm is of order O(logT)O(\log T). The performance is measured w.r.t. to the ratio of seeded welfare.

The seeded integral greedy algorithm is identical to Algorithm 2, except that all agents are given an initial seed utility δ\delta, which is taken into account when deciding the winning agent:

it:=min(argmaxi[n]vitδ+Uit1),xit=𝟏(i=it).\displaystyle i^{t}:=\min\bigg{(}\mathrm{arg}\max_{i\in[n]}\frac{v_{i}^{t}}{\delta+U_{i}^{t-1}}\bigg{)},\ \ x_{i}^{t}=\bm{1}(i=i^{t}). (5)

The full pseudocode for the seeded algorithm is presented in Algorithm 3.

Input: number of agents nn, time horizon TT, seed utility δ\delta
Initialization: Ui0=0U^{0}_{i}=0 for all ii.
for t=1,,Tt=1,\cdots,T, do
       Observe agent values for item tt, and allocates the whole item to agent iti^{t}:
it:=min(argmaxi[n]vitδ+Uit1),xit=𝟏(i=it).\displaystyle i^{t}:=\min\bigg{(}\mathrm{arg}\max_{i\in[n]}\frac{v_{i}^{t}}{\delta+U_{i}^{t-1}}\bigg{)},\ \ x_{i}^{t}=\bm{1}(i=i^{t}). (6)
Agent ii updates current utility
Uit=Uit1+xitvit.U_{i}^{t}=U_{i}^{t-1}+x_{i}^{t}v_{i}^{t}.
Output: Allocations {xit}i,t\{x_{i}^{t}\}_{i,t}
Algorithm 3 Seeded Integral Greedy Algorithm (Equal Weights)

For the seeded algorithm, we study the criterion Rδ(𝒗)R_{\delta}(\bm{v}) which is also defined w.r.t. the seeded utility:

Rδ(v)=supU~{1ni=1nU~i+δUi+δ},\displaystyle R_{\delta}(v)=\sup_{\widetilde{U}}\Bigg{\{}\frac{1}{n}\sum_{i=1}^{n}\frac{\widetilde{U}_{i}+\delta}{{U}_{i}+\delta}\Bigg{\}}\;,

where the supremum runs over all feasible hindsight allocations, with resulting utility U~1,,U~n.\widetilde{U}_{1},\cdots,\widetilde{U}_{n}. Notice that by the AM-GM inequality, RδR_{\delta} is also competitive w.r.t. the geometric mean. However, due to the presence of the seed utility δ\delta, it is not directly comparable to the criterion (i=1nUi/Ui)1/n(\prod_{i=1}^{n}{U^{\star}_{i}}/{U_{i}})^{1/n}.

Theorem 3.5 (Upper Bound of RδR_{\delta} for Seeded Integral Greedy).

Run the seeded integral greedy algorithm with seed utility δ\delta. For any input 𝐯\bm{v} satisfying vit[0,1]v_{i}^{t}\in[0,1],

Rδ(v)3+4δ+2log(1+1δ)+2logT.\displaystyle R_{\delta}(v)\leq 3+\frac{4}{\delta}+2\log\Big{(}1+\frac{1}{\delta}\Big{)}+2\log T\;. (7)

The ratio is of order logT\log T, similar to the result given by Azar et al. (2016). Unlike our previous results, here the ratio is unbounded as TT\to\infty since we are considering a broader range of inputs beyond Section 3.1. Also, there is a trade-off in choosing seed utility δ\delta. Although a larger δ\delta brings a better bound in (7), it is less able to tell us how the algorithm compares to the offline optimal, since Rδ(v)1R_{\delta}(v)\rightarrow 1 as δ\delta grows large.

4 Analysis of PACE

We continue to analyzing the performance of PACE. Missing proofs in this section can be found in Appendix 0.C.

4.1 Assumptions on the Input

To analyze PACE in the adversarial setting, it is necessary to adopt more assumptions than for the integral greedy algorithm. This is because PACE projects the average utility up to round tt to [,r][\ell,r] in its decision, or equivalently, projects the total utility in round tt to [t,rt][\ell t,rt]. In the stationary and non-stationary setting of Liao et al. (2022), the expected utility of each agent grows uniformly with time; in that case the projection helps PACE achieve theoretical guarantees on its performance. However, in the adversarial setting, the input may vary drastically over time, which makes the projection operation more problematic. In the worst case, adversarial input might cause extreme behavior of the projection, such as not allocating any item to a certain agent.

In their most generic form, our assumptions require that each agent achieves infinite utility as TT\rightarrow\infty and bounded non-zero valuation ratios; see Assumption 4.1. Notice that unbounded utility is necessary if we hope to derive meaningful convergence guarantees for worst-case envy and Nash welfare. {assumption} For each integer T>0T>0, 𝒱εT(,r){\mathcal{V}}_{\varepsilon}^{T}(\ell,r) is the set of inputs which satisfy:

  • The utility of each agent under PACE is infinite when the projection bounds are set to be [,r][\ell,r].

  • For each i[n]i\in[n] and t[T]t\in[T], vit{0}[ε,1]v_{i}^{t}\in\{0\}\cup[\varepsilon,1].

Since section 4.1 is potentially hard to verify, we next identify sufficient conditions under which PACE guarantees infinite utility for each agent. {assumption} For integer T>0T>0, and c(0,1]c\in(0,1], 𝒱ε,cT{\mathcal{V}}_{\varepsilon,c}^{T} is the set of inputs which satisfy:

  • For each i[n]i\in[n], VicTV_{i}\geq cT, that is, the monopolistic utility of each agent under PACE is Θ(T)\Theta(T).

  • For each i[n]i\in[n] and t[T]t\in[T], vit{0}[ε,1]v_{i}^{t}\in\{0\}\cup[\varepsilon,1].

Assumption 4.1 strengthens Assumption 3.1 by requiring the monopolistic utility of each agent to be linear in TT. Intuitively, this matches the linear projection bounds [t,rt][\ell t,rt] on utility. We will show in Lemma 6 that, with appropriate initialization of the projection bounds, Assumption 4.1 with any c>0c>0 leads to infinite agent utilities, and thus implies Assumption 4.1 for some \ell and rr.

Conversely, if an agent’s inputs have o(T)o(T) agent monopolistic utility, the following example shows that problems may occur, see Example 2.

Example 2.

Cases with t=1T𝟏{vit>0}=f(T)=o(T)\sum_{t=1}^{T}\bm{1}\{v_{i}^{t}>0\}=f(T)=o(T) for i=1,2i=1,2 in an nn-agent case (n>2n>2). For any (0,ε)\ell\in(0,\varepsilon), there exists a sufficiently large T1T_{1} such that f(T1)<εT1f(T_{1})<\frac{\ell}{\varepsilon}T_{1}. We can construct an instance where agents 11 and 22 have value zero for the first (1ε)T1\left(1-\frac{\ell}{\varepsilon}\right)T_{1} items and value ε\varepsilon for the remaining items. One can check that both utilities never reach t\ell t, and thus U^1t=U^2t=t\hat{U}_{1}^{t}=\hat{U}_{2}^{t}=\ell t always. Under lexicographical tie-breaking, agent 11 will receive all items. Agent 22 will receive nothing at all.

We remark that in order to initialize \ell and rr, PACE requires knowing a lower bound on ε\varepsilon. In contrast, the integral greedy algorithm does not require such knowledge. A minimal requirement is to set ε\ell\leq\varepsilon:

Example 3.

Case with >ε\ell>\varepsilon: Consider an input where every agent has value ε\varepsilon in each round. The projection is then activated in every round for every agent, and the algorithm allocates all items to the first agent.

4.2 Envy Analysis for PACE

Next we show that the PACE algorithm, as long as it is appropriately initialized and section 4.1 holds, achieves 1+2log1/ε1+2\log 1/\varepsilon multiplicative envy bound asymptotically. Our proof is performed by reducing the adversarial envy-maximizing problem in PACE to the canonical program (3) by showing that PACE with our assumptions lead to a weakly harder problem than (3) for the adversary.

Theorem 4.1 (Upper Bound of Multiplicative Envy for PACE).

Under Assumption 4.1 with

<ε2ε+1+(n1)(1+log(1/ε)),r=1,\ell<\frac{\varepsilon^{2}}{\varepsilon+1+(n-1)\left(1+\log(1/\varepsilon)\right)},\ r=1,

PACE achieves

limTsup𝒗𝒱εT(,r)Envyij1+2log1ε,i,j[n].\lim_{T\rightarrow\infty}\sup_{\bm{v}\in{\mathcal{V}}_{\varepsilon}^{T}(\ell,r)}\mathrm{Envy}_{ij}\leq 1+2\log\frac{1}{\varepsilon},\ \ \forall i,j\in[n].
Proof.

Due to symmetry, we only focus on the envy of agent 11. Similar to the proof of Theorem 3.1, we first transform the input 𝒗\bm{v} into 𝒗\bm{v}^{\prime}, such that the allocation of PACE is preserved. Notice that in this proof we use a different transformation, since the problem cannot be reduced to pairwise 2-agent cases. It is no longer obvious that the transformation preserves the PACE allocation; we will prove this later. The transformation is as follows:

  1. 1.

    For agent j{2,3,,n}j\in\{2,3,\cdots,n\}, set agent jj’s value to 0 for all items that are not allocated to jj.

  2. 2.

    Move all items on which agent 11 has value 0 to the end of the input sequence.

  3. 3.

    Move all items in A1A_{1} to the beginning of the input sequence.

Allocation for items in Step 2. is clearly preserved under PACE and does not affect the envy of agent 1, so we can assume such items do not exist without loss of generality.

Denote S=|A1|.S=|A_{1}|. For item ss in the original input 𝒗\bm{v}, let s>ss^{\prime}>s be the round it appears in 𝒗\bm{v}^{\prime}. We define t:=εSε(1+1ε(n1)(1+log1ε)).t^{*}:=\frac{\varepsilon S}{\varepsilon-\ell}\left(1+\frac{1}{\varepsilon}(n-1)\left(1+\log\frac{1}{\varepsilon}\right)\right). The proof is divided into three parts.

For the first part, Lemma 4 shows that in the first tt^{*} rounds in 𝒗\bm{v}^{\prime}, the transformation preserves the PACE allocation.

Lemma 4.

If s<ts^{\prime}<t^{*}, then for agent j{2,,n}j\in\{2,\cdots,n\},

Ujs(𝒗)<sU^js(𝒗)U^1s(𝒗)<ε.U_{j}^{s^{\prime}}(\bm{v}^{\prime})<\ell s^{\prime}\implies\frac{\hat{U}_{j}^{s^{\prime}}(\bm{v}^{\prime})}{\hat{U}_{1}^{s^{\prime}}(\bm{v}^{\prime})}<\varepsilon.

Also, all items in the first tt^{*} rounds of 𝐯\bm{v}^{\prime} has the same allocation result as 𝐯\bm{v}.

For the second part, we show that in the first tt^{*} rounds, the problem of maximizing envy under PACE can be relaxed into a canonical program defined as (3). Lemma 5 gives upper bounds on multiplicative envy in the first tt^{*} rounds, as well as the number of rounds within tt^{*} for agent j{2,,n}j\in\{2,\cdots,n\} to have non-zero values.

Lemma 5.

For any agent j{2,,n}j\in\{2,\cdots,n\}, let BjB_{j} be the set of items that belongs to AjA_{j} in the original sequence 𝐯\bm{v}, and are permuted to the first tt^{*} position in the transformed sequence. Then it holds that

limSU1(Bj)U1(A1)1+2log1ε,limS|Bj|S1ε(1+log1ε).\lim_{S\rightarrow\infty}\frac{U_{1}(B_{j})}{U_{1}(A_{1})}\leq 1+2\log\frac{1}{\varepsilon},\ \lim_{S\rightarrow\infty}\frac{|B_{j}|}{S}\leq\frac{1}{\varepsilon}\left(1+\log\frac{1}{\varepsilon}\right).

For the final part, we show that the length of the transformed sequence cannot exceed tt^{*}. Let m:=Sε(1+log1ε)m:=\frac{S}{\varepsilon}\left(1+\log\frac{1}{\varepsilon}\right) be the asymptotic upper bound of |Bj||B_{j}| given in Lemma 5. It is straightforward to show with some simple algebra that S+(n1)mtS+(n-1)m\leq t^{*}. Notice that S+(n1)mS+(n-1)m is an asymptotic upper bound on the number of non-empty items (an empty item is one that is evaluated zero by all buyers) within tt^{*} rounds in the transformed sequence. Therefore, S+(n1)mtS+(n-1)m\leq t^{*} implies that the transformed sequence has to end within tt^{*} rounds. Thus tt^{*} is an upper bound of both the original and transformed sequence, whose allocation results are guaranteed to be the same by Lemma 4. The upper bound for envy in Lemma 5 is then a valid upper bound for the original sequence.

\hfill\square

4.3 Nash Welfare Analysis for PACE

Next we focus on establishing a worst-case guarantee for the asymptotic competitive ratio of Nash Welfare for PACE under Assumption 4.1. To begin with, we show that Assumption 4.1 implies infinite agent utilities with appropriate initialization.

Lemma 6.

For any input 𝐯𝒱ε,cT\bm{v}\in\mathcal{V}_{\varepsilon,c}^{T} (defined in section 4.1), if

<cε21+(n1)(1+log1/ε),r=1,\ell<\frac{c\varepsilon^{2}}{1+(n-1)(1+\log 1/\varepsilon)},\ r=1,

then there exists a constant d>0d>0 (which depends on \ell), such that for each i[n]i\in[n], PACE satisfies limTUi(Ai)dT.\lim_{T\rightarrow\infty}U_{i}({A_{i}})\geq dT. Furthermore, dc=Ω(1n)\frac{d}{c}=\Omega\left(\frac{1}{n}\right) as nn\rightarrow\infty.

The proof of Lemma 6 is also by a reduction from the canonical problem (3). We remark that Lemma 6 yields more than infinite agent utilities: it also tells us that the utilities are linear in TT. Moreover, from d/c=Ω(n1)d/c=\Omega\left(n^{-1}\right) we know that PACE computes an asymptotic approximate proportional allocation, which helps to derive a bounded competitive ratio w.r.t. Nash welfare. The bound can be furthermore refined using the envy results.

Theorem 4.2 (Upper Bound of Competitive Ratio for PACE).

Under Assumption 4.1, for any input 𝐯𝒱ε,cT\bm{v}\in\mathcal{V}_{\varepsilon,c}^{T}, if

<cε21+ε+(n1)(1+log1/ε),r=1,\ell<\frac{c\varepsilon^{2}}{1+\varepsilon+(n-1)(1+\log 1/\varepsilon)},\ r=1,

PACE achieves

limTsupv𝒱εT(i=1nUii=1nUi(Ai))1/n(1+2log1ε)1c.\lim_{T\rightarrow\infty}\sup_{v\in\mathcal{V}^{T}_{\varepsilon}}\left(\frac{\prod_{i=1}^{n}U_{i}^{\star}}{\prod_{i=1}^{n}U_{i}(A_{i})}\right)^{1/n}\leq\left(1+2\log\frac{1}{\varepsilon}\right)\cdot\frac{1}{c}.

Theorem 4.2 gives an upper bound depends only on parameters ε\varepsilon and cc, which are both independent of TT. We remark that the constant 1/c1/c might not be tight.

We also remark that in both Theorem 4.1 and Theorem 4.2 \ell is at most the order of 1/n1/n, which is aligned with the stationary and non-stationary setting (projecting utilities to an Ω(1/n)\Omega(1/n) bound is unreasonable since there are nn agents). However with adversarial input \ell decreases as ε0\varepsilon\rightarrow 0, which means that PACE requires a wider projection interval when its input becomes potentially more extreme.

5 Conclusion

We proved horizon-independent bounds for envy and Nash welfare for the integral greedy algorithm and PACE under adversarial inputs with mild assumptions. Our results complete the first best-of-many-worlds result for online fair allocation, since PACE thus achieves guarantees under stochastic (Gao et al., 2021), stochastic but nonstationary (Liao et al., 2022), and adversarial inputs. Moreover, our results on integral greedy are of independent interest, as they characterize assumptions needed to achieve guarantees for that algorithm.

It remains open whether the constant in Theorem 4.2 can be improved. A more general open question is to explore more best-of-many-worlds online fair allocation algorithms, with potentially different performance measures and assumptions.

References

  • Anari et al. [2017] Nima Anari, Shayan Oveis Gharan, Amin Saberi, and Mohit Singh. Nash social welfare, matrix permanent, and stable polynomials. In 8th Innovations in Theoretical Computer Science Conference (ITCS 2017). Schloss Dagstuhl-Leibniz-Zentrum fuer Informatik, 2017.
  • Anari et al. [2018] Nima Anari, Tung Mai, Shayan Oveis Gharan, and Vijay V Vazirani. Nash social welfare for indivisible items under separable, piecewise-linear concave utilities. In Proceedings of the Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms, pages 2274–2290. SIAM, 2018.
  • Azar et al. [2016] Yossi Azar, Niv Buchbinder, and Kamal Jain. How to allocate goods in an online market? Algorithmica, 74(2):589–601, 2016.
  • Bach and Levy [2019] Francis Bach and Kfir Y Levy. A universal algorithm for variational inequalities adaptive to smoothness and noise. In Conference on learning theory, pages 164–194. PMLR, 2019.
  • Balseiro et al. [2023] Santiago R Balseiro, Haihao Lu, and Vahab Mirrokni. The best of many worlds: Dual mirror descent for online allocation problems. Operations Research, 71(1):101–119, 2023.
  • Banerjee et al. [2022] Siddhartha Banerjee, Vasilis Gkatzelis, Artur Gorokh, and Billy Jin. Online nash social welfare maximization with predictions. In Proceedings of the 2022 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 1–19. SIAM, 2022.
  • Barman et al. [2018] Siddharth Barman, Sanath Kumar Krishnamurthy, and Rohit Vaish. Finding fair and efficient allocations. In Proceedings of the 2018 ACM Conference on Economics and Computation, pages 557–574, 2018.
  • Barman et al. [2020] Siddharth Barman, Umang Bhaskar, Anand Krishna, and Ranjani G Sundaram. Tight approximation algorithms for p-mean welfare under subadditive valuations. arXiv preprint arXiv:2005.07370, 2020.
  • Benade et al. [2018] Gerdus Benade, Aleksandr M Kazachkov, Ariel D Procaccia, and Christos-Alexandros Psomas. How to make envy vanish over time. In Proceedings of the 2018 ACM Conference on Economics and Computation, pages 593–610, 2018.
  • Bogomolnaia et al. [2022] Anna Bogomolnaia, Hervé Moulin, and Fedor Sandomirskiy. On the fair division of a random object. Management Science, 68(2):1174–1194, 2022.
  • Caragiannis et al. [2019] Ioannis Caragiannis, David Kurokawa, Hervé Moulin, Ariel D Procaccia, Nisarg Shah, and Junxing Wang. The unreasonable fairness of maximum nash welfare. ACM Transactions on Economics and Computation (TEAC), 7(3):1–32, 2019.
  • Castiglioni et al. [2022] Matteo Castiglioni, Andrea Celli, and Christian Kroer. Online learning under budget and ROI constraints and applications to bidding in non-truthful auctions. arXiv preprint arXiv:2302.01203, 2022.
  • Celli et al. [2022] Andrea Celli, Matteo Castiglioni, and Christian Kroer. Best of many worlds guarantees for online learning with knapsacks. arXiv preprint arXiv:2202.13710, 2022.
  • Chaudhury et al. [2021] Bhaskar R Chaudhury, Jugal Garg, and Ruta Mehta. Fair and efficient allocations under subadditive valuations. In Proceedings of the AAAI Conference on Artificial Intelligence, 2021.
  • Cole and Gkatzelis [2018] Richard Cole and Vasilis Gkatzelis. Approximating the nash social welfare with indivisible items. SIAM Journal on Computing, 47(3):1211–1236, 2018.
  • Cole et al. [2017] Richard Cole, Nikhil Devanur, Vasilis Gkatzelis, Kamal Jain, Tung Mai, Vijay V Vazirani, and Sadra Yazdanbod. Convex program duality, fisher markets, and nash social welfare. In Proceedings of the 2017 ACM Conference on Economics and Computation, pages 459–460, 2017.
  • Conitzer et al. [2022] Vincent Conitzer, Christian Kroer, Debmalya Panigrahi, Okke Schrijvers, Nicolas E Stier-Moses, Eric Sodomka, and Christopher A Wilkens. Pacing equilibrium in first price auction markets. Management Science, 68(12):8515–8535, 2022.
  • Eisenberg and Gale [1959] Edmund Eisenberg and David Gale. Consensus of subjective probabilities: The pari-mutuel method. The Annals of Mathematical Statistics, 30(1):165–168, 1959.
  • Gao and Kroer [2023] Yuan Gao and Christian Kroer. Infinite-dimensional fisher markets and tractable fair division. Operations Research, 71(2):688–707, 2023.
  • Gao et al. [2021] Yuan Gao, Alex Peysakhovich, and Christian Kroer. Online market equilibrium with application to fair division. Advances in Neural Information Processing Systems, 34:27305–27318, 2021.
  • Garg et al. [2018] Jugal Garg, Martin Hoefer, and Kurt Mehlhorn. Approximating the nash social welfare with budget-additive valuations. In Proceedings of the Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms, pages 2326–2340. SIAM, 2018.
  • Garg et al. [2020] Jugal Garg, Pooja Kulkarni, and Rucha Kulkarni. Approximating nash social welfare under submodular valuations through (un) matchings. In Proceedings of the fourteenth annual ACM-SIAM symposium on discrete algorithms, pages 2673–2687. SIAM, 2020.
  • He et al. [2019] Jiafan He, Ariel D Procaccia, CA Psomas, and David Zeng. Achieving a fairer future by changing the past. IJCAI’19, 2019.
  • Huang et al. [2022] Zhiyi Huang, Minming Li, Xinkai Shu, and Tianze Wei. Online nash welfare maximization without predictions. arXiv preprint arXiv:2211.03077, 2022.
  • Kaneko and Nakamura [1979] Mamoru Kaneko and Kenjiro Nakamura. The nash social welfare function. Econometrica: Journal of the Econometric Society, pages 423–435, 1979.
  • Lee [2017] Euiwoong Lee. Apx-hardness of maximizing nash social welfare with indivisible items. Information Processing Letters, 122:17–20, 2017.
  • Li and Vondrák [2022] Wenzheng Li and Jan Vondrák. A constant-factor approximation algorithm for nash social welfare with submodular valuations. In 2021 IEEE 62nd Annual Symposium on Foundations of Computer Science (FOCS), pages 25–36. IEEE, 2022.
  • Liao et al. [2022] Luofeng Liao, Yuan Gao, and Christian Kroer. Nonstationary dual averaging and online fair allocation. Advances in Neural Information Processing Systems, 35:37159–37172, 2022.
  • McGlaughlin and Garg [2020] Peter McGlaughlin and Jugal Garg. Improving nash social welfare approximations. Journal of Artificial Intelligence Research, 68:225–245, 2020.
  • Moulin [2004] Hervé Moulin. Fair division and collective welfare. MIT press, 2004.
  • Nash Jr [1950] John F Nash Jr. The bargaining problem. Econometrica: Journal of the econometric society, pages 155–162, 1950.
  • Varian [1974] Hal R Varian. Equity, envy, and efficiency. Journal of Economic Theory, 9(1):63–91, 1974.
  • Xiao [2009] Lin Xiao. Dual averaging method for regularized stochastic learning and online optimization. Advances in Neural Information Processing Systems, 22, 2009.
  • Zeng and Psomas [2020] David Zeng and Alexandros Psomas. Fairness-efficiency tradeoffs in dynamic fair division. In Proceedings of the 21st ACM Conference on Economics and Computation, pages 911–912, 2020.

Appendix 0.A Additional Related Work

Offline Nash Welfare Approximation

To explore PACE’s behavior under adversarial inputs, we focus on the approximation of the EG objective, which is equivalent to maximizing Nash welfare. Nash welfare was introduced by Nash Jr [1950], and is one of the ideal proxies for balancing fairness and efficiency in allocation problems  [Kaneko and Nakamura, 1979]. Maximizing Nash welfare is well-studied in the offline settings. For divisible items, it is equivalent to the Eisenberg-Gale convex program  [Eisenberg and Gale, 1959]. For indivisible items, computing a Nash-welfare-maximizing allocation is APX-hard Lee [2017] for additive utilities. Cole and Gkatzelis [2018] gave the first approximation algorithm, with 2e1/e2e^{1/e} competitive ratio. Anari et al. [2017] applied matrix permanent and stable polynomials to get an ee-approximation. Cole et al. [2017] improved the ratio to 22. The state-of-the-art algorithm, given subsequently by Barman et al. [2018], has a ratio of e1/ee^{1/e}. More general utility classes have been considered beyond additive ones. Garg et al. [2018] achieved a 2e1/e+o(1)2e^{1/e}+o(1) ratio under budget-additive values. Anari et al. [2018] considered separable, piecewise-linear and concave valuations and gave an e2e^{2}-competitive algorithm. For submodular utilities, Garg et al. [2020] achieved an O(nlogn)O(n\log n)-approximation, which was recently improved to a constant ratio [Li and Vondrák, 2022]. For more generalized forms of submodular utilities, an O(n)O(n) ratio was reached by Barman et al. [2020] and Chaudhury et al. [2021] independently.

Online Allocation with Envy Guarantees

Besides Nash welfare maximization, our analysis on multiplicative envy is also related to the line of works focused on achieving (possibly approximate) envy-freeness in the online setting. Bogomolnaia et al. [2022] assume stochastic input and enforce envy-freeness as a soft constraint, while maximizing social welfare. He et al. [2019] further allow reallocating previous items, and show that O(T)O(T) reallocation is enough to achieve envy-freeness up to an item. Benade et al. [2018] considers envy minimization in the stochastic, indivisible setting, and show that allocating each item to each agent uniformly at random is near-optimal up to logarithmic factors. Zeng and Psomas [2020] considers the indivisible setting with non-adaptive adversary, showing that nontrivial approximation of envy-freeness and Pareto-optimality is hard to achieve simultaneously. In contrast to the above works, we focus on multiplicative envy instead of additive envy. While it is shown by Caragiannis et al. [2019] that Nash welfare maximizing allocation has approximate envy-freeness, our analysis on multiplicative envy is independent of other fairness measures. Finally note that the PACE algorithm, while focused on the stronger guarantee of asymptotically maximizing Nash welfare in the stochastic setting, actually achieves asymptotic envy-freeness as well, since it converges to the equilibrium allocation of the underlying Fisher market.

Appendix 0.B Missing Proofs in Section 3

0.B.1 Proof of Theorem 3.1

According to Lemma 2, it suffices to bound the envy of all 22–agent instances. We will show that for any input 𝒗𝒱εT\bm{v}\in\mathcal{V}_{\varepsilon}^{T},

B2U2(A1)B1U2(A2)1+2log1ε+(1+ε4)(1+ε2)ε61T+(1+ε2)2ε81T2\frac{B_{2}U_{2}(A_{1})}{B_{1}U_{2}(A_{2})}\leq 1+2\log\frac{1}{\varepsilon}+\frac{(1+\varepsilon^{4})(1+\varepsilon^{2})}{\varepsilon^{6}}\frac{1}{T}+\frac{(1+\varepsilon^{2})^{2}}{\varepsilon^{8}}\frac{1}{T^{2}} (8)

Consider the following transformation operation from instance 𝒗\bm{v} into 𝒗{\bm{v}}^{\prime}:

  1. 1.

    Set agent 11’s valuation for all items in A2A_{2} to 0.

  2. 2.

    Move all items in A2A_{2} to the beginning of the input sequence, and all items in A1A_{1} to the end of the input sequence (in arbitrary order).

We claim that the allocation result under the budget-adapted greedy allocation algorithm is invariant to this transformation. Clearly, all items in A2A_{2} will still be allocated to agent 22, since we’ve set agent 11’s valuation on them to zero. For item tA1t\in A_{1}, let tt^{\prime} be the round it appears in 𝒗\bm{v}^{\prime}. One can check

B1v1tB2v2tU1t(𝒗)U2t(𝒗)U1t(𝒗)U2t(𝒗),\frac{B_{1}v_{1}^{t}}{B_{2}v_{2}^{t}}\geq\frac{U_{1}^{t}(\bm{v})}{U_{2}^{t}(\bm{v})}\geq\frac{U_{1}^{t^{\prime}}(\bm{v}^{\prime})}{U_{2}^{t^{\prime}}(\bm{v}^{\prime})},

where the first inequality is due to tA1t\in A_{1}, and the second inequality is because U1t(𝒗)=U1t(𝒗)U_{1}^{t^{\prime}}(\bm{v}^{\prime})=U_{1}^{t}(\bm{v}) and U2t(𝒗)=s=1Tx2sv2sU1t(𝒗)U_{2}^{t^{\prime}}(\bm{v}^{\prime})=\sum_{s=1}^{T}x_{2}^{s}v_{2}^{s}\geq U_{1}^{t}(\bm{v}).

Therefore, it suffices to consider only the transformed sequences, i.e., ones with agent 22 getting every item in the first RR rounds, and then nothing thereafter. To find the transformed sequence of maximum envy, we next note that we can increase agent 22’s valuation on the items occurring after round RR as much as possible, while ensuring that none of them are allocated to agent 22 by the algorithm, while weakly increasing envy. This is equivalent to the following optimization problem, with the multiplicative envy of agent 22 upper bounded by its objective:

maxv1t,v2t\displaystyle\max_{v_{1}^{t},v_{2}^{t}} B2B1V20t=R+1v2t\displaystyle\ \ \frac{B_{2}}{B_{1}V_{2}^{0}}\sum_{t=R+1}^{\infty}v_{2}^{t}
s.t. v1t,v2t{0}[ε,1]\displaystyle\ \ v_{1}^{t},v_{2}^{t}\in\{0\}\cup[\varepsilon,1]
v1tv2tB2U1t1B1V20\displaystyle\ \ \frac{v_{1}^{t}}{v_{2}^{t}}\geq\frac{B_{2}U_{1}^{t-1}}{B_{1}V_{2}^{0}}
U1t=s=R+1tv1s\displaystyle\ \ U_{1}^{t}=\sum_{s=R+1}^{t}v_{1}^{s}

Notice the constraint that all items after round RR are allocated to agent 11. Together with the {0}[ε,1]\{0\}\cup[\varepsilon,1] range for values, this gives an upper bound of v2t/v1tv_{2}^{t}/v_{1}^{t} on each round:

v2tv1tq(U1t1):={min{B1V20B2U1t1,1ε}U1t1B1B2εV200U1t1>B1B2εV20\frac{v_{2}^{t}}{v_{1}^{t}}\leq q(U_{1}^{t-1}):=\begin{cases}\min\left\{\frac{B_{1}V_{2}^{0}}{B_{2}U_{1}^{t-1}},\frac{1}{\varepsilon}\right\}&U_{1}^{t-1}\leq\frac{B_{1}}{B_{2}\varepsilon}V_{2}^{0}\\ 0&U_{1}^{t-1}>\frac{B_{1}}{B_{2}\varepsilon}V_{2}^{0}\end{cases}

The multiplicative envy can then be further upper bounded by

B2U2(A1)B1U2(A2)B2B1V20t=R+1R+Sv1tq(U1t1)\frac{B_{2}U_{2}(A_{1})}{B_{1}U_{2}(A_{2})}\leq\frac{B_{2}}{B_{1}V_{2}^{0}}\sum_{t=R+1}^{R+S}v_{1}^{t}q(U_{1}^{t-1}) (9)

where SS is the number of rounds in the above optimization problem with v1t,v2t>0v_{1}^{t},v_{2}^{t}>0. Since we should not allocate anything to agent 22 after round RR, we have U1S(0,B1V20B2ε+1]U_{1}^{S}\in\left(0,\frac{B_{1}V_{2}^{0}}{B_{2}\varepsilon}+1\right].

Next, we show that the right hand side of (9) can be approximated with an integration. We have

B2B1V20(t=R+1R+Sv1tq(U1t1)0U1R+Sq(u)du)\displaystyle\frac{B_{2}}{B_{1}V_{2}^{0}}\left(\sum_{t=R+1}^{R+S}v_{1}^{t}q(U_{1}^{t-1})-\int_{0}^{U_{1}^{R+S}}q(u)\mathrm{d}u\right)
=\displaystyle= B2B1V20t=R+1R+S(v1tq(U1t1)U1t1U1tq(u)du)\displaystyle\frac{B_{2}}{B_{1}V_{2}^{0}}\sum_{t=R+1}^{R+S}\left(v_{1}^{t}q(U_{1}^{t-1})-\int_{U_{1}^{t-1}}^{U_{1}^{t}}q(u)\mathrm{d}u\right)
=\displaystyle= B2B1V20t=R+1R+SU1t1U1t(q(U1t1)q(u))du\displaystyle\frac{B_{2}}{B_{1}V_{2}^{0}}\sum_{t=R+1}^{R+S}\int_{U_{1}^{t-1}}^{U_{1}^{t}}\left(q(U_{1}^{t-1})-q(u)\right)\mathrm{d}u
(a)\displaystyle\overset{\text{(a)}}{\leq} B2B1V20t=R+1R+S(v1t)(q(U1t1)q(U1t))\displaystyle\frac{B_{2}}{B_{1}V_{2}^{0}}\sum_{t=R+1}^{R+S}(v_{1}^{t})\left(q(U_{1}^{t-1})-q(U_{1}^{t})\right)
(b)\displaystyle\overset{\text{(b)}}{\leq} B2B1V20t=R+1R+SB1B2ε2V20(v1t)2\displaystyle\frac{B_{2}}{B_{1}V_{2}^{0}}\sum_{t=R+1}^{R+S}\frac{B_{1}}{B_{2}\varepsilon^{2}V_{2}^{0}}(v_{1}^{t})^{2}
=\displaystyle= 1ε2(V20)2t=R+1R+S(v1t)2\displaystyle\frac{1}{\varepsilon^{2}(V_{2}^{0})^{2}}\sum_{t=R+1}^{R+S}(v_{1}^{t})^{2}
(c)\displaystyle\overset{\text{(c)}}{\leq} U1R+Sε2(V20)2\displaystyle\frac{U_{1}^{R+S}}{\varepsilon^{2}(V_{2}^{0})^{2}}

where (a) is because q(u)q(u) is non–increasing, (b) is because the right–side derivative of q(u)q(u) is upper bounded by B1B2ε2V20\frac{B_{1}}{B_{2}\varepsilon^{2}V_{2}^{0}}, and (c) is because v1t1v_{1}^{t}\leq 1. Combined with 9 we have

B2U2(A1)B1U2(A2)B2B1V200U1R+Sq(u)du+U1R+Sε2(V20)2\frac{B_{2}U_{2}(A_{1})}{B_{1}U_{2}(A_{2})}\leq\frac{B_{2}}{B_{1}V_{2}^{0}}\int_{0}^{U_{1}^{R+S}}q(u)\mathrm{d}u+\frac{U_{1}^{R+S}}{\varepsilon^{2}(V_{2}^{0})^{2}}\\ (10)

We then deal with U1R+SU_{1}^{R+S} using the relation TR+ST\geq R+S. This is true because we allow no “null” items in a TT–round instance. From εSU1R+SV20/ε+1R/ε+1\varepsilon S\leq U_{1}^{R+S}\leq V_{2}^{0}/\varepsilon+1\leq R/\varepsilon+1 we know V20ε31+ε2TV_{2}^{0}\geq\frac{\varepsilon^{3}}{1+\varepsilon^{2}}T. Hence,

B2U2(A1)B1U2(A2)\displaystyle\frac{B_{2}U_{2}(A_{1})}{B_{1}U_{2}(A_{2})} B2B1V200V20/ε+1q(u)du+1ε2V20(1ε+1V20)\displaystyle\leq\frac{B_{2}}{B_{1}V_{2}^{0}}\int_{0}^{V_{2}^{0}/\varepsilon+1}q(u)\mathrm{d}u+\frac{1}{\varepsilon^{2}V_{2}^{0}}\left(\frac{1}{\varepsilon}+\frac{1}{V_{2}^{0}}\right)
=B2B1V200V20/εq(u)du+B2B1V20V20/εV20/ε+1q(u)du+1ε2V20(1ε+1V20)\displaystyle=\frac{B_{2}}{B_{1}V_{2}^{0}}\int_{0}^{V_{2}^{0}/\varepsilon}q(u)\mathrm{d}u+\frac{B_{2}}{B_{1}V_{2}^{0}}\int_{V_{2}^{0}/\varepsilon}^{V_{2}^{0}/\varepsilon+1}q(u)\mathrm{d}u+\frac{1}{\varepsilon^{2}V_{2}^{0}}\left(\frac{1}{\varepsilon}+\frac{1}{V_{2}^{0}}\right)
1+2log1ε+εV20+1ε2V20(1ε+1V20)\displaystyle\leq 1+2\log\frac{1}{\varepsilon}+\frac{\varepsilon}{V_{2}^{0}}+\frac{1}{\varepsilon^{2}V_{2}^{0}}\left(\frac{1}{\varepsilon}+\frac{1}{V_{2}^{0}}\right)
1+2log1ε+(1+ε4)(1+ε2)ε61T+(1+ε2)2ε81T2\displaystyle\leq 1+2\log\frac{1}{\varepsilon}+\frac{(1+\varepsilon^{4})(1+\varepsilon^{2})}{\varepsilon^{6}}\frac{1}{T}+\frac{(1+\varepsilon^{2})^{2}}{\varepsilon^{8}}\frac{1}{T^{2}}
=1+2log1ε+O(1T)\displaystyle=1+2\log\frac{1}{\varepsilon}+O\left(\frac{1}{T}\right)

This finishes our proof.

0.B.2 Proof of Lemma 3

For the same reason as Theorem 3.1, it suffices to consider j=1,i=2j=1,i=2 in 22-agent instances, with all inputs transformed as described in the proof of Theorem 3.1. UU is the short for U2(A2)U_{2}(A_{2}), and also the parameter for the canonical optimization problem.

Consider the two phases in the horizon each with length T1T_{1} and T2T_{2}. We upper bound them respectively:

  1. 1.

    Rounds with U1tUU_{1}^{t}\leq U. Since the increment of v1tv_{1}^{t} is at least ε\varepsilon whenever agent 11 receives an item, there are at most U/εU/\varepsilon rounds in this phase.

    limUT1U1ε.\lim_{U\rightarrow\infty}\frac{T_{1}}{U}\leq\frac{1}{\varepsilon}.
  2. 2.

    Rounds with U<U1tU/εU<U_{1}^{t}\leq U/\varepsilon. Assume this phase begins at round ss. By the analysis of the canonical optimization problem (3), agent 22’s envy in this phase can be upper bounded asymptotically.

    1Ut=ss+T21v2t1Ut=sv1tq(U1t1)1UUU/εq(U1)dU1=1+log1ε.\frac{1}{U}\sum_{t=s}^{s+T_{2}-1}v_{2}^{t}\leq\frac{1}{U}\sum_{t=s}^{\infty}v_{1}^{t}q(U_{1}^{t-1})\rightarrow\frac{1}{U}\int_{U}^{U/\varepsilon}q(U_{1})\mathrm{d}U_{1}=1+\log\frac{1}{\varepsilon}.

    Since v2tv_{2}^{t} is at least ε\varepsilon in each round,

    limUT2U1εlimUt=ss+T21v2tU1ε1ε(1+log1ε).\lim_{U\rightarrow\infty}\frac{T_{2}}{U}\leq\frac{1}{\varepsilon}\leq\lim_{U\rightarrow\infty}\frac{\sum_{t=s}^{s+T_{2}-1}v_{2}^{t}}{U}\cdot\frac{1}{\varepsilon}\leq\frac{1}{\varepsilon}\left(1+\log\frac{1}{\varepsilon}\right).

Combining the bounds for T1T_{1} and T2T_{2} proves Lemma 3.

0.B.3 Proof of Theorem 3.2

We prove Theorem 3.2 by constructing a hard 22–agent instance that reaches 1+2log1ε1+2\log\frac{1}{\varepsilon} envy asymptotically. The construction follows the spirit of the hard instances described in the proof of Theorem 3.1, where agent 22 receives items in the beginning T0T_{0} rounds, but nothing afterwards.

We construct a hard instance for B1=B2B_{1}=B_{2}. For unequal weights the construction is similar. For T0,a>0T_{0},a>0, consider an instance with 2k+22k+2 phases. In each phase, the same item appears for many rounds. The agent’s valuations on these items are described in Table 2.

Phase Length Valuation 1 Valuation 2 U1U_{1} (after the phase) U2U_{2} (after the phase)
A1 T0T_{0} 0 11 0 T0T_{0}
A2 T0T_{0} ε\varepsilon 11 εT0\varepsilon\cdot T_{0} T0T_{0}
B1 (11a)T0\left(1-\frac{1}{a}\right)T_{0} εa1\varepsilon\cdot a^{1} 11 εa1T0\varepsilon\cdot a^{1}\cdot T_{0} T0T_{0}
B2 (11a)T0\left(1-\frac{1}{a}\right)T_{0} εa2\varepsilon\cdot a^{2} 11 εa2T0\varepsilon\cdot a^{2}\cdot T_{0} T0T_{0}
Bkk (11a)T0\left(1-\frac{1}{a}\right)T_{0} εak=1\varepsilon\cdot a^{k}=1 11 εakT0=T0\varepsilon\cdot a^{k}\cdot T_{0}=T_{0} T0T_{0}
C1 1ε(11a)T0\frac{1}{\varepsilon}\left(1-\frac{1}{a}\right)T_{0} εa1\varepsilon\cdot a^{1} ε\varepsilon a1T0a^{1}\cdot T_{0} T0T_{0}
C2 1ε(11a)T0\frac{1}{\varepsilon}\left(1-\frac{1}{a}\right)T_{0} εa2\varepsilon\cdot a^{2} ε\varepsilon a2T0a^{2}\cdot T_{0} T0T_{0}
Ckk 1ε(11a)T0\frac{1}{\varepsilon}\left(1-\frac{1}{a}\right)T_{0} εak=1\varepsilon\cdot a^{k}=1 ε\varepsilon akT0=1εT0a^{k}\cdot T_{0}=\frac{1}{\varepsilon}T_{0} T0T_{0}
Table 2: A worst-case instance for envy

One can check that the boundary condition v1t/v2t=U1t/U2tv_{1}^{t}/v_{2}^{t}=U_{1}^{t}/U_{2}^{t} holds at the end of each phase, and all items after phase A1 is allocated to agent 11. Agent 22’s monopolistic utility is

V2=2(1+(11a)loga1ε)T0V_{2}=2\left(1+\left(1-\frac{1}{a}\right)\cdot\log_{a}{\frac{1}{\varepsilon}}\right)T_{0}

As T0T_{0}\rightarrow\infty, we consider a1a\rightarrow 1, we have

(11a)loga1εlog1ε,V2T0=2(1+log1ε).\left(1-\frac{1}{a}\right)\cdot\log_{a}{\frac{1}{\varepsilon}}\rightarrow\log\frac{1}{\varepsilon},\ \frac{V_{2}}{T_{0}}=2\left(1+\log\frac{1}{\varepsilon}\right).

Out of his monopolistic utility, agent 22 only gets utility T0T_{0}. This gives 1+2log1/ε1+2\log 1/\varepsilon multiplicative envy asymptotically.

0.B.4 Proof of Theorem 3.3, ε=1\varepsilon=1

Without loss of generality, we assume U1(A1)U2(A2)Un(An)U_{1}(A_{1})\leq U_{2}(A_{2})\leq\cdots\leq U_{n}(A_{n}). Our main proof strategy is to show that for any input instance 𝒗𝒱1T\bm{v}\in\mathcal{V}_{1}^{T}, we have for i[n]i\in[n],

limTUiUi(Ai)ni+1.\lim_{T\rightarrow\infty}\frac{U_{i}^{\star}}{U_{i}(A_{i})}\leq n-i+1. (11)

For i[n]i\in[n], define Si[T]S_{i}\subseteq[T] as the set of items on which at least one agent in {1,2,,i}\{1,2,\cdots,i\} has nonzero valuation,

Si={t:j[i],vjt=1}.S_{i}=\{t:\ \exists\ j\in[i],\ v_{j}^{t}=1\}.

Consider agent kk where k>ik>i. Notice that the integral greedy algorithm guarantees that the number of items in SiS_{i} which are allocated to agent kk is strictly upper bounded by Ui(Ai)+1U_{i}(A_{i})+1. If not, then the last such item tt should have been allocated to some agent jij\leq i since Ukt1>Ui(Ai)>Uj(Aj)Ujt1U_{k}^{t-1}>U_{i}(A_{i})>U_{j}(A_{j})\geq U_{j}^{t-1}, which is a contradiction. Therefore, we get an upper bound on the amount of items that buyer k>ik>i receives in SiS_{i}:

|AkSi|Ui(Ai)(1in1,k>i).|A_{k}\cap S_{i}|\leq U_{i}(A_{i})\ \ (1\leq i\leq{n-1},k>i).

Define the item set Di=ki(AkSi)D_{i}=\bigcup_{k\geq i}(A_{k}\cap S_{i}). Then,

|Di|(ni+1)Ui(Ai)+(ni).|D_{i}|\leq(n-i+1)U_{i}(A_{i})+(n-i). (12)

Intuitively, it is possible that in an offline optimal allocation, all items in DiD_{i} should be completely shared among the first ii agents, i.e., nothing in DiD_{i} is allocated to agent k>ik>i. To verify this, simply consider a case where |Si||S_{i}| is negligibly small compared with |Si+1||S_{i+1}|. Further, it is also possible that the all items in DD are given to agent ii in an optimal allocation, for example, in the case where buyer 1,2,,i11,2,\cdots,i-1 have zero valuation on these items. This will lead to Ui=Ui(Ai)+(ni)(Ui(Ai)+1){U_{i}^{\star}}=U_{i}(A_{i})+(n-i)(U_{i}(A_{i})+1).

However, we will show that UiU_{i}^{\star} can not be any larger by contradiction. Suppose Ui(Ai)(ni+1)(Ui(Ai)+1)U_{i}(A_{i}^{\star})\geq(n-i+1)(U_{i}(A_{i})+1). By inequality (12), at least one item in D¯i=j<i(AjSi1)\bar{D}_{i}=\bigcup_{j<i}(A_{j}\cap S_{i-1}) is partly allocated to agent ii in 𝒙i\bm{x}_{i}^{\star}. This means that an item that belongs to agent jj (j<i)(j<i) in the integral greedy allocation should be re-allocated to agent ii in the offline optimal allocation, even with a large Ui>UjU_{i}^{\star}>U_{j}^{\star}. This is impossible for an optimal allocation, since allocating this item to agent jj strictly increases the geometric mean.

Therefore,

Ui(Ai)/Ui(Ai)(ni+1)+1Ui(Ai).U_{i}(A_{i}^{\star})/U_{i}(A_{i})\leq(n-i+1)+\frac{1}{U_{i}(A_{i}^{\star})}.

Combining this with the fact that the optimal offline allocation has Ui(Ai)(T)U_{i}(A_{i})\rightarrow\infty\ (T\rightarrow\infty) for all i[n]i\in[n], this proves (11). \hfill\square

0.B.5 Proof of Theorem 3.3, ε(0,1)\varepsilon\in(0,1)

Without loss of generality, we assume the integral greedy algorithm gives us U1(A1)U2(A2)Un(An)U_{1}(A_{1})\leq U_{2}(A_{2})\leq\cdots\leq U_{n}(A_{n}). To prove Theorem 3.3, it suffices to show that for any α>0\alpha>0 there exists a constant λ\lambda such that for all i[n]i\in[n] and any input instance in 𝒱εT\mathcal{V}_{\varepsilon}^{T},

limTUiUi(Ai)λiα(1+(ni)(1+2log1ε)).\lim_{T\rightarrow\infty}\frac{U_{i}^{\star}}{U_{i}(A_{i})}\leq\lambda\cdot i^{\alpha}\cdot\left(1+(n-i)\left(1+2\log\frac{1}{\varepsilon}\right)\right). (13)

We introduce two lemmas to prove (13). Lemma 7 shows that for a set of item AAiA\subset A_{i} which belongs to agent ii in the greedy algorithm, there is an upper bound on the increment of welfare if one transfers AA to other agents. It is a generalization of the envy result. The proof of 7 is similar to 3.1, and is deferred to Section 0.B.6.

Lemma 7.

For i[n]i\in[n], let AAiA\subseteq A_{i} be a set of items that are allocated to agent ii by the integral greedy algorithm, such that |A||A|\rightarrow\infty as TT\rightarrow\infty. Let J[n]\{i}J\subseteq[n]\backslash\{i\} be a set of agents not including ii, and U=maxjJUj(Aj)U=\max_{j\in J}U_{j}(A_{j}). Then for any 𝐯𝒱εT\bm{v}\in\mathcal{V}_{\varepsilon}^{T},

limTsAmaxjJvjsU0min{1ε,U1(A)U}r(u)du,\lim_{T\rightarrow\infty}\frac{\sum_{s\in A}\max_{j\in J}v_{j}^{s}}{U}\leq\int_{0}^{\min\left\{\frac{1}{\varepsilon},\frac{U_{1}(A)}{U}\right\}}r(u)\mathrm{d}u,

where

r(u)={1ε0uε1uε<u1ε0u>1ε.r(u)=\begin{cases}\frac{1}{\varepsilon}&0\leq u\leq\varepsilon\\ \frac{1}{u}&\varepsilon<u\leq\frac{1}{\varepsilon}\\ 0&u>\frac{1}{\varepsilon}\end{cases}.

The next lemma guarantees that asymptotically, the utility of any single agent in the offline optimal allocation can be at most λ\lambda times as large as the utility of the “wealthiest” agent in the integral greedy allocation. λ\lambda is a positive constant which does not depend on TT and nn. The proof of Lemma 8 can be found in Section 0.B.7.

Lemma 8.

For any given input 𝐯𝒱εT\bm{v}\in\mathcal{V}_{\varepsilon}^{T} and α>0\alpha>0, there exists a constant λ>0\lambda>0 such that for any k[n]k\in[n],

limTUkmaxj[n]Uj(Aj)λnα.\lim_{T\rightarrow\infty}\frac{U_{k}^{\star}}{\max_{j\in[n]}U_{j}(A_{j})}\leq\lambda n^{\alpha}.

After introducing the two auxiliary lemmas, we return to the proof of Theorem 3.3.

For i[n]i\in[n], define Si[T]S_{i}\subseteq[T] as the set of items on which at least one agent in {1,2,,i}\{1,2,\cdots,i\} has nonzero valuation.

Si={t:j[i],vjt=1}.S_{i}=\{t:\ \exists\ j\in[i],\ v_{j}^{t}=1\}.

For agent kik\geq i, we consider AkSiA_{k}\cap S_{i}, which is the set of items in SiS_{i} that are allocated to agent kk.

In the proof sketch, we give the intuition that, it is possible that the optimal allocation gives all items in AkSi,(k>i)A_{k}\cap S_{i},(k>i) to agent the first ii agents, which might cause a large ratio of Ui/Ui(Ai)U_{i}^{\star}/U_{i}(A_{i}). To bound the ratio, we first apply Lemma 7 to bound the maximum welfare of the first ii agents on item set AkSiA_{k}\cap S_{i}. For k>ik>i,

limTsSiAkmaxjivjsUi1+2log1ε.\lim_{T\rightarrow\infty}\frac{\sum_{s\in S_{i}\cap A_{k}}\max_{j\leq i}v_{j}^{s}}{U_{i}}\leq 1+2\log\frac{1}{\varepsilon}. (14)

Then, consider the input 𝒗\bm{v}^{\prime}, which is defined as setting agent kk’s value to 0 on every item in SiS_{i}. Because agent 1,2,,i1,2,\cdots,i are given strictly more items in 𝒗\bm{v}^{\prime}, both the integral greedy algorithms should have non-decreasing agent utilities when TT is large enough. (The amount of decrease should be negligible in TT).

limTUj(𝒗)Ui(𝒗)limTUj(𝒗)Ui(𝒗),j[i]\displaystyle\lim_{T\rightarrow\infty}\frac{U_{j}^{\star}(\bm{v})}{U_{i}(\bm{v})}\leq\lim_{T\rightarrow\infty}\frac{U_{j}^{\star}(\bm{v}^{\prime})}{U_{i}(\bm{v})},\forall j\in[i] (15)
limTUj(𝒗)Ui(𝒗)limTUj(𝒗)Ui(𝒗),j[i]\displaystyle\lim_{T\rightarrow\infty}\frac{U_{j}(\bm{v})}{U_{i}(\bm{v})}\leq\lim_{T\rightarrow\infty}\frac{U_{j}(\bm{v}^{\prime})}{U_{i}(\bm{v})},\forall j\in[i]

Consider the integral greedy allocation after transforming from 𝒗\bm{v} to 𝒗\bm{v}^{\prime}. By equation (14), the total utility increase of the ii agents is asymptotically upper bounded by (ni)(1+2log1ε)(n-i)\left(1+2\log\frac{1}{\varepsilon}\right). Then the utility of the “wealthiest” agent after the transformation should be upper bounded by

limTmaxj[i]Uj(𝒗)Ui(𝒗)1+(ni)(1+2log1ε).\lim_{T\rightarrow\infty}\frac{\max_{j\in[i]}U_{j}(\bm{v}^{\prime})}{U_{i}(\bm{v})}\leq 1+(n-i)\left(1+2\log\frac{1}{\varepsilon}\right). (16)

If (16) does not hold, this will contradict the the non-decreasing property in (15).

Combining (16) and the first inequality in (15),

limTUi(𝒗)Ui(𝒗)\displaystyle\lim_{T\rightarrow\infty}\frac{U_{i}^{\star}(\bm{v})}{U_{i}(\bm{v})} limTUi(𝒗)Ui(𝒗)\displaystyle\leq\lim_{T\rightarrow\infty}\frac{U_{i}^{\star}(\bm{v}^{\prime})}{U_{i}(\bm{v})}
=limTmaxj[i]Uj(𝒗)Ui(𝒗)limTUi(𝒗)maxj[i]Uj(𝒗)\displaystyle=\lim_{T\rightarrow\infty}\frac{\max_{j\in[i]}U_{j}(\bm{v}^{\prime})}{U_{i}(\bm{v})}\cdot\lim_{T\rightarrow\infty}\frac{U_{i}^{\star}(\bm{v}^{\prime})}{\max_{j\in[i]}U_{j}(\bm{v}^{\prime})}
(1+(ni)(1+2log1ε))limTUi(𝒗)maxj[i]Uj(𝒗)\displaystyle\leq\left(1+(n-i)\left(1+2\log\frac{1}{\varepsilon}\right)\right)\lim_{T\rightarrow\infty}\frac{U_{i}^{\star}(\bm{v}^{\prime})}{\max_{j\in[i]}U_{j}(\bm{v}^{\prime})}
(a)λiα(1+(ni)(1+2log1ε)).\displaystyle\overset{\text{(a)}}{\leq}\lambda\cdot i^{\alpha}\cdot\left(1+(n-i)\left(1+2\log\frac{1}{\varepsilon}\right)\right).

In the last step (a), we applied Lemma 8 to 𝒗\bm{v}^{\prime}, since 𝒗\bm{v}^{\prime} can be regarded as an input sequence with ii agents. This proves (13).

0.B.6 Proof of Lemma 7

The proof of 7 follows the same pattern as the proof of Theorem 3.1. First, transform the input while preserving the result of integral greedy algorithm. Then, use an optimization program to characterize that gives maximum value to the left-hand side of 7.

Thanks to Lemma 2, we do not need to consider agents that are not in J{i}J\cup\{i\}, thus assume without loss of generality that i=ni=n, J={1,2,,n1}J=\{1,2,\cdots,n-1\} and U1(A1)Un1(An1)=UU_{1}(A_{1})\leq\cdots\leq U_{n-1}(A_{n-1})=U.

For any given input 𝒗\bm{v}, we transform it into 𝒗\bm{v}^{\prime} without changing the result of integral greedy allocation:

  1. 1.

    For each j{1,2,,n1}j\in\{1,2,\cdots,n-1\} and each item in AjA_{j}, set all agents’ value to 0 except agent jj.

  2. 2.

    Put all items in AnA_{n} to the end of the sequence.

Therefore, it suffices to only consider sequences that only allocate items to agent nn after other agents have received all items in the beginning RR rounds.

maxvit\displaystyle\max_{v_{i}^{t}} 1Ut=R+1maxj<nvjt\displaystyle\ \ \frac{1}{U}\sum_{t=R+1}^{\infty}\max_{j<n}v_{j}^{t}
s.t. vit{0}[ε,1],i[n]\displaystyle\ \ v_{i}^{t}\in\{0\}\cup[\varepsilon,1],\ \ \forall i\in[n]
vntvjtUnt1Uj(Aj),j[n1]\displaystyle\ \ \frac{v_{n}^{t}}{v_{j}^{t}}\geq\frac{U_{n}^{t-1}}{U_{j}(A_{j})},\ \ \forall j\in[n-1]
Unt=s=R+1tvns\displaystyle\ \ U_{n}^{t}=\sum_{s=R+1}^{t}v_{n}^{s}

An upper bound of (maxj<nvjt)/vnt(\max_{j<n}v_{j}^{t})/v_{n}^{t} is given by min{1ε,UUnt1}\min\left\{\frac{1}{\varepsilon},\frac{U}{U_{n}^{t-1}}\right\}. The rest of the analysis is identical to the analysis of the canonical optimization program (3) in Theorem 3.1.

0.B.7 Proof of Lemma 8

For simplicity of notation, in this proof we denote Uj=Uj(Aj)U_{j}=U_{j}(A_{j}).

Consider modifying the integral greedy allocation into an optimal one: for each i,j[n]i,j\in[n], some items that belong to ii in the integral greedy allocation might be partly re-allocated to jj in the optimal allocation, and vice versa. To characterize this procedure of modification, we define the following variables:

  • Zi:=tAi(1xi,t)vitZ_{i}:=\sum_{t\in A_{i}}(1-x_{i}^{\star,t})\cdot v_{i}^{t} is agent ii’s value on the items that belong to him in the integral greedy allocation, but not in the optimal allocation.

  • Yji:=tAjxi,tvitY_{ji}:=\sum_{t\in A_{j}}x_{i}^{\star,t}v_{i}^{t} as agent ii’s value on the items that belong to agent jj in the integral greedy allocation, but now belong to agent ii in the optimal allocation. Specifically, we let Yjj=0Y_{jj}=0.

From the above definition,

Ui=UiZi+jiYji,i[n].U_{i}^{\star}=U_{i}-Z_{i}+\sum_{j\neq i}Y_{ji},\ i\in[n]. (17)

Divide all variables in (17) by maxj[n]Uj(Aj)\max_{j\in[n]}U_{j}(A_{j}) to get ui,ui,ziu_{i}^{\star},u_{i},z_{i} and yjiy_{ji}. Then (17) becomes

ui=uizi+jiyji,i[n].u_{i}^{\star}=u_{i}-z_{i}+\sum_{j\neq i}y_{ji},\ i\in[n]. (18)

All above variables should be non-negative. Applying Lemma 7, we get further constraints on yjiy_{ji}:

yji\displaystyle y_{ji} ui0min{1ε,zjui}r(u)du+oT(1)c1min{ui,uj}+oT(1)\displaystyle\leq u_{i}\int_{0}^{\min\{\frac{1}{\varepsilon},\frac{z_{j}}{u_{i}}\}}r(u)\mathrm{d}u+o_{T}(1)\leq c_{1}\min\{u_{i},u_{j}\}+o_{T}(1) (19)
ijyji\displaystyle\sum_{i\neq j}y_{ji} maxijui0min{1ε,zjui}r(u)du+oT(1)c2min{maxijui,uj}+oT(1)\displaystyle\leq\max_{i\neq j}u_{i}\int_{0}^{\min\{\frac{1}{\varepsilon},\frac{z_{j}}{u_{i}}\}}r(u)\mathrm{d}u+o_{T}(1)\leq c_{2}\min\{\max_{i\neq j}u_{i},u_{j}\}+o_{T}(1) (20)

In the above conditions we use oTo_{T} to explicitly indicate that the asymptotic notation is on TT. c1,c21/εc_{1},c_{2}\leq 1/\varepsilon are constants.

Notice that so far we have not yet used the condition that (u1,,un)(u_{1}^{\star},\cdots,u_{n}^{\star}) is an optimal allocation. We use a necessary condition (not sufficient) to characterize optimality:

uj<εuiyji=0u_{j}^{\star}<\varepsilon u_{i}^{\star}\implies y_{ji}=0 (21)

Condition (21) holds because when uj<εuiu_{j}^{\star}<\varepsilon u_{i}^{\star} in an optimal allocation, agent jj must have no envy towards agent ii. If not, then allocating any envied item to agent jj will strictly increase the geometric mean, since all non-negative values are in [ε,1][\varepsilon,1].

Next, we show that uk=o(nα)u_{k}^{\star}=o(n^{\alpha}), for any α>0\alpha>0. If uku_{k}^{\star} has a constant upper bound, this is obviously true. Next, we show that if ukg(n)u_{k}^{\star}\geq g(n) as TT goes to infinity, where g(n)g(n)\rightarrow\infty as nn\rightarrow\infty, we can derive g(n)=o(nα)g(n)=o(n^{\alpha}) for any positive α\alpha. Notice that the asymptotic notion is on nn when we refer to g(n)g(n).

Consider a directed graph DD with [n][n] as its vertex set, and the set of edges is defined as:

E(D)={(i,j):yji>0}E(D)=\left\{(i,j):y_{ji}>0\right\}

Define J()[n]J(\ell)\subseteq[n] as the set of all vertices jj such that the distance from ii to jj is \ell, i.e., the shortest directed path from ii to jj has length \ell. J(0)={i}J(0)=\{i\}. We denote m:=|J()|m_{\ell}:=|J(\ell)|.

For jJ()j\in J(\ell), there exists a path j0j1jj_{0}j_{1}\cdots j_{\ell} where j0=k,j=jj_{0}=k,j_{\ell}=j. By condition (21), the existence of edge (js,jr)(j_{s},j_{r}) implies ujrεujsu_{j_{r}}^{\star}\geq\varepsilon u_{j_{s}}^{\star}. Thus ujεuj1εuk=g(n)u_{j}^{\star}\geq\varepsilon u_{j-1}^{\star}\geq\cdots\geq\varepsilon^{\ell}u_{k}^{\star}=g(n). This gives a lower bound on the sum of uju_{j}^{\star} where jJ()j\in J(\ell):

jJ()ujεg(n)m.\sum_{j\in J(\ell)}u_{j}^{\star}\geq\varepsilon^{\ell}\cdot g(n)\cdot m_{\ell}. (22)

Next, we show inductively that m=Ω(g(n)m1).m_{\ell}=\Omega(g(n)\cdot m_{\ell-1}). For =1\ell=1, this is true because as TT\rightarrow\infty, by condition (19),

g(n)ukuk+jJ(1)yjkuk+c1m1.g(n)\leq u_{k}^{\star}\leq u_{k}+\sum_{j\in J(1)}y_{jk}\leq u_{k}+c_{1}m_{1}.

For 1\ell\geq 1, suppose m=Ω(g(n)m1)=Ω(g(n))m_{\ell^{\prime}}=\Omega(g(n)\cdot m_{\ell^{\prime}-1})=\Omega(g^{\ell^{\prime}}(n)) is true for =1,2,,\ell^{\prime}=1,2,\cdots,\ell. Consider again the sum of all uju_{j}^{\star} where jJ()j\in J(\ell), as TT\rightarrow\infty,

jJ()uj\displaystyle\sum_{j\in J(\ell)}u_{j}^{\star} jJ()uj+jJ()ijyij\displaystyle\leq\sum_{j\in J(\ell)}u_{j}+\sum_{j\in J(\ell)}\sum_{i\neq j}y_{ij}
=(a)jJ()uj+jJ()=0+1iJ()yij\displaystyle\overset{\text{(a)}}{=}\sum_{j\in J(\ell)}u_{j}+\sum_{j\in J(\ell)}\sum_{\ell^{\prime}=0}^{\ell+1}\sum_{i\in J(\ell^{\prime})}y_{ij}
=jJ()uj+=0+1iJ()jJ()yij\displaystyle=\sum_{j\in J(\ell)}u_{j}+\sum_{\ell^{\prime}=0}^{\ell+1}\sum_{i\in J(\ell^{\prime})}\sum_{j\in J(\ell)}y_{ij}
(b)jJ()uj+=0+1iJ()c2ui\displaystyle\overset{\text{(b)}}{\leq}\sum_{j\in J(\ell)}u_{j}+\sum_{\ell^{\prime}=0}^{\ell+1}\sum_{i\in J(\ell^{\prime})}c_{2}u_{i}
(1+c2)m+c2m+1+=0+1c2m\displaystyle\leq(1+c_{2})m_{\ell}+c_{2}m_{\ell+1}+\sum_{\ell^{\prime}=0}^{\ell+1}c_{2}{m_{\ell}^{\prime}}
=(c)O(m)+c2m+1,\displaystyle\overset{\text{(c)}}{=}O(m_{\ell})+c_{2}m_{\ell+1}, (23)

where (a) is because yij=0y_{ij}=0 for iJ()i\in J(\ell^{\prime}) if +2\ell^{\prime}\geq\ell+2, (b) is a direct application of condition 20, and (c) is by the induction hypothesis.

Combining (22) and (23), c2m+1+O(m)εg(n)m=Ω(mg(n))c_{2}m_{\ell+1}+O(m_{\ell})\geq\varepsilon^{\ell}g(n)m_{\ell}=\Omega(m_{\ell}g(n)). This tells us m+1=Ω(g(n)m)m_{\ell+1}=\Omega(g(n)\cdot m_{\ell}) and thus completes the induction.

Define :=max{:J() is non-empty}\ell^{*}:=\max\{\ell:J(\ell)\text{ is non-empty}\}, i.e., \ell^{*} is the length of the longest path starting from SS. Because m=Ω(g(n))nm_{\ell^{*}}=\Omega\left(g^{\ell^{*}}(n)\right)\leq n, we have

=O(lognlogg(n)).\ell^{*}=O\left(\frac{\log n}{\log g(n)}\right). (24)

Meanwhile, by (23), m=Ω(εg(n)m)m_{\ell^{*}}=\Omega(\varepsilon^{\ell^{*}}g(n)m_{\ell^{*}}). We then know that,

g(n)=O(1ε).g(n)=O\left(\frac{1}{\varepsilon^{\ell^{*}}}\right). (25)

Combining (24) and (25) one can derive g(n)=O(elogn)=o(nα)g(n)=O(e^{\sqrt{\log n}})=o(n^{\alpha}) for any α>0\alpha>0.

0.B.8 Proof of Theorem 3.4

Consider dividing the horizon of length TT into nn different phases, with length T1,,TnT_{1},\cdots,T_{n}, satisfying

j=1nTj=T,limTTj1Tj=0(j={1,2,,n})\sum_{j=1}^{n}T_{j}=T,\ \lim_{T\rightarrow\infty}\frac{T_{j-1}}{T_{j}}=0\ \ (\forall j=\{1,2,\cdots,n\})

For the first phase (i.e. for t{1,2,,T1}t\in\{1,2,\cdots,T_{1}\}), vit=1v_{i}^{t}=1 for all i[n]i\in[n]. Because there are T1T_{1} items and nn buyers, for any online integral algorithm there exists a buyer i1i_{1} whose utility after phase 1 satisfies Ui1T1T1n{U}_{i_{1}}^{T_{1}}\leq\frac{T_{1}}{n}.

Recursively, for each phase k2k\geq 2, we set for all t=Tk1+1,,Tkt=T_{k-1}+1,\cdots,T_{k}:

vit={0i{i1,,ik1}1o.w.,v_{i}^{t}=\begin{cases}0&i\in\{i_{1},\cdots,i_{k-1}\}\\ 1&\text{o.w.}\end{cases},

and let ik{i1,,ik1}i_{k}\not\in\{i_{1},\cdots,i_{k-1}\} be a buyer whose utility after round TkT_{k} satisfies UikTk1nj=1kTj{U}_{i_{k}}^{T_{k}}\leq\frac{1}{n}\sum_{j=1}^{k}T_{j}.

Notice that the optimal offline allocation will allocate everything in phase kk to buyer iki_{k}. We have limTUi=Tni+1\lim_{T\rightarrow\infty}U_{i}^{\star}=T_{n-i+1} and

limT(Ui),TUiT1i.\lim_{T\rightarrow\infty}\frac{({U}_{i})^{\star,T}}{{U}_{i}^{T}}\geq\frac{1}{i}.

This leads to a ratio at least (n!)1/n(n!)^{1/n}.

0.B.9 Proof of Theorem 3.5

Define vmax=maxi,τviτ{v_{\mathrm{max}}}=\max_{i,\tau}v^{\tau}_{i}, vi,max=maxτviτ{v_{i,\mathrm{max}}}=\max_{\tau}{v^{\tau}_{i}}, and v¯max=1ni=1nvi,max\bar{v}_{\mathrm{max}}=\frac{1}{n}{\sum_{i=1}^{n}}v_{i,\mathrm{max}}. In this proof, we study the general case with viτ[0,vmax]{v_{i}^{\tau}}\in[0,v_{\mathrm{max}}]. We will prove that

Rδ(v)3+4δv¯max+2log(1+vmaxδ)+2logT.\displaystyle R_{\delta}(v)\leq 3+\frac{4}{\delta}\bar{v}_{\mathrm{max}}+2\log\Big{(}1+\frac{{v_{\mathrm{max}}}}{\delta}\Big{)}+2\log T\;. (26)

The proof is divided into two major steps.

0.B.9.1 Step 1. Ratio bounded by prices.

We first introduce notations for quantities that appear during the run of Algorithm 3.

Let pτ=maxiβiτviτ=viττ/(Uiττ1+δ){p^{\tau}}=\max_{i}\beta^{\tau}_{i}{v^{\tau}_{i}}={v^{\tau}_{i^{\tau}}}/({U^{\tau-1}_{i^{\tau}}+\delta}) be the price of item τ\tau, and βiτ=1δ+Uiτ1\beta^{\tau}_{i}=\frac{1}{\delta+{U^{\tau-1}_{i}}}. Then Algorithm 3 Equation 5 can be written as iτ=minargmaxi{βiτviτ}{i^{\tau}}=\min\arg\max_{i}\{\beta_{i}^{\tau}v_{i}^{\tau}\}. Also let uiτ=viτ1(i=iτ){u^{\tau}_{i}}={v_{i}^{\tau}}1(i=i^{\tau}) be the utility of agent ii from item τ\tau according to the allocation of the seeded integral greedy algorithm. Let Uiτ=s=1,,τuisU^{\tau}_{i}=\sum_{s=1,\dots,\tau}u^{s}_{i} be the cumulative utilities for τ1\tau\geq 1 and Ui0=0U^{0}_{i}=0 and Ui=UiTU_{i}=U_{i}^{T}.

We begin with the following claim.

Claim.

For any feasible hindsight allocation xx^{*} and its resulting utilities U~i\widetilde{U}_{i},

1ni=1nU~i+δUi+δ1+1nτ=1Tpτ=1+1ni=1nτ=1TuiτUiτ1+δ.\displaystyle\frac{1}{n}{\sum_{i=1}^{n}}\frac{\widetilde{U}_{i}+\delta}{U_{i}+\delta}\leq 1+\frac{1}{n}{\sum_{\tau=1}^{T}}{p^{\tau}}=1+\frac{1}{n}{\sum_{i=1}^{n}}{\sum_{\tau=1}^{T}}\frac{{u^{\tau}_{i}}}{{U^{\tau-1}_{i}}+\delta}\;. (27)

Notice that 1ni=1nU~i+δUi+δ1ni=1nU~iUi+δ+1\frac{1}{n}{\sum_{i=1}^{n}}\frac{\widetilde{U}_{i}+\delta}{U_{i}+\delta}\leq\frac{1}{n}{\sum_{i=1}^{n}}\frac{\widetilde{U}_{i}}{U_{i}+\delta}+1. Then,

1ni=1nU~iUi+δ\displaystyle\frac{1}{n}{\sum_{i=1}^{n}}\frac{\widetilde{U}_{i}}{U_{i}+\delta}
=1ni=1n(τ=1Txi,τviτUi+δ)\displaystyle=\frac{1}{n}{\sum_{i=1}^{n}}\bigg{(}\frac{{\sum_{\tau=1}^{T}}{x^{*,\tau}_{i}}{v^{\tau}_{i}}}{U_{i}+\delta}\bigg{)}
=1ni=1nτ=1T(viτUi+δxi,τ)\displaystyle=\frac{1}{n}{\sum_{i=1}^{n}}{\sum_{\tau=1}^{T}}\bigg{(}\frac{{v^{\tau}_{i}}}{U_{i}+\delta}\cdot{x^{*,\tau}_{i}}\bigg{)}
1ni=1nτ=1T(viτUiτ1+δxi,τ)=:(A).\displaystyle\leq\frac{1}{n}{\sum_{i=1}^{n}}{\sum_{\tau=1}^{T}}\bigg{(}\frac{{v^{\tau}_{i}}}{{{U^{\tau-1}_{i}}}+\delta}\cdot{x^{*,\tau}_{i}}\bigg{)}=:(\text{A})\;.

By the definition of pτ{p^{\tau}}, it holds for all agent ii,

viτUiτ1+δpτ.\displaystyle\frac{{v^{\tau}_{i}}}{{{U^{\tau-1}_{i}}}+\delta}\leq{p^{\tau}}\;.

Applying the supply feasibility constraint,

(A)1ni=1nτ=1Tpτxi,τ1nτ=1Tpτ.\displaystyle(\mathrm{A})\leq\frac{1}{n}{\sum_{i=1}^{n}}{\sum_{\tau=1}^{T}}{p^{\tau}}\cdot{x^{*,\tau}_{i}}\leq\frac{1}{n}{\sum_{\tau=1}^{T}}{p^{\tau}}\;.

Next, use integrality of Algorithm 3 and rewrite the sum of prices as the sum of running ratios:

1nτ=1Tpτ\displaystyle\frac{1}{n}{\sum_{\tau=1}^{T}}{p^{\tau}}
=1nτ=1TviττUiττ1+δxiττ\displaystyle=\frac{1}{n}{\sum_{\tau=1}^{T}}\frac{v^{\tau}_{i^{\tau}}}{U^{\tau-1}_{i^{\tau}}+\delta}\cdot x^{\tau}_{{i^{\tau}}}
=1nτ=1T(viττUiττ1+δxiττ+iiτviτUiτ1+δxiτ)\displaystyle=\frac{1}{n}{\sum_{\tau=1}^{T}}\bigg{(}\frac{v^{\tau}_{i^{\tau}}}{U^{\tau-1}_{i^{\tau}}+\delta}\cdot x^{\tau}_{{i^{\tau}}}+\sum_{i\neq{i^{\tau}}}\frac{{v^{\tau}_{i}}}{{U^{\tau-1}_{i}}+\delta}\cdot{x^{\tau}_{i}}\bigg{)}
=1nτ=1Ti=1n(viτUiτ1+δxiτ).\displaystyle=\frac{1}{n}{\sum_{\tau=1}^{T}}{\sum_{i=1}^{n}}\bigg{(}\frac{{v^{\tau}_{i}}}{{U^{\tau-1}_{i}}+\delta}\cdot{x^{\tau}_{i}}\bigg{)}\;.

0.B.9.2 Step 2. Introducing a basic inequality.

Lemma 9 (Bach and Levy [2019]).

For nonnegative numbers a1,,an[0,a]a_{1},\ldots,a_{n}\in[0,a] and any a00a_{0}\geq 0, it holds

i=1naia0+j=1i1aj2+4aa0+2log(1+i=1n1ai/a0).\displaystyle\sum_{i=1}^{n}\frac{a_{i}}{a_{0}+\sum_{j=1}^{i-1}a_{j}}\leq 2+\frac{4a}{a_{0}}+2\log\left(1+\sum_{i=1}^{n-1}a_{i}/a_{0}\right)\;.

We use the above inequality to bound the right-hand side of (27). For a fixed agent ii, let vi,max=maxτviτ{v_{i,\mathrm{max}}}=\max_{\tau}{v^{\tau}_{i}}, then

τ=1T(viτUiτ1+δxiτ)\displaystyle{\sum_{\tau=1}^{T}}\bigg{(}\frac{{v^{\tau}_{i}}}{{U^{\tau-1}_{i}}+\delta}\cdot{x^{\tau}_{i}}\bigg{)}
=τ=1T(viτxiτs=1τ1visxis+δ)\displaystyle={\sum_{\tau=1}^{T}}\bigg{(}\frac{{v^{\tau}_{i}}{x^{\tau}_{i}}}{{\sum_{s=1}^{\tau-1}}{v^{s}_{i}}{x^{s}_{i}}+\delta}\bigg{)}
2+4vi,maxδ+2log(1+τ=1T1uiτδ)\displaystyle\leq 2+4\frac{{v_{i,\mathrm{max}}}}{\delta}+2\log\bigg{(}1+\sum_{\tau=1}^{T-1}\frac{{u^{\tau}_{i}}}{\delta}\bigg{)} (Invoking Lemma 9)
=2+4vi,maxδ+2logT+2log(1T+1Tτ=1T1uiτδ)\displaystyle=2+4\frac{{v_{i,\mathrm{max}}}}{\delta}+2\log T+2\log\bigg{(}\frac{1}{T}+\frac{1}{T}\sum_{\tau=1}^{T-1}\frac{{u^{\tau}_{i}}}{\delta}\bigg{)}
2+4vi,maxδ+2logT+2log(1+vmaxδ).\displaystyle\leq 2+4\frac{{v_{i,\mathrm{max}}}}{\delta}+2\log T+2\log\bigg{(}1+\frac{{v_{\mathrm{max}}}}{\delta}\bigg{)}\;.

Finally, putting together we have the desired result.

1ni=1nU~i+δUi+δ\displaystyle\frac{1}{n}{\sum_{i=1}^{n}}\frac{\widetilde{U}_{i}+\delta}{U_{i}+\delta}
1+1ni=1n(2+4vi,maxδ+2logT+2log(1+vmaxδ))\displaystyle\leq 1+\frac{1}{n}{\sum_{i=1}^{n}}\Bigg{(}2+4\frac{{v_{i,\mathrm{max}}}}{\delta}+2\log T+2\log\bigg{(}1+\frac{{v_{\mathrm{max}}}}{\delta}\bigg{)}\Bigg{)}
=3+4δv¯max+2log(1+vmaxδ)+2logT.\displaystyle=3+\frac{4}{\delta}\bar{v}_{\mathrm{max}}+2\log\bigg{(}1+\frac{{v_{\mathrm{max}}}}{\delta}\bigg{)}+2\log T\;.

Appendix 0.C Missing Proofs in Section 4

0.C.1 Proof of Lemma 4

By the definition of tt^{*} and the bound on \ell one can check:

U1(A1)S>1εt.U_{1}(A_{1})\cdot S>\frac{1}{\varepsilon}\ell t^{*}.

If Ujs1(𝒗)<(s1)U_{j}^{s^{\prime}-1}(\bm{v}^{\prime})<\ell(s^{\prime}-1), then

U^js1(𝒗)U^1s1(𝒗)=tU1(A1)S<ε.\frac{\hat{U}_{j}^{s^{\prime}-1}(\bm{v}^{\prime})}{\hat{U}_{1}^{s^{\prime}-1}(\bm{v}^{\prime})}=\frac{\ell t^{*}}{U_{1}(A_{1})\cdot S}<\varepsilon.

Now we show that the PACE allocation is preserved for the first tt^{*} items in the transformed sequence. For an original item sAjs\in A_{j} and corresponding item in ss^{\prime} in 𝒗\bm{v}^{\prime} (sts^{\prime}\leq t^{*}), we check the two cases:

  1. 1.

    When Ujs1(𝒗)(s1)U_{j}^{s^{\prime}-1}(\bm{v}^{\prime})\geq\ell(s^{\prime}-1), we have

    U^1s1(𝒗)\displaystyle\hat{U}_{1}^{s^{\prime}-1}(\bm{v}^{\prime}) =max{U1(A1),(s1)}max{U1s(𝒗),(s1)}U^1s(𝒗),\displaystyle=\max\{U_{1}(A_{1}),\ell(s^{\prime}-1)\}\geq\max\{U_{1}^{s}(\bm{v}),\ell(s-1)\}\geq\hat{U}_{1}^{s}(\bm{v}),
    U^js1(𝒗)\displaystyle\hat{U}_{j}^{s^{\prime}-1}(\bm{v}^{\prime}) =max{Ujs1(𝒗),s}=Ujs1(𝒗)=max{Ujs1(𝒗),(s1)}=U^js(𝒗).\displaystyle=\max\{U_{j}^{s^{\prime}-1}(\bm{v}^{\prime}),\ell s^{\prime}\}=U_{j}^{s^{\prime}-1}(\bm{v}^{\prime})=\max\{U_{j}^{s-1}(\bm{v}),\ell(s-1)\}=\hat{U}_{j}^{s}(\bm{v}).

    Hence U^js1(𝒗)/U^1s1(𝒗)U^js(𝒗)/U^1s(𝒗)>vjs/v1s\hat{U}_{j}^{s^{\prime}-1}(\bm{v}^{\prime})/\hat{U}_{1}^{s^{\prime}-1}(\bm{v}^{\prime})\geq\hat{U}_{j}^{s}(\bm{v})/\hat{U}_{1}^{s}(\bm{v})>v_{j}^{s}/v_{1}^{s}.

  2. 2.

    When Ujt1(𝒗)<(t1)U_{j}^{t^{\prime}-1}(\bm{v}^{\prime})<\ell(t^{\prime}-1), by Lemma 4,

    U^js1(𝒗)U^1s1(𝒗)<εvjsv1s\displaystyle\frac{\hat{U}_{j}^{s^{\prime}-1}(\bm{v}^{\prime})}{\hat{U}_{1}^{s^{\prime}-1}(\bm{v}^{\prime})}<\varepsilon\leq\frac{v_{j}^{s}}{v_{1}^{s}}

Because we have set all agents’ valuations to 0 except agent 11 and jj, in both cases the item is indeed allocated to agent jj, which is the same as the original input 𝒗\bm{v}.

0.C.2 Proof of Lemma 5

In this proof, we let a=U1(A1)/|A1|[ε,1]a={U_{1}(A_{1})}/{|A_{1}|}\in[\varepsilon,1] be the per-round utility gain of agent 11 in the beginning SS rounds.

To identify the sequence that gives maximum envy of agent 11 on agent jj’s bundle BjB_{j}, we use the same technique as the proof of Theorem 3.1. The problem of maximizing agent 11’s envy is equivalent to the following program:

maxv1ti,vjti,ti\displaystyle\max_{v_{1}^{t_{i}},v_{j}^{t_{i}},t_{i}} 1aSi=1kv1ti\displaystyle\ \ \frac{1}{aS}\sum_{i=1}^{k}v_{1}^{t_{i}} (28)
s.t. S<t1<<tkt\displaystyle\ \ S<t_{1}<\cdots<t_{k}\leq t^{*}
v1ti,vjti{0}[ε,1]\displaystyle\ \ v_{1}^{t_{i}},v_{j}^{t_{i}}\in\{0\}\cup[\varepsilon,1]
vjtiv1tiU^jti1U^1ti1\displaystyle\ \ \frac{v_{j}^{t_{i}}}{v_{1}^{t_{i}}}\geq\frac{\hat{U}_{j}^{t_{i-1}}}{\hat{U}_{1}^{t_{i-1}}}
U^1ti=max{ti,aS}\displaystyle\ \ \hat{U}_{1}^{t_{i}}=\max\{\ell{t_{i}},aS\}
U^jti=max{ti,Ujti}\displaystyle\ \ \hat{U}_{j}^{t_{i}}=\max\{\ell{t_{i}},{U}_{j}^{t_{i}}\}
Ujti=i=1kvjti\displaystyle\ \ {U}_{j}^{t_{i}}=\sum_{i=1}^{k}v_{j}^{t_{i}}

{t1,,tk}\{t_{1},\cdots,t_{k}\} are kk positions where we put the items of BjB_{j}.

Recall the technique we have used when proving Theorem 3.1, where we introduced the function qq such that q(u)q(u) is an upper bound of v1t+1/vjt+1v_{1}^{t+1}/v_{j}^{t+1} when Ujt=uU_{j}^{t}=u. When the horizon is infinitely long, each choice of v1tv_{1}^{t} makes up only a small step, and the envy is upper–bounded by the integral 1U1S0q(u)du\frac{1}{U_{1}^{S}}\int_{0}^{\infty}q(u)\mathrm{d}u in the asymptotic sense.

There is a major difference when it comes to PACE. In the integral greedy algorithm, the upper bound of v1t+1/vjt+1v_{1}^{t+1}/v_{j}^{t+1} is clearly given by min{1ε,U1Su}\min\{\frac{1}{\varepsilon},\frac{U_{1}^{S}}{u}\} once we know that Ujt=uU_{j}^{t}=u. However, in PACE, knowing only Ujt=uU_{j}^{t}=u is not sufficient to bound on v1t+1/v2t+1v_{1}^{t+1}/v_{2}^{t+1}. The bound may also depend on tt, the position of the item.

v1tv2tmin{1ε,max{(t1),aS}max{(t1),U2t1}}.\frac{v_{1}^{t}}{v_{2}^{t}}\leq\min\left\{\frac{1}{\varepsilon},\frac{\max\{\ell(t-1),aS\}}{\max\{\ell(t-1),{U}_{2}^{t-1}\}}\right\}.

Thanks to Lemma 4 and the fact that t<aS\ell t^{*}<aS, we can simplify the above equation and remove the dependence on tt.

v1tv2t{1εU2t1/aSεaSU2t1ε<U2t1/aS1/ε0U2t1/aS1/ε.\frac{v_{1}^{t}}{v_{2}^{t}}\leq\begin{cases}\frac{1}{\varepsilon}&U_{2}^{t-1}/aS\leq\varepsilon\\ \frac{aS}{U_{2}^{t-1}}&\varepsilon<U_{2}^{t-1}/aS\leq 1/\varepsilon\\ 0&U_{2}^{t-1}/aS\geq 1/\varepsilon\end{cases}. (29)

The upper bound (29) is identical to the constraint in a canonical optimization program of an envy maximizer parametrized by aSaS, defined as (3) in the proof of Theorem 3.1. Hence, both asymptotic upper bounds for envy and |BjC1||B_{j}\cap C_{1}| in the canonical optimization program still hold.

0.C.3 Proof of Lemma 6

To prove Lemma 6, we show that for any given cc, by choosing ¯\ell\leq\bar{\ell} where ¯<cε1+(n1)(1+log1/ε)\bar{\ell}<\frac{c\varepsilon}{1+(n-1)(1+\log 1/\varepsilon)}, we have for any i[n]i\in[n],

limTUi(Ai)T¯.\lim_{T\rightarrow\infty}\frac{U_{i}(A_{i})}{T}\geq{\bar{\ell}}.

We show this by contradiction. Assume there exists i[n]i\in[n] such that Ui(Ai)<TU_{i}(A_{i})<T. Let Ci={t:vit}C_{i}=\{t:v_{i}^{t}\} be the set of items which agent ii has non-zero value. Consider item ss, the last item in AjCiA_{j}\cap C_{i}. To construct a contradiction, we show that Ui(Ai)<¯TU_{i}(A_{i})<\bar{\ell}T implies an upper bound on each |AjCi||A_{j}\cap C_{i}|, and thus |Ci||C_{i}| (with a union bound). This might contradict the lower bound ViV_{i}.

Consider an adversarial setting where the adversary attempts to maximize |AjCi||A_{j}\cap C_{i}|, subject to the assumptions on the input and the allocation rule of PACE. Suppose AjCi={t1,,tm}A_{j}\cap C_{i}=\{t_{1},\cdots,t_{m}\} where t1<<tmt_{1}<\cdots<t_{m}. We relax the constraint on the adversary as follows:

  1. 1.

    For tAjCit\in A_{j}\cap C_{i}, the adversary subject to constraint

    vjtU^jt1maxk[n]vktU^kt1.\frac{v_{j}^{t}}{\hat{U}_{j}^{t-1}}\geq\max_{k\in[n]}\frac{v_{k}^{t}}{\hat{U}_{k}^{t-1}}.

    We relax this into

    vjtU^jt1vitU^it1.\frac{v_{j}^{t}}{\hat{U}_{j}^{t-1}}\geq\frac{v_{i}^{t}}{\hat{U}_{i}^{t-1}}. (30)
  2. 2.

    Notice that U^jtUjt\hat{U}_{j}^{t}\geq U_{j}^{t} and U^it¯T\hat{U}_{i}^{t}\leq\bar{\ell}T by assumption. Then we can further relax (30) into

    vjtUjt1vit¯T.\frac{v_{j}^{t}}{{U}_{j}^{t-1}}\geq\frac{v_{i}^{t}}{\bar{\ell}T}. (31)
  3. 3.

    The constraint on UjtU_{j}^{t} is minimum:

    UjtsAjCi,s<tvjs.U_{j}^{t}\geq\sum_{s\in A_{j}\cap C_{i},s<t}v_{j}^{s}. (32)

The relaxed adversarial setting consists of constraint (31), (32) and the value domain requirement vit,vjt[0,1]v_{i}^{t},v_{j}^{t}\in[0,1]. The value it concerns are vit,vjtv_{i}^{t},v_{j}^{t} where tCiAjt\in C_{i}\cap A_{j}.

Recall the definition of the canonical optimization program for an envy-maximizer in (3). We notice that the above relaxed constraint is identical to a canonical optimization program parametrized by ¯T\bar{\ell}T. By Lemma 3, the length of the value sequence in the relaxed setting is asymptotically bounded,

limT|CiAj|T¯ε(1+log1ε).\lim_{T\rightarrow\infty}\frac{|C_{i}\cap A_{j}|}{T}\leq\frac{\bar{\ell}}{\varepsilon}\left(1+\log\frac{1}{\varepsilon}\right). (33)

Since (33) is an upper bound for the relaxed setting, it must also hold for the true adversarial setting. We then know that

limT|Ci|T\displaystyle\lim_{T\rightarrow\infty}\frac{|C_{i}|}{T} =limT|Ai|T+jilimT|CiAj|T\displaystyle=\lim_{T\rightarrow\infty}\frac{|A_{i}|}{T}+\sum_{j\neq i}\lim_{T\rightarrow\infty}\frac{|C_{i}\cap A_{j}|}{T}
¯ε(1+(n1)(1+log1ε))\displaystyle\leq\frac{\bar{\ell}}{\varepsilon}\left(1+(n-1)\left(1+\log\frac{1}{\varepsilon}\right)\right)
<c,\displaystyle<c,

where the last step is by the definition of ¯\bar{\ell}.

However, this is impossible, since |Ci|<cT|C_{i}|<cT implies Vi<cTV_{i}<cT. By contradiction we must have Ui(Ai)¯TU_{i}(A_{i})\geq\bar{\ell}T for sufficiently large TT.

0.C.4 Proof of Theorem 4.2

Since <cε1+(n1)(1+log1/ε),r=1\ell<\frac{c\varepsilon}{1+(n-1)\left(1+\log 1/\varepsilon\right)},r=1, by Lemma 6,

𝒗𝒱ε,cT𝒗𝒱εT(,r).\bm{v}\in\mathcal{V}_{\varepsilon,c}^{T}\implies\bm{v}\in\mathcal{V}_{\varepsilon}^{T}(\ell,r).

We can then apply the envy bound in Theorem 4.1 to 𝒗𝒱ε,cT\bm{v}\in\mathcal{V}_{\varepsilon,c}^{T} to get that for all iji\neq j,

limTUi(Aj)Ui(Ai)1+2log1ε.\lim_{T\rightarrow\infty}\frac{U_{i}(A_{j})}{U_{i}(A_{i})}\leq{1+2\log\frac{1}{\varepsilon}}.

Summing over all j[n]j\in[n] and using Vi=j=1nUi(Aj)V_{i}=\sum_{j=1}^{n}U_{i}(A_{j}), we have

limTViUi(Ai)=limTj=1nUi(Aj)Ui(Ai)1+(n1)(1+2log1ε).\lim_{T\rightarrow\infty}\frac{V_{i}}{U_{i}(A_{i})}=\lim_{T\rightarrow\infty}\sum_{j=1}^{n}\frac{U_{i}(A_{j})}{U_{i}(A_{i})}\leq 1+(n-1)\left({1+2\log\frac{1}{\varepsilon}}\right).

Applying the condition VicTV_{i}\geq cT,

limTUi(Ai)Tc(1+(n1)(1+2log1ε))1cn(1+2log1ε)1.\lim_{T\rightarrow\infty}\frac{U_{i}(A_{i})}{T}\geq c\left(1+(n-1)\left(1+2\log\frac{1}{\varepsilon}\right)\right)^{-1}\geq\frac{c}{n}\left(1+2\log\frac{1}{\varepsilon}\right)^{-1}.

On the other hand, by the AM-GM inequality, the optimal value of Nash Welfare can be bounded as follows:

i=1nUi(Tn)n.\prod_{i=1}^{n}{U_{i}^{\star}}\leq\left(\frac{T}{n}\right)^{n}.

Following from the above two inequalities,

limTi=1nUiUi(Ai)limTi=1nUiTUi(Ai)T1c(1+2log1ε).\displaystyle\lim_{T\rightarrow\infty}\prod_{i=1}^{n}\frac{U_{i}^{\star}}{U_{i}(A_{i})}\leq\lim_{T\rightarrow\infty}\prod_{i=1}^{n}U_{i}^{\star}\cdot\frac{T}{U_{i}(A_{i})}\cdot T\leq\frac{1}{c}\left(1+2\log\frac{1}{\varepsilon}\right).

This proves Theorem 4.2.