This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Convergence of Multi-Issue Iterative Voting under Uncertainty

Joshua Kavner Reshef Meir Rensselaer Polytechnic Institute
{[email protected],[email protected]}
Francesca Rossi Technion Israel Institute of Technology
[email protected]
Lirong Xia
Abstract

We study the effect of strategic behavior in iterative voting for multiple issues under uncertainty. We introduce a model synthesizing simultaneous multi-issue voting with Meir et al. (2014)’s local dominance theory and determine its convergence properties. After demonstrating that local dominance improvement dynamics may fail to converge, we present two sufficient model refinements that guarantee convergence from any initial vote profile for binary issues: constraining agents to have 𝒪\mathcal{O}-legal preferences and endowing agents with less uncertainty about issues they are modifying than others. Our empirical studies demonstrate that although cycles are common when agents have no uncertainty, introducing uncertainty makes convergence almost guaranteed in practice.

1 Introduction

Consider a wedding planner who is deciding a wedding’s banquet and wants to accommodate the party invitees’ preferences. Suppose they have three issues: the main course (chicken, beef, or salmon), the paired wine (red or white), and the dessert cake’s flavor (chocolate or vanilla). How should the planner proceed? On the one hand, the planner could request each attendee’s (agent’s) full preference ranking over the dpd^{p} possibilities (alternatives), for pp issues with dd candidates each. However, this is computationally prohibitive and imposes a high cognitive cost for agents.

On the other hand, the planner could only request agents’ most preferred alternatives and decide each issue simultaneously. Although simpler, this option admits multiple election paradoxes whereby agents can collectively select each of their least favored outcomes. For example, suppose three agents prefer (1,1,0)(1,1,0), (1,0,1)(1,0,1), and (0,1,1)(0,1,1) first respectively on three binary issues and all prefer (1,1,1)(1,1,1) last Lacy and Niou (2000). Then the agents select (1,1,1(1,1,1) by majority rule on each issue independently.

A third approach is to decide the issues in sequence and have agents vote for their preferred alternative on the current issue given the previously chosen outcomes Lang and Xia (2009); Xia et al. (2011). Still, the joint outcome may depend on the voting agenda and agents may be uneasy voting on the current issue if their preference depends on the outcomes of later issues Conitzer et al. (2009).

In this work, we study iterative voting (IV) as a different yet natural method for deciding multiple issues Meir et al. (2010). We elicit agents’ most preferred alternatives and, given information about others’ votes, allow agents to update their reports before finalizing the group decision. This approach combines the efficiency of simultaneous voting with the dynamics of sequential voting, thus incorporating information about agents’ lower-ranked preferences without directly eliciting them. Like the former approach, agents only report their most preferred alternative. Like the latter approach, agents only update one issue at a time but are unrestricted in the order of improvements.

IV is an effective framework for its adaptability to various information and behavioral schemes. For example, agents may be fully aware of each other’s votes, such as in online Doodle polls Zou et al. (2015), and update to the myopic best response of all others. Alternatively, agents may only have partial information, such as from imprecise opinion polls Reijngoud and Endriss (2012) or latency if they can only periodically retrieve vote counts on a networked system. We then take a strict uncertainty approach in which agents assume that any vote profile in a set is possible without assigning probabilities to them. Agents update to votes that locally dominate their prior reports Meir et al. (2014).

We ask two primary questions: (1) Under what conditions does multi-issue IV converge? (2) How does introducing and increasing uncertainty affect the rate of convergence?

The conclusion from previous research regarding convergence in single-issue IV is mixed, as the plurality and veto voting rules have strong guarantees, yet many other rules admit cycles Meir et al. (2017). This leaves us with mixed hope that multi-issue IV can converge, and if so, that it can solve other problems like multiple election paradoxes Xia et al. (2011). Furthermore, in contrast to the single-issue setting Meir (2015), uncertainty for multiple issues plays a double role. First, like the single-issue case, agents consider themselves as possibly pivotal on any issue that is sufficiently close to a tie. Second—and this part is new—agents may be uncertain regarding their own preferences on a particular issue, as this may depend on the outcomes of other uncertain issues.

1.1 Our Contribution

On the conceptual side, we introduce a model that synthesizes prior work in local dominance strategic behavior, iterative voting, and simultaneous voting over multiple issues. This generalized model naturally captures both types of uncertainty discussed above.

On the technical side, we first show that IV for either best response, without uncertainty, or local dominance improvement dynamics, with uncertainty, may not converge. We then present two model refinements that prove sufficient to guarantee convergence for binary issues: restricting agent preferences to be 𝒪\mathcal{O}-legal converges because agents’ preferences on issues are not interdependent; alternating uncertainty, in which agents are more certain about the current issue than others, converges because fewer preference rankings yield improvement steps in this case. These convergence results do not extend to the multi-candidate issues setting, as LDI plurality dynamics may cycle if agents have partial order preference information.

Our convergence results for binary issues also hold for a nonatomic variant of IV, in which agents are part of a very large population and arbitrary subsets of agents may change their vote simultaneously, establishing more general convergence results (see Appendix C).

We conclude with empirical evidence corroborating our findings that introducing uncertainty eliminates almost all cycles in IV for multiple binary issues. Our experiments further suggest IV improves the quality of equilibrium vote profiles relative to their respective truthful profiles, thus reducing multiple election paradoxes. Increasing uncertainty yields faster convergence but degrades this welfare improvement.

1.2 Related Work

Our model pulls insights from research across multi-issue voting, IV, and local dominance strategic behavior. Multi-issue voting is an extensively studied problem in economics and computer science with applications in direct democracy referendums, group planning and committee elections (see e.g., Lang and Xia (2016) for a survey). Our work follows research in agent strategic behavior by Lang (2007), Lang and Xia (2009), Conitzer et al. (2009), Xia et al. (2011).

Single-issue IV was initially studied by Meir et al. (2010) for best response dynamics and the plurality social choice rule, whose authors bounded its guaranteed convergence rate. Subsequent work demonstrated that iterative veto converges Reyhani and Wilson (2012); Lev and Rosenschein (2012), although many other voting rules do not Koolyk et al. (2017) unless agents are additionally restricted in their behavior Reijngoud and Endriss (2012); Grandi et al. (2013); Obraztsova et al. (2015); Endriss et al. (2016).

This review already narrows down the possibility of convergence in the multi-issue setting. We therefore restrict our attention to plurality, extending the models of Meir et al. (2014), Meir (2015). Their research finds broad conditions for voting equilibrium to exist and guaranteed convergence for iterative plurality. In particular, Meir (2015) studies a nonatomic model variation where agents have negligible impact on the outcome but multiple agents update their votes simultaneously, greatly simplifying the dynamics.

Bowman et al. (2014) and Grandi et al. (2020) empirically show for multiple binary issues that IV improves the social welfare of voting outcomes using computer simulations and human experiments respectively. Our work augments this research by characterizing convergence in settings where agents do not have complete information.

A related line of research studies strategic behavior in epistemic voting games, when agents have uncertainty about the preferences or votes of others (see e.g., Meir (2018) for a survey). Notably, Chopra et al. (2004), Conitzer et al. (2011), Reijngoud and Endriss (2012), Van Ditmarsch et al. (2013) each study the susceptibility and computational complexity of voting rules to local dominance improvement steps. Game-theoretic properties of strategic behavior for Gibbard-Satterthwaite games were studied by Myerson and Weber (1993), Grandi et al. (2019), Elkind et al. (2020). Other forms of IV include work from Airiau and Endriss (2009), Desmedt and Elkind (2010), Xia and Conitzer (2010).

2 Preliminaries

Basic model.

Let 𝒫={1,2,,p}\mathcal{P}=\{1,2,\ldots,p\} be the set of pp issues over the joint domain 𝒟=×i=1pDi\mathcal{D}=\times_{i=1}^{p}D_{i}, where DiD_{i} is the finite value domain of candidates for issue ii. We call the issues binary if Di={0,1}D_{i}=\{0,1\} for each i𝒫i\in\mathcal{P} or multi-candidate otherwise. Each of nn agents is endowed with a preference ranking Rj(𝒟)R_{j}\in\mathcal{L}(\mathcal{D}), the set of strict linear orders over the ×i=1p|Di|\times_{i=1}^{p}|D_{i}| alternatives. We call the collection of agents’ preferences P=(R1,,Rn)P=(R_{1},\ldots,R_{n}) a preference profile and each agent’s most preferred alternative their truthful vote. A vote profile a=(a1,,an)𝒟na=(a_{1},\ldots,a_{n})\in\mathcal{D}^{n} is a collection of votes, where aj=(aj1,,ajp)𝒟a_{j}=(a^{1}_{j},\ldots,a^{p}_{j})\in\mathcal{D} collects agent jj’s single-candidate vote per issue. A resolute voting rule f:𝒟n𝒟f:\mathcal{D}^{n}\rightarrow\mathcal{D} maps vote profiles to a unique outcome. We call a𝒟a\in\mathcal{D} and aiDia^{i}\in D_{i} for i𝒫i\in\mathcal{P} an alternative or outcome synonymously.

Simultaneous plurality voting.

A local voting rule, applied to each issue independently, is simultaneous if issues’ outcomes are revealed to agents at the same time. It is sequential according to the order 𝒪={o1,,op}\mathcal{O}=\{o_{1},\ldots,o_{p}\} if outcomes of each issue oio_{i} are revealed to agents prior to voting on the next issue oi+1o_{i+1} Lacy and Niou (2000). We focus on simultaneous plurality voting and adapt the framework of Xia et al. (2011).

The plurality rule fi(a)f^{i}(a) applied to vote profile aa on issue ii only depends on the total number of votes for each of its candidates. We thus take the score of a candidate cDic\in D_{i} as si(c;a)=|{jn:aji=c}|s^{i}(c;a)=|\{j\leq n~{}:~{}a^{i}_{j}=c\}| and define the score tuple s(a)=(si(a))i𝒫s(a)=(s^{i}(a))_{i\in\mathcal{P}} as a collection of score vectors si(a)=(si(c;a))cDis^{i}(a)=(s^{i}(c;a))_{c\in D_{i}}. We use the plurality rule f(a)=(fi(a))i𝒫𝒟f(a)=(f^{i}(a))_{i\in\mathcal{P}}\in\mathcal{D}, where fi(a)=argmaxcDisi(c;a)f^{i}(a)=\operatorname*{arg\,max}_{c\in D_{i}}s^{i}(c;a), breaking ties lexicographically on each issue. We often use ss, s(a)s(a), and aa interchangeably for ease of notation.

Let aja_{-j} denote the vote profile without agent jj and (aj,a^j)(a_{-j},\hat{a}_{j}) the profile aa by replacing jj’s vote with the prospective vote a^j\hat{a}_{j}. Then sjs_{-j} and sj+a^js_{-j}+\hat{a}_{j} denote the corresponding adjusted score tuples.

Preferential dependence.

A preference ranking is called separable if its relative ordering of candidates in DiD_{i}, for each issue i𝒫i\in\mathcal{P}, are consistent across all outcomes of the other issues. This has the advantage that such agents may express their preferences for any issue without knowing other issues’ outcomes and avoid multiple-election paradoxes (see e.g., Xia et al. (2011)), but it is a very demanding assumption Hodge (2002). Relaxing agents preferences to be 𝒪\mathcal{O}-legal maintains representation compactness without permitting arbitrary preferential dependencies Lang and Xia (2009).

Formally, for some order 𝒪={o1,,op}\mathcal{O}=\{o_{1},\ldots,o_{p}\} over the issues, the preference ranking RR is called 𝒪\mathcal{O}-legal if, given the outcomes of the prior issues {fo1,fo2,,foi1}\{f^{o_{1}},f^{o_{2}},\ldots,f^{o_{i-1}}\}, the relative ordering of candidates in DoiD_{o_{i}} is constant for any combination of outcomes of the later issues {foi+1,,fop}\{f^{o_{i+1}},\ldots,f^{o_{p}}\}. Hence, the ranking of candidates DoiD_{o_{i}} in RR depends only, if at all, on the outcomes of issues prior to it in 𝒪\mathcal{O}.

The preference profile PP is called 𝒪\mathcal{O}-legal if every ranking is 𝒪\mathcal{O}-legal for the same order 𝒪\mathcal{O}. Thus, the ranking RR is separable if it is 𝒪\mathcal{O}-legal for any order 𝒪\mathcal{O}.

Example 1.

Let there be p=2p=2 binary issues and n=3n=3 agents with preferences P=(R1,R2,R3)P=(R_{1},R_{2},R_{3}) such that:

R1:(1,0)1(0,0)1(0,1)1(1,1)R_{1}:(1,0)\succ_{1}(0,0)\succ_{1}(0,1)\succ_{1}(1,1),

R2:(1,1)2(0,0)2(0,1)2(1,0)R_{2}:(1,1)\succ_{2}(0,0)\succ_{2}(0,1)\succ_{2}(1,0), and

R3:(0,0)3(0,1)3(1,0)3(1,1)R_{3}:(0,0)\succ_{3}(0,1)\succ_{3}(1,0)\succ_{3}(1,1).

The truthful vote profile a=((1,0),(1,1),a=((1,0),(1,1), (0,0))(0,0)) consists of each agent’s most preferred alternative. The score tuple is then s(a)={(1,2),(2,1)}s(a)=\{(1,2),(2,1)\}, so with plurality f(a)=(1,0)f(a)=(1,0).

Note that R1R_{1} is 𝒪\mathcal{O}-legal for 𝒪={2,1}\mathcal{O}=\{2,1\}: the agent always prefers 010\succ 1 on the second issue, yet their preference for the first issue depends on f2f^{2}. R3R_{3} is separable, as the agent prefers 010\succ 1 on each issue independent of the other issue’s outcome. R2R_{2} is neither separable nor 𝒪\mathcal{O}-legal for any 𝒪\mathcal{O}.

Finally, agent 22 can improve the outcome for themselves by voting for a^2=(0,1)\hat{a}_{2}=(0,1) instead of a2=(1,1)a_{2}=(1,1). The adjusted score tuple is s2={(1,1),(2,0)}s_{-2}=\{(1,1),(2,0)\}, so s2+a^2={(2,1),(2,1)}s_{-2}+\hat{a}_{2}=\{(2,1),(2,1)\} and f(s2+a^2)=(0,0)2(1,0)=f(a)f(s_{-2}+\hat{a}_{2})=(0,0)\succ_{2}(1,0)=f(a).

Improvement dynamics.

We implement the iterative voting (IV) model introduced by Meir et al. (2010) and refined for uncertainty by Meir et al. (2014), Meir (2015). Given agents’ truthful preferences PP and an initial vote profile a(0)a(0), we consider an iterative process of vote profiles a(t)=(a1(t),,an(t))a(t)=(a_{1}(t),\ldots,a_{n}(t)) that describe agents’ reported votes over time t0t\geq 0. For each round tt, a scheduler ϕ\phi chooses an agent jj to make an improvement step over their prior vote aj(t)a_{j}(t) by applying a specified response function gj:𝒟n𝒟g_{j}:\mathcal{D}^{n}\rightarrow\mathcal{D}. Each agent’s response implicitly depends on their preferences and belief about the current real vote profile, but they are not aware of others’ private preferences. All other votes besides jj’s remain unchanged.

A scheduler is broadly defined as a mapping from sequences of vote profiles to an agent with an improvement step in the latest vote profile Apt and Simon (2015). An improvement step must be selected if one exists, and a vote profile where no improvement step exists (i.e., gj(a)=ajjng_{j}(a)=a_{j}~{}\forall j\leq n) is called an equilibrium. The literature on game dynamics considers different types of response functions, schedulers, initial profiles, and other assumptions (see e.g., Fudenberg and Levine (2009), Marden et al. (2007), Bowling (2005), Young (1993)). This means that there are multiple levels in which a voting rule may guarantee convergence to an equilibrium Meir et al. (2017). In this work, we study two response functions: best response (BR), without uncertainty, and local dominance improvements (LDI), with uncertainty. For both dynamics, we restrict agents to only changing their vote on a single current issue i𝒫i\in\mathcal{P} per round, as determined by the scheduler ϕ\phi. We therefore have the following form of convergence, as described by Kukushkin (2011), Monderer and Shapley (1996), Milchtaich (1996):

Definition 1.

An IV dynamic has the restricted-finite improvement property if every improvement sequence is finite from any initial vote profile for a given response function.

Under BR dynamics, agents know the real score tuple s(a)s(a).

Definition 2 (Best response).

Given the vote profile aa, gj(a):=a^jg_{j}(a):=\hat{a}_{j} which yields agent jj’s most preferred outcome of the set {f(aj,a~j):a~jiDi,a~jk=ajkki}\{f(a_{-j},\tilde{a}_{j}):\tilde{a}^{i}_{j}\in D_{i},\tilde{a}^{k}_{j}=a^{k}_{j}~{}\forall k\neq i\}, defaulting as aja_{j} if the best outcome is the same.

LDI dynamics are based on the notions of strict uncertainty and local dominance Conitzer et al. (2011); Reijngoud and Endriss (2012). Let S×i=1p|Di|S\subseteq\times_{i=1}^{p}\mathbb{N}^{|D_{i}|} be a set of score tuples that, informally, describes agent jj’s uncertainty about the real score tuple s(a)s(a). An LDI step to a prospective vote a^j\hat{a}_{j} is (1) weakly better off than their original aja_{j} for every vSv\in S, (2) strictly better off for some tuple in vSv\in S, and (3) not worse off for any vSv\in S. This is formally defined as follows.

Definition 3.

The vote a^j\hat{a}_{j} SS-beats aja_{j} if there is at least one score tuple vSv\in S such that f(v+a^j)jf(v+aj)f(v+\hat{a}_{j})\succ_{j}f(v+a_{j}). The vote a^j\hat{a}_{j} SS-dominates aja_{j} if both (I) a^j\hat{a}_{j} SS-beats aja_{j}; and (II) aja_{j} does not SS-beat a^j\hat{a}_{j}.

Definition 4 (Local dominance improvement).

Given the vote profile aa and agent jj, let LDjiLD^{i}_{j} be the set of votes that SS-dominate aja_{j}, only differ from aja_{j} on the ithi^{th} issue, and are not themselves SS-dominated by votes differing from aja_{j} only on the ithi^{th} issue. Then gj(a)=ajg_{j}(a)=a_{j} if LDji=LD^{i}_{j}=\emptyset and a^jLDji\hat{a}_{j}\in LD^{i}_{j} otherwise.

This definition distinguishes from (weak) LDI in Meir (2015), in that agents may change their votes consecutively but only on different issues. Note that SS-dominance is transitive and antisymmetric, but not complete, so an agent jj may not have an improvement step. To fully define the model, we need to specify SS for every aa. For example, if S={s(aj)}S=\{s(a_{-j})\} and each jj has no uncertainty about the real score tuple, then LDI coincides with BR and an equilibrium coincides with Nash equilibrium. Therefore, LDI broadens BR dynamics.

Distance-based uncertainty.

Agents in the single-issue model constructed their uncertainty sets using distance-based uncertainty, in which all score vectors close enough to the current profile were believed possible Meir et al. (2014); Meir (2015). We adapt this to the multi-issue setting by assuming agents uphold candidate-wise distance-based uncertainty over score vectors for each issue independently.

For any issue i𝒫i\in\mathcal{P}, let δ(si(a),s~i(a))\delta(s^{i}(a),\tilde{s}^{i}(a)) be a distance measure for score vectors for any vote profile aa. This measure is candidate-wise if it can be written as δ(si(a),s~i(a))=maxcDiδ^(si(c;a),s~i(c;a))\delta(s^{i}(a),\tilde{s}^{i}(a))=\max_{c\in D_{i}}\hat{\delta}(s^{i}(c;a),\tilde{s}^{i}(c;a)) for some monotone function δ^\hat{\delta}. For example, the multiplicative distance (where δ^(s,s~)=max{s~/s,s/s~}1\hat{\delta}(s,\tilde{s})=\max\{\tilde{s}/s,s/\tilde{s}\}-1) and the \ell_{\infty} metric (where δ^(s,s~)=|ss~|\hat{\delta}(s,\tilde{s})=|s-\tilde{s}|) are candidate-wise.

Given the vote profile aa and issue i𝒫i\in\mathcal{P}, we model agent jj’s uncertainty about the adjusted score vector sjis^{i}_{-j} by the uncertainty score set S~ji(s;rji):={vi:δ(vi,sji)rji}\tilde{S}^{i}_{-j}(s;r^{i}_{j}):=\left\{v^{i}:\delta(v^{i},s^{i}_{-j})\leq r^{i}_{j}\right\} with an uncertainty parameter rjir^{i}_{j}. That is, given all other agents’ votes ajia_{-j}^{i}, agent jj is not sure what the real score vector is viS~ji(s;rji)v^{i}\in\tilde{S}_{-j}^{i}(s;r^{i}_{j}). We take S~j(s,rj):=×i=1pS~ji(s;rji)\tilde{S}_{-j}(s,r_{j}):=\times_{i=1}^{p}\tilde{S}^{i}_{-j}(s;r^{i}_{j}) for rj=(rji)i𝒫r_{j}=(r_{j}^{i})_{i\in\mathcal{P}}, and drop the parameters if the context is clear.

Example 2.

Let there be p=2p=2 binary issues and n=13n=13 agents. Define the vote profile aa such that s(a)={(8,5),(9,4)}s(a)=\{(8,5),(9,4)\} and aj=(0,1)a_{j}=(0,1) for some agent jj. Suppose jj uses the \ell_{\infty} uncertainty metric with (rj1,rj2)=(1,1)(r^{1}_{j},r^{2}_{j})=(1,1). Then the uncertainty score set without jj’s vote is:

S~j={(6,7,8)×(4,5,6)}×{(8,9,10)×(2,3,4)}\tilde{S}_{-j}=\{(6,7,8)\times(4,5,6)\}\times\{(8,9,10)\times(2,3,4)\}

Consider the prospective vote a^j=(1,1)\hat{a}_{j}=(1,1). Then the set of possible outcomes is {f(v+a^j):vS~j}={0,1}×{0}.\{f(v+\hat{a}_{j})~{}:~{}v\in\tilde{S}_{-j}\}=\{0,1\}\times\{0\}.

Remark.

Notice that in any vote profile reachable via BR or LDI dynamics from the truthful vote profile, no agent with fully separable preferences will have an improvements step. This follows from the definitions of BR and LDI.

3 Best-Response Dynamics

Given the vote profile aa, consider agent jj changing their vote aja_{j} on issue ii to the prospective vote a^j\hat{a}_{j}. Under BR dynamics, without uncertainty, jj changes their vote only if they can feasibly improve the outcome f(a)f(a) to one more favorable with respect to RjR_{j}. This happens under two conditions. First, jj must be pivotal on the ithi^{th} issue, meaning that changing their vote will necessarily change the outcome. Second, jj must be preferential to change ii by voting for a^ji\hat{a}^{i}_{j} over ajia^{i}_{j} given the outcomes of the other issues 𝒫\{i}\mathcal{P}\backslash\{i\}. Agent jj’s preferences are always well-defined since they know every issue’s real outcome. Thus BR dynamics behave similar to the single-issue setting, which we recall converges Meir et al. (2010). However, in the multi-issue setting, agents’ preferences on each issue may change as other issues’ outcomes change. This entails the possibility of a cycle, as declared in the following proposition and proved with the subsequent example.

Proposition 1.

BR dynamics for multiple issues may not converge, even if issues are binary.

Example 3.

Let there be p=2p=2 binary issues and n=3n=3 agents without uncertainty and the following preferences:

R1:(0,1)1(1,1)1(1,0)1(0,0)R_{1}:(0,1)\succ_{1}(1,1)\succ_{1}(1,0)\succ_{1}(0,0),

R2:(0,0)2(0,1)2(1,1)2(1,0)R_{2}:(0,0)\succ_{2}(0,1)\succ_{2}(1,1)\succ_{2}(1,0), and

R3:(1,0)3(1,1)3(0,0)3(0,1)R_{3}:(1,0)\succ_{3}(1,1)\succ_{3}(0,0)\succ_{3}(0,1).

Agent jj aj(0)a_{j}(0) aj(1)a_{j}(1) aj(2)a_{j}(2) aj(3)a_{j}(3)
11 (0,1)(0,1) (1,1)(1,1) (1,1)(1,1) (0,1)(0,1)
22 (0,0)(0,0) (0,0)(0,0) (0,1)(0,1) (0,1)(0,1)
33 (1,0)(1,0) (1,0)(1,0) (1,0)(1,0) (1,0)(1,0)
f(a)f(a) (0,0)(0,0) (1,0)(1,0) (1,1)(1,1) (0,1)(0,1)
Table 1: Agents’ votes for a(0)a(0) (truthful), a(1)a(1), a(2)a(2), and a(3)a(3).

Table 1 demonstrates a cycle generated by BR dynamics from the truthful vote profile a(0)a(0). The order of improvement steps is j=(1,2,1,2)j=(1,2,1,2). No other BR step is possible from any profile in the cycle, demonstrating that no agent scheduler can lead to convergence. Notice that agent 33 never changes their vote since their ranking is fully separable.

4 Local Dominance Improvement Dynamics

4.1 Non-Convergence

LDI dynamics broadens best-response, but it is initially unclear how increasing agents’ uncertainty parameters will affect their improvement steps and whether this eliminates the possibility of cycles. Under LDI dynamics, with uncertainty, agent jj changes their vote aja_{j} on issue ii to the prospective vote a^j\hat{a}_{j} from vote profile aa with two conditions similar to BR dynamics: if they (I) believe they may be pivotal and (II) can improve the outcome with respect to RjR_{j} (corresponding to SS-dominance). First, intuitively, it would seem that uncertainty over jj’s current issue ii affects whether they believe they can change ii’s outcome. Notice that if issue ii has a unique winning candidate and no runner-up, then jj has no best-response but may have an LDI step only if their uncertainty parameter is large enough. Second, unlike BR dynamics and the single-issue setting Meir et al. (2014); Meir (2015), jj’s preference over candidates of issue ii may depend on the outcomes of other issues, which jj may be uncertain about. It stands to reason that the more uncertainty jj has over other issues, the less clarity the agent has over their own preference for issue ii’s candidates.

Despite these two apparent parallels, we observe for multi-candidate issues that the relationship between uncertainty parameter and existence of LDI steps is not monotonic. Example 5 in Appendix B demonstrates that different sets of prospective votes LDjiLD^{i}_{j} in the same vote profile aa are non-comparable for the same agent with different uncertainty parameters. On the other hand, LDI dynamics have the aforementioned structure with binary issues, as we prove in Theorem 1. Still, this relationship has counter-acting forces: increasing uncertainty on issue ii may only add LDI steps over issue ii but may only eliminate steps on all other issues.

Theorem 1.

Given binary issues, consider agent jj changing their vote on issue ii in vote profile aa with one of three uncertainty parameters: rjr_{j}, qjq_{j}, or q^j\hat{q}_{j}. The parameters rjr_{j} and qjq_{j} only differ on issue kik\neq i such that rjk<qjkr_{j}^{k}<q_{j}^{k}; rjr_{j} and q^j\hat{q}_{j} only differ on issue ii such that rji<q^jir_{j}^{i}<\hat{q}_{j}^{i}. Let LDjiLD^{i}_{j}, LQjiLQ^{i}_{j}, and LQ^ji\hat{LQ}^{i}_{j} denote jj’s possible LDI steps according to each of the parameters. Then LQjiLDjiLQ^jiLQ_{j}^{i}\subseteq LD_{j}^{i}\subseteq\hat{LQ}^{i}_{j}.

We prove the theorem in two parts by demonstrating that if a vote a^jS~j(a;qj)\hat{a}_{j}~{}\tilde{S}_{-j}(a;q_{j})-dominates aja_{j}, then it must a^jS~j(a;rj)\hat{a}_{j}~{}\tilde{S}_{-j}(a;r_{j})-dominates aja_{j}; likewise, this implies that a^jS~j(a;q^j)\hat{a}_{j}~{}\tilde{S}_{-j}(a;\hat{q}_{j})-dominates aja_{j}. Each of these relationships arise as a result of S~j(a;rj)S~j(a;qj)\tilde{S}_{-j}(a;r_{j})\subseteq\tilde{S}_{-j}(a;q_{j}) and S~j(a;rj)S~j(a;q^j)\tilde{S}_{-j}(a;r_{j})\subseteq\tilde{S}_{-j}(a;\hat{q}_{j}). For binary issues, this is sufficient and we do not need to show that a^j\hat{a}_{j} is not S~j(a;rj)\tilde{S}_{-j}(a;r_{j})- or S~j(a;q^j)\tilde{S}_{-j}(a;\hat{q}_{j})-dominated. This may not be the case for multi-candidate issues, as previously stated in Example 5. The full proof may be found in Appendix A.

In sum, we find that enabling agents to have uncertainty over issues does not eliminate the possibility of a cycle, unlike the single-issue setting, as declared in the following proposition and proved with Example 6 in Appendix B.

Proposition 2.

LDI dynamics with multiple issues may not converge, even if agents have the same constant uncertainty parameters and issues are binary.

4.2 Strategic responses and 𝒪\mathcal{O}-legal preferences

We next present two model refinements that prove sufficient to guarantee convergence for binary issues: 𝒪\mathcal{O}-legal preferences and a specific form of dynamic uncertainty.

We are motivated by observing in Examples 3 and 6 that cycles appear due to agents’ interdependent preferences among the issues. For simplicity, consider two issues ii and kk in which no agent has an LDI step on issue ii. LDI steps on issue kk may change the set of possible winning candidates (see Definition 5), thus transforming agents’ preferences over the candidates in ii. This enables agents to have LDI steps over issue ii, even if they may no longer have improvement steps over kk. It stands to reason that eliminating interdependent preferences by restricting agents to have 𝒪\mathcal{O}-legal preference rankings would guarantee convergence.

We prove this is the case in Theorem 2. We first introduce a characterization about agents’ strategic responses, extending a lemma from Meir (2015) to the multi-issue setting.

Definition 5.

Agent jj believes a candidate cc on issue ii is a possible winner if there is some score vector where cc wins. Formally,

Wji(a)={cDi:vS~j(a;rj)s.t.fi(v+aj)=c}W^{i}_{j}(a)=\{c\in D_{i}~{}:~{}\exists v\in\tilde{S}_{-j}(a;r_{j})~{}\text{s.t.}~{}f^{i}(v+a_{j})=c\}.

In contrast, jj calls cc a potential winner if there is some score vector in which they can vote to make cc win:

Hji(a)={cDi:vS~j(a;rj)anda^js.t.a^ji=c,a^jk=ajkki,s.t.fi(v+a^j)=c}H^{i}_{j}(a)=\{c\in D_{i}~{}:~{}\exists v\in\tilde{S}_{-j}(a;r_{j})~{}\text{and}~{}\hat{a}_{j}~{}\text{s.t.}~{}\hat{a}^{i}_{j}=c,\hat{a}^{k}_{j}=a^{k}_{j}~{}\forall k\neq i,~{}\text{s.t.}~{}f^{i}(v+\hat{a}_{j})=c\}.

The set of real potential winners is denoted: H0i(a)={cDi:fi(sj+a^j)=cwherea^ji=c,a^jk=ajkki}H^{i}_{0}(a)=\{c\in D_{i}~{}:~{}f^{i}(s_{-j}+\hat{a}_{j})=c~{}\text{where}~{}\hat{a}^{i}_{j}=c,\hat{a}^{k}_{j}=a^{k}_{j}~{}\forall k\neq i\}.

By this definition, Wji(a)Hji(a)W^{i}_{j}(a)\subseteq H^{i}_{j}(a). 111 Without uncertainty, Hji(a)H^{i}_{j}(a) (or Hji(a){aji}H^{i}_{j}(a)\cup\{a^{i}_{j}\} if adding a vote to ajia^{i}_{j} makes it win) is also known as the chasing set (excluding f(a)f(a)) Rabinovich et al. (2015) or potential winner set (including f(a)f(a)) Kavner and Xia (2021) on issue ii. Hji(a)H^{i}_{j}(a) coincides with Meir et al. (2014) and Meir (2015)’s definition of a possible winner “Wj(s)W_{j}(s).” Denote by 𝒲i(a;rj)=×k𝒫\{i}Wjk(a)\mathcal{W}^{-i}(a;r_{j})=\times_{k\in\mathcal{P}\backslash\{i\}}W^{k}_{j}(a) the set of possible winning candidates on all issues besides ii, from agent jj’s perspective with uncertainty parameter rjr_{j}.

Lemma 1.

Consider an LDI step aj𝑗a^ja_{j}\xrightarrow{j}\hat{a}_{j} over issue ii from vote profile aa by agent jj with uncertainty parameter rjr_{j}. Then either (1) ajiHji(a)a^{i}_{j}\notin H^{i}_{j}(a); or for every combination of possible winners in 𝒲i(a;rj)\mathcal{W}^{-i}(a;r_{j}), either (2) ajijba^{i}_{j}\prec_{j}b for all bHji(a)b\in H^{i}_{j}(a); or (3) rji=0r^{i}_{j}=0, {aji,a^ji}H0(a)\{a^{i}_{j},\hat{a}^{i}_{j}\}\subseteq H_{0}(a) and a^jijaji\hat{a}^{i}_{j}\succ_{j}a^{i}_{j}.

The proof of this lemma directly follows that of Lemma 3 in Meir (2015); see Appendix A.

Theorem 2.

Suppose all agents have 𝒪\mathcal{O}-legal preferences for the common order 𝒪\mathcal{O} over binary issues. Then LDI dynamics converge.

Proof.

Fix an initial vote profile a(0)a(0). Suppose for contradiction that there is a cycle among the vote profiles 𝒞={a(t1),,a(tT)}\mathcal{C}=\{a(t_{1}),\ldots,a(t_{T})\}, where a(tT+1)=a(t1)a(t_{T}+1)=a(t_{1}) and a(t1)a(t_{1}) is reachable from a(0)a(0) via LDI dynamics. Let ii be the highest order issue in 𝒪\mathcal{O} for which any agent changes their vote in 𝒞\mathcal{C}.

Let t[t1,tT)t^{*}\in[t_{1},t_{T}) be the first round that some agent jj takes an LDI step on issue ii, where aj𝑗a^ja_{j}\xrightarrow{j}\hat{a}_{j} from vote profile a(t)a(t^{*}); let t(t,tT]t^{**}\in(t^{*},t_{T}] be the last round that jj switches their vote on ii back to ajia^{i}_{j}. It must be the case that ajiHji(a(t))a^{i}_{j}\in H^{i}_{j}(a(t^{*})), since issues are binary and otherwise, |Hji(a(t))|=1|H^{i}_{j}(a(t^{*}))|=1 and jj would not have an improvement step. Hence by Lemma 1, a^jijaji\hat{a}^{i}_{j}\succ_{j}a^{i}_{j} for every combination of possible winners in 𝒲i(a(t);rj)\mathcal{W}^{-i}(a(t^{*});r_{j}). Likewise, on round tt^{**}, ajija^jia^{i}_{j}\succ_{j}\hat{a}^{i}_{j} for every combination of possible winners in 𝒲i(a(t);rj)\mathcal{W}^{-i}(a(t^{**});r_{j}). Thus for some issue kk and outcomes x,y{0,1}x,y\in\{0,1\}, xyx\neq y, we have Wjk(a(t))={x}W^{k}_{j}(a(t^{*}))=\{x\} and Wjk(a(t))={y}W^{k}_{j}(a(t^{**}))=\{y\}.

Since jj has 𝒪\mathcal{O}-legal preferences, kk must be prior to issue ii in the order 𝒪\mathcal{O}. However, no agent changed their vote on issue kk between rounds tt^{*} and tt^{**} so it must be that xWjk(a(t))x\in W^{k}_{j}(a(t^{**})), even if jj’s uncertainty parameters changed. This forms a contradiction, so no such cycle can exist.

The intuition behind Theorem 2 is that as an LDI sequence develops, there is some “foremost” issue ii in which no LDI step takes place on any issue prior to ii in the order 𝒪\mathcal{O}. Agents’ relative preferences for the candidates in ii are fixed because their preferences are 𝒪\mathcal{O}-legal: score vectors for issues prior to ii in 𝒪\mathcal{O} do not change scores of issues afterward do not affect agents’ preferences for ii. Hence, agents’ improvement steps over the issue ii converge, whereas any cycle must have a sub-sequence of vote profile whose votes for issue ii cycles.

Note that 𝒪\mathcal{O}-legality is not necessary for convergence, as BR dynamics induced from the truthful vote profile in Example 1 converge. Although 𝒪\mathcal{O}-legality is a strict assumption, loosening this even slightly may lead to cycles. Example 3 demonstrates a cycle in which each agent has an 𝒪\mathcal{O}-legal ranking but orders differ between agents.

Furthermore, the theorem describes that LDI steps over the issue ii eventually terminate, thus enabling each subsequent issue in 𝒪\mathcal{O} to converge. This seems to suggest that IV under 𝒪\mathcal{O}-legal preferences is the same as truthful sequential voting, where agents vote for their preferred alternative on each issue oio_{i} given the known previous outcomes of {o1,,oi1}\{o_{1},\ldots,o_{i-1}\} Lang and Xia (2009). Although the procedures’ outcomes could be the same, there are two notable differences. First, the initial vote profile could have an issue with a single possible winning candidate that differs from the truthful sequential outcome; then no agent has an improvement step on that issue. Second, based on the scheduler, there may be a round with multiple possible winning candidates but no subsequent improvement steps on an issue. This may terminate IV with outcomes different than the truthful sequential vote.

This convergence result does not extend to the multi-candidate case, as declared in the following proposition and proved with the subsequent example.

Proposition 3.

LDI dynamics may not converge for multiple issues, even if agents have the same constant uncertainty parameters and 𝒪\mathcal{O}-legal preferences for the common order 𝒪\mathcal{O}.

Example 4.

Let there be p=2p=2 issues and n=15n=15 agents who each use the \ell_{\infty} uncertainty metric with common fixed uncertainty parameters (rj1,rj2)=(2,1)jn(r^{1}_{j},r^{2}_{j})=(2,1)~{}\forall j\leq n. Label the candidates {0,1}\{0,1\} and {a,b,c,d}\{a,b,c,d\} respectively. Agent jj has preferences: if f1=0f^{1}=0 then bjcjajdb\succ_{j}c\succ_{j}a\succ_{j}d on the second issue; otherwise cjbjajdc\succ_{j}b\succ_{j}a\succ_{j}d. Agent kk always prefers akdkbkca\succ_{k}d\succ_{k}b\succ_{k}c on the second issue. These preferences are 𝒪\mathcal{O}-legal for 𝒪={1,2}\mathcal{O}=\{1,2\}.

Define a(0)a(0) so s(a(0))={(7,8),(3,5,5,2)}s(a(0))=\{(7,8),(3,5,5,2)\} and aj(0)=ak(0)=(0,a)a_{j}(0)=a_{k}(0)=(0,a). There are four LDI steps involved in this cycle: (i) (0,a)𝑗(0,d)(0,a)\xrightarrow{j}(0,d), (ii) (0,a)𝑘(0,d)(0,a)\xrightarrow{k}(0,d), (iii) (0,d)𝑗(0,a)(0,d)\xrightarrow{j}(0,a), and (iv) (0,d)𝑘(0,a)(0,d)\xrightarrow{k}(0,a). We prove these steps are valid in Appendix B.

Note that Hj2(a(0))={a,b,c}H^{2}_{j}(a(0))=\{a,b,c\} and Hj2(a(2))={b,c,d}H^{2}_{j}(a(2))=\{b,c,d\}. In contrast to the single-issue setting (see Lemma 4 of Meir (2015)), agent jj takes LDI steps to candidates not in the potential winning set. This results from jj’s uncertainty over whether bb or cc is most-preferred, even as both are preferable to aa and dd. Hence, we get the following corollary:

Corollary 1.

LDI dynamics may not converge for plurality over a single issue for agents with partial order preferences.

4.3 Alternating Uncertainty

In Theorem 1 we found that for binary issues, agents may have fewer LDI steps over an issue ii if that issue has less uncertainty and other issues have more. This suggests that LDI steps occur from a relative lack of information about the current issue’s score vector than for other issues. If agents can gather more information about the current issue before changing their vote, thereby decreasing its uncertainty relative to other issues, then they may not have an LDI step.

We therefore consider a specific form of dynamics over agents’ uncertainty parameters where agents can gather this information and consider themselves pivotal only with respect to the lowered uncertainty. Agents are assumed to subsequently forget this relative information since it may be outdated by the time they change their vote again. We show in the following theorem that this eliminates cycles.

Definition 6.

(Alternating Uncertainty.) Fix two parameters rjcr^{c}_{j}, rjor^{o}_{j} for each agent jj such that rjc<rjor^{c}_{j}<r^{o}_{j}. Define each agent jj’s uncertainty parameters such that whenever they are scheduled to change their vote on issue ii, jj’s uncertainty for ii is rjcr^{c}_{j} and for each other issue kik\neq i is rjor^{o}_{j}.

Theorem 3.

Given binary issues, LDI dynamics converges for agents with alternating uncertainty.

Proof.

Fix an initial vote profile a(0)a(0) and uncertainty parameters rjc,rjor^{c}_{j},r^{o}_{j} for each agent jnj\leq n. Suppose for contradiction that there is a cycle among the vote profiles 𝒞={a(t1),,a(tT)}\mathcal{C}=\{a(t_{1}),\ldots,a(t_{T})\}, where a(tT+1)=a(t1)a(t_{T}+1)=a(t_{1}) and a(t1)a(t_{1}) is reachable from a(0)a(0) via LDI dynamics. Without loss of generality, suppose all issues and agents are involved in the cycle.

Consider the agent jj with the largest rjo=maxunruor^{o}_{j}=\max_{u\leq n}r^{o}_{u}. Let t[t1,tT)t^{*}\in[t_{1},t_{T}) be the first round that jj takes an LDI step on issue ii, where aj𝑗a^ja_{j}\xrightarrow{j}\hat{a}_{j} from vote profile a(t)a(t^{*}); let t(t,tT]t^{**}\in(t^{*},t_{T}] be the last round that jj switches their vote on ii back to ajia^{i}_{j}. It must be the case that ajiHji(a(t))a^{i}_{j}\in H^{i}_{j}(a(t^{*})), since issues are binary and otherwise, |Hji(a(t))|=1|H^{i}_{j}(a(t^{*}))|=1 and jj would not have an improvement step. Hence by Lemma 1, a^jijaji\hat{a}^{i}_{j}\succ_{j}a^{i}_{j} for every combination of possible winners in 𝒲i(a(t);rj)\mathcal{W}^{-i}(a(t^{*});r_{j}). Likewise, on round tt^{**}, ajija^jia^{i}_{j}\succ_{j}\hat{a}^{i}_{j} for every combination of possible winners in 𝒲i(a(t);rj)\mathcal{W}^{-i}(a(t^{**});r_{j}). Thus for some issue kk and outcomes x,y{0,1}x,y\in\{0,1\}, xyx\neq y, we have Wjk(a(t))={x}W^{k}_{j}(a(t^{*}))=\{x\} and Wjk(a(t))={y}W^{k}_{j}(a(t^{**}))=\{y\}.

Let t(t,t)t^{\prime}\in(t^{*},t^{**}) be the first round since tt^{*} that some agent hh changes their vote on issue kk. Then Hhk(a(t))={0,1}H^{k}_{h}(a(t^{\prime}))=\{0,1\}. Since Wjk(a(t))=Wjk(a(t))={x}{0,1}W^{k}_{j}(a(t^{\prime}))=W^{k}_{j}(a(t^{*}))=\{x\}\subsetneq\{0,1\} and distance functions are candidate-wise, rhcrjor^{c}_{h}\geq r^{o}_{j}. This entails rho>rjor^{o}_{h}>r^{o}_{j} by definition of alternating uncertainty, which contradicts the assertion that jj is the agent uu with the largest ruor^{o}_{u}.

This convergence result does not extend to the multi-candidate case, as Example 4 also covers this setting.

5 Experiments

Refer to caption
Figure 1: Percentage of non-equilibrium profiles as nn
increases.
Refer to caption
Figure 2: Number of profiles whose sequences cycle as nn increases.
Refer to caption
Figure 3: Log average number steps to converge as nn
increases; 95% CI (too small to show).
Refer to caption
Figure 4: Average percent change in welfare as nn
increases; 95% CI (too small to show).

Our computational experiments investigate the effects of uncertainty, number of issues, and number of agents on LDI dynamics. Specifically, we ask how often truthful vote profiles are themselves in equilibrium, how often LDI dynamics does not converge, and the path length to equilibrium when it converges. Our inquiry focuses on whether cycles are commonplace in practice even as convergence is not guaranteed.

We answer these questions for a broad cross-section of inputs, with n{7,11,15,19}n\in\{7,11,15,19\} agents, p{2,3,4,5}p\in\{2,3,4,5\} binary issues, and r{0,1,2,3}r\in\{0,1,2,3\} uncertainty that is constant for all agents, issues, and rounds. We generate 10,00010,000 preference profiles for each combination by sampling agents’ preferences uniformly and independently at random. We simulate LDI dynamics from the truthful vote profile using a scheduler that selects profiles uniformly at random from the set of valid LDI steps from all agents and issues. If there are no such steps, we say the sequence has converged. Otherwise, we take 50,00050,000 rounds as a sufficiently large stopping condition to declare the sequence has cycled.

Our results are presented in Figures 44 with respect to nn and rr for p=5p=5. We find that as uncertainty is introduced and rr increases, the availability of LDI steps diminishes significantly (Figure 4) to the point of eliminating (almost) all cycles and shortening the path length to convergence (Figure 4). Figure 4 presents the number of profiles whose LDI sequence cycles for no uncertainty, while only 55 of our sampled r1r\geq 1 profiles’ sequences cycle. Therefore, cycles with uncertainty are the exception rather than the norm.

These findings corroborate our theoretical analysis. As uncertainty increases, more issues are perceived by agents to have more than one possible winner. Since issues are interdependent for many preference rankings, fewer agents have LDI steps. On the other hand, as nn increases, more agents have rankings without these interdependencies, thus increasing the availability of LDI steps.

As an additional inquiry, we studied how IV affects the quality of outcomes by comparing the social welfare of equilibrium to truthful vote profiles.222Measured by the percent change in Borda welfare. The Borda utility of outcome aa for ranking RR is 2p2^{p} minus the index of aa’s position in RR; the Borda welfare is the sum of utilities across agents. We find in Figure 4 that IV improves average welfare, but at a rate decreasing in rr. This finding agrees with experiments by Bowman et al. (2014), Grandi et al. (2020), suggesting that IV may reduce multiple-election paradoxes by helping agents choose better outcomes. However, further work will be needed to generalize this conclusion, as it contrasts experiments of single-issue IV by Meir et al. (2020), Koolyk et al. (2017).

6 Discussion and Open Questions

We have introduced a novel model of strategic behavior in IV for multiple issues under uncertainty. We find that for binary issues, the existence of cycles hinges on the interdependence of issues in agents’ preference rankings. Specifically, once an agent jj takes an LDI step on an issue ii, they only subsequently revert their vote when their preference for ii changes. This occurs when the set of possible winning candidates of other issues, that affect jj’s preference for ii, have changed.

Without this interdependence, agents’ preference over issues change a finite number of times, so LDI dynamics converge (Theorem 2). As uncertainty increases over issues other than the one agents are changing, fewer preference rankings admit LDI steps which eliminates cycles (Theorem 3). Convergence does not extend to multi-candidate issues, as LDI plurality dynamics may cycle if agents only have partial order preference information (Corollary 1). Our experiments confirm that convergence is practically guaranteed with uncertainty, despite its possibility, and suggests IV improves agents’ social welfare over their truthful vote outcome.

There are several open directions for future work. First, our empirical study was limited by assuming agents’ preferences were sampled from the impartial culture preference distribution and had additive social welfare (see e.g., Tsetlin et al. (2003), Sen (1999)). Proving IV’s welfare properties for more realistic preference distributions and welfare functions may follow the dynamic price of anarchy work of Brânzei et al. (2013), Kavner and Xia (2021) and smoothed analysis techniques of Xia (2020).

Second, IV is useful, in part, for protecting agents’ privacy, as they do not explicitly reveal their truthful preferences. However, agents that make improvement steps implicitly reveal partial preference information. Studying IV when agents account for others’ preferences based on current information is an interesting open direction. A third direction is to detail the properties, such as anonymity and Condorcet-consistency, of multi-issue IV rules that may be inherited by the decision rule used locally on each issue, as studied for sequential voting rules by Xia et al. (2011).

References

  • Airiau and Endriss [2009] Stéphane Airiau and Ulle Endriss. Iterated majority voting. In International Conference on Algorithmic Decision Theory, pages 38–49. Springer, 2009.
  • Apt and Simon [2015] Krzysztof R Apt and Sunil Simon. A classification of weakly acyclic games. Theory and Decision, 78(4):501–524, 2015.
  • Bowling [2005] Michael Bowling. Convergence and no-regret in multiagent learning. In Proceedings of the Annual Conference on Neural Information Processing Systems (NIPS), pages 209–216, Vancouver, Canada, 2005.
  • Bowman et al. [2014] Clark Bowman, Jonathan K Hodge, and Ada Yu. The potential of iterative voting to solve the separability problem in referendum elections. Theory and decision, 77(1):111–124, 2014.
  • Brânzei et al. [2013] Simina Brânzei, Ioannis Caragiannis, Jamie Morgenstern, and Ariel D. Procaccia. How bad is selfish voting? In Proceedings of the 27th AAAI Conference on Artificial Intelligence, pages 138–144, 2013.
  • Chopra et al. [2004] Samir Chopra, Eric Pacuit, and Rohit Parikh. Knowledge-theoretic properties of strategic voting. In European Workshop on Logics in Artificial Intelligence, pages 18–30. Springer, 2004.
  • Conitzer et al. [2009] Vincent Conitzer, Jérôme Lang, and Lirong Xia. How hard is it to control sequential elections via the agenda? In Proceedings of the Twenty-First International Joint Conference on Artificial Intelligence (IJCAI), pages 103–108, Pasadena, CA, USA, 2009.
  • Conitzer et al. [2011] Vincent Conitzer, Toby Walsh, and Lirong Xia. Dominating manipulations in voting with partial information. In Proceedings of the National Conference on Artificial Intelligence (AAAI), pages 638–643, San Francisco, CA, USA, 2011.
  • Desmedt and Elkind [2010] Yvo Desmedt and Edith Elkind. Equilibria of plurality voting with abstentions. In Proceedings of the 11th ACM conference on Electronic commerce, pages 347–356, 2010.
  • Elkind et al. [2020] Edith Elkind, Umberto Grandi, Francesca Rossi, and Arkadii Slinko. Cognitive hierarchy and voting manipulation in k-approval voting. Mathematical Social Sciences, 108:193–205, 2020.
  • Endriss et al. [2016] Ulle Endriss, Svetlana Obraztsova, Maria Polukarov, and Jeffrey S. Rosenschein. Strategic voting with incomplete information. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, IJCAI’16, page 236–242. AAAI Press, 2016.
  • Fudenberg and Levine [2009] Drew Fudenberg and David K Levine. Learning and equilibrium. Annu. Rev. Econ., 1(1):385–420, 2009.
  • Grandi et al. [2013] Umberto Grandi, Andrea Loreggia, Francesca Rossi, Kristen Brent Venable, and Toby Walsh. Restricted manipulation in iterative voting: Condorcet efficiency and borda score. In International Conference on Algorithmic Decision Theory, pages 181–192. Springer, 2013.
  • Grandi et al. [2019] Umberto Grandi, Daniel Hughes, Francesca Rossi, and Arkadii Slinko. Gibbard–satterthwaite games for k-approval voting rules. Mathematical Social Sciences, 99:24–35, 2019.
  • Grandi et al. [2020] Umberto Grandi, Jérôme Lang, Ali Ozkes, and Stéphane Airiau. Voting behavior in one-shot and iterative multiple referenda. Available at SSRN, 2020.
  • Hodge [2002] Jonathan K Hodge. Separable preference orders. Western Michigan University, 2002.
  • Kavner and Xia [2021] Joshua Kavner and Lirong Xia. Strategic behavior is bliss: Iterative voting improves social welfare. In Proceedings of NeurIPS, 2021.
  • Koolyk et al. [2017] Aaron Koolyk, Tyrone Strangway, Omer Lev, and Jeffrey S. Rosenschein. Convergence and quality of iterative voting under non-scoring rules. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, IJCAI-17, pages 273–279, 2017.
  • Kukushkin [2011] Nikolai S Kukushkin. Acyclicity of improvements in finite game forms. International Journal of Game Theory, 40(1):147–177, 2011.
  • Lacy and Niou [2000] D. Lacy and E. Niou. A problem with referenda. Journal of Theoretical Politics, 12(1):5–31, 2000.
  • Lang [2007] Jérôme Lang. Vote and aggregation in combinatorial domains with structured preferences. In Proceedings of the Twentieth International Joint Conference on Artificial Intelligence (IJCAI), pages 1366–1371, Hyderabad, India, 2007.
  • Lang and Xia [2009] Jérôme Lang and Lirong Xia. Sequential composition of voting rules in multi-issue domains. Mathematical Social Sciences, 57(3):304–324, 2009.
  • Lang and Xia [2016] Jérôme Lang and Lirong Xia. Voting in Combinatorial Domains. In Felix Brandt, Vincent Conitzer, Ulle Endriss, Jérôme Lang, and Ariel Procaccia, editors, Handbook of Computational Social Choice, chapter 9. Cambridge University Press, 2016.
  • Lev and Rosenschein [2012] Omer Lev and Jeffrey S Rosenschein. Convergence of iterative voting. In Proceedings of the 11th International Conference on Autonomous Agents and Multiagent Systems-Volume 2, pages 611–618, 2012.
  • Marden et al. [2007] Jason R. Marden, Gürdal Arslan, and Jeff S. Shamma. Regret based dynamics: Convergence in weakly acyclic games. In Proceedings of the 6th International Joint Conference on Autonomous Agents and Multiagent Systems, AAMAS ’07, New York, NY, USA, 2007. Association for Computing Machinery. ISBN 9788190426275. doi: 10.1145/1329125.1329175. URL https://doi.org/10.1145/1329125.1329175.
  • Meir [2015] Reshef Meir. Plurality Voting under Uncertainty. In Proceedings of the 29th AAAI on Artificial Intelligence, 2015.
  • Meir [2018] Reshef Meir. Strategic Voting. Synthesis Lectures on Artificial Intelligence and Machine Learning. Morgan&Claypool, 2018.
  • Meir et al. [2010] Reshef Meir, Maria Polukarov, Jeffrey S. Rosenschein, and Nicholas R. Jennings. Convergence to Equilibria of Plurality Voting. In Proceedings of the National Conference on Artificial Intelligence (AAAI), pages 823–828, Atlanta, GA, USA, 2010.
  • Meir et al. [2014] Reshef Meir, Omer Lev, and Jeffrey S. Rosenschein. A Local-Dominance Theory of Voting Equilibria. In Proceedings of the 15th ACM Conference on Electronic Commerce, pages 313–330, Palo Alto, CA, USA, 2014.
  • Meir et al. [2017] Reshef Meir, Maria Polukarov, Jeffrey S Rosenschein, and Nicholas R Jennings. Iterative voting and acyclic games. Artificial Intelligence, 252:100–122, 2017.
  • Meir et al. [2020] Reshef Meir, Kobi Gal, and Maor Tal. Strategic voting in the lab: compromise and leader bias behavior. Autonomous Agents and Multi-Agent Systems, 34(1):1–37, 2020.
  • Milchtaich [1996] Igal Milchtaich. Congestion games with player-specific payoff functions. Games and economic behavior, 13(1):111–124, 1996.
  • Monderer and Shapley [1996] Dov Monderer and Lloyd S. Shapley. Potential games. Games and Economic Behavior, 14(1):124–143, 1996.
  • Myerson and Weber [1993] Roger B Myerson and Robert J Weber. A theory of voting equilibria. American Political science review, 87(1):102–114, 1993.
  • Obraztsova et al. [2015] Svetlana Obraztsova, Evangelos Markakis, Maria Polukarov, Zinovi Rabinovich, and Nicholas R. Jennings. On the convergence of iterative voting: How restrictive should restricted dynamics be? In Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, AAAI’15, page 993–999, 2015.
  • Rabinovich et al. [2015] Zinovi Rabinovich, Svetlana Obraztsova, Omer Lev, Evangelos Markakis, and Jeffrey Rosenschein. Analysis of equilibria in iterative voting schemes. In Proceedings of the 29th AAAI Conference on Artificial Intelligence. AAAI Press, 2015.
  • Reijngoud and Endriss [2012] Annemieke Reijngoud and Ulle Endriss. Voter response to iterated poll information. In Proceedings of the Eleventh International Joint Conference on Autonomous Agents and Multi-Agent Systems (AAMAS), volume 1, pages 635–644, 06 2012.
  • Reyhani and Wilson [2012] Reyhaneh Reyhani and Mark C. Wilson. Best reply dynamics for scoring rules. In Proceedings of the 20th European Conference on Artificial Intelligence, ECAI’12, page 672–677, NLD, 2012. IOS Press.
  • Sen [1999] Amartya Sen. The Possibility of Social Choice. American Economic Review, 89(3):349–378, 1999.
  • Tsetlin et al. [2003] Ilia Tsetlin, Michel Regenwetter, and Bernard Grofman. The impartial culture maximizes the probability of majority cycles. Social Choice and Welfare, 21(3):387–398, 2003.
  • Van Ditmarsch et al. [2013] Hans Van Ditmarsch, Jérôme Lang, and Abdallah Saffidine. Strategic voting and the logic of knowledge. arXiv preprint arXiv:1310.6436, 2013.
  • Xia [2020] Lirong Xia. The Smoothed Possibility of Social Choice. In Proceedings of NeurIPS, 2020.
  • Xia and Conitzer [2010] Lirong Xia and Vincent Conitzer. Stackelberg voting games: Computational aspects and paradoxes. In Proceedings of the National Conference on Artificial Intelligence (AAAI), pages 921–926, Atlanta, GA, USA, 2010.
  • Xia et al. [2011] Lirong Xia, Vincent Conitzer, and Jérôme Lang. Strategic sequential voting in multi-issue domains and multiple-election paradoxes. In Proceedings of the ACM Conference on Electronic Commerce (EC), pages 179–188, San Jose, CA, USA, 2011.
  • Young [1993] H Peyton Young. The Evolution of Conventions. Econometrica, 61(1):57–84, January 1993.
  • Zou et al. [2015] James Zou, Reshef Meir, and David Parkes. Strategic voting behavior in doodle polls. In Proceedings of the 18th ACM Conference on Computer Supported Cooperative Work & Social Computing, CSCW ’15, page 464–472, New York, NY, USA, 2015.

Appendix

Appendix A Deferred Proofs

Theorem 1.

Given binary issues, consider agent jj changing their vote on issue ii in vote profile aa with one of three uncertainty parameters: rjr_{j}, qjq_{j}, or q^j\hat{q}_{j}. The parameters rjr_{j} and qjq_{j} only differ on issue kik\neq i such that rjk<qjkr_{j}^{k}<q_{j}^{k}; rjr_{j} and q^j\hat{q}_{j} only differ on issue ii such that rji<q^jir_{j}^{i}<\hat{q}_{j}^{i}. Let LDjiLD^{i}_{j}, LQjiLQ^{i}_{j}, and LQ^ji\hat{LQ}^{i}_{j} denote jj’s possible LDI steps according to each of the parameters. Then LQjiLDjiLQ^jiLQ_{j}^{i}\subseteq LD_{j}^{i}\subseteq\hat{LQ}^{i}_{j}.

Proof.

We prove the theorem by proving a weaker claim that holds for the more general multi-candidate issues setting. Specifically, let DjiD^{i}_{j}, QjiQ^{i}_{j}, and Q^ji\hat{Q}^{i}_{j} denote the set of votes that SS-dominate aja_{j} and only differ on the ithi^{th} issue, corresponding to S=S~j(a;rj)S=\tilde{S}_{-j}(a;r_{j}), S=S~j(a;qj)S=\tilde{S}_{-j}(a;q_{j}), or S=S~j(a;q^j)S=\tilde{S}_{-j}(a;\hat{q}_{j}) respectively. We show that QjiDjiQ^jiQ^{i}_{j}\subseteq D^{i}_{j}\subseteq\hat{Q}^{i}_{j}. The theorem follows because LDji=DjiLD^{i}_{j}=D^{i}_{j} (respectively, LQji=QjiLQ^{i}_{j}=Q^{i}_{j} and LQ^ji=Q^ji\hat{LQ}^{i}_{j}=\hat{Q}^{i}_{j}) for binary issues.

(Part 1: QjiDjiQ^{i}_{j}\subseteq D^{i}_{j}.) Without loss of generality let DjiD^{i}_{j}\neq\emptyset. Suppose a^jQji\hat{a}_{j}\in Q^{i}_{j}. Then a^j\hat{a}_{j} S~j(a;qj)\tilde{S}_{-j}(a;q_{j})-dominates aja_{j}, entailing:

  1. 1.

    s~S~j(a;qj)\exists\tilde{s}\in\tilde{S}_{-j}(a;q_{j}) such that f(s~+a^j)jf(s~+aj)f(\tilde{s}+\hat{a}_{j})\succ_{j}f(\tilde{s}+a_{j}), and

  2. 2.

    s~S~j(a;qj)\nexists\tilde{s}^{\prime}\in\tilde{S}_{-j}(a;q_{j}) such that f(s~+a^j)jf(s~+aj)f(\tilde{s}^{\prime}+\hat{a}_{j})\prec_{j}f(\tilde{s}^{\prime}+a_{j}).

First, by construction of the uncertainty sets, S~jk(a;rj)S~jk(a;qj)\tilde{S}^{k}_{-j}(a;r_{j})\subseteq\tilde{S}^{k}_{-j}(a;q_{j}) and S~jh(a;rj)=S~jh(a;qj)\tilde{S}^{h}_{-j}(a;r_{j})=\tilde{S}^{h}_{-j}(a;q_{j}) for all hkh\neq k; therefore S~j(a;rj)S~j(a;qj)\tilde{S}_{-j}(a;r_{j})\subseteq\tilde{S}_{-j}(a;q_{j}). It follows that s~S~j(a;rj)\nexists\tilde{s}^{\prime}\in\tilde{S}_{-j}(a;r_{j}) such that f(s~+a^j)jf(s~+aj)f(\tilde{s}^{\prime}+\hat{a}_{j})\prec_{j}f(\tilde{s}^{\prime}+a_{j}) by point (2) above. Hence, aja_{j} does not S~j(a;rj)\tilde{S}_{-j}(a;r_{j})-beat a^j\hat{a}_{j}.

Second, define a score tuple v~\tilde{v} such that v~h=s~h\tilde{v}^{h}=\tilde{s}^{h} for each hkh\neq k and v~kS~k(a;rj)\tilde{v}^{k}\in\tilde{S}^{k}(a;r_{j}) arbitrarily. It is the case that v~S~(a;rj)\tilde{v}\in\tilde{S}(a;r_{j}) since S~jh(a;rj)=S~jh(a;qj)\tilde{S}^{h}_{-j}(a;r_{j})=\tilde{S}^{h}_{-j}(a;q_{j}) for all hkh\neq k. Since fi(s~+a^j)fi(s~+aj)f^{i}(\tilde{s}+\hat{a}_{j})\neq f^{i}(\tilde{s}+a_{j}) and fh(s~+a^j)=fh(s~+aj)f^{h}(\tilde{s}+\hat{a}_{j})=f^{h}(\tilde{s}+a_{j}) for all hih\neq i, we have f(v~+a^j)f(v~+aj)f(\tilde{v}+\hat{a}_{j})\neq f(\tilde{v}+a_{j}). It follows from the above argument that f(v~+a^j)jf(v~+aj)f(\tilde{v}+\hat{a}_{j})\succ_{j}f(\tilde{v}+a_{j}). Therefore ajS~j(a;rj)a_{j}~{}\tilde{S}_{-j}(a;r_{j})-beats a^j\hat{a}_{j} and a^jDji\hat{a}_{j}\in D^{i}_{j}.

(Part 2: DjiQ^jiD^{i}_{j}\subseteq\hat{Q}^{i}_{j}.) Without loss of generality let Q^ji\hat{Q}^{i}_{j}\neq\emptyset. Suppose a^jDji\hat{a}_{j}\in D^{i}_{j}. Then a^j\hat{a}_{j} S~j(a;rj)\tilde{S}_{-j}(a;r_{j})-dominates aja_{j}, entailing:

  1. 1.

    s~S~j(a;rj)\exists\tilde{s}\in\tilde{S}_{-j}(a;r_{j}) such that f(s~+a^j)jf(s~+aj)f(\tilde{s}+\hat{a}_{j})\succ_{j}f(\tilde{s}+a_{j}), and

  2. 2.

    s~S~j(a;rj)\nexists\tilde{s}^{\prime}\in\tilde{S}_{-j}(a;r_{j}) such that f(s~+a^j)jf(s~+aj)f(\tilde{s}^{\prime}+\hat{a}_{j})\prec_{j}f(\tilde{s}^{\prime}+a_{j}).

By construction of the uncertainty sets, S~ji(a;rj)S~ji(a;q^j)\tilde{S}^{i}_{-j}(a;r_{j})\subseteq\tilde{S}^{i}_{-j}(a;\hat{q}_{j}) and S~jh(a;rj)=S~jh(a;q^j)\tilde{S}^{h}_{-j}(a;r_{j})=\tilde{S}^{h}_{-j}(a;\hat{q}_{j}) for all hih\neq i; therefore S~j(a;rj)S~j(a;q^j)\tilde{S}_{-j}(a;r_{j})\subseteq\tilde{S}_{-j}(a;\hat{q}_{j}). It immediately follows that s~S~j(a;q^j)\tilde{s}\in\tilde{S}_{-j}(a;\hat{q}_{j}), so a^j\hat{a}_{j} S~j(a;rj)\tilde{S}_{-j}(a;r_{j})-beats aja_{j}. Furthermore, s~S~j(a;q^j)\nexists\tilde{s}^{\prime}\in\tilde{S}_{-j}(a;\hat{q}_{j}) such that f(s~+a^j)jf(s~+aj)f(\tilde{s}^{\prime}+\hat{a}_{j})\prec_{j}f(\tilde{s}^{\prime}+a_{j}) by point (2) above. Hence, aja_{j} does not S~j(a;rj)\tilde{S}_{-j}(a;r_{j})-beat a^j\hat{a}_{j}. Therefore a^jQ^ji\hat{a}_{j}\in\hat{Q}^{i}_{j}, concluding our proof.

Lemma 1.

Consider an LDI step aj𝑗a^ja_{j}\xrightarrow{j}\hat{a}_{j} over issue ii from vote profile aa by agent jj with uncertainty parameter rjr_{j}. Then either (1) ajiHji(a)a^{i}_{j}\notin H^{i}_{j}(a); or for every combination of possible winners in 𝒲i(a;rj)\mathcal{W}^{-i}(a;r_{j}), either (2) ajijba^{i}_{j}\prec_{j}b for all bHji(a)b\in H^{i}_{j}(a); or (3) rji=0r^{i}_{j}=0, {aji,a^ji}H0(a)\{a^{i}_{j},\hat{a}^{i}_{j}\}\subseteq H_{0}(a) and a^jijaji\hat{a}^{i}_{j}\succ_{j}a^{i}_{j}.

Proof.

The proof follows that of Lemma 3 in Meir [2015]. Suppose that aji,bHji(a)a^{i}_{j},b\in H^{i}_{j}(a) and ajijba^{i}_{j}\succ_{j}b whenever some combination of possible winners (c1,c2,,ci1,ci+1,,cp)𝒲i(a;rj)(c^{1},c^{2},\ldots,c^{i-1},c^{i+1},\ldots,c^{p})\in\mathcal{W}^{-i}(a;r_{j}) wins (i.e., (1) and (2) are violated). Assume first that a^jiH0(a)\hat{a}^{i}_{j}\notin H_{0}(a). By Lemma 2 of Meir [2015], s~S~j(a;rj)\exists\tilde{s}\in\tilde{S}_{-j}(a;r_{j}) such that:

  • aji,ba^{i}_{j},b have maximal score (possibly with other candidates), strictly above a^ji\hat{a}^{i}_{j}, in s~i\tilde{s}^{i};

  • for each kik\neq i, s~k\tilde{s}^{k} is such that fk(s~+aj)=ckf^{k}(\tilde{s}+a_{j})=c^{k} wins.

W.l.o.g. assume bb is prior to ajia^{i}_{j} in tie-breaking (otherwise adjust s~\tilde{s} so that fi(s~)=bf^{i}(\tilde{s})=b). Thus fi(s~+aj)=ajif^{i}(\tilde{s}+a_{j})=a^{i}_{j} while fi(s~+a^j)=bf^{i}(\tilde{s}+\hat{a}_{j})=b. Since ajijba^{i}_{j}\succ_{j}b given s~\tilde{s}, this implies ajS~j(a;rj)a_{j}~{}\tilde{S}_{-j}(a;r_{j})-beats a^j\hat{a}_{j}.

The remaining case is where a^jiH0(a)\hat{a}^{i}_{j}\in H_{0}(a) and ajija^jia^{i}_{j}\succ_{j}\hat{a}^{i}_{j} whenever some (c1,c2,,ci1,ci+1,,cp)𝒲i(a;rj)(c^{1},c^{2},\ldots,c^{i-1},c^{i+1},\ldots,c^{p})\in\mathcal{W}^{-i}(a;r_{j}) wins. Then in s~\tilde{s} where aji,a^jia^{i}_{j},\hat{a}^{i}_{j} are tied and (ck)k𝒫\{i}(c^{k})_{k\in\mathcal{P}\backslash\{i\}} wins, it is better for jj to vote for aia_{i}.

In both cases we get a^jLDji\hat{a}_{j}\notin LD^{i}_{j}, which is a contradiction. ∎

Appendix B Supplementary Examples

Example 4.

(Cont’d.) Recall the agents jj and kk with the following preferences:

  • If f1=0f^{1}=0 then jj prefers bcadb\succ c\succ a\succ d on the second issue;

  • if f1=1f^{1}=1 then jj prefers cbadc\succ b\succ a\succ d on the second issue;

  • agent kk prefers adbca\succ d\succ b\succ c on the second issue regardless of the first issue’s outcome.

Initialize the vote profile a(0)a(0) such that s(a(0))={(7,8),(3,5,5,2)}s(a(0))=\{(7,8),(3,5,5,2)\} and aj(0)=ak(0)=(0,a)a_{j}(0)=a_{k}(0)=(0,a). There are four LDI steps involved in this cycle: (i) (0,a)𝑗(0,d)(0,a)\xrightarrow{j}(0,d), (ii) (0,a)𝑘(0,d)(0,a)\xrightarrow{k}(0,d), (iii) (0,d)𝑗(0,a)(0,d)\xrightarrow{j}(0,a), and (iv) (0,d)𝑘(0,a)(0,d)\xrightarrow{k}(0,a). We will prove that these are valid LDI steps in turn, demonstrating that (i) the new vote SS-beats the old vote, (ii) the old vote does not SS-beat the new vote, and (iii) the new vote is not SS-dominated. For any candidate e{a,b,c,d}e\in\{a,b,c,d\}, denote the vote switching the second candidate to ee by e^=(0,e)\hat{e}=(0,e). Recall that ties are broken in lexicographical order.

(Step 1: s2(a(0))=(3,5,5,2)s^{2}(a(0))=(3,5,5,2).) Since rj2=1r^{2}_{j}=1 and aj2=aa^{2}_{j}=a, we have Hj2(a(0))={a,b,c}H^{2}_{j}(a(0))=\{a,b,c\}.

(i) Let s~={(6,8),(3,4,4,2)}S~j(a(0);rj)\tilde{s}=\{(6,8),(3,4,4,2)\}\in\tilde{S}_{-j}(a(0);r_{j}). Then f(s~+d^)=(1,b)j(1,a)=f(s~+a^)f(\tilde{s}+\hat{d})=(1,b)\succ_{j}(1,a)=f(\tilde{s}+\hat{a}), so d^S~j\hat{d}~{}\tilde{S}_{-j}-beats a^\hat{a}.

(ii) For any s~S~j(a(0);rj)\tilde{s}\in\tilde{S}_{-j}(a(0);r_{j}), we have that if f2(s~+d^)=af^{2}(\tilde{s}+\hat{d})=a then f2(s~+a^)=af^{2}(\tilde{s}+\hat{a})=a; otherwise, if f2(s~+a^){b,c}f^{2}(\tilde{s}+\hat{a})\in\{b,c\} then f2(s~+d^)=f2(s~+a^)f^{2}(\tilde{s}+\hat{d})=f^{2}(\tilde{s}+\hat{a}). Hence, it is never preferable for jj to vote for a^\hat{a} than d^\hat{d}.

(iii) Neither b^\hat{b} nor c^\hat{c} S~j\tilde{S}_{-j}-dominate d^\hat{d} since for s~={(6,8),(2,4,5,2)}\tilde{s}^{\prime}=\{(6,8),(2,4,5,2)\} and s~′′={(7,7),(2,4,4,2)}\tilde{s}^{\prime\prime}=\{(7,7),(2,4,4,2)\} it is preferable for jj to vote for d^\hat{d} than b^\hat{b} and c^\hat{c}, respectively.

(Step 2: s2(a(1))=(2,5,5,3)s^{2}(a(1))=(2,5,5,3).) Since rk2=1r^{2}_{k}=1 and aj2=aa^{2}_{j}=a, we have Hk2(a(1))={b,c,d}H^{2}_{k}(a(1))=\{b,c,d\}.

(i) Let s~={(6,8),(2,4,4,4)}S~k(a(1);rk)\tilde{s}=\{(6,8),(2,4,4,4)\}\in\tilde{S}_{-k}(a(1);r_{k}). Then f(s~+d^)=(1,d)k(1,b)=f(s~+a^)f(\tilde{s}+\hat{d})=(1,d)\succ_{k}(1,b)=f(\tilde{s}+\hat{a}), so d^S~k\hat{d}~{}\tilde{S}_{-k}-beats a^\hat{a}.

(ii) For any s~S~k(a(1);rk)\tilde{s}\in\tilde{S}_{-k}(a(1);r_{k}), we have that if f2(s~+d^)=df^{2}(\tilde{s}+\hat{d})=d then f2(s~+a^)=bf^{2}(\tilde{s}+\hat{a})=b; otherwise f2(s~+d^)=f2(s~+a^)f^{2}(\tilde{s}+\hat{d})=f^{2}(\tilde{s}+\hat{a}). Hence, it is never preferable for kk to vote for a^\hat{a} than d^\hat{d}.

(iii) Neither b^\hat{b} nor c^\hat{c} S~k\tilde{S}_{-k}-dominate d^\hat{d} since for s~={(6,8),(2,4,4,4)}\tilde{s}=\{(6,8),(2,4,4,4)\} (the same as in (i)), it is preferable for jj to vote for d^\hat{d} than either b^\hat{b} or c^\hat{c}.

(Step 3: s2(a(2))=(1,5,5,4)s^{2}(a(2))=(1,5,5,4).) Since rj2=1r^{2}_{j}=1 and aj2=da^{2}_{j}=d, we have Hj2(a(2))={b,c,d}H^{2}_{j}(a(2))=\{b,c,d\}.

(i) Let s~={(6,8),(1,4,4,4)}S~j(a(2);rj)\tilde{s}=\{(6,8),(1,4,4,4)\}\in\tilde{S}_{-j}(a(2);r_{j}). Then f(s~+a^)=(1,b)j(1,d)=f(s~+d^)f(\tilde{s}+\hat{a})=(1,b)\succ_{j}(1,d)=f(\tilde{s}+\hat{d}), so a^S~j\hat{a}~{}\tilde{S}_{-j}-beats d^\hat{d}.

(ii) For any s~S~j(a(2);rj)\tilde{s}\in\tilde{S}_{-j}(a(2);r_{j}), we have that if f2(s~+a^)=df^{2}(\tilde{s}+\hat{a})=d then f2(s~+d^)=df^{2}(\tilde{s}+\hat{d})=d; otherwise, if f2(s~+d^){b,c}f^{2}(\tilde{s}+\hat{d})\in\{b,c\} then f2(s~+a^)=f2(s~+d^)f^{2}(\tilde{s}+\hat{a})=f^{2}(\tilde{s}+\hat{d}). Hence, it is never preferable for jj to vote for d^\hat{d} than a^\hat{a}.

(iii) Neither b^\hat{b} nor c^\hat{c} S~j\tilde{S}_{-j}-dominate a^\hat{a} since for s~={(6,8),(2,4,5,3)}\tilde{s}^{\prime}=\{(6,8),(2,4,5,3)\} and s~′′={(7,7),(2,4,4,3)}\tilde{s}^{\prime\prime}=\{(7,7),(2,4,4,3)\} it is preferable for jj to vote for a^\hat{a} than b^\hat{b} and c^\hat{c}, respectively.

(Step 4: s2(a(3))=(2,5,5,3)s^{2}(a(3))=(2,5,5,3).) Since rk2=1r^{2}_{k}=1 and aj2=da^{2}_{j}=d, we have Hk2(a(3))={a,b,c}H^{2}_{k}(a(3))=\{a,b,c\}.

(i) Let s~={(6,8),(3,4,4,2)}S~k(a(3);rk)\tilde{s}=\{(6,8),(3,4,4,2)\}\in\tilde{S}_{-k}(a(3);r_{k}). Then f(s~+a^)=(1,a)k(1,b)=f(s~+d^)f(\tilde{s}+\hat{a})=(1,a)\succ_{k}(1,b)=f(\tilde{s}+\hat{d}), so a^S~k\hat{a}~{}\tilde{S}_{-k}-beats d^\hat{d}.

(ii) For any s~S~k(a(3);rk)\tilde{s}\in\tilde{S}_{-k}(a(3);r_{k}), we have that if f2(s~+a^)=af^{2}(\tilde{s}+\hat{a})=a then f2(s~+d^)=bf^{2}(\tilde{s}+\hat{d})=b; otherwise f2(s~+a^)=f2(s~+d^)f^{2}(\tilde{s}+\hat{a})=f^{2}(\tilde{s}+\hat{d}). Hence, it is never preferable for kk to vote for d^\hat{d} than a^\hat{a}.

(iii) Neither b^\hat{b} nor c^\hat{c} S~k\tilde{S}_{-k}-dominate a^\hat{a} since for s~={(6,8),(3,4,4,2)}\tilde{s}=\{(6,8),(3,4,4,2)\} (the same as in (i)), it is preferable for jj to vote for a^\hat{a} than either b^\hat{b} or c^\hat{c}.

Example 5.

Consider p=2p=2 issues with three candidates each, labeled {Ai,Bi,Ci,Di}\{A^{i},B^{i},C^{i},D^{i}\} for i2i\leq 2. Let agent jj have \ell_{\infty} uncertainty metric and 𝒪=(2,1)\mathcal{O}=(2,1)-legal preferences that satisfy:

  • If f2=C2f^{2}=C^{2}, then over issue 11: A1jB1jD1jC1A^{1}\succ_{j}B^{1}\succ_{j}D^{1}\succ_{j}C^{1}

  • If f2=B2f^{2}=B^{2}, then over issue 11: B1jA1jD1jC1B^{1}\succ_{j}A^{1}\succ_{j}D^{1}\succ_{j}C^{1}

  • If f2=A2f^{2}=A^{2}, then over issue 11: C1jA1jB1jD1C^{1}\succ_{j}A^{1}\succ_{j}B^{1}\succ_{j}D^{1}

Suppose aj=(D1,C2)a_{j}=(D^{1},C^{2}) in vote profile aa with score tuple s(a)={(6,6,3,5),(2,4,6,0)}s(a)=\{(6,6,3,5),(2,4,6,0)\}.

(Part 1: increasing rj1r^{1}_{j}.) This part of the example is motivated by Lemma 3b of Meir [2015].

If rj=(0,0)r_{j}=(0,0) then it is easy to see that jj has no BR steps: the unique winner is f(a)=(A1,C2)f(a)=(A^{1},C^{2}) and, although jj can make B1B^{1} win, they prefer A1B1A^{1}\succ B^{1} when f2(a)=C2f^{2}(a)=C^{2}. Hence LDj1=LD^{1}_{j}=\emptyset.

If rj=(1,0)r_{j}=(1,0) then

S~j1={(5,6,7)×(5,6,7)×(2,3,4)×(3,4,5)}\tilde{S}^{1}_{-j}=\{(5,6,7)\times(5,6,7)\times(2,3,4)\times(3,4,5)\}

First, when s~={(5,5,4,5),(2,4,6,0)}S~j\tilde{s}=\{(5,5,4,5),(2,4,6,0)\}\in\tilde{S}_{-j}, it is preferable for jj to vote for either a^j=(A1,C2)\hat{a}_{j}=(A^{1},C^{2}) or a^j=(B1,C2)\hat{a}^{\prime}_{j}=(B^{1},C^{2}) instead of aja_{j}. Recall that ties are broken by lexicographical order. Hence both a^j\hat{a}_{j} and a^jS~j\hat{a}^{\prime}_{j}~{}\tilde{S}_{-j}-beat aja_{j}. Second, it is easy to see that there is no s~S~j\tilde{s}^{\prime}\in\tilde{S}_{-j} where it is preferable for jj to vote for aja_{j} instead of either a^j\hat{a}_{j} or a^j\hat{a}^{\prime}_{j}. Thus, aja_{j} does not S~j\tilde{S}_{-j}-beat either a^j\hat{a}_{j} or a^j\hat{a}^{\prime}_{j}. Third, we notice that a^jS~\hat{a}_{j}~{}\tilde{S}-dominates a^j\hat{a}^{\prime}_{j}, which which follows from A1jB1A^{1}\succ_{j}B^{1} whenever f2(a)=C2f^{2}(a)=C^{2}. It is easy to see that a^j\hat{a}_{j} is not S~\tilde{S}-dominated. Thus LDj1={(A1,C2)}LD^{1}_{j}=\{(A^{1},C^{2})\}.

Now consider rj=(2,0)r_{j}=(2,0). Then

S~j1={\displaystyle\tilde{S}^{1}_{-j}=\{ (4,5,6,7,8)×(4,5,6,7,8)\displaystyle(4,5,6,7,8)\times(4,5,6,7,8)
×\displaystyle\times (1,2,3,4,5)×(2,3,4,5,6)}\displaystyle(1,2,3,4,5)\times(2,3,4,5,6)\}

Since s~={(5,5,3,5),(2,4,6,0)}S~j\tilde{s}=\{(5,5,3,5),(2,4,6,0)\}\in\tilde{S}_{-j}, both a^j\hat{a}_{j} and a^j\hat{a}^{\prime}_{j} still S~j\tilde{S}_{-j}-beat aja_{j}. However, as s~={(4,4,5,5),(2,4,6,0)}S~j\tilde{s}=\{(4,4,5,5),(2,4,6,0)\}\in\tilde{S}_{-j} we have f1(s~+aj)=(D1,C2)j(C1,C2)=f1(s~+a^j)=f1(s~+a^j)f^{1}(\tilde{s}+a_{j})=(D^{1},C^{2})\succ_{j}(C^{1},C^{2})=f^{1}(\tilde{s}+\hat{a}_{j})=f^{1}(\tilde{s}+\hat{a}^{\prime}_{j}), so ajS~ja_{j}~{}\tilde{S}_{-j}-beats both a^j\hat{a}_{j} and a^j\hat{a}^{\prime}_{j}. Hence, LDj1=LD^{1}_{j}=\emptyset.

(Part 2: increasing rj2r^{2}_{j}.) Recall that when rj=(1,0)r_{j}=(1,0), we found that a^j=(A1,C2)S~j\hat{a}_{j}=(A^{1},C^{2})~{}\tilde{S}_{-j}-dominated aj=(D1,C2)a_{j}=(D^{1},C^{2}), while a^j=(B1,C2)S~j\hat{a}^{\prime}_{j}=(B^{1},C^{2})~{}\tilde{S}_{-j}-dominated aja_{j} and a^jS~j\hat{a}_{j}~{}\tilde{S}_{-j}-dominated a^j\hat{a}^{\prime}_{j}; thus LDj1={(A1,C2)}LD^{1}_{j}=\{(A^{1},C^{2})\}.

If rj=(1,1)r_{j}=(1,1), then

S~j2={(1,2,3)×(3,4,5)×(4,5,6)×(0,1)}\tilde{S}^{2}_{-j}=\{(1,2,3)\times(3,4,5)\times(4,5,6)\times(0,1)\}

For s~={(5,5,4,5),(2,4,5,0)}S~j\tilde{s}=\{(5,5,4,5),(2,4,5,0)\}\in\tilde{S}_{-j} it is preferable for jj to vote for a^j\hat{a}_{j} rather than a^j\hat{a}^{\prime}_{j}, whereas for s~={(5,5,4,5),(2,5,4,0)}S~j\tilde{s}=\{(5,5,4,5),(2,5,4,0)\}\in\tilde{S}_{-j} it is preferable for jj to vote for a^j\hat{a}^{\prime}_{j} rather than a^j\hat{a}_{j}. Then both a^j\hat{a}_{j} and a^j\hat{a}^{\prime}_{j} S~j\tilde{S}_{-j}-dominate aja_{j} but neither are S~j\tilde{S}_{-j}-dominated. Therefore LDj1={(A1,C2),(B1,C2)}LD^{1}_{j}=\{(A^{1},C^{2}),(B^{1},C^{2})\}.

If rj=(1,2)r_{j}=(1,2), then

S~j2={\displaystyle\tilde{S}^{2}_{-j}=\{ (0,1,2,3,4)×(2,3,4,5,6)\displaystyle(0,1,2,3,4)\times(2,3,4,5,6)
×\displaystyle\times (3,4,5,6,7)×(0,1,2)}\displaystyle(3,4,5,6,7)\times(0,1,2)\}

Then there are score vectors s~S~j\tilde{s}\in\tilde{S}_{-j} where f2(s~+aj)=f2(s~+a^j)=f2(s~+a^j)=C2f^{2}(\tilde{s}+a_{j})=f^{2}(\tilde{s}+\hat{a}_{j})=f^{2}(\tilde{s}+\hat{a}^{\prime}_{j})=C^{2}, so that neither a^j\hat{a}_{j} nor a^jS~j\hat{a}^{\prime}_{j}~{}\tilde{S}_{-j}-dominate aja_{j}. Thus LDj1=LD^{1}_{j}=\emptyset.

Example 6.

Let there be p=2p=2 binary issues and n=13n=13 agents who each use the \ell_{\infty} uncertainty metric with common fixed uncertainty parameters (rj1,rj2)=(1,2)jn(r^{1}_{j},r^{2}_{j})=(1,2)~{}\forall j\leq n. Suppose that agents’ preferences abide by the following four types:

  • (Type 1) three agents have rankings: (0,1)(1,1)(1,0)(0,0)(0,1)\succ(1,1)\succ(1,0)\succ(0,0);

  • (Type 2) five agents have rankings: (0,0)(0,1)(1,1)(1,0)(0,0)\succ(0,1)\succ(1,1)\succ(1,0);

  • (Type 3) four agents have rankings: (1,0)(1,1)(0,0)(0,1)(1,0)\succ(1,1)\succ(0,0)\succ(0,1); and

  • (Type 4) one agent has ranking: (1,1)(1,0)(0,1)(0,0)(1,1)\succ(1,0)\succ(0,1)\succ(0,0).

There is a cycle passing through the four vote profiles a(0)a(0) (which is truthful), a(3)a(3), a(8)a(8), and a(11)a(11) listed in Table 2, in which every agent of the same type has the same vote. There are four parts of the cycle between these profiles:

  • between a(0)a(0) and a(3)a(3), all agents of Type 1 make LDI steps on the first issue (0,1)1(1,1)(0,1)\xrightarrow{1}(1,1);

  • between a(3)a(3) and a(8)a(8), all agents of Type 2 make LDI steps on the second issue (0,0)2(0,1)(0,0)\xrightarrow{2}(0,1);

  • between a(8)a(8) and a(11)a(11), all agents of Type 1 make LDI steps on the first issue (1,1)1(0,1)(1,1)\xrightarrow{1}(0,1);

  • between a(11)a(11) and a(16)=a(0)a(16)=a(0), all agents of Type 2 make LDI steps on the second issue (0,1)2(0,0)(0,1)\xrightarrow{2}(0,0).

Notice that no agent of Types 3 and 4 make LDI steps since they are voting truthfully and have separable preferences.

Agent Type jj aj(0)a_{j}(0) aj(3)a_{j}(3) aj(8)a_{j}(8) aj(11)a_{j}(11)
11 (0,1)(0,1) (1,1)(1,1) (1,1)(1,1) (0,1)(0,1)
22 (0,0)(0,0) (0,0)(0,0) (0,1)(0,1) (0,1)(0,1)
33 (1,0)(1,0) (1,0)(1,0) (1,0)(1,0) (1,0)(1,0)
44 (1,1)(1,1) (1,1)(1,1) (1,1)(1,1) (1,1)(1,1)
Table 2: Agents’ votes for a(0)a(0) (truthful), a(3)a(3), a(8)a(8), and a(11)a(11).

We claim these are valid LDI steps as follows. For any vote profile in {a(0),a(1),a(2)}\{a(0),a(1),a(2)\}, let tt be the number of Type 1 agents who have made LDI steps. Then s(a(t))={(8t,5+t),(9,4)}s(a(t))=\{(8-t,5+t),(9,4)\} and for any Type 1 agent jj who has not made their first LDI step yet,

S~j1={(6t,7t,8t)×(4+t,5+t,6+t)}\tilde{S}^{1}_{-j}=\{(6-t,7-t,8-t)\times(4+t,5+t,6+t)\}

S~j2={(7,8,9,10,11)×(2,3,4,5,6)}\tilde{S}^{2}_{-j}=\{(7,8,9,10,11)\times(2,3,4,5,6)\}

Define a^j=(1,1)\hat{a}_{j}=(1,1). Then s~={(6,6),(9,4)S~j(a(t);rj)t{0,1,2}\tilde{s}=\{(6,6),(9,4)\in\tilde{S}_{-j}(a(t);r_{j})~{}\forall t\in\{0,1,2\}, so f(s~+a^j)=(1,0)j(0,0)=f(s~+aj)f(\tilde{s}+\hat{a}_{j})=(1,0)\succ_{j}(0,0)=f(\tilde{s}+a_{j}) and a^jS~j\hat{a}_{j}~{}\tilde{S}_{-j}-beats aja_{j}. It is easy to see that aja_{j} does not S~j\tilde{S}_{-j}-beat a^\hat{a} and that a^\hat{a} is not S~j\tilde{S}_{-j}-dominated since issues are binary. Thus LDj1={a^j}LD^{1}_{j}=\{\hat{a}_{j}\}.

The other LDI steps follow similar reasoning, yielding the cycle presented in the table. It can be verified that this represents all possible LDI sequences from the truthful vote profile.

Appendix C Nonatomic model

Each of the results presented for binary issues and atomic agents, where each agent contributes one unit of influence to the population of nn votes, also generalize to a nonatomic variant of IV, in which agents are part of a very large population and have negligible influence over the outcome. Our model extends Meir [2015]’s iterative plurality voting for nonatomic agents to the multi-issue setting. Like this setting, our convergence results permit arbitrary subsets of agents to change their vote simultaneously. This differs from the finite case which may not converge if multiple agents change their votes simultaneously Meir et al. [2010].

In this section, we provide the necessary definitions for the nonatomic model, using our existing notation wherever possible. There are two major differences: (i) with the identify of an agent, and (ii) with the uncertainty score set. First, rather than treating each agent jnj\leq n individually, there are groups of identical agents with ϵ\epsilon mass; “an agent of type jj” then refers to a representative agent in this group. Second, as agents have negligible influence, type jj agents consider any score tuple in S~(a;rj)\tilde{S}(a;r_{j}) possible; all agents agree of the same type agree what real score tuples are possible. Lemma 1 and Theorems 1, 2, and 3 are each upheld after applying this model redefinition. Similar non-convergence results also apply for the nonatomic variant of IV.

Basic notations.

We do not have a finite set of agents. Rather, a preference profile QΔ((D))Q\in\Delta(\mathcal{L}(D)) is a distribution over rankings that specifies the fraction of agents Q(R)Q(R) for each R(D)R\in\mathcal{L}(D). We only consider moves by subsets of agents since each agent has negligible influence. To avoid infinite improvement sequence paths of sequentially smaller subsets of agents, we assume there is a minimum resolution ϵ\epsilon, such that sets of agents of the same type with mass ϵ\epsilon always move together (although in an uncoordinated manner; see Appendix IX of Meir [2015]). We denote the collection of these 1/ϵ1/\epsilon sets by JJ. Since all agents in set jJj\in J are indistinguishable, we refer to “agent jj” as an arbitrary agent in the set jj. Then Rj(D)R_{j}\in\mathcal{L}(D) is the preference, rjr_{j} is the uncertainty parameter, and aja_{j} is the vote of an arbitrary agent in the set of vote profile aa.

Winner determination.

For any issue i𝒫i\in\mathcal{P}, we define the score vector si(a)=(si(c;a))cDis^{i}(a)=(s^{i}(c;a))_{c\in D_{i}} induced by the vote profile aa such that si(c;a)=|{f:aji=c}|ϵ[0,1]s^{i}(c;a)=|\{f:a^{i}_{j}=c\}|\epsilon\in[0,1]. Winner determination is exactly as in the atomic model with lexicographical tie-breaking.

Uncertainty and local dominance.

We assume agents utilize the same candidate-wise distance uncertainty as the atomic model. Like the atomic model, each agent of type jJj\in J selected by the scheduler to change their vote has uncertainty parameters rj=(rji)i𝒫r_{j}=(r^{i}_{j})_{i\in\mathcal{P}}. Unlike the atomic model, agents have negligible influence in the score vector. Hence, they take their uncertainty score sets with respect to the real score tuple: S~i(a;rji)={vi:δ(vi,si(a))rji}\tilde{S}^{i}(a;r^{i}_{j})=\{v^{i}:\delta(v^{i},s^{i}(a))\leq r^{i}_{j}\}. As before, S~(a;rj)=×i𝒫S~i(a;rji)\tilde{S}(a;r_{j})=\times_{i\in\mathcal{P}}\tilde{S}^{i}(a;r^{i}_{j}).

Given vS~(a;rj)v\in\tilde{S}(a;r_{j}), we define f(v+aj)f(v+a_{j}) to be the outcome an agent of type jj expects by voting for aja_{j} according to score vector ss. That is, for each issue ii, the extra vote ajia^{i}_{j} decides the winner if several candidates are tied with maximal score in ss, overriding the default tie-breaker (see Appendix X of Meir [2015]).

The definition of a local dominance improvement (LDI) step for a nonatomic agent of type jJj\in J from vote aja_{j} to a^j\hat{a}_{j} on issue ii then is the same as in the atomic model of Definitions 3 and 4, applying this redefinition and using S~(a;rj)\tilde{S}(a;r_{j}) in lieu of S~j(a;rj)\tilde{S}_{-j}(a;r_{j}). The response function gjg_{j} is also the same, except that its domain (all possible profiles) is now 𝒟|J|\mathcal{D}^{|J|} rather than 𝒟n\mathcal{D}^{n}. The definition of equilibrium does not change.