Optimality Gap of Decentralized Submodular Maximization under Probabilistic Communication
Abstract
This paper considers the problem of decentralized submodular maximization subject to partition matroid constraint using a sequential greedy algorithm with probabilistic inter-agent message-passing. We propose a communication-aware framework where the probability of successful communication between connected devices is considered. Our analysis introduces the notion of the probabilistic optimality gap, highlighting its potential influence on determining the message-passing sequence based on the agent’s broadcast reliability and strategic decisions regarding agents that can broadcast their messages multiple times in a resource-limited environment. This work not only contributes theoretical insights but also has practical implications for designing and analyzing decentralized systems in uncertain communication environments. A numerical example demonstrates the impact of our results.
Keywords: Decentralized Submodular Maximization; Probabilistic Message-Passing; Optimality Gap Analysis.
I Introduction
This paper considers the problem of decentralized submodular maximization subject to a partition matroid when agents implement a sequential greedy algorithm with probabilistic inter-agent communication. We study a network of agents, denoted as , each equipped with communication and computation capabilities and interconnected via a directed chain graph (see Fig. 1). The goal for each agent is to select up to strategies from its local distinct discrete strategy set in way that a normal monotone increasing and submodular utility function , , evaluated at collective strategy set of the agents is maximized. The optimization problem is expressed as
(1a) | |||
(1b) |
where is a partition matroid, restricting the number of strategies each agent can choose to . The agents access the utility function via a black box that returns for any set (value oracle model).
Submodular function maximization problems such as (1) are often NP-hard [1]. However, submodularity, a property of set functions with deep theoretical implications, enables the establishment of constant factor approximate (suboptimal) solutions for submodular maximization problems. A fundamental result by Nemhauser et al. [1] establishes that the simple sequential greedy algorithm shown in Algorithm 1 is guaranteed to provide a constant -approximation factor solution for the submodular maximization problem (1).

Submodular maximization problems subject to matroid constraints appear in many data-centric problems, such as agglomerative clustering, exemplar-based clustering [2], categorical feature compression [3], and data subset selection [4]. To alleviate the burden on central sequential greedy solvers handling massive data, various parallel computing methods such as MapReduce-based solvers or federated learning frameworks [5] have been proposed.
In the domain of cyber-physical systems, submodular maximization problems of the form (1) appear in various optimal resource allocation and scheduling tasks, such as optimal sensor placement [6], energy storage placement [7, 8], measurement scheduling [9], voltage control in smart grids [10], and persistent monitoring via mobile robots [11]. These applications often require decentralized solutions, where agents achieve a global optimal solution through local computations and inter-agent interactions/communication.
Distributed solutions based on continuous relaxations [12, 13] are proposed in the literature to solve problem (1) over graph integration topologies, however, they often come with significant communication and computation costs. The sequential decision-making structure of Algorithm 1 lends itself naturally to decentralized implementation over a directed chain graph using sequential message-passing, as described in Algorithm 2 [11]. Starting with the first agent in the chain, each agent, after receiving —the compound choices of the preceding agents—from its in-neighbor, runs a local sequential greedy algorithm over its local dataset to make its choices considering the choices already made. After determining its local selections and incorporating them into , the agent forwards this updated set to its out-neighbor. This method of sequential message-passing can also be facilitated through cloud access.
In practice, message delivery reliability between an in-neighbor and its out-neighbor is not always assured. The effect of delivery failures on a decentralized sequential greedy algorithm has been explored in [14]; refer to Section II for more information. This research examines how a specific message passing sequence impacts the optimality gap from a deterministic point of view. However, practical message delivery success is subject to unpredictable factors such as communication channel strength, agent reliability, and network congestion. The impact of probabilistic communication on convergence in various optimization problems is well documented in the literature, such as in [15, 16]. These studies offer theoretical insights and have practical significance for the design and analysis of decentralized systems in uncertain communication settings. Considering these uncertainties, this paper addresses the pivotal research question of how probabilistic message-passing affects the guaranteed optimality gap of a decentralized sequential greedy algorithm. We frame this challenge as the following problem statement.
Problem 1 (Probabilistic Sequential Message-Passing)
The main objective of this paper is to investigate and address Problem 1, focusing on characterizing the probabilistic optimality gap, . Understanding is crucial for decentralized sequential greedy algorithms operating under probabilistic message-passing conditions. This work not only contributes theoretical insights but also has practical implications for designing and analyzing decentralized systems in uncertain communication environments.
Notation and definitions: For a discrete ground set , is its power set, the set that contains all the subsets of . A set function is submodular if and only if for any and for all we have that
(3) |
Function is normal if and is monotone increasing if for any we have . For any and any , is the marginal gain of adding to the set .
II Sequential greedy under probabilistic inter-agent message-passing
To set the stage to address Problem 1, it is pivotal to first examine the consequences of failed message deliveries and the ensuing disruption in the flow of information during the algorithm’s operation. For this purpose, we use two specific graph topologies: the communication graph , which describes the sequence of message transfers, and the information graph , which describes the dissemination of information as a result of the message-passing process; see Fig. 2. In the information graph, an arrow from agent to agent indicates that agent has successfully received information from agent , either directly or through message-passing by other agents preceding agent .
![]() |
![]() |
(a) Information graph when the chain is fully connected. | (b) Information graph when the chain is disconnected. |
When implementing Algorithm 2, a fully successful message transmission sequence, depicted in Fig. 2(a), equates to the formation of a complete graph in the undirected version of the communication graph . Conversely, interruptions in the message-passing path, as shown in Fig. 2(b), result in an undirected version of that lacks full connectivity. To evaluate the degree to which approaches a complete graph configuration, we employ the clique number of the information graph, denoted by , as a metric. This measure, indicative of the size of the largest clique (complete sub-graph) within the graph, serves to quantify the graph’s connectivity level. In the context of Algorithm 2, a complete yields a clique number equal to . Interruptions in message passage, however, diminish this number, signifying reduced information connectivity. One can expect that the larger the clique number, indicating a larger fully connected component in the information graph, the better the optimality gap should be for the sequential greedy algorithm. This hypothesis has been formally validated in [14], which demonstrates that the optimality gap of a sequential greedy algorithm tackling problem (1)–under the condition that agents’ policy sets do not overlap–can be effectively quantified as follows.
Lemma II.1
This bound recovers the well-known optimality gap of when the information graph is complete, i.e., .
When message-passing success is probabilistic, both the outcome of the algorithm’s execution and the clique number of the information graph become random variables. To deduce the expected optimality gap, we propose employing a decision tree representation to account for all possible message-passing outcomes. The upper diagram in Fig. 3 shows this decision tree for the network of five agents shown in Fig. 2, where the red dashed lines represent unsuccessful messages and the black arrows denote successful ones. The lower diagram in Fig. 3 displays the associated clique number for each message-passing sequence outcome. For instance, the leftmost path on the decision tree, indicating a scenario where no messages are successfully passed, results in a clique number of one. In contrast, the rightmost path, which signifies a scenario in which all messages are successfully delivered, resulting in . With a comprehensive overview of all possible message-passing scenarios,the expected optimality gap of Algorithm 2 can be accurately characterized as follows.
Theorem II.1
Proof:
Since is the expected probabilistic optimality gap, we calculate it by summing over all the possible optimality gaps computed from Lemma II.1 based on the clique numbers within the range from to , each weighted by its probability . ∎

As shown in Fig. 3, a probabilistic communication chain of length has possible message-passing sequences. Denote the collection of all such sequence outcomes as , and the random variable associated with any given sequence outcome as . The probability of each can be computed from a combination of Bernoulli distributions, each representing a possible connection between agents, that is, . Consequently, computing the clique number and associated probability for each sequence individually, will lead to an exponential time computation to determine . However, as seen in Fig. 3, multiple sequence outcomes can have the same clique number. In what follows, we propose a methodology that groups sequences based on their underlying structure. Specifically, to determine the probability of each clique number, we analyze the structure common to sequences that result in the same clique numbers and derive an exact formula for these probabilities. Our approach reduces the complexity of computing to polynomial time. The critical information in our study is that the clique number is defined by the length of the longest directed path in the message-passing graph, plus one. We start with the following result. In this result and what follows, to simplify the notation, we write as .
Lemma II.2
(Probability of Achieving a Specific Clique Number in Probabilistic Sequential Message-Passing): The probability that message-passing sequences successfully achieve a clique number of is given by
Proof:
represents the probability that message-passing sequences achieve a clique number that is at least . These sequences, which satisfy , include those with exact clique numbers ranging from to , and these sequences are mutually exclusive. Thus, we express
Similarly, . The proof then follows by deducting from . ∎
In what follows, we explain how to compute to fully describe from (4). is equivalent to determining the probability of a family of message-passing sequences achieving a maximum connected component length of . To this end, we introduce some essential definitions. A family of message-passing sequences, denoted by , refers to a group of sequences that exhibit a specific pattern of connectivity. Thus, a family of message-passing sequences is an event associated with the random variable . With a slight abuse of terminology, we define a generative sequence as the guideline in which a connectivity structure is imposed on some of the edges of the massage-passing sequence, e.g., the first two edges are connected, but the fourth is not. Thus, a generative sequence becomes basis for a family of sequences in which all members have the first two edges connected but the fourth not. We denote a generative sequence by , where , respectively, specify the set of connected edges and disconnected edges of the generative sequence. For examples of generative sequences, see Fig. 4. We let to denote the family generated by .
![]() |
(a) Generative sequences of size . |
![]() |
(b) Generative sequences of size . |
Lemma II.3 (Probability of a family)
Given a massage passing sequence of agents with the corresponding probability of successful message-passing of , the probability of family generated by is
(5) |
Lemma II.3’s proof follows trivially by writing all the outcomes generated by and adding their probabilities.
We call two generative sequences independent when the families they initiate are distinct, in the sense that there exists no that is a member of the families generated by these generative sequences. For example, in top plot of Fig. 4 generative sequences A and B, and A and C are independent, but generative sequences A and D are not independent.
Next, we walk through the process of creating a family of message-passing sequences that feature connected components with a minimum length of . For a specified , we consider the generative sequences , where , with representing the pre-specified connected components, and indicating the pre-specified disconnected components, noting that . These generative sequences each contribute to the creation of families of message-passing sequences that ensure the presence of connected components at least units long. Collectively, these sequences constitute , capturing the full set of message-passing sequences with connected components of at least in length, as depicted in Fig. 4 through examples involving a chain of agents. It is noteworthy that
(6) |
Given the construction where for any and , as illustrated in Fig. 4, the families generated by for are independent and thus mutually exclusive events. Therefore,
However, families from to have dependencies that we discuss next. With the same statement used for , a family with will be dependent with families with as . Therefore, we are seeking for the independent set of families in which can defined as the sequences in that do not belong to previous families , in set notation . Note that by invoking a general probability rule 111For two independent events and , we have [17]. we obtain
(7) |
As a generalization for the previous expression, for the sake of understanding, let us consider the families that creates dependency to , , such as . Then, notice that
To obtain a close expression for (6) it is needed to find the independent expression for each family. By a simple set operation it can be seen that
where is the independent expressions for families . Therefore
(8) |
Each is computed from (5). Notice that equation (8) can be computed in oracle calls to (5) and there is no need of computing all of the the possible sequences. With equation (8) at hand, from Lemma II.2 we arrive at
(9) |
Now, using expression (9), the proposed optimality gap in Theorem II.1 can be easily computed in polynomial time. That not only allows the computation of the gap itself, but creates the possibility of pursuing an analysis of how the algorithm would perform in the worst-case scenario as a function of the parameters of the system.
III Extra resources assignment effect on the probabilistic gap
![]() |
![]() |
![]() |
By allowing repeated communication, we can increase the probability of message delivery and subsequently improve the optimality gap. However, multiple message-passing increases the end-to-end delay of executing the algorithm and may not be possible in resource limited scenarios. Therefore, the natural research question arises pertaining to the allocation of a finite quantity of multiple message-passing rights in an optimal manner.
Problem 2 (multiple message-passing right allocation)
The key observation is that the best agent to reinforce is not solely the one with the lowest communication probability but a combination of communication reliability and the agent’s position in the communication chain. This is demonstrated in our empirical study in Section IV.
The message-passing process from agent to agent follows a Bernoulli distribution
where is the probability of agent’s successful delivery in one communication, is the number of trials (messages sent) and is the number of successes (messages delivered). Since communication between two nodes is established when at least one success occurs, the probability that the recipient gets the message is
(10) |
Note that is the new communication probability from agent to agent , for the sake of understanding, we will keep referring to them as . By considering a Bernoulli measure, expression in Lemma II.3 can be re-defined as
enabling us to analyze the behavior of message-passing reinforcement for . Note that when the message-passing count for a device increases, the expected optimality gap needs to be recalculated. As previously mentioned, for each , calculating its probability requires a time complexity of . Consequently, the gap can be determined within , indicating that optimizing the communication strategy is a polynomial-time process. Further examination of (9) reveals that doubling the message-passing trail for agent does not influence families where . Thus, by equipping the system with a certain memory capacity, computation time of for all possible reinforcements can be significantly reduced.
IV Empirical study of the probabilistic gap
Consider a multi-sensor deployment problem where a group of heterogeneous agents , each with two sensors to deploy at some prespecified deployment points of interest. There are total locations where sensors can be deployed, but each agent has access to only a subset of locations, denoted by . The sets are not distinct, but using a simple trick, we can create the local selection set of the agents in a distinct way as for . The environment includes a set of randomly generated sampled points, denoted by . The sensor deployment objective is to cover as many points from as possible, which is accomplished by solving a maximization problem of the form (1) with and the utility function given by , where if there exists at least one element such that ; otherwise . Here, is the coverage radius of sensor . This utility function is known to be submodular and monotone increasing. This problem is illustrated in Fig. 5.

For visualization of the communication chain, we label the agents alphabetically. Now consider two different sequences of communication: lexicographic order ABCDEFGH and a random shuffle DBHGFCAE, as shown in Fig. 6. Note that with agents, there are up to possible communication chains that can be established. In this section, we consider only two sequences to illustrate the particular behavior under the allowance for extra message-passing. It is important to note that due to the probabilistic nature of the problem, the results presented in this section are the mean values from 10,000 iterations for each scenario.
Table I shows the results of our study for the communication chains illustrated in Fig. 6. The second column presents the average utility value over 10,000 iterations when agents use Algorithm 2 to determine their deployment locations. The third column shows computed from (4). The fourth column identifies the agent whose communication is reinforced by allowing two back-to-back communication trials. The agent to be reinforced is determined by computing for all possible reinforcements, with these values shown in Table II. The fifth column indicates the edge label that is reinforced (the link coming out of ). Finally, the sixth and seventh columns present the average utility value and computed from (4) under communication reinforcement. As expected, we observe that , indicating that reinforcement improves the expected optimality gap. This improvement is also reflected in the utility values. Another noteworthy observation is that the reinforced agent is not necessarily the one with the lowest communication probability, and the reinforced edge differs between sequences.
Sequence | ||||||
---|---|---|---|---|---|---|
ABCDEFGH | 1324.21 | 0.2692 | F | 6 | 1610.88 | 0.3243 |
DBHGFCAE | 1554.54 | 0.2788 | C | 4 | 1749.21 | 0.3313 |
Sequence | 1 | 2 | 3 | 4 | 5 | 6 | 7 |
---|---|---|---|---|---|---|---|
ABCDEFGH | 0.2974 | 0.3192 | 0.3006 | 0.3012 | 0.3218 | 0.3243 | 0.3030 |
DBHGFCAE | 0.2998 | 0.3286 | 0.3028 | 0.3313 | 0.3166 | 0.3218 | 0.3004 |
V Conclusions
This paper addressed the problem of decentralized submodular maximization subject to partition matroid constraint using a sequential greedy algorithm with probabilistic inter-agent message-passing. We proposed a communication-aware framework that considers the reliability of communication between connected devices, emphasizing its impact on the optimality gap. Our analysis introduced the notion of the probabilistic optimality gap , highlighting its crucial role in understanding the expected performance of the sequential greedy algorithm under probabilistic communication. By characterizing as an explicit function of communication probabilities, we created a framework to answer critical questions such as which agent should be reinforced (allowed multiple communications) to improve the optimality gap or how to compare the optimality gaps of various communication chains, where the order of agents in the chain changes while their reliability remains the same. In our empirical study, we specifically focused on the case where only one agent is allowed to communicate twice. However, the methodology we presented is generalizable and can be extended to scenarios where multiple agents are allowed multiple message-passing opportunities. This extension can be framed as a set function maximization problem subject to a uniform matroid constraint, with as the utility function. What agents to reinforce can be efficiently decided using a sequential greedy approach, given that is monotone increasing and, if shown to be submodular, would benefit from the known optimality gap of . Future work will focus on formally proving the submodularity of and developing strategies to optimize communication reinforcement in more complex scenarios involving multiple agents. We aim also to leverage insights from this study to design more robust and efficient decentralized systems for real-world applications, particularly in environments where communication reliability is variable and resources are limited.
References
- [1] G. L. Nemhauser, L. A. Wolsey, and M. L. Fisher, “An analysis of approximations for maximizing submodular set functions—i,” Mathematical Programming, vol. 14, no. 1, pp. 265–294, 1978.
- [2] P.-J. Honysz, A. Schulze-Struchtrup, S. Buschjäger, and K. Morik, “Providing meaningful data summarizations using examplar-based clustering in industry 4.0,” ArXiv, vol. abs/2105.12026, 2021.
- [3] A. Rostamizadeh, H. Esfandiari, L. Chen, MohammadHossein, Bateni, T. Fu, and V. Mirrokni, “Categorical feature compression via submodular optimization,” in icml, pp. 515–523, 2019.
- [4] K. Wei, R. Iyer, and J. Bilmes, “Max-sum diversification, monotone submodular functions and dynamic updates,” in icml, vol. 37, pp. 1954–1963, 2015.
- [5] A. Rafiey, “Decomposable submodular maximization in federated setting,” ArXiv, vol. abs/2402.00138, 2024.
- [6] N. Mehr and R. Horowitz, “A submodular approach for optimal sensor placement in traffic networks,” in American Control Conference, pp. 6353–6358, 2018.
- [7] J. Qin, I. Yang, and R. Rajagopal, “Submodularity of storage placement optimization in power networks,” IEEE Tran. on Automatic Control, vol. 64, no. 8, pp. 3268–3283, 2019.
- [8] M. Bucciarelli, S. Paoletti, E. Dall’Anese, and A. Vicino, “On the greedy placement of energy storage systems in distribution grids,” in American Control Conference, 2020.
- [9] S. T. Jawaid and S. Smith, “Submodularity and greedy algorithms in sensor scheduling for linear dynamical systems,” Automatica, vol. 61, pp. 282–288, 2015.
- [10] Z. Liu, A. Clark, P. Lee, L. Bushnell, D. Kirschen, and R. Poovendran, “Towards scalable voltage control in smart grid: A submodular optimization approach,” in ACM/IEEE 7th International Conference on Cyber-Physical Systems, 2016.
- [11] N. Rezazadeh and S. S. Kia, “A sub-modular receding horizon solution for mobile multi-agent persistent monitoring,” Automatica, vol. 127, p. 109460, 2021.
- [12] A. Robey, A. Adibi, B. Schlotfeldt, G. Pappas, and H. Hassani, “Optimal algorithms for submodular maximization with distributed constraints,” ArXiv, vol. abs/1909.13676, 2019.
- [13] N. Rezazadeh and S. S. Kia, “Distributed strategy selection: A submodular set function maximization approach,” Automatica, vol. 153, p. 111000, 2023.
- [14] B. Gharesifard and S. L. Smith, “Distributed submodular maximization with limited information,” IEEE Transactions on Control of Network Systems, vol. 5, no. 4, pp. 1635–1645, 2018.
- [15] J. Perazzone, S. Wang, M. Ji, and K. Chan, “Communication-efficient device scheduling for federated learning using stochastic optimization,” in IEEE INFOCOM 2022-IEEE Conference on Computer Communications, (Piscataway, N,J), pp. 1449–1458, IEEE, 2022.
- [16] M. Rostami and S. S. Kia, “Federated learning using variance reduced stochastic gradient for probabilistically activated agents,” in American Control Conference, pp. 861–866, IEEE, 2023.
- [17] L. Bassham, A. Rukhin, J. Soto, J. Nechvatal, M. Smid, S. Leigh, M. Levenson, M. Vangel, N. Heckert, and D. Banks, “A statistical test suite for random and pseudorandom number generators for cryptographic applications,” 2010-09-16 2010.