This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Group Complete-{s}\{s\} Pliable Index Coding

Badri N. Vellambi, Sina Eghbal, Lawrence Ong, and Parastoo Sadeghi    Sina Eghbal, Badri N. Vellambi University of Cincinnati
[email protected] [email protected]
   Lawrence Ong University of Newcastle
[email protected]
   Parastoo Sadeghi University of New South Wales (Canberra)
[email protected]
Abstract

This paper introduces a novel class of PICOD(tt) problems referred to as gg-group complete-SS PICOD(tt) problems. It constructs a multi-stage achievability scheme to generate pliable index codes for group complete PICOD problems when S={s}S=\{s\} is a singleton set. Using the maximum acyclic induced subgraph bound, lower bounds on the broadcast rate are derived for singleton SS, which establishes the optimality of the achievability scheme for a range of values for tt and for any gg and ss. For all other values, it is shown that the achievability scheme is optimal among the restricted class of broadcast codes.

I Introduction

Index Coding [1] is a multicast problem where a single server conveys messages to several receivers over a noiseless broadcast channel. The problem is studied under the premise that each receiver has a subset of messages as side information and demands a specific message that is not in its side information. In a flexible variant of the problem called pliable index coding [2] (abbreviated usually as PICOD(tt)), receivers do not demand specific messages. Instead, each receiver is satisfied upon decoding any tt messages not in its side information. Both problem formulations seek the shortest normalized broadcast length and a corresponding optimal code.

Finding optimal linear codes for index coding and PICOD(tt) problems is NP-hard [1, 2]. Despite this pessimistic result, index coding remains an active research area with results in two broad clusters: the design of index codes, and establishing fundamental limits on optimal index code length.

An index-coding problem is completely characterized by the side information of the receivers and the message(s) they demand. A problem can thus be captured by a directed graph where the nodes represent the message indices. Using this graph representation, index codes have been designed based on the graph components (cycles [3], cliques [4], partial cliques [4], interlinked cycles [5]) or its coloring [6]. Coloring of confusion graphs, where each node represents a realization of the messages, has also been used to design codes [4, 7]; it must be remarked that the size of confusion graphs grows exponentially with the message size. Besides graph-based approaches, codes based on random-coding arguments have also been devised [8]. In [9], matroid theory was used to show impossibility of optimal linear codes over fields with characteristic three, and optimal non-linear codes for small instances of the problem were explicitly constructed.

The graph representation of an index-coding problem allows us to not only design codes, but also converse results. Specifically, the size of a maximum acyclic induced subgraph (MAIS) is a lower bound to the optimal index code length [1]. Another approach to deriving converse results uses Shannon-type information inequalities, and this is referred to as the polymatroid lower bound [8].

Designing codes for the PICOD problem, on the other hand, took very different approaches: the code construction is mostly algorithmic. Two algorithmic coding schemes with proven bounds are as follows: Brahma and Fragouli [2] proposed a randomized coding scheme which constructs codes of length O(min{tlog2n,tlogn+log3n})O(\min\{t\log^{2}{n},t\log{n}+\log^{3}{n}\}) if m=O(nδ)m=O(n^{\delta}), where nn is the number of receivers and mm is the number of messages; Song et al. [10] utilized a greedy approach to construct codes whose length is bounded from above by O(tlogn+log2n)O(t\log{n}+\log^{2}{n}).

Further, algorithmic coding schemes without established theoretical upper bounds have been developed. Such algorithms typically produce codes of shorter average lengths than those produced by the previously mentioned algorithms. For instance, the Greedy Cover algorithm [2] constructs subsets of messages (that are jointly coded) by greedily and sequentially selecting a message that maximizes its utility. An improved variation of this algorithm was recently proposed in [11]. Krishnan et al. [12] used a graph-based method to construct codes for PICOD(tt) problems. Their coding scheme involves a tt-fold conflict-free coloring of the vertices of a hypergraph associated with the PICOD(tt).

Lower bounds for PICOD(1) were derived based on receivers that are absent in the problems [13]. Although this approach builds on the concept of MAIS lower bound for index coding, the result is expressed in terms of how a side-information set intersects or contains another. The lower bounds are tight for special side-information configurations and for all problems with up to four absent receivers [14].

Other studies focused on devising optimal coding schemes for specific classes of PICOD(tt) problems. Liu and Tuninetti [15] investigated a subset of PICOD(tt) problems known as complete-SS problems. In these problems, every size ss subset of mm messages where sSs\in S serves as the side information for a receiver. Optimal code length has been found for two classes of complete-SS problems: S={smin,,smax}S=\{s_{\text{min}},\dotsc,s_{\text{max}}\} or S={0,,smin1}{smax+1,,m1}S=\{0,\dotsc,s_{\text{min}}-1\}\cup\{s_{\text{max}}+1,\dotsc,m-1\}, for any sminsmaxs_{\text{min}}\leq s_{\text{max}}. For these two classes, an optimal code either conveys messages in an uncoded fashion or is an MDS code (based on partial clique for index coding). The matching converse is mainly based on MAIS.

In this work, we consider a generalization of complete-SS PICOD(tt). Our generalization is termed gg-group complete-SS PICOD(tt) problem where there are mm groups each of gg messages, and every receiver that has ss groups where sSs\in S is present in the problem. In total, there are sS(ms)\sum_{s\in S}\binom{m}{s} receivers. Our work focuses on the singleton S={s}S=\{s\} case where 1sm1\leq s\leq m. We first present a multi-stage coding scheme that uses the ingredients of code design that are optimal in the ungrouped complete-SS setting. The scheme is shown to be optimal for t>(g1)(ms)t>(g-1)(m-s). For t(g1)(ms)t\leq(g-1)(m-s), we establish that the scheme is optimal among all broadcast codes. Our results builds on and subsumes those of Liu and Tuninetti for the singleton S when we set g=1g=1 [15].

This paper is organized as follows. Section II formally presents the group complete PICOD problem, Section III presents the notation used in the problem and discusses the tools and relevant graph-theoretical results used in this work, Section IV presents our achievability, converse, and pertinent ancillary results. Most proofs are relegated to the supplementary appendix document.

II Problem Formulation

The Pliable Index Coding (PICOD(tt)) problem consists of the following components:

  • A server possesses mm i.i.d messages X1,X2,,XmX_{1},X_{2},\ldots,X_{m} each of which take values in 𝔽L\mathbbm{F}^{L}, where 𝔽\mathbbm{F} is a finite field, and LL is the number of symbols of 𝔽\mathbbm{F} in each message.

  • nn receivers that each have a subset of mm messages (known as side information), and request any additional tt messages not in their side information. Without loss of generality, we may assume that each receiver has at most mtm-t messages in the side information.

The goal of the PICOD(tt) problem is to design strategies for the server to compress the mm messages into an index that is broadcast over a noiseless channel to meet the receiver demands. A specific class of PICOD(tt) problems known as complete-SS PICOD(tt) problem was introduced in [15]. In this formulation, S{1,2,,mt}S\subseteq\{1,2,\ldots,m-t\} defines the receivers and their side information as follows:

  • For each sSs\in S, there are (ms)\binom{m}{s} receivers each with a distinct subset of size ss of messages as side information. We denote the side information of node rr as r\mathcal{M}_{r}.

  • In total, there are sS(ms)\sum_{s\in S}\binom{m}{s} receivers.

For example, if m=3m=3, t=1t=1 and S={1,2}S=\{1,2\}, then the complete-SS PICOD(tt) problem with 33 messages consists of (31)+(32)=6\binom{3}{1}+\binom{3}{2}=6 receivers r1,,r6r_{1},\ldots,r_{6} with side-information subsets r1={X1}\mathcal{M}_{r_{1}}=\{X_{1}\}, r2={X2}\mathcal{M}_{r_{2}}=\{X_{2}\}, r3={X3}\mathcal{M}_{r_{3}}=\{X_{3}\}, r4={X1,X2}\mathcal{M}_{r_{4}}=\{X_{1},X_{2}\}, r5={X1,X3}\mathcal{M}_{r_{5}}=\{X_{1},X_{3}\}, and r6={X2,X3}\mathcal{M}_{r_{6}}=\{X_{2},X_{3}\}, respectively. Since t=1t=1, each receiver wants an additional message that it does not already know.

This work focuses on a generalization of the complete-SS PICOD(tt) problem, referred to as gg-Group Complete-SS PICOD(tt) problem with mm groups defined as follows:

  • There are mgmg messages partitioned into mm groups of gg messages each. The groupings are given by 𝒢i={X(i1)g+1,,Xig}\mathcal{G}_{i}=\{X_{(i-1)g+1},\ldots,X_{ig}\}, i=1,,mi=1,\ldots,m.

  • S{1,,mgtg}S\subseteq\{1,\ldots,\lfloor\frac{mg-t}{g}\rfloor\} defines the number and side information of the receivers in the problem instance. For every sSs\in S, there are (ms)\binom{m}{s} receivers each with ss groups as side information. The set of all sS(ms)\sum_{s\in S}\binom{m}{s} receivers in the problem is denoted by \mathfrak{R}.

    For example, if m=3m=3, g=2g=2 S={1,2}S=\{1,2\}, then the problem has (31)+(32)\binom{3}{1}+\binom{3}{2} receivers. The side information of the six receivers r1,,r6r_{1},\ldots,r_{6} are r1=𝒢1\mathcal{M}_{r_{1}}=\mathcal{G}_{1}, r2=𝒢2\mathcal{M}_{r_{2}}=\mathcal{G}_{2}, r3=𝒢3\mathcal{M}_{r_{3}}=\mathcal{G}_{3}, r4=𝒢1𝒢2\mathcal{M}_{r_{4}}=\mathcal{G}_{1}\cup\mathcal{G}_{2}, r5=𝒢1𝒢3\mathcal{M}_{r_{5}}=\mathcal{G}_{1}\cup\mathcal{G}_{3}, and r6=𝒢2𝒢3\mathcal{M}_{r_{6}}=\mathcal{G}_{2}\cup\mathcal{G}_{3} respectively. Observe that the receivers either have 2 messages (1 group) or 4 messages (2 groups) in their side information.

  • Each receiver requests tt extra messages that is not in its side information.

Note that if we choose g=1g=1, our gg-group complete-SS formulation reverts back to the complete-SS setup of [15]. Even though the receiver demands are pliable, the server has to select which tt messages to communicate to which receiver. This is done through decoding choice (function) 𝒟:2mg\mathcal{D}:\mathfrak{R}\rightarrow 2^{\llbracket mg\rrbracket} with |𝒟(r)|=t|\mathcal{D}(r)|=t for each rr\in\mathfrak{R}. Throughout this paper, we use the notation that m\llbracket m\rrbracket as a shorthand for {1,2,,m}\{1,2,\ldots,m\} for any natural number mm, and for a set XX, 2X2^{X} denotes the power set of XX.

For the sake of analysis, it is sometimes useful to relax the condition that the decoding choice for each receiver is precisely tt messages. It is therefore convenient to define a generalized decoding choice that assigns at least tt messages to each receiver. Thus, a generalized decoding choice is a function 𝒟:2mg\mathcal{D}:\mathfrak{R}\rightarrow 2^{\llbracket mg\rrbracket}, where |𝒟(r)|t|\mathcal{D}(r)|\geq t for each rr\in\mathfrak{R}.

Given a PICOD problem, the server selects a decoding choice 𝒟\mathcal{D}, and a pliable index code CC that works for this choice, whose details are as follows:

  • CC comprises of an encoding function 𝔼:(𝔽L)mg𝔽\mathbbm{E}:\left(\mathbb{F}^{L}\right)^{mg}\rightarrow\mathbbm{F}^{\ell} that the server uses to jointly encode the mgmg messages and convey ll symbols of 𝔽\mathbbm{F} to each receiver. The encoder transmits 𝔼(X1,,Xmg)\mathbbm{E}(X_{1},\ldots,X_{mg}) to all receivers noiselessly. The normalized size L\frac{\ell}{L} is termed the rate RCR_{C} of the code.

  • Each receiver rr\in\mathfrak{R}, CC comprises of a decoding function 𝔻r:𝔽×(𝔽L)κr(𝔽L)t\mathbbm{D}_{r}:\mathbbm{F}^{\ell}\times\left(\mathbbm{F}^{L}\right)^{\kappa_{r}}\rightarrow\left(\mathbbm{F}^{L}\right)^{t}, where κr\kappa_{r} denotes the number of messages in the side information of receiver rr. Each receiver rr\in\mathfrak{R}, uses the decoding function 𝔻r\mathbbm{D}_{r} with the broadcast encoded message 𝔼(X1,,Xmg)\mathbbm{E}(X_{1},\ldots,X_{mg}) along with its side information to decode the tt messages in 𝒟(r)\mathcal{D}(r), i.e., {Xj:j𝒟(r)}\{X_{j}:j\in\mathcal{D}(r)\}.

In this work, we focus solely on the canonical singleton setting of gg-group complete-SS PICOD(tt) where S={s}S=\{s\} for smgtgs\in\llbracket\lfloor\frac{mg-t}{g}\rfloor\rrbracket. In this problem setting, there are (ms)\binom{m}{s} receivers each with ss groups of messages in its side information. The goal is to identify the optimal broadcast rate R(m,s,g,t)R^{*}(m,s,g,t) that is the least among all pliable index codes, i.e.,

R(m,s,g,t)=min𝒟(minC works for 𝒟RC).R^{*}(m,s,g,t)=\min_{\mathcal{D}}\left(\min_{C\textrm{ works for }\mathcal{D}}R_{C}\right).

III Required Graph-theoretic Notions and Tools

As in previous works, non-achievability (lower bound) results are shown predominantly using graph-theoretic arguments with a heavy use of the maximum acyclic induced subgraph (MAIS) bound [1]. We therefore first describe our notation for the graph representation. Given a gg-group complete-{s}\{s\} PICOD(tt) problem with mm groups, and once a particular decoding choice 𝒟\mathcal{D} is selected, we can represent the resulting index coding problem as a directed graph G𝒟G_{\mathcal{D}} with t(ms)t\binom{m}{s} nodes using the following two rules:

  • R1

    A receiver rr with side information r==1s𝒢i\mathcal{M}_{r}=\cup_{\ell=1}^{s}\mathcal{G}_{i_{\ell}} that requests messages 𝒟(r)={j1,,jt}\mathcal{D}(r)=\{j_{1},\ldots,j_{t}\} is represented by tt separate nodes labelled as (r,j1),(r,j2),,(r,jt)(r,j_{1}),(r,j_{2}),\ldots,(r,j_{t}).

  • R2

    A directed edge from node (r,j)(r,j) to (r,j)(r^{\prime},j^{\prime}) exists in G𝒟G_{\mathcal{D}} iff XjrX_{j^{\prime}}\in\mathcal{M}_{r}, i.e., the node (r,j)(r,j) represents a receiver whose side information contains XjX_{j^{\prime}}.

In this digraph representation, a node (r,j)(r,j) is said to have receiver label rr, and (decoded) message label jj. Conversely, given a receiver rr and a message index j𝒟(r)j\in\mathcal{D}(r), we view (r,j)(r,j) as one of the tt distinct node representations of rr.

In the ungrouped (g=1g=1) case, a receiver r={i1,,is}r=\{i_{1},\ldots,i_{s}\} that recovers tt messages has the side information of (t+ss)1\binom{t+s}{s}-1 other receivers, and can therefore emulate them. Here, by emulate, we mean the receiver rr can possibly decode more messages by use of the decoding function of any other receiver whose side information it possesses after decoding its tt messages. This can be done iteratively until the receiver rr can emulate no more receivers. In the grouped case, it is not necessary that a node can emulate another upon decoding its tt messages. Instead, a receiver rr can emulate another receiver rr^{\prime} upon decoding its tt messages iff the messages exclusive to the side information of rr^{\prime} is decoded by rr, i.e., rr𝒟(r)\mathcal{M}_{r^{\prime}}\setminus\mathcal{M}_{r}\in\mathcal{D}(r). Therefore, a necessary and sufficient condition for rr to emulate another receiver is that 𝒟(r)\mathcal{D}(r) contain at least another full group.

One of the key benefits of this digraph representation is that it allows us to focus on the sizes of acyclic induced subgraphs of G𝒟G_{\mathcal{D}} to bound from below the rate for the PICOD(t)(t) problem. In fact, the best such lower bound is given by the size of the maximum acyclic induced subgraph with distinct message labels [1]. In this work, we denote MAIS(G𝒟)\texttt{MAIS}(G_{\mathcal{D}}) to be any one maximum acyclic induced subgraph of G𝒟G_{\mathcal{D}} with distinct message labels, and let |MAIS(G𝒟)||\texttt{MAIS}(G_{\mathcal{D}})| to be its size. Given a PICOD problem, a decoding choice 𝒟\mathcal{D}, and an MAIS GG in G𝒟G_{\mathcal{D}}, we will find it handy to organize GG in layers based on the following rules:

  • R1

    Layer 1, denoted by 1\mathcal{L}_{1}, consists of nodes of GG that are sources (i.e., nodes with no incoming edges);

  • R2

    For i>1i>1, Layer ii, denoted by i\mathcal{L}_{i}, consists of all nodes whose in-neighbors consists of nodes in Layers jj for j<ij<i with at least one in-neighbor in Layer i1i-1; and

  • R3

    The last layer consists only of sinks (but not necessarily all the sinks).

We say that a group 𝒢\mathcal{G} of messages is fully present in a layer x\mathcal{L}_{x}, if 𝒢{Xj:(r,j)x}\mathcal{G}\subseteq\{X_{j}:(r,j)\in\mathcal{L}_{x}\}. We say a group 𝒢\mathcal{G} of messages is partially present in a layer x\mathcal{L}_{x}, if 0<|𝒢{Xj:(r,j)x}|<g0<|\mathcal{G}\cap\{X_{j}:(r,j)\in\mathcal{L}_{x}\}|<g. Lastly, we say that a group 𝒢\mathcal{G} of messages is absent in a layer x\mathcal{L}_{x}, if 𝒢{Xj:(r,j)x}=\mathcal{G}\cap\{X_{j}:(r,j)\in\mathcal{L}_{x}\}=\emptyset. Conversely, a layer x\mathcal{L}_{x} is perfectly covered by groups if there exists a collection of groups 𝒢i1,,𝒢ik\mathcal{G}_{i_{1}},\ldots,\mathcal{G}_{i_{k}} whose kgkg message labels together form the message labels of the nodes in the layer, i.e., 𝒢i1𝒢ik={Xj:(r,j)x}\mathcal{G}_{i_{1}}\cup\ldots\cup\mathcal{G}_{i_{k}}=\{X_{j}:(r,j)\in\mathcal{L}_{x}\}. We extend the above notions to say a group 𝒢\mathcal{G} of messages is partially present or fully present in a MAIS GG of G𝒟G_{\mathcal{D}} if it is partially present or fully present in any one layer. Note that from Property P5 below, a group cannot be partially present in more than one layer. Similarly, we say that a group 𝒢\mathcal{G} of messages is absent from a MAIS GG of G𝒟G_{\mathcal{D}} if it is absent from every layer of GG.

Lemma 1

The following properties are true of any MAIS GG of any digraph representation G𝒟G_{\mathcal{D}}.

  • P1

    An edge starting from a node of Layer ii can only end in a node in a layer of larger index.

  • P2

    If jj is the message label of a node in Layer ii, and rr is receiver label of a node in Layer kk for kik\geq i, then the side information of rr does not contain XjX_{j}.

  • P3

    Two nodes that have the same receiver label have the same out-neighborhood.

  • P4

    Two nodes whose message labels are from the same group have the same in-neighborhood.

  • P5

    A group can be partially present in only one layer.

  • P6

    The set of message labels in an MAIS and the subset of messages in the side information of the sinks are disjoint.

Proof:

The proof of the properties are as follows:

P1: An edge from a node in Layer kk to a node in Layer ii for k>ik>i would violate Rule R2 of the construction of layers.

P2: Suppose XjrX_{j}\in\mathcal{M}_{r}. Then there must be an edge from the node with receiver label rr in Layer kk to the node in Layer ii with message label jj. This then violates P1 since kik\geq i.

P3: The condition for an edge to start in a node (r,j)(r,j) and end in a node (r,j)(r^{\prime},j^{\prime}) depends only on the receiver label rr and the message label jj^{\prime}. Hence any two nodes that have the same receiver label must have the same out-neighborhood.

P4: From Rule R2 of digraph construction, it is clear that the in-neighborhood of a node (r,j)(r,j) is uniquely determined by the group where XjX_{j} lies. Thus, two nodes with message labels from the same group have the same in-neighborhood.

P5: Suppose that there are (r,j)(r,j) and (r,j)(r^{\prime},j^{\prime}) in GG such that: (a) Xj,XjX_{j},X_{j^{\prime}} belong to the same group; and (b) the nodes (r,j)(r,j) and (r,j)(r^{\prime},j^{\prime}) lie in layers kk and kk^{\prime}, respectively, with kkk\leq k^{\prime}. From P4, we see that the in-neighborhood of (r,j)(r,j) and (r,j)(r^{\prime},j^{\prime}) are the same. Hence, they cannot lie in different layers.

P6: If they are not disjoint, a message index is present both as message label and in the side-information of a sink node. By Rule R2, this would mean that the sink has an outgoing edge, which is a contradiction. ∎

Corollary 1

From Property P1 of Lemma 1, it follows that there are no edges between nodes of the same layer.

Corollary 2

From Property P2 of Lemma 1, it follows that if jj is a message label of a source node in an MAIS GG of G𝒟G_{\mathcal{D}}, and rr is a receiver label of any node in GG, then XjrX_{j}\notin\mathcal{M}_{r}; and if jj is a message label of a node in GG, and rr is a receiver label of a sink node in GG, then again XjrX_{j}\notin\mathcal{M}_{r}.

The last fact that we will need for our non-achievability results characterizes the size of an MAIS for generalized decoding choice obtained by appending to decoding choices messages that a receiver decodes by emulation.

Lemma 2

Let 𝒟\mathcal{D} be a decoding choice for a gg-group complete-{s}\{s\} PICOD(tt) problem with mm groups. Suppose a receiver rr uses 𝔻r\mathbb{D}_{r} to decode its tt messages, and is then able to emulate r1r_{1} to decode a new message XjX_{j}, where j𝒟(r1)j\in\mathcal{D}(r_{1}). Let 𝒟app\mathcal{D}_{\textrm{app}} be a generalized decoding choice such that:

𝒟app(x)={𝒟(x)xr𝒟(x){j}x=r.\displaystyle\mathcal{D}_{\textrm{app}}(x)=\left\{\begin{array}[]{cc}\mathcal{D}(x)&x\neq r\\ \mathcal{D}(x)\cup\{j\}&x=r\end{array}\right..

Then, |MAIS(G𝒟app)|=|MAIS(G𝒟)|.|\texttt{MAIS}(G_{\mathcal{D}_{\textrm{app}}})|=|\texttt{MAIS}(G_{\mathcal{D}})|.

Proof:

Proof is given in the supplementary document. ∎

Remark 1

Given a decoding choice 𝒟\mathcal{D}, we can sequentially append for each receiver every additional messages it will decode by emulating other receivers. We repeat this until we can add no new messages to any receiver. Let 𝒟¯\overline{\mathcal{D}} be the generalized decoding choice upon termination of this process. By repeated application of the Lemma 2, we can show that

|MAIS(G𝒟¯)|=|MAIS(G𝒟)|.|\texttt{MAIS}(G_{\overline{\mathcal{D}}})|=|\texttt{MAIS}(G_{\mathcal{D}})|.
Corollary 3

Given a decoding choice 𝒟\mathcal{D}, let 𝒯={(rk,jk)}k=1\mathcal{T}=\{(r_{k},j_{k})\}_{k=1}^{\ell} be an enumeration of the additional messages decoded by receivers by successive emulation. Suppose we append 𝒟\mathcal{D} with a subset 𝒮𝒯\mathcal{S}\subseteq\mathcal{T} of extra messages decoded by receivers to obtain a generalized decoding choice 𝒟\mathcal{D}^{*}. Then, by a similar argument, we can show that

|MAIS(G𝒟¯)|=|MAIS(G𝒟)|=|MAIS(G𝒟)|.|\texttt{MAIS}(G_{\overline{\mathcal{D}}})|=|\texttt{MAIS}(G_{{\mathcal{D}^{*}}})|=|\texttt{MAIS}(G_{\mathcal{D}})|.

IV New Results

We begin this section with the achievability result, which is an extension of the one proposed in [15].

Theorem 1

R(m,s,g,t)min{s+t,t(ms)(ms)}R^{*}(m,s,g,t)\leq\min\{s+t,\lceil\frac{t}{(m-s)}\rceil(m-s)\}.

Proof:

The code construction for a gg-group complete-{s}\{s\} PICOD(t)(t) problem with mm groups is as follows:

  • For each i=1,2,,tmsi=1,2,\ldots,\lfloor\frac{t}{m-s}\rfloor, the sender conveys {Xi,Xg+i,,X(m1)g+i}\{X_{i},X_{g+i},\ldots,X_{(m-1)g+i}\} (one new message from each of the mm groups) via an MDS code of size msm-s. Each receiver already possess ss of these mm encoded messages, and will see the ss messages to decode the remaining msm-s messages. At the end of these rounds, every receiver will have decoded a total of tms(ms)\lfloor\frac{t}{m-s}\rfloor(m-s) messages with tms\lfloor\frac{t}{m-s}\rfloor messages from each group.

  • If t=ttms(ms)>0t^{\prime}=t-\lfloor\frac{t}{m-s}\rfloor(m-s)>0, then sender does one of the following:

    • If s+t<mss+t^{\prime}<m-s, the sender transmits messages Xj,Xg+j,,X(s+t1)g+jX_{j},X_{g+j},\ldots,X_{(s+t^{\prime}-1)g+j} uncoded, where j=tms+1j=\lfloor\frac{t}{m-s}\rfloor+1.

    • If s+tmss+t^{\prime}\geq m-s, the sender encodes Xj,Xg+j,,X(m1)g+jX_{j},X_{g+j},\ldots,X_{(m-1)g+j} (one new message from each of the mm groups) via an MDS code of size msm-s. Each receiver then receives msm-s new messages, which may be more than the remaining tt^{\prime} they need.

The following figure presents the plot of the achievable rate offered by Theorem 1. Additionally, the applicable matching non-achievability results for various ranges of tt are also indicated therein. Note from the proof of achievability that in the last round, an MDS code will be used if and only if smss\geq m-s, in which case the achievable rate is given by tms(ms)\lceil\frac{t}{m-s}\rceil(m-s); alternately, the last round will comprise of uncoded messages iff s<mss<m-s, which introduces the unity slope portions. In both cases, the achievable rate region is non-convex due to the absence of a time-sharing argument that is usually used to convexify rate regions.

Refer to caption
Figure 1: The achievable rate given in Theorem 1

We now present three ancillary results that are needed to derive our non-achievability results. The first two connect the size of MAIS of G𝒟G_{\mathcal{D}} to the number of messages decoded by a node in G𝒟G_{\mathcal{D}}. The third result argues that under certain conditions if there is a decoding choice whose size of a MAIS is s+tis+t-i for i>0i>0, then there must exist another decoding choice whose size of MAIS is s+ti+1s+t-i+1. Our non-achievability results prove the non-existence of a decoding choice whose MAIS size is s+t1s+t-1, and therefore, using the third result, we are guaranteed that the smallest size of any MAIS of any decoding choice is s+ts+t.

Lemma 3

If for a decoding choice 𝒟\mathcal{D}, there exists a receiver rr that decodes s+ts+t messages by sequentially emulating other receivers, then |MAIS(G𝒟)|s+t|\texttt{MAIS}(G_{\mathcal{D}})|\geq s+t.

Proof:

Let 𝒟(r)={j1,,jt}\mathcal{D}(r)=\{j_{1},\ldots,j_{t}\}. Then rr uses the decoding function 𝔻r\mathbbm{D}_{r} of the receiver rr to decode Xj1,,XjtX_{j_{1}},\ldots,X_{j_{t}}. Since, rr decodes s+ts+t messages by emulating other receivers, these tt messages must contain at least one full group allowing it to emulate another receiver. This process repeats until rr decodes ss additional messages. Let Xj1,,Xjs+tX_{j_{1}},\ldots,X_{j_{s+t}} be an enumeration of the s+ts+t distinct messages rr decodes sequentially. For i=1,,s+ti=1,\ldots,s+t, let rkir_{k_{i}} be the receiver that rr emulates to decode XjiX_{j_{i}} using 𝔻ri\mathbbm{D}_{r_{i}}. Then, the following statements hold.

  • F1

    For any is+ti\in\llbracket s+t\rrbracket, (rki,ji)(r_{k_{i}},j_{i}) is a node in G𝒟G_{\mathcal{D}}.

  • F2

    For any is+ti\in\llbracket s+t\rrbracket, rkir{Xj1,,Xji1}\mathcal{M}_{r_{k_{i}}}\subseteq\mathcal{M}_{r}\cup\{X_{j_{1}},\ldots,X_{j_{i-1}}\}.

  • F3

    For any is+ti\in\llbracket s+t\rrbracket, XjirX_{j_{i}}\notin\mathcal{M}_{r}.

  • F4

    For any 0<<s+t0<\ell<\ell^{\prime}\leq s+t, XrkX_{\ell^{\prime}}\notin\mathcal{M}_{r_{k_{\ell}}}. Otherwise, XX_{\ell^{\prime}} must either be decoded earlier in the enumeration of messages decoded by rr, or is in the side information of rr. The former inference violates the fact that the s+ts+t messages are distinct, the latter violates the fact that XX_{\ell^{\prime}} is innovative to rr.

Consider the subgraph induced by {vi=(rki,ji):i=1,,s+t}\{v_{i}=(r_{k_{i}},j_{i}):i=1,\ldots,s+t\} in G𝒟G_{\mathcal{D}}. From F4 above, we see that this induced subgraph has edges from viv_{i} to vjv_{j} only if i<ji<j. Consequently, this subgraph of s+ts+t nodes is acyclic with distinct message labels, and therefore, the claim follows. ∎

Lemma 4

Let m,s,g,tm,s,g,t be such that t>(g1)(ms)t>(g-1)(m-s) and s+t=g(ms)s+t=g(m-s). If for a decoding choice 𝒟\mathcal{D}, |MAIS(G𝒟)|=s+t|\texttt{MAIS}(G_{\mathcal{D}})|=s+t, then there exists a receiver rr that decodes s+ts+t messages.

Proof:

Let vk=(rik,jk)v_{k}=(r_{i_{k}},j_{k}), k=1,,s+tk=1,\ldots,s+t, be the labels of an MAIS G0G_{0} of G𝒟G_{\mathcal{D}}. Since this subgraph is acyclic, we may assume that v1,v2,,vs+tv_{1},v_{2},\ldots,v_{s+t} is in a topologically sorted order with sources at the beginning and the sinks at the end. Now, consider the sink vs+t=(ris+t,js+t)v_{s+t}=(r_{i_{s+t}},j_{{s+t}}). Observe that the mgmg messages in the problem are partitioned into two: the gsgs messages that are in the side information of receiver ris+tr_{i_{s+t}}, and the (s+t)=g(ms)(s+t)=g(m-s) message labels of the nodes in G0G_{0}. Therefore, it follows that all sinks of G0G_{0} must have the same receiver label ris+tr_{i_{s+t}}. Now, since ris+tr_{i_{s+t}} decodes t>(g1)(ms)t>(g-1)(m-s) messages, it must decode at least one full group. Now, consider the acyclic subgraph G1G_{1} obtained by deleting all nodes in G0G_{0} whose message label is decoded by ris+tr_{i_{s+t}} by use of 𝔻r\mathbbm{D}_{r}. Consider the sinks of G1G_{1}. As before, the message labels in G1G_{1} is precisely ris+tc𝒟(ris+t)\mathcal{M}_{r_{i_{s+t}}}^{c}\setminus\mathcal{D}(r_{i_{s+t}}) (due to Property P6 of Lemma 1). Thus, the sinks of G1G_{1} whose side information is a subset of the complement of the message labels in G1G_{1}, can each be emulated by ris+tr_{i_{s+t}} to uncover more messages.

This process of emulating and recovering more messages can be followed iteratively until the receiver labels of all nodes in G0G_{0} are emulated by ris+tr_{i_{s+t}}. Thus, at each iteration k>0k>0, we can define a new acyclic subgraph GkG_{k} by deleting all nodes in Gk1G_{k-1} whose message label is decoded by the sink nodes of Gk1G_{k-1} by use of their decoding functions. Since ris+tr_{i_{s+t}} can emulate the sink nodes of Gk1G_{k-1}, ris+tr_{i_{s+t}} uncovers all messages decoded by the sink nodes of Gk1G_{k-1}. Thus, we see that the receiver ris+tr_{i_{s+t}} will eventually decode the s+ts+t messages corresponding to the message labels of the nodes in G0G_{0}. ∎

Lemma 5

Let m,s,g,tm,s,g,t be such that t>(g1)(ms)t>(g-1)(m-s) and s+tg(ms)s+t\leq g(m-s). Then, if there exists a decoding choice 𝒟\mathcal{D} with |MAIS(G𝒟)|=s+ti|\texttt{MAIS}(G_{\mathcal{D}})|=s+t-i for i>0i>0 then there also exists 𝒟\mathcal{D}^{\prime} with |MAIS(G𝒟)|=s+ti+1|\texttt{MAIS}(G_{\mathcal{D}^{\prime}})|=s+t-i+1.

Proof:

The proof proceeds in three cases depending on the number of sinks kk in G𝒟G_{\mathcal{D}}. The case that k>tk>t is shown to lead to contradiction, whereas in the other two cases, a specific node (r,j)(r,j^{*}) in G𝒟G_{\mathcal{D}} and a message label j^𝒟(r)\hat{j}\notin\mathcal{D}(r) are identified. The decoding choice 𝒟\mathcal{D} is then altered to obtain 𝒟^\hat{\mathcal{D}} by replacing jj^{*} in 𝒟(r)\mathcal{D}(r) with j^\hat{j}. It is then shown that |MAIS(G𝒟^)|=|MAIS(G𝒟)|+1|\texttt{MAIS}(G_{\hat{\mathcal{D}}})|=|\texttt{MAIS}(G_{\mathcal{D}})|+1. The proof is detailed in the supplementary appendix document. ∎

Corollary 4

Let m,s,g,tm,s,g,t be such that t>(g1)(ms)t>(g-1)(m-s) and s+tg(ms)s+t\leq g(m-s). From Lemma 5 Case 1, it is clear that precisely ss groups are absent from an MAIS(G𝒟)\texttt{MAIS}(G_{\mathcal{D}}) for any 𝒟\mathcal{D}.

We are now ready to present our non-achievability results.

Theorem 2

Let m,s,g,tm,s,g,t be such that t>(g1)(ms)t>(g-1)(m-s) and s+t=g(ms)s+t=g(m-s). Then, R(m,s,g,t)=s+tR^{*}(m,s,g,t)=s+t.

Proof:

The proof starts with a decoding choice 𝒟\mathcal{D} for which the size of MAIS is s+t1s+t-1. It identifies a special receiver r^\hat{r} that can decode an additional message by emulating other receivers. Constructing a generalized decoding choice by appending to 𝒟(r^)\mathcal{D}(\hat{r}) this additional message allows to then establish that an MAIS corresponding to decoding choice 𝒟\mathcal{D} has to have a size at least s+ts+t. By invoking Lemma 5, we establish that s+ts+t is a lower bound to the size of an MAIS corresponding to any decoding choice. Complete proof can be found in the supplementary appendix document. ∎

Theorem 3

Let m,s,g,tm,s,g,t be such that t>(g1)(ms)t>(g-1)(m-s) and s+t<g(ms)s+t<g(m-s). Then, R(m,s,g,t)=s+tR^{*}(m,s,g,t)=s+t.

Proof:

The proof follows closely that of Theorem 2, and can be found in the supplementary appendix document. ∎

Theorem 4

Let m,s,g,tm,s,g,t be such that s+t>g(ms)s+t>g(m-s) and g1<tmsgg-1<\frac{t}{m-s}\leq g. Then, R(m,s,g,t)=g(ms)R^{*}(m,s,g,t)=g(m-s).

Proof:

As in the proof of Proposition 7 of [15], this proof selects a sub-problem that is a gg-group complete-{s}\{s^{\prime}\} problem over mm^{\prime} groups for suitable mm^{\prime} and gg^{\prime} for which Theorem 2 is applicable. Complete proof can be found in the supplementary appendix document. ∎

Theorem 5

Let 𝒞m,s,g,t\mathcal{C}_{m,s,g,t} denote the family of broadcast codes for the gg-group complete-{s}\{s\} PICOD(t)t) problem with mm groups where if a message is conveyed to one receiver, every receiver that does not have that message in its side information also decodes that message. The optimal rate among codes from 𝒞m,s,g,t\mathcal{C}_{m,s,g,t} for any (f1)(ms)<t<f(ms)(f-1)(m-s)<t<f(m-s) for fgf\in\llbracket g\rrbracket is given by R(m,s,g,t)=min{s+t,f(ms)}.R^{\prime}(m,s,g,t)=\min\{s+t,f(m-s)\}.

Proof:

The proof establishes that among broadcast codes, there exists an optimal one where the sender ignores gfg-f messages and communicates at most ff messages per group. This then allows us to reduce the problem to a ff-group complete-{s}\{s\} PICOD(tt) problem whose rate can be deduced from Theorems 3 and 4. Complete proof can be found in the supplementary appendix document. ∎

References

  • [1] Z. Bar-Yossef, Y. Birk, T. Jayram, and T. Kol, “Index coding with side information,” IEEE Transactions on Information Theory, vol. 57, no. 3, pp. 1479–1494, 2011.
  • [2] S. Brahma and C. Fragouli, “Pliable index coding,” IEEE Transactions on Information Theory, vol. 61, no. 11, pp. 6192–6203, 2015.
  • [3] M. A. R. Chaudhry, Z. Asad, A. Sprintson, and M. Langberg, “On the complementary index coding problem,” in Proc. IEEE Int. Symp. Inf. Theory (ISIT), St Petersburg, Russia, July 31–Aug. 5 2011, pp. 244–248.
  • [4] Y. Birk and T. Kol, “Coding on demand by an informed source (ISCOD) for efficient broadcast of different supplemental data to caching clients,” IEEE Trans. Inf. Theory, vol. 52, no. 6, pp. 2825–2830, June 2006.
  • [5] C. Thapa, L. Ong, and S. J. Johnson, “Interlinked cycles for index coding: Generalizing cycles and cliques,” IEEE Trans. Inf. Theory, vol. 63, no. 6, pp. 3692–3711, June 2017.
  • [6] K. Shanmugam, A. G. Dimakis, and M. Langberg, “Local graph coloring and index coding,” in Proc. IEEE Int. Symp. Inf. Theory (ISIT), Istanbul, Turkey, July 7–12 2013, pp. 1152–1156.
  • [7] C. Thapa, L. Ong, and S. J. Johnson, “Structural characteristics of two-sender index coding,” Entropy, vol. 21, no. 6:615, June 2019.
  • [8] F. Arbabjolfaei, B. Bandemer, Y.-H. Kim, E. Şaşoğlu, and L. Wang, “On the capacity region for index coding,” in Proc. IEEE Int. Symp. Inf. Theory (ISIT), Istanbul, Turkey, July 7–12 2013, pp. 962–966.
  • [9] A. Sharififar, P. Sadeghi, and N. Aboutorab, “On the optimality of linear index coding over the fields with characteristic three,” IEEE Open Journal of the Communications Society, vol. 3, pp. 2024–2050, 2022.
  • [10] L. Song and C. Fragouli, “A polynomial-time algorithm for pliable index coding,” IEEE Transactions on Information Theory, vol. 64, no. 2, pp. 979–999, 2017.
  • [11] S. Eghbal, B. N. Vellambi, L. Ong, and P. Sadeghi, “An improved greedy cover algorithm for pliable index coding,” in 2023 59th Annual Allerton Conference on Communication, Control, and Computing (Allerton).   IEEE, 2023, pp. 1–8.
  • [12] P. Krishnan, R. Mathew, and S. Kalyanasundaram, “Pliable index coding via conflict-free colorings of hypergraphs,” IEEE Transactions on Information Theory, 2024.
  • [13] L. Ong, B. N. Vellambi, and J. Kliewer, “Optimal-rate characterisation for pliable index coding using absent receivers,” in IEEE Int. Symp. Inf. Theory (ISIT), Paris, France, July 7–12 2019.
  • [14] L. Ong, B. N. Vellambi, J. Kliewer, and P. Sadeghi, “Improved lower bounds for pliable index coding using absent receivers,” in Int. Zurich Semin. Commun. (IZS), Zurich, Switzerland, Feb. 26–28 2020.
  • [15] T. Liu and D. Tuninetti, “Tight information theoretic converse results for some pliable index coding problems,” IEEE Transactions on Information Theory, vol. 66, no. 5, pp. 2642–2657, 2019.
  • [16] M. Langberg and M. Effros, “Network coding: Is zero error always possible?” in 2011 49th Annual Allerton Conference on Communication, Control, and Computing (Allerton).   IEEE, 2011, pp. 1478–1485.

Appendix

Proof of Lemma 2

Note that, by construction, G𝒟G_{\mathcal{D}} is a subgraph of G𝒟appG_{\mathcal{D}_{\textrm{app}}} since the latter contains all nodes and edges of G𝒟G_{\mathcal{D}} and additionally one extra node (r,j)(r,j) and its incident incoming and outgoing edges. Furthermore, G𝒟G_{\mathcal{D}} is an induced subgraph of G𝒟appG_{\mathcal{D}_{\textrm{app}}} since G𝒟appG_{\mathcal{D}_{\textrm{app}}} does not have any additional edges among nodes already present in G𝒟G_{\mathcal{D}}. Therefore, it follows that any (acyclic) induced subgraph of G𝒟G_{\mathcal{D}} is also an (acyclic) induced subgraph of G𝒟appG_{\mathcal{D}_{\textrm{app}}}. However, G𝒟appG_{\mathcal{D}_{\textrm{app}}} could potentially have acyclic induced subgraphs that involve the node (r,j)(r,j). Thus it follows that |MAIS(G𝒟app)||MAIS(G𝒟)||\texttt{MAIS}(G_{\mathcal{D}_{\textrm{app}}})|\geq|\texttt{MAIS}(G_{\mathcal{D}})|. Next, to show |MAIS(G𝒟app)|=|MAIS(G𝒟)||\texttt{MAIS}(G_{\mathcal{D}_{\textrm{app}}})|=|\texttt{MAIS}(G_{\mathcal{D}})|, it suffices to establish |MAIS(G𝒟app)||MAIS(G𝒟)||\texttt{MAIS}(G_{\mathcal{D}_{\textrm{app}}})|\leq|\texttt{MAIS}(G_{\mathcal{D}})|.

The only extra components G𝒟appG_{\mathcal{D}_{\textrm{app}}} contains is the node (r,j)(r,j) and its incident edges. Now, let GappG_{\textrm{app}} be an MAIS for G𝒟appG_{\mathcal{D}_{\textrm{app}}}. If GappG_{\textrm{app}} does not contain the node (r,j)(r,j) it is also an MAIS for G𝒟G_{\mathcal{D}}, and the claim follows. We may therefore assume that (r,j)(r,j) is present in GappG_{\textrm{app}}. We now proceed in cases.

Case A: (r,j) is a source (i.e., in Layer 1\mathcal{L}_{1}) of GappG_{\textrm{app}}. In this case, consider another induced subgraph GappG^{\prime}_{\textrm{app}} constructed from GappG_{\textrm{app}} by deleting (r,j)(r,j) and its incident edges from GappG_{\textrm{app}}, and adding (r1,j)(r_{1},j) and required incident edges (to make the subgraph induced). Since originally, (r,j)(r,j) was a source in GappG_{\textrm{app}}, the receiver label of every other node in GappG_{\textrm{app}} did not contain XjX_{j} in its side information. Therefore, (r1,j)(r_{1},j) has to be a source node in GappG^{\prime}_{\textrm{app}}. This further means that GappG_{\textrm{app}}^{\prime} is acyclic, since any newly introduced loop in GappG_{\textrm{app}}^{\prime} must traverse (r1,j)(r_{1},j), but that is impossible since (r1,j)(r_{1},j) is a source node in GappG^{\prime}_{\textrm{app}}. Thus, GappG^{\prime}_{\textrm{app}} is an acyclic induced subgraph of G𝒟appG_{\mathcal{D}_{\textrm{app}}} whose size is the same as GappG_{\textrm{app}}. Thus, GappG^{\prime}_{\textrm{app}} is also an MAIS of G𝒟appG_{\mathcal{D}_{\textrm{app}}}. However, since GappG^{\prime}_{\textrm{app}} does not contain (r,j)(r,j), it is also an MAIS of G𝒟G_{\mathcal{D}}, and therefore, |MAIS(G𝒟app)|=|MAIS(G𝒟)||\texttt{MAIS}(G_{\mathcal{D}_{\textrm{app}}})|=|\texttt{MAIS}(G_{\mathcal{D}})|.

Case B: (r,j) is not a source (i.e., in Layer \mathcal{L}_{\ell} for >1\ell>1) of GappG_{\textrm{app}}. In this case we construct an alternate MAIS for G𝒟appG_{\mathcal{D}_{\textrm{app}}} as follows. As before, we start with GappG_{\textrm{app}} and delete (r,j)(r,j) and its incident edges, and add (r1,j)(r_{1},j) and required incident edges (to make the subgraph induced). Let the resultant graph be GappG^{\prime}_{\textrm{app}}. Now, Since (r,j)(r,j) originally was in layer \mathcal{L}_{\ell}, it follows that all its in-neighbors are in layers k\mathcal{L}_{k}, k<k<\ell. Therefore, all incoming edges into (r1,j)(r_{1},j) after replacement will also be from earlier layers j\mathcal{L}_{j}, j<j<\ell. However, it is possible that after replacement GappG^{\prime}_{\textrm{app}} contains edges from (r1,j)(r_{1},j) to some kk nodes in the previous layers. Let (r1,j1),(r2,j2),,(rk,jk)(r_{1}^{\prime},j_{1}),(r_{2}^{\prime},j_{2}),\ldots,(r_{k}^{\prime},j_{k}) denote these out-neighbors of (r1,j)(r_{1},j) in layers k\mathcal{L}_{k}, k<k<\ell. These back edges have the potential to induce cycles. Note that j1,,jkj_{1},\ldots,j_{k} are messages that r1r_{1} possesses as side information that rr does not have in its side information. However, since rr emulates r1r_{1} upon decoding its tt messages, d1,,dk𝒟(r)d_{1},\ldots,d_{k}\in\mathcal{D}(r). Hence, we can now delete the kk nodes (ri,ji)(r_{i}^{\prime},j_{i}) and their incident edges, and add (r,ji)(r,j_{i}), iki\in\llbracket k\rrbracket and their incident edges instead. Let Gapp′′G{{}^{\prime\prime}}_{\textrm{app}} denote the resultant graph.

Thus far, the modifications from GappG_{\textrm{app}} to Gapp′′G_{\textrm{app}}^{{}^{\prime\prime}} have preserved the size and the induced nature of the subgraphs. The last step is to establish its acyclicity. This follows since any cycle in Gapp′′G{{}^{\prime\prime}}_{\textrm{app}} must involve (r1,j)(r_{1},j), back edges and at least one of (r,ji)(r,j_{i})’s. But note that the out-neighbor of (r,ji)(r,j_{i})’s are nodes whose message labels are present in the side information of rr, and such nodes are in layers k\mathcal{L}_{k}, k>k>\ell, and since these layers are untouched, there are no back edges from nodes in these layers back to (r1,j)(r_{1},j). Therefore, Gapp′′G_{\textrm{app}}^{{}^{\prime\prime}} is an MAIS of G𝒟appG_{\mathcal{D}_{\textrm{app}}}, and since the nodes of Gapp′′G_{\textrm{app}}^{{}^{\prime\prime}} are all present in G𝒟G_{\mathcal{D}}, it follows that Gapp′′G_{\textrm{app}}^{{}^{\prime\prime}} is also an MAIS of G𝒟G_{\mathcal{D}}, and therefore, |MAIS(G𝒟app)|=|MAIS(G𝒟)||\texttt{MAIS}(G_{\mathcal{D}_{\textrm{app}}})|=|\texttt{MAIS}(G_{\mathcal{D}})|.

Proof of Lemma 5

We have three cases depending on G=MAIS(G𝒟)G=\texttt{MAIS}(G_{\mathcal{D}}).

Case 1: There are k>tk>t sinks in GG. In this case, since there are only tt node representations per receiver, there has to be at least two distinct sink node receiver labels. It is therefore the case that there are at least s+1s+1 groups, say 𝒢i1,,𝒢is+1\mathcal{G}_{i_{1}},\ldots,\mathcal{G}_{i_{s+1}} whose messages are present in the side information of the receivers represented by the sink nodes. Note that there are t(s+1s)=t(s+1)>s+tit\binom{s+1}{s}=t(s+1)>s+t-i node representations of receivers whose side information is a subset of k=1s+1𝒢ik\cup_{k=1}^{s+1}\mathcal{G}_{i_{k}}. Let (r,j)(r,j) be one such node representation that is absent from MAIS(G𝒟)\texttt{MAIS}(G_{\mathcal{D}}). Consider the subgraph GG^{\prime} induced by vertices already in GG along with (r,j)(r,j). Note that GG^{\prime} is also acyclic, since any cycle that is newly introduced has to traverse (r,j)(r,j); however, the out-degree of (r,j)(r,j) in GG^{\prime} is zero, since neither is XjrX_{j}\in\mathcal{M}_{r} nor is a message label of a node in GG a message in r\mathcal{M}_{r}. Hence, GG^{\prime} is a larger acyclic induced subgraph subsuming GG, which contradicts the maximality of GG. Hence, this case cannot arise.

Case 2: There are k<tk<t sinks in GG. If there are two distinct receiver labels among sink nodes, the union of the side information of sink nodes contains at least s+1s+1 groups, which leads to a contradiction just as in Case 1. Thus, all sink nodes must have the same receiver label. Let (r,j1),,(r,jk)(r,j_{1}),\ldots,(r,j_{k}) denote the sink nodes. Let j𝒟(r){j1,,jk}j^{*}\in\mathcal{D}(r)\setminus\{j_{1},\ldots,j_{k}\}. Pick j^r\hat{j}\notin\mathcal{M}_{r} such that j^\hat{j} does not appear as a message label in GG. There are at least i>0i>0 choices for j^\hat{j}, since the number of messages in r\mathcal{M}_{r} is gsgs, and the number of message labels is s+tig(ms)is+t-i\leq g(m-s)-i. Define a new decoding choice 𝒟^\hat{\mathcal{D}} by:

𝒟^(x)={𝒟(x)xr(𝒟(r){j}){j^}x=r.\hat{\mathcal{D}}(x)=\left\{\begin{array}[]{cc}\mathcal{D}(x)&x\neq r\\ \left(\mathcal{D}(r)\setminus\{j^{*}\}\right)\cup\{\hat{j}\}&x=r\end{array}\right..

The decoding choice 𝒟^\hat{\mathcal{D}} is the same for all receivers, but for rr; and for rr, only one message jj^{*} that possibly appears as a node label in GG is replaced by j^\hat{j} that does not. Note that GG is also an acyclic induced subgraph of G𝒟^G_{\hat{\mathcal{D}}}; however, the subgraph G^\hat{G} induced by vertices already in GG along with (r,j^)(r,\hat{j}) is also an acyclic induced subgraph of G𝒟^G_{\hat{\mathcal{D}}} since (r,j^)(r,\hat{j}) is a sink node in G𝒟^G_{\hat{\mathcal{D}}}. In fact, G^\hat{G} is maximal since changing only one message label for the decoding choice of only one receiver can alter the size of MAIS by at most one. The claim then follows since |G^|=|MAIS(G𝒟^)|=s+ti+1|\hat{G}|=|\texttt{MAIS}(G_{\hat{\mathcal{D}}})|=s+t-i+1.

Case 3: There are precisely tt sinks in GG. In this case, we look at GG through the layering given in Section III. We search GG to find the layer with the smallest index \ell that contains a partially covered group, say 𝒢\mathcal{G}^{*}. Now, four cases arise:

  • 3a

    GG has no partially covered group. In this case, the message labels of nodes in GG is precisely a union of message labels of groups 𝒢i1,,𝒢ik\mathcal{G}_{i_{1}},\ldots,\mathcal{G}_{i_{k}}, i.e., {j:(r,j)G}={j:Xj𝒢i1𝒢ik}\{j:(r,j)\in G\}=\{j:X_{j}\in\mathcal{G}_{i_{1}}\cup\ldots\cup\mathcal{G}_{i_{k}}\}. Then, since kg=|G|=s+tig(ms)ikg=|G|=s+t-i\leq g(m-s)-i, it follows that k<msk<m-s. Then, the mk>sm-k>s groups are absent from the message labels of nodes in GG. As in Case 1, there then are t(s+1s)t\binom{s+1}{s} node representations of receivers whose side information is a subset of these s+1s+1 receivers. Since t(s+1s)>s+tit\binom{s+1}{s}>s+t-i, one can find a node representation that is absent in GG, and add to GG the representation and necessary incident edges to derive a larger acyclic induced subgraph leading then to a contradiction. So, this case does not arise.

  • 3b

    Partially covered group 𝒢\mathcal{G}^{*} is in layer \mathcal{L}_{\ell}, and \mathcal{L}_{\ell} has a non-sink node with a message label in 𝒢\mathcal{G}^{*}. In this case, pick a non-sink node (r,j)(r,j^{*}) in this layer such that Xj𝒢X_{j^{*}}\in\mathcal{G}^{*}. From P3 of Lemma 1, it follows that every node with receiver label rr is also a non-sink node. There cannot be tt nodes in GG with receiver label rr. Otherwise, GG has at least tt non-sink nodes and tt sink nodes, and s+ti=|G|2t>2(g1)(ms)g(ms)s+t-i=|G|\geq 2t>2(g-1)(m-s)\geq g(m-s), which violates the hypothesis s+tg(ms)s+t\leq g(m-s). Here, we have used the fact that g2g\geq 2 if a group is present partially. Let (r,j~)(r,\tilde{j}) be a node in G𝒟G_{\mathcal{D}} that is not in GG.

    Now, pick j^\hat{j} such that Xj^𝒢X_{\hat{j}}\in\mathcal{G}^{*} and j^\hat{j} is not a message label in layer \mathcal{L}_{\ell}. Clearly, j^𝒟(r)\hat{j}\notin\mathcal{D}(r). Since, if it were, the induced subgraph with nodes of GG and (r,j^)(r,\hat{j}) will be acyclic. This is because if there is a cycle involving (r,j^)(r,\hat{j}), there will also be a cycle involving (r,j)(r,j^{*}) as their in- and out-neighborhoods are identical (due to Properties P3 and P4 of Lemma 1), which contradicts the acyclicity of GG. Now, define a new decoding choice 𝒟^\hat{\mathcal{D}} as follows:

    𝒟^(x)={𝒟(x)xr(𝒟(r){j~}){j^}x=r.\hat{\mathcal{D}}(x)=\left\{\begin{array}[]{cc}\mathcal{D}(x)&x\neq r\\ \left(\mathcal{D}(r)\setminus\{\tilde{j}\}\right)\cup\{\hat{j}\}&x=r\end{array}\right..

    Note that 𝒟\mathcal{D} and 𝒟^\hat{\mathcal{D}} differ at most in one decoded message for precisely one receiver alone. Consider G^\hat{G} to be the induced subgraph with nodes of GG and (r,j^)(r,\hat{j}). By the above argument, G^\hat{G} is also an acyclic induced subgraph of G𝒟^G_{\hat{\mathcal{D}}}; G^\hat{G} is maximal since changing only one message label for the decoding choice of only one receiver can alter the size of MAIS by at most one. The claim then follows since |G^|=|MAIS(G𝒟^)|=s+ti+1|\hat{G}|=|\texttt{MAIS}(G_{\hat{\mathcal{D}}})|=s+t-i+1.

  • 3c

    Partially covered group 𝒢\mathcal{G}^{*} is in layer \mathcal{L}_{\ell}, >1\ell>1, and all nodes with message labels in 𝒢\mathcal{G}^{*} are sink nodes.

    Pick a node in layer \mathcal{L}_{\ell}, whose message label is in 𝒢\mathcal{G}^{*}, and select one of its in-neighbor, say (r,j)(r^{*},j^{*}), from layer 1\mathcal{L}_{\ell-1}. By an argument similar to Case 3b, we can infer that: (a) all nodes with receiver label rr^{*} are non-sink nodes, and (b) there cannot be tt node representations of the receiver rr^{*}. Thus, there must be a representation (r,j~)(r^{*},\tilde{j}) in G𝒟G_{\mathcal{D}} that is not in GG. Let j^\hat{j} be a message label such that Xj^𝒢X_{\hat{j}}\in\mathcal{G}^{*}, and j^\hat{j} is not a message label in GG. Note that j^𝒟(r)\hat{j}\notin\mathcal{D}(r^{*}) using the same argument in Case 3b. Now, define a new decoding choice 𝒟^\hat{\mathcal{D}} as follows:

    𝒟^(x)={𝒟(x)xr(𝒟(r){j~}){j^}x=r.\hat{\mathcal{D}}(x)=\left\{\begin{array}[]{cc}\mathcal{D}(x)&x\neq r^{*}\\ \left(\mathcal{D}(r^{*})\setminus\{\tilde{j}\}\right)\cup\{\hat{j}\}&x=r^{*}\end{array}\right..

    Note that 𝒟\mathcal{D} and 𝒟^\hat{\mathcal{D}} differ at most in one decoded message for precisely one receiver alone. Consider the induced subgraph G^\hat{G} with nodes of GG and (r,j^)(r^{*},\hat{j}). G^\hat{G} is acyclic, since (r,j^)(r^{*},\hat{j}) is a sink in G^\hat{G}. Further, G^\hat{G} is maximal since changing only one message label for the decoding choice of only one receiver can alter the size of MAIS by at most one. The claim then follows since |G^|=|MAIS(G𝒟^)|=s+ti+1|\hat{G}|=|\texttt{MAIS}(G_{\hat{\mathcal{D}}})|=s+t-i+1.

  • 3d

    Partially covered group 𝒢\mathcal{G}^{*} is the first layer, and all nodes whose message labels are in 𝒢\mathcal{G}^{*} are sink nodes. As in previous cases, the tt sink nodes must have the same receiver label, say rr^{*}. Let 𝒢i1,,𝒢is\mathcal{G}_{i_{1}},\ldots,\mathcal{G}_{i_{s}} denote the groups that receiver rr^{*} possesses. Now, one of three situations arises:

    • (i)

      Layer 1\mathcal{L}_{1} has only sink nodes. In this case, this layer has tt nodes, and since tms>g1\frac{t}{m-s}>g-1, there must be a group, say 𝒢~\tilde{\mathcal{G}}, that is fully present in layer 1\mathcal{L}_{1}.

    • (ii)

      There is a non-sink node whose message label lies in a group, say 𝒢~\tilde{\mathcal{G}}, that is fully present in layer 1\mathcal{L}_{1}.

    • (iii)

      There is a non-sink node whose message label belongs to a group partially present in this layer. This situation is subsumed under Case 3b.

    There are t(s+1s)>s+tit\binom{s+1}{s}>s+t-i node representations of receivers that have as side information ss groups among 𝒢i1,,𝒢is\mathcal{G}_{i_{1}},\ldots,\mathcal{G}_{i_{s}} and 𝒢~\tilde{\mathcal{G}}. Therefore, one such representation (r,j)(r,j^{*}) is absent from GG. Now, consider a label j^\hat{j} such that Xj^𝒢X_{\hat{j}}\in\mathcal{G}^{*}, but is not a message label in GG. Note that j^𝒟(r)\hat{j}\notin\mathcal{D}(r), since if it were, the induced subgraph consisting of the nodes of GG along with (r,j^)(r,\hat{j}) would be a larger acyclic induced subgraph of G𝒟G_{\mathcal{D}}111This is because (r,j^)(r,\hat{j}) must have the same in-neighborhood as the nodes in 1\mathcal{L}_{1} with message labels in 𝒢\mathcal{G}^{*}, due to Property P4 of Lemma 1. But nodes in 1\mathcal{L}_{1} has no incoming edge, and so neither does (r,j^)(r,\hat{j})., contradicting the fact that GG is an MAIS of G𝒟G_{\mathcal{D}}. Therefore, define a new decoding choice 𝒟^\hat{\mathcal{D}} as follows:

    𝒟^(x)={𝒟(x)xr(𝒟(r){j}){j^}x=r.\hat{\mathcal{D}}(x)=\left\{\begin{array}[]{cc}\mathcal{D}(x)&x\neq r\\ \left(\mathcal{D}(r)\setminus\{j^{*}\}\right)\cup\{\hat{j}\}&x=r\end{array}\right..

    As in other cases, the subgraph G^\hat{G} induced by vertices already in GG along with (r,j^)(r,\hat{j}) is an acyclic induced subgraph graph of G𝒟^G_{\hat{\mathcal{D}}}. As before, G^\hat{G} is maximal thereby establishing the existence of a decoding choice 𝒟^\hat{\mathcal{D}} with |MAIS(G𝒟^)|=s+ti+1|\texttt{MAIS}(G_{\hat{\mathcal{D}}})|=s+t-i+1.

Proof of Theorem 2

By repeated use of Lemma 5, we argue that if there is a decoding choice 𝒟\mathcal{D} for which |MAIS(G𝒟)|<s+t|\texttt{MAIS}(G_{\mathcal{D}})|<s+t, there must exist a decoding choice 𝒟\mathcal{D}^{\prime} for which |MAIS(G𝒟)|=s+t1|\texttt{MAIS}(G_{\mathcal{D}^{\prime}})|=s+t-1. In this theorem, we will establish that there is no such 𝒟\mathcal{D}^{\prime}, thereby establishing that for every valid decoding choice 𝒟\mathcal{D}, |MAIS(G𝒟)|s+t|\texttt{MAIS}(G_{\mathcal{D}})|\geq s+t. The proof is then complete by the achievability result in Theorem 1.

Let us assume the existence of a decoding choice 𝒟\mathcal{D}^{\prime} for which |MAIS(G𝒟)|=s+t1|\texttt{MAIS}(G_{\mathcal{D}^{\prime}})|=s+t-1. From Corollary 4, it follows that there are precisely ss groups absent from MAIS(G𝒟)\texttt{MAIS}(G_{\mathcal{D}^{\prime}}); these are precisely the side information of the receiver label of the sink nodes. Let 1,,s\mathcal{F}_{1},\ldots,\mathcal{F}_{s} denote these ss groups. Note that MAIS(G𝒟)\texttt{MAIS}(G_{\mathcal{D}^{\prime}}) has exactly s+t1=g(ms)1s+t-1=g(m-s)-1 distinct message labels. So there is precisely one group partially present in MAIS(G𝒟)\texttt{MAIS}(G_{\mathcal{D}^{\prime}}). Let s+1\mathcal{F}_{s+1} denote this group. Note that g1g-1 message indices from s+1\mathcal{F}_{s+1} are present as message labels in MAIS(G𝒟)\texttt{MAIS}(G_{\mathcal{D}^{\prime}}). Let the remaining ms1m-s-1 groups be denoted by s+2,,m.\mathcal{F}_{s+2},\ldots,\mathcal{F}_{m}. By hypothesis,

s=s+tt<g(ms)(g1)(ms)=(ms).s=s+t-t<g(m-s)-(g-1)(m-s)=(m-s).

Therefore, 2s+1m2s+1\leq m. Construct an (s+1)×(2s+1)(s+1)\times(2s+1) binary matrix BB that is tied to 𝒟\mathcal{D}^{\prime} as follows:

  • For is+1i\in\llbracket s+1\rrbracket, row ii corresponds to receiver with side information 1,,i1,i+1,,s+1\mathcal{F}_{1},\ldots,\mathcal{F}_{i-1},\mathcal{F}_{i+1},\ldots,\mathcal{F}_{s+1} (i.e., the first s+1s+1 groups but i\mathcal{F}_{i}). We let Bi,j=1B_{i,j}=1 and Bi,i=0B_{i,i}=0 for is+1i\in\llbracket s+1\rrbracket and js+1{i}j\in\llbracket s+1\rrbracket\setminus\{i\}.

  • For is+1i\in\llbracket s+1\rrbracket, and j=s+1,,2s+1j=s+1,\ldots,2s+1, we set Bi,j=1B_{i,j}=1 if the receiver corresponding to this row decodes all messages in group j\mathcal{F}_{j} by using its decoding function or by repeatedly emulating other receivers.

Now, an application of Lemma 6 to the last ss columns of BB yields a non-empty subset Ps+1P\subseteq\llbracket s+1\rrbracket of rows of BB with the following property:

  • |𝒥|=|P|1|\mathcal{J}|=|P|-1, where 𝒥={j{s+2,,2s+1}:Bi,j=1 for all iP}\mathcal{J}=\{j\in\{s+2,\ldots,2s+1\}:B_{i,j}=1\textrm{ for all }i\in P\}, i.e., 𝒥\mathcal{J} collects the indices of groups s+2,,2s+1\mathcal{F}_{s+2},\ldots,\mathcal{F}_{2s+1} that can be fully decoded by each receiver corresponding to row iPi\in P.

Now consider the receiver rr^{*} with side information {k:k𝒥(s+1P)}.\left\{\mathcal{F}_{k}:k\in\mathcal{J}\cup\big{(}\llbracket s+1\rrbracket\setminus P\big{)}\right\}. Clearly, this is a bona fide receiver since it has s+1|P|+|P|1=ss+1-|P|+|P|-1=s groups as side information. Now, consider the following facts and reasoning:

  • F5

    By construction, the receiver corresponding to the row iPi\in P decodes all messages in group k\mathcal{F}_{k}, k𝒥k\in\mathcal{J}, and therefore can emulate rr^{*}.

  • F6

    rr^{*} decodes at most (g1)(g-1) messages from each group k\mathcal{F}_{k} where k{s+2,,2s+1}𝒥k\in\{s+2,\ldots,2s+1\}\setminus\mathcal{J}. If rr^{*} decodes all messages of any such group by emulating, then by fact F1, every receiver corresponding to the row iPi\in P will have also decoded this group, which would necessitate that the index of this group be present in 𝒥\mathcal{J}, a contradiction. This is at most (g1)(s+1|P|)(g-1)(s+1-|P|) messages.

  • F7

    Among the other groups, i.e., k\mathcal{F}_{k} for k{2s+2,,m}k\in\{2s+2,\ldots,m\}, rr^{*} can decode at most g(m(2s+2)+1)=g(m2s1)g(m-(2s+2)+1)=g(m-2s-1) messages.

Combining facts F6 and F7, we see that rr^{*} can decode at most

(g1)(s+1|P|)+g(m2s1)\displaystyle(g-1)(s+1-|P|)+g(m-2s-1)
=g(ms)s1(g1)|P|\displaystyle=g(m-s)-s-1-(g-1)|P|
=t1(g1)|P|\displaystyle=t-1-(g-1)|P|

messages from k\mathcal{F}_{k}, k{s+2,,m}𝒥k\in\{s+2,\ldots,m\}\setminus\mathcal{J}. Since rr^{*} must decode at least tt messages, rr^{*} must decode (g1)|P|+1(g-1)|P|+1 messages from the groups k\mathcal{F}_{k}, kPk\in P. Note that these |P||P| groups are precisely those missing from its side information. Thus, rr^{*} must decode, on average, (g1)+1|P|(g-1)+\frac{1}{|P|} messages per group from groups k\mathcal{F}_{k}, kPk\in P. Thus, rr^{*} must decode all messages from a group, say i\mathcal{F}_{i^{*}} with iPi^{*}\in P. Now, consider the receiver r^\hat{r} that has as side information 1,,i1,i+1,,s+1\mathcal{F}_{1},\ldots,\mathcal{F}_{i^{*}-1},\mathcal{F}_{i^{*}+1},\ldots,\mathcal{F}_{s+1}; this receiver corresponds to row iPi^{*}\in P. By fact F5, r^\hat{r} can decode i\mathcal{F}_{i^{*}} by emulating rr^{*}. Now let generalized decoding choice 𝒟app\mathcal{D}^{\prime}_{\textrm{app}} be defined by

𝒟app(x)={𝒟(x)xr^𝒟(x){j:ji}x=r^.\displaystyle\mathcal{D}^{\prime}_{\textrm{app}}(x)=\left\{\begin{array}[]{cc}\mathcal{D}^{\prime}(x)&x\neq\hat{r}\\ \mathcal{D}^{\prime}(x)\cup\{j:j\in\mathcal{F}_{i^{*}}\}&x=\hat{r}\end{array}\right..

Consider an induced subgraph of 𝒟app\mathcal{D}^{\prime}_{\textrm{app}} defined as follows:

  • 1.

    Start with MAIS(G𝒟)\texttt{MAIS}(G_{\mathcal{D}^{\prime}}) that has s+t1s+t-1 nodes.

  • 2.

    Identify g1g-1 nodes from this subgraph whose message labels belong to the message indices in the partially present group s+1\mathcal{F}_{s+1}. Delete these g1g-1 nodes and their incident edges.

  • 3.

    Add gg nodes (r^,j)(\hat{r},j), jij\in\mathcal{F}_{i^{*}} to the subgraph. Add any incoming/outgoing edges to make this graph an induced subgraph of G𝒟appG_{\mathcal{D}^{\prime}_{\textrm{app}}}.

Note that the newly added nodes are sinks in the resultant graph because the message labels originally did not contain any labels that were indices of messages in 1,,s\mathcal{F}_{1},\ldots,\mathcal{F}_{s}, and those that were indices of messages in s+1\mathcal{F}_{s+1} were deleted in step 2. Hence, the modification does not induce any cycles. It then follows that the resultant graph, which has s+ts+t nodes is an acyclic induced subgraph of G𝒟appG_{\mathcal{D}^{\prime}_{\textrm{app}}}. Thus, it follows from Corollary 3 that |MAIS(𝒟)|=|MAIS(𝒟app)|=s+t|\texttt{MAIS}(\mathcal{D}^{\prime})|=|\texttt{MAIS}(\mathcal{D}^{\prime}_{\textrm{app}})|=s+t, which is a contradiction. Therefore, no decoding choice 𝒟\mathcal{D}^{\prime} has |MAIS(𝒟)|=s+t1|\texttt{MAIS}(\mathcal{D}^{\prime})|=s+t-1. Therefore, we infer that

R(m,s,g,t)min𝒟|MAIS(𝒟)|s+t=g(ms).\displaystyle R^{*}(m,s,g,t)\geq\min_{\mathcal{D}}|\texttt{MAIS}(\mathcal{D})|\geq s+t=g(m-s).

The achievability follows from Theorem 1.

Proof of Theorem 3

Just as in Theorem 2, the goal here is to show that there is no decoding choice 𝒟\mathcal{D} such that |MAIS(𝒟)|=s+t1|\texttt{MAIS}(\mathcal{D})|=s+t-1. Let us therefore assume 𝒟\mathcal{D} is such that |MAIS(𝒟)|=s+t1|\texttt{MAIS}(\mathcal{D})|=s+t-1. From Corollary 4 it follows that there are precisely ss groups absent from MAIS(G𝒟)\texttt{MAIS}(G_{\mathcal{D}}); these are precisely the side information of the receiver label of the sink nodes. Let 𝒜1,,𝒜s\mathcal{A}_{1},\ldots,\mathcal{A}_{s} denote these ss groups. It then follows that there are msm-s partially or fully present groups. Since |MAIS(𝒟)|=s+t1|\texttt{MAIS}(\mathcal{D})|=s+t-1, there must be at least one partially present group. Let’s call this 𝒫1\mathcal{P}_{1}. Since s+t1s+(g1)(ms)s+t-1\geq s+(g-1)(m-s), at least ss groups must be fully present in MAIS(G𝒟)\texttt{MAIS}(G_{\mathcal{D}}). Let 1,,s\mathcal{F}_{1},\ldots,\mathcal{F}_{s} denote these groups. Let the remaining m2s1m-2s-1 groups be 𝒫2,,𝒫m2s\mathcal{P}_{2},\ldots,\mathcal{P}_{m-2s}. Let yiy_{i}, i=1,,m2si=1,\ldots,m-2s denote the number of message indices in group 𝒫i\mathcal{P}_{i} that are missing from the message labels in MAIS(G𝒟)\texttt{MAIS}(G_{\mathcal{D}}). By choice, y1>0y_{1}>0, and

k=1m2syk=g(ms)(s+ti).\displaystyle\sum_{k=1}^{m-2s}y_{k}=g(m-s)-(s+t-i). (1)

Now, just as in Theorem 2, let us define an (s+1)×(2s+1)(s+1)\times(2s+1) binary matrix BB that is tied to 𝒟\mathcal{D} as follows:

  • For isi\in\llbracket s\rrbracket, row ii corresponds to receiver with side information 𝒜1,,𝒜i1,𝒜i+1,,𝒜s\mathcal{A}_{1},\ldots,\mathcal{A}_{i-1},\mathcal{A}_{i+1},\ldots,\mathcal{A}_{s} and 𝒫1\mathcal{P}_{1}. We let Bi,j=1B_{i,j}=1 and Bi,i=0B_{i,i}=0 for isi\in\llbracket s\rrbracket and js+1{i}j\in\llbracket s+1\rrbracket\setminus\{i\}.

  • Row s+1s+1 corresponds to receiver with side information 𝒜1,,𝒜s\mathcal{A}_{1},\ldots,\mathcal{A}_{s}. We let Bs+1,j=1B_{s+1,j}=1 iff jsj\in\llbracket s\rrbracket.

  • For is+1i\in\llbracket s+1\rrbracket, and j>s+1j>s+1, we set Bi,j=1B_{i,j}=1 if the receiver corresponding to this row decodes all messages in group j(s+1)\mathcal{F}_{j-(s+1)} by using its decoding function or by repeatedly emulating other receivers.

Now, an application of Lemma 6 to the last ss columns of BB will yield a non-empty subset Ps+1P\subseteq\llbracket s+1\rrbracket of rows of BB with the following property:

  • |𝒥|=|P|1|\mathcal{J}|=|P|-1, where 𝒥={js:Bi,j+s+1=1 for all iP}\mathcal{J}=\{j\in\llbracket s\rrbracket:B_{i,j+s+1}=1\textrm{ for all }i\in P\}, i.e., 𝒥\mathcal{J} collects the indices of groups 1,,s\mathcal{F}_{1},\ldots,\mathcal{F}_{s} that can be fully decoded by each receiver corresponding to row iPi\in P.

Now, if s+1Ps+1\in P, define rr^{*} to be the receiver with side information 𝒜i\mathcal{A}_{i} for isPi\in\llbracket s\rrbracket\setminus P, and j\mathcal{F}_{j} for j𝒥j\in\mathcal{J}; and if s+1Ps+1\notin P, define rr^{*} to be the receiver with side information 𝒫1\mathcal{P}_{1}, 𝒜i\mathcal{A}_{i} for isPi\in\llbracket s\rrbracket\setminus P, and j\mathcal{F}_{j} for j𝒥j\in\mathcal{J}. Clearly, this is a bona fide receiver since it has s+1|P|+|P|1=ss+1-|P|+|P|-1=s groups as side information. Now, consider the following:

  • F8

    By construction, every receiver corresponding to the row iPi\in P recovers all messages in the groups in 𝒥\mathcal{J}, and therefore can emulate rr^{*}.

  • F9

    rr^{*} decodes at most (g1)(g-1) messages from each group k\mathcal{F}_{k} for ks𝒥k\in\llbracket s\rrbracket\setminus\mathcal{J}. If rr^{*} decodes all messages of any such group by emulating, then by fact F8, every receiver corresponding to the row iPi\in P will have also decoded this group, which would necessitate that the index of this group be present in 𝒥\mathcal{J}, a contradiction. So rr^{*} can decode at most (g1)(g-1) messages from each group k\mathcal{F}_{k} for ks𝒥k\in\llbracket s\rrbracket\setminus\mathcal{J}. This is at most (g1)(s+1|P|)(g-1)(s+1-|P|) messages.

We now divide the proof into two cases.

  • Case 1: s+1Ps+1\in P. In this case, rr^{*} has 𝒜i\mathcal{A}_{i} for iPi\in P, and j\mathcal{F}_{j} for j𝒥j\in\mathcal{J} as side information. Let us count the number of messages rr^{*} can decode from messages in 1,,s\mathcal{F}_{1},\ldots,\mathcal{F}_{s} and only those message indices in 𝒫1,,𝒫m2s\mathcal{P}_{1},\ldots,\mathcal{P}_{m-2s} that are present in |MAIS(𝒟)||\texttt{MAIS}(\mathcal{D})|. The former computation is already given in fact F9. The latter is at most k=1m2s(gyi).\sum_{k=1}^{m-2s}(g-y_{i}). Together, the sum comes to

    (s+1|P|)(g1)\displaystyle(s+1-|P|)(g-1) +i=1m2s(gyi)\displaystyle+\sum_{i=1}^{m-2s}(g-y_{i})
    =(1)t(1+(|P|1)(g1)).\displaystyle\stackrel{{\scriptstyle\eqref{eqn:ysum}}}{{=}}t-\Big{(}1+(|P|-1)(g-1)\Big{)}.

    Since rr^{*} must decode tt messages, rr^{*} must decode 1+(|P|1)(g1)11+(|P|-1)(g-1)\geq 1 messages from 𝒜i\mathcal{A}_{i} for iPsi\in P\cap\llbracket s\rrbracket or from the message indices in 𝒫1,,𝒫m2s\mathcal{P}_{1},\ldots,\mathcal{P}_{m-2s} that are missing in MAIS(𝒟)\texttt{MAIS}(\mathcal{D}). Now, two cases arise:

    • 1a:

      rr^{*} decodes a message Xj^X_{\hat{j}} that is in 𝒫1,,𝒫m2s\mathcal{P}_{1},\ldots,\mathcal{P}_{m-2s} but is missing in MAIS(𝒟)\texttt{MAIS}(\mathcal{D}). Since in this setting s+1Ps+1\in P, by fact F8, the receiver rsinkr_{\textrm{sink}} with side information 𝒜1,,𝒜s\mathcal{A}_{1},\ldots,\mathcal{A}_{s} can mimic rr^{*} to also decode j^\hat{j}. Then, we can append j^\hat{j} to 𝒟(rsink)\mathcal{D}(r_{\textrm{sink}}) to obtain a generalized decoding choice DD^{*}. The induced subgraph of G𝒟G_{\mathcal{D}^{*}} consisting of the nodes of MAIS(𝒟)\texttt{MAIS}(\mathcal{D}) along with (rsink,j^)(r_{\textrm{sink}},\hat{j}) has a size of s+ts+t and is also acyclic. Therefore, it follows that |MAIS(𝒟)|s+t|\texttt{MAIS}(\mathcal{D}^{*})|\geq s+t. However, by Corollary 3, s+t|MAIS(𝒟)|=|MAIS(𝒟)|=s+t1s+t\leq|\texttt{MAIS}(\mathcal{D})|=|\texttt{MAIS}(\mathcal{D}^{*})|=s+t-1, which is a contradiction.

    • 1b:

      rr^{*} decodes no messages in 𝒫1,,𝒫m2s\mathcal{P}_{1},\ldots,\mathcal{P}_{m-2s} that are missing in MAIS(𝒟)\texttt{MAIS}(\mathcal{D}). In this case, rr^{*} must decode 1+(|P|1)(g1)1+(|P|-1)(g-1) messages from the |P|1|P|-1 groups not in its side information, i.e., from 𝒜i\mathcal{A}_{i} for iPsi\in P\cap\llbracket s\rrbracket. This would mean that it must decode at least one group, say 𝒜i\mathcal{A}_{i^{*}} fully by emulating other nodes. By invoking F8, we can argue that receiver r^\hat{r} with side information 𝒜1,,𝒜i1,𝒜i+1,𝒜s,𝒫1\mathcal{A}_{1},\ldots,\mathcal{A}_{i^{*}-1},\mathcal{A}_{i^{*}+1},\mathcal{A}_{s},\mathcal{P}_{1} can mimic rr^{*} and decode messages in 𝒜i\mathcal{A}_{i^{*}} by successive emulation. In this case, let generalized decoding choice 𝒟\mathcal{D}^{*} append to 𝒟\mathcal{D} the gg new messages of 𝒜i\mathcal{A}_{i^{*}} that r^\hat{r} decodes. Now, consider the induced subgraph of G𝒟G_{\mathcal{D}^{*}} obtained by deleting from MAIS(𝒟)\texttt{MAIS}(\mathcal{D}) at most g1g-1 nodes whose message labels belong to 𝒫1\mathcal{P}_{1}, and then adding gg nodes {(r^,k):Xk𝒜i}\{(\hat{r},k):X_{k}\in\mathcal{A}_{i^{*}}\} (and any incident edges). This subgraph is acyclic and has a size of at least s+ts+t. Again, by Corollary 3, s+t|MAIS(𝒟)|=|MAIS(𝒟)|=s+t1s+t\leq|\texttt{MAIS}(\mathcal{D})|=|\texttt{MAIS}(\mathcal{D}^{*})|=s+t-1, which is a contradiction.

  • Case 2: s+1Ps+1\notin P. In this case, rr^{*} has 𝒫1\mathcal{P}_{1}, 𝒜i\mathcal{A}_{i} for iPi\in P, and j\mathcal{F}_{j} for j𝒥j\in\mathcal{J} as side information. As before, let us count the number of messages rr^{*} can decode from messages in 1,,s\mathcal{F}_{1},\ldots,\mathcal{F}_{s} and only those message indices in 𝒫2,,𝒫m2s\mathcal{P}_{2},\ldots,\mathcal{P}_{m-2s} that are present in MAIS(𝒟)\texttt{MAIS}(\mathcal{D}). The former computation is already given in fact F9. The latter is given by k=2m2s(gyi).\sum_{k=2}^{m-2s}(g-y_{i}). Together, the sum comes to

    (s+1|P|)(g1)\displaystyle(s+1-|P|)(g-1) +i=2m2s(gyi)\displaystyle+\sum_{i=2}^{m-2s}(g-y_{i})
    =(1)t((g1)|P|+2y1).\displaystyle\stackrel{{\scriptstyle\eqref{eqn:ysum}}}{{=}}t-\big{(}(g-1)|P|+2-y_{1}\big{)}.

    Thus, rr^{*} has to decode (g1)|P|+2y1(g-1)|P|+2-y_{1} messages from 𝒜i\mathcal{A}_{i} for iPi\in P or from the message indices in 𝒫1,,𝒫m2s\mathcal{P}_{1},\ldots,\mathcal{P}_{m-2s} that are missing in MAIS(𝒟)\texttt{MAIS}(\mathcal{D}). Suppose that rr^{*} decodes xx messages whose indices are in 𝒫1,,𝒫m2s\mathcal{P}_{1},\ldots,\mathcal{P}_{m-2s}, but are missing in MAIS(𝒟)\texttt{MAIS}(\mathcal{D}). Then, rr^{*} decodes (g1)|P|+2y1x(g-1)|P|+2-y_{1}-x messages from 𝒜i\mathcal{A}_{i}, iPi\in P. Thus, there must exist a group 𝒜i\mathcal{A}_{i^{*}}, iPi^{*}\in P such that rr^{*} decodes at least (g1)|P|+2y1x|P|\frac{(g-1)|P|+2-y_{1}-x}{|P|} messages in 𝒜i\mathcal{A}_{i^{*}}.

    From F8, we see that the receiver r^\hat{r} with side information 𝒜1,,𝒜i1,𝒜i+1,𝒜s,𝒫1\mathcal{A}_{1},\ldots,\mathcal{A}_{i^{*}-1},\mathcal{A}_{i^{*}+1},\ldots\mathcal{A}_{s},\mathcal{P}_{1} emulates rr^{*} to decode

    (g1)|P|+2y1x|P|+x>gy1\displaystyle\frac{(g-1)|P|+2-y_{1}-x}{|P|}+x>g-y_{1}

    since (|P|1)(x+y1)|P|+2>0(|P|-1)(x+y_{1})-|P|+2>0 because x+y1>0x+y_{1}>0. Thus, we see that the receiver r^\hat{r} decodes at least gy1+1g-y_{1}+1 messages, say Xk1,,Xkgy1+1X_{k_{1}},\ldots,X_{k_{g-y_{1}+1}}, that are not part of 𝒟(r^)\mathcal{D}(\hat{r}). Consider the generalized decoding choice 𝒟\mathcal{D}^{*} by appending to 𝒟\mathcal{D} the new messages that r^\hat{r} decodes. Next, consider the induced subgraph of G𝒟G_{\mathcal{D}^{*}} obtained by deleting from MAIS(𝒟)\texttt{MAIS}(\mathcal{D}) gy1g-y_{1} nodes whose message labels belong to 𝒫1\mathcal{P}_{1}, and then adding gy1+1g-y_{1}+1 nodes {(r^,ki):1igy1+1}\{(\hat{r},k_{i}):1\leq i\leq g-y_{1}+1\} (and any incident edges). This subgraph is acyclic and has a size of s+ts+t. Again, by Corollary 3, it follows that s+t|MAIS(𝒟)|=|MAIS(𝒟)|=s+t1s+t\leq|\texttt{MAIS}(\mathcal{D})|=|\texttt{MAIS}(\mathcal{D}^{*})|=s+t-1, which is a contradiction.

Thus, in all cases, we reach a contradiction, which implies that our assumption of the existence of a decoding choice DD with |MAIS(𝒟)|=s+t1|\texttt{MAIS}(\mathcal{D})|=s+t-1 is untenable. Therefore,

R(m,s,g,t)min𝒟|MAIS(𝒟)|s+t.\displaystyle R^{*}(m,s,g,t)\geq\min_{\mathcal{D}}|\texttt{MAIS}(\mathcal{D})|\geq s+t.

The proof is complete due to the achievability in Theorem 1.

Proof of Theorem 4

As in the proof of Proposition 7 of [15], we choose α=s+tg(ms)\alpha=s+t-g(m-s). Note that 0<αs0<\alpha\leq s. Suppose we look at a subproblem of the PICOD problem where we focus on delivering tt messages to receivers that have 𝒢1,,𝒢α\mathcal{G}_{1},\ldots,\mathcal{G}_{\alpha} and any sαs-\alpha other groups as side information. This subproblem has (mαsα)\binom{m-\alpha}{s-\alpha} receivers. Note that a lower bound on the optimal rate of this subproblem is also a lower rate for R(m,s,g,t).R^{*}(m,s,g,t). Further, any code for the subproblem need only focus on the mαm-\alpha groups that are not common to all receivers. Hence, this is effectively an (mα,sα,g,t)(m-\alpha,s-\alpha,g,t) PICOD problem. But note that (sα)+t=g((mα)(sα))(s-\alpha)+t=g((m-\alpha)-(s-\alpha)) and t>(g1)((mα)(sα))t>(g-1)((m-\alpha)-(s-\alpha)). Hence, from Theorem 2, we see that a lower bound to this sub-problem, and hence a lower bound for R(m,s,g,t)R^{*}(m,s,g,t) is given by g(ms)g(m-s). The upper bound follows from the achievability in Theorem 1.

Proof of Theorem 5

From Theorem 1, it is clear that by sending f1f-1 MDS sequentially at a rate of msm-s each followed by either an uncoded transmission of s+(t(f1)(ms))s+(t-(f-1)(m-s)) messages (when s+t<f(ms)s+t<f(m-s)) or another MDS code at a rate of msm-s (when s+tf(ms)s+t\geq f(m-s)) suffices to meet all receiver demands. Hence, the optimal broadcast rate among codes from 𝒞m,s,g,t\mathcal{C}_{m,s,g,t} is no more than f(ms)f(m-s).

Consider an asymptotically vanishing error (lossless) variant of the PICOD problem where the goal is to jointly encode nn i.i.d. copies of the mgmg messages to deliver the nn realizations of some tt messages to each receiver. By a simple information-theoretic argument, we can establish that the optimal rate (normalized by message size) for this relaxed variant when restricted only to broadcast codes is identical to the maximum number of messages any receiver recovers. Additionally, from Theorem 2 of [16], it follows that the optimal rate of the relaxed variant of the PICOD problem and optimal rate among codes from 𝒞m,s,g,t\mathcal{C}_{m,s,g,t} for the zero-error PICOD formulation in this work are identical. Thus, the optimal rate among codes from 𝒞m,s,g,t\mathcal{C}_{m,s,g,t} for the PICOD formulation in this work is identical to the maximum number of messages communicated to any receiver.

Now, let CC be a code of optimal rate from this family. For this code, let the encoder communicate 0aig0\leq a_{i}\leq g messages from the group 𝒢i\mathcal{G}_{i}, imi\in\llbracket m\rrbracket. Without loss of generality, we can relabel the groups so that a1a2ama_{1}\geq a_{2}\geq\cdots\geq a_{m}. Consider the receiver that has the last ss groups and therefore decodes the most messages among all receivers. Then, this receiver decodes a1+a2++amsa_{1}+a_{2}+\cdots+a_{m-s} messages. Hence CC’s broadcast rate must equal j=1msaj\sum_{j=1}^{m-s}a_{j}. From the achievability argument, we see that i=1msaiR(m,s,g,t)f(ms)\sum_{i=1}^{m-s}a_{i}\leq R^{\prime}(m,s,g,t)\leq f(m-s). Further, the least number of messages that any receiver decodes is by the receiver that has the first ss groups (after relabeling) as side information, and therefore, i=s+1mait\sum_{i=s+1}^{m}a_{i}\geq t.

If a1am>1a_{1}-a_{m}>1, then consider another assignment that conveys b1=a11b_{1}=a_{1}-1, bi=aib_{i}=a_{i} for 1<i<m1<i<m, and bm=am+1b_{m}=a_{m}+1 messages from each group. The least number of messages that any receiver decodes under CC^{\prime} is the sum of the msm-s smallest bb’s; this sum is either as,,am1a_{s},\ldots,a_{m-1} or as+1,,am1,am+1a_{s+1},\ldots,a_{m-1},a_{m}+1. In any case, the sum is larger than that of CC. Therefore, every receiver will be satisfied under CC^{\prime} as well. Further, the most number of messages that any receiver decodes is given by the sum of msm-s largest bb’s, which is either a11,a2,,amsa_{1}-1,a_{2},\ldots,a_{m-s} or a2,,ams+1a_{2},\ldots,a_{m-s+1}. In either case, the sum of msm-s largest bb’s is at most i=1msait\sum_{i=1}^{m-s}a_{i}\geq t. We can relabel the groups to sort the bb’s in non-increasing order, and repeat this approach of balancing where we attempt to reduce/increase the maximum/minimum number of messages conveyed from a group, respectively. This balancing will end with a configuration that conveys α1α2αm\alpha_{1}\geq\alpha_{2}\geq\cdots\geq\alpha_{m} messages from each group (after relabeling) where α1αm1\alpha_{1}-\alpha_{m}\leq 1. Thus, we see that

i=1msαii=1msaiR(m,s,g,t)f(ms).\sum_{i=1}^{m-s}\alpha_{i}\leq\sum_{i=1}^{m-s}a_{i}\leq R^{\prime}(m,s,g,t)\leq f(m-s).

Thus it follows that αi{f1,f}\alpha_{i}\in\{f-1,f\} for each imi\in\llbracket m\rrbracket. Since no more than ff messages are conveyed per group, we can remove/ignore gfg-f messages per group, and consider an ff-group complete-{s}\{s\} PICOD(tt) problem with mm groups. Since tt now meets the hypothesis of Theorems 3 and 4, we see that the lowest rate for any code (broadcast or not) is at least min{s+t,f(ms)}\min\{s+t,f(m-s)\}. Therefore, it follows that

min{s+t,f(ms)}j=1msαiR(m,s,g,t)\min\{s+t,f(m-s)\}\leq\sum_{j=1}^{m-s}\alpha_{i}\leq R^{\prime}(m,s,g,t)

The achievability follows from Theorem 1 by noting that the code construction results in a broadcast code.

A Technical Result

The following is an equivalent formulation of Lemma 4 of [15] stated here without proof. It is invoked in the proofs of Theorems 2 and 3.

Lemma 6

Let BB be an (s+1)×s(s+1)\times s binary matrix for some s>0s>0. Then there exists a non-empty subset Ps+1P\subseteq\llbracket s+1\rrbracket such that {j:Bi,j=1 for every iP}=|P|1\{j:B_{i,j}=1\textrm{ for every }i\in P\}=|P|-1.