You are on page 1of 6

Information Leakage in Index Coding

Yucheng Liu† , Lawrence Ong† , Phee Lep Yeoh§ , Parastoo Sadeghi‡ , Joerg Kliewer∗ , and Sarah Johnson† ,
† TheUniversity of Newcastle, Australia (emails: {yucheng.liu, lawrence.ong, sarah.johnson}@newcastle.edu.au)
§ University of Sydney, Australia (email: phee.yeoh@sydney.edu.au)
‡ University of New South Wales, Canberra, Australia (email: p.sadeghi@unsw.edu.au)
∗ New Jersey Institute of Technology, USA (email: jkliewer@njit.edu)

Abstract—We study the information leakage to a guessing Receiver 1


arXiv:2205.10821v1 [cs.IT] 22 May 2022

X1 =?
adversary in index coding with a general message distribution. knows X2
X1 Y = X1 ⊕ X2
Under both vanishing-error and zero-error decoding assump- Server
tions, we develop lower and upper bounds on the optimal X2
leakage rate, which are based on the broadcast rate of the Receiver 2 X2 =?
knows X1
subproblem induced by the set of messages the adversary tries
to guess. When the messages are independent and uniformly
distributed, the lower and upper bounds match, establishing an Guessing
equivalence between the two rates. (X1, X2) =?
adversary

I. I NTRODUCTION Figure 1. There are two correlated binary messages X{1,2} =


(X1 , X2 ) with distribution: PX{1,2} (0, 0) = 0.1, PX{1,2} (0, 1) = 0.2,
Index coding [1], [2] studies the communication problem PX{1,2} (1, 0) = 0.3, and PX{1,2} (1, 1) = 0.4. A binary codeword Y is
where a server broadcasts messages via a noiseless channel generated by the server as Y = X1 ⊕ X2 and broadcast to the receivers.
to multiple receivers with side information. Due to its simple Every receiver can decode its wanted message based on the codeword
yet fundamental model, index coding has been recognized and its side information. An adversary eavesdrops the codeword Y and
makes a single guess on X{1,2} . If Y = 0, the adversary’s guess will
as a canonical problem in network information theory, and is be (1, 1) as PX{1,2} |Y (1, 1|0) = 0.8 > PX{1,2} |Y (0, 0|0) = 0.2 >
closely connected with many other problems such as network PX{1,2} |Y (0, 1|0) = PX{1,2} |Y (1, 0|0) = 0. Similarly, if Y = 1, the
coding, distributed storage, and coded caching. Despite sub- adversary’s guess will be (1, 0).
stantial progress achieved so far (see [3] and the references
therein), the index coding problem remains open in general. Recently we have studied [15] information leakage to a
In secure index coding [4]–[7], the server must simulta- guessing adversary in zero-error source coding defined by
neously satisfy the legitimate receivers’ decoding require- a family of confusion graphs [16]. While the index coding
ments and protect the content of some messages from being problem can also be characterized by a confusion graph
obtained by an eavesdropping adversary. A variant of this family [17], the study of information leakage in index coding
setup puts security constraints on the receivers themselves is intrinsically different from that of source coding in the
against some messages [4], [8], [9]. Instead of protecting following aspects. The most significant difference comes
the messages, another variant of index coding has also been from the different internal structures within the two confusion
studied from a privacy-preserving perspective, where the goal graph families. More specifically, for the source coding model
is to limit the information that a receiver can infer about the we considered [15], the relationship among the confusion
identities of the requests of other receivers [10]. The privacy- graphs of different sequence lengths is characterized by the
utility tradeoff in a multi-terminal data publishing problem disjunctive product [18]. On the other hand, for the index
inspired by index coding was investigated in [11]. coding problem, the relationship among the confusion graphs
In this work, we study the information leakage to a cannot be characterized by any previously defined graph
guessing adversary in index coding, which, to the best of product. Another difference is that while our previous work
our knowledge, has not been considered in the literature. [15] requires zero-error decoding at the legitimate receiver
The adversary eavesdrops the broadcast codeword and tries to assuming worst-case source distribution, this paper considers
guess the message tuple via maximum likelihood estimation both zero-error and vanishing-error scenarios and assume a
within a certain number of trials. Our aim is to characterize general message distribution. Furthermore, in this work we
the information leakage to the adversary, which is defined take into account the adversary’s side information which can
as the ratio between the adversary’s probability of successful include any message in the system.
guessing after and before observing the codeword [12]–[14]. Our main contribution is developing lower and upper
For a visualization of the problem setup, see Figure 1. bounds (i.e., converse and achievability results) on the op-
timal information leakage rate, for both vanishing-error and
This work was supported by the ARC Discovery Scheme DP190100770;
the US National Science Foundation Grant CNS-1815322; and the ARC zero-error scenarios. The converse bound is derived using
Future Fellowship FT190100429. graph-theoretic techniques based on the notion of confusion
graphs for index coding [17]. The achievability result is • n deterministic decoders g = (gi , i ∈ [n]), one for each
established by constructing a deterministic coding scheme receiver i ∈ [n], such that gi : {1, 2, . . . , M }×X t|Ai | →
as a composite of the coding schemes for two subproblems, X t maps the codeword y and the side information xtAi
one induced by the messages the adversary knows as side to some estimated sequence x̂ti .
information and the other induced by the messages the For any  > 0, we say a (t, M, f, g) index code is valid
adversary does not know and thus tries to guess. (with respect to ) if and only if (iff) the average probability
Moreover, we show that when the messages are uniformly .
of error satisfies Pe = P{(X̂[n] ) 6= (X[n] )} ≤ . We say a
distributed and independent of each other (as in most exist- compression rate R is achievable iff for every  > 0, there
ing works for index coding), the lower and upper bounds exists a valid (t, M, f, g) code such that R ≥ (log M )/t.
developed match. This establishes an equivalence between
The optimal compression rate R, also referred to as the
the optimal leakage rate of the problem and the optimal
broadcast rate, can be defined as
compression rate of the subproblem induced by the messages
log M
the adversary tries to guess. R = lim lim inf . (1)
→0 t→∞ valid (t, M, f, g) code t
II. S YSTEM M ODEL AND P ROBLEM F ORMULATION We say a (t, M, f, g) code is valid with respect to zero-
. error decoding iff the average probability of error is zero.
Notation: For any a ∈ Z+ , [a] = {1, 2, · · · , a}. For any
discrete random variable Z with probability distribution PZ , The zero-error broadcast rate ρ can then be defined as
we denote its alphabet by Z with realizations z ∈ Z. . log M
ρ = lim inf . (2)
There are n discrete memoryless stationary messages t→∞ valid (t, M, f, g) code w.r.t. t
zero-error decoding
(sources), Xi , i ∈ [n], of some common finite alphabet X . For
. . Clearly, by definition, we always have R ≤ ρ.
any S ⊆ [n], set XS = (Xi , i ∈ S), xS = (xi , i ∈ S), and
. |S|
XS = X . Thus X[n] denotes the tuple of all n messages, The side information availability at receivers for a specific
and x[n] ∈ X[n] denotes a realization of the message n-tuple. index coding instance can be represented by a sequence
By convention, X∅ = x∅ = X∅ = ∅. We consider an arbitrary, (i|j ∈ Ai ), i ∈ [n]. Alternatively, it can be characterized
but fixed distribution PX[n] on X[n] , assuming without loss by a family of confusion graphs, (Γt , t ∈ Z+ ) [17]. For
of generality that it has full support1 . a given sequence length t, the confusion graph Γt is an
There is a server containing all messages. It encodes undirected graph defined on the message sequence tuple
t t
the tuple of n message sequences X[n] t
= (Xit , i ∈ [n]) alphabet X[n] . That is, V (Γt ) = X[n] . Vertex xt[n] in Γ
t
according to some (possibly randomized) encoding function corresponds to the realization x[n] . Any two different vertices
f to some codeword Y that takes values in the code al- xt[n] , z[n]
t
are adjacent in Γt iff xti 6= zit and xtAi = zA t
i
for
phabet Y = {1, 2, . . . , M }. Each message sequence Xit = some receiver i ∈ [n]. We call any pair of vertices satisfying
(Xi,1 , Xi,2 , . . . , Xi,t ) is of length t symbols. The server this condition confusable. Hence, E(Γt ) = {{xt[n] , z[n] t
} :
then transmits the codeword to n receivers via a noiseless t t t t
xi 6= zi and xAi = zAi for some i ∈ [n]}.
broadcast channel of normalized unit capacity. Let PY,X[n] t
For correct decoding at all receivers, any two realizations
denote the joint distribution of the message sequence tuple xt[n] , z[n]
t
can be mapped to the same codeword y with
t
X[n] and the codeword Y . For any S ⊆ [n], we define the nonzero probabilities iff they are not confusable [17]. See
following notation for message sequence tuples. Figure 2 below for a toy example of an index coding instance
• XSt = X t|S| ; and its confusion graph. For the definitions for basic graph-
• XSt = (Xit , i ∈ S) = (XS,1 , XS,2 , . . . , XS,t ), where theoretic notions, see any textbook on graph theory (e.g.,
XS,j = (Xi,j , i ∈ S) for every j ∈ [t]. Note that Xi,j Scheinerman and Ullman [18]).
denotes the j-th symbol of message sequence Xit . Consider any set S ⊆ [n]. The subproblem induced by
• Similarly, xtS = (xti , i ∈ S) = (xS,j , j ∈ [t]), where S is jointly characterized by the distribution PXS and the
xS,j = (xi,j , i ∈ S) for every j ∈ [t]. sequence (i|Ai ∩ S), i ∈ S. Let Γt (S) denote the confusion
Also, as the messages are memoryless, for any xt[n] = graph of sequence length t of the subproblem induced by S.
(x[n],1 , x[n],2 , . . . , x[n],t ), PX[n]
t (x
t
Q Let R(S) and ρ(S) denote the broadcast rate and zero-error
[n] ) = j∈[t] PX[n] (x[n],j ).
broadcast rate of the subproblem induced by S, respectively.
On the receiver side, we assume that receiver i ∈ [n]
wishes to obtain message Xit and knows XA t
as side in- Preliminaries on R and ρ: Consider any index coding
formation for some Ai ⊆ [n] \ {i}.
i
problem characterized by confusion graphs (Γt , t ∈ Z+ ) and
distribution PX[n] . We start with the following lemma.
More formally, a (t, M, f, g) index code can be defined by
Lemma 1: To characterize the broadcast rate R and zero-
• One stochastic encoder f : X nt → {1, 2, . . . , M } at the error broadcast rate ρ, it suffices to only consider index codes
server that maps each message sequence tuple xt[n] ∈ with deterministic encoding function f .
X nt to a codeword y ∈ {1, 2, . . . , M }, and
The above lemma can be simply proved by showing that
1 While a common assumption in most index coding literatures is that the
given any valid index code with a stochastic encoding func-
messages are independent and uniformly distributed, here we consider the tion, one can construct another valid code with a deterministic
more general case with arbitrary joint distribution for the messages. encoding function and same or smaller compression rate.
(0, 0, 0) (0, 0, 1) adversary successfully guessing xtQ with c(t) number of
guesses is
 
(1, 0, 0) (0, 1, 0) X
Ps (XPt ) = EXPt  max
t
PXQt |XPt (xtQ |XPt ) ,
K⊆XQ :|K|≤c(t)
xtQ ∈K

(1, 0, 1) (0, 1, 1)
and the expected successful guessing probability after ob-
serving y is
 
X
(1, 1, 0) (1, 1, 1) Ps (XPt , Y ) = EY,XPt  maxt PXQt |Y,XPt (xtQ |Y, XPt ).
 
K⊆XQ : t
x ∈K
Figure 2. The confusion graph Γ1 with t = 1 for the 3-message index |K|≤c(t) Q
coding instance (1|−), (2|3), (3|2). Note that, for example, x[n] = (0, 0, 0)
The leakage to the adversary, denoted by L, is defined as
and z[n] = (0, 0, 1) are confusable because x3 = 0 6= z3 = 1 and xA3 =
x2 = 0 = z2 = zA3 . Suppose (0, 0, 0) and (0, 0, 1) are mapped to the the logarithm of the ratio between the expected probabilities
same codeword y with certain nonzero probabilities. Then upon receiving of the adversary successfully guessing xQ after and before
this y, receiver 3 will not be able to tell whether the value for X3 is 0 or observing the transmitted codeword y. That is,
1 based on its side information of X2 = 0. For this graph, it can be easily
verified that the independence number is 2, and that the chromatic number . Ps (XPt , Y )
equals to the fractional chromatic number, both of which equal to 4. We L = log . (4)
have drawn an optimal coloring scheme with 4 colors in the graph. Ps (XPt )
The (optimal) leakage rate can then be defined as
Most existing results in the literature on the optimal .
L = lim lim t−1 inf L. (5)
compression rate (vanishing or zero error) of index coding →0 t→∞ valid (t, M, f, g) code

were established assuming deterministic encoding functions. Remark 1: It can be readily verified that the leakage metric
Lemma 1 indicates that those results can be directly applied L is always non-negative. When c(t) = 1 (i.e., the adversary
to characterizing R and ρ. only makes a single guess after each observation), L reduces
Since we are considering fixed-length codes (rather than to the min-entropy leakage [12]. When c(t) = 1 and the
variable-length codes), the zero-error broadcast rate ρ does messages are uniformly distributed, L is equal to the maximal
not depend on PX[n] and can be characterized solely by the leakage [14] and the maximum min-entropy leakage [13].
confusion graphs (Γt , t ∈ Z+ ) [17] as If we require zero-error decoding at receivers, the zero-
1 (a) 1 error (optimal) leakage rate λ can be similarly defined as
ρ = lim log χ(Γt ) = lim log χf (Γt ), (3) .
t→∞ t t→∞ t λ = lim t−1 inf L. (6)
t→∞ valid (t, M, f, g) code w.r.t.
where χ(·) and χf (·) respectively denote the chromatic zero-error decoding
number and fractional chromatic number of a graph, and the
By definition, we always have L ≤ λ.
proof of (a) can be found in [3, Section 3.2].
It has been shown [19] that, with the messages X[n] being III. I NFORMATION L EAKAGE IN I NDEX C ODING
uniformly distributed and independent of each other, the
A. Leakage Under A General Message Distribution
vanishing-error broadcast rate R equals to the zero-error
broadcast rate ρ. Such equivalence does not hold for a general Consider any index coding problem (i|j ∈ Ai ), i ∈ [n]
distribution PX[n] as it has been shown in [20] that the with confusion graphs (Γt , t ∈ Z+ ) and distribution PX[n] .
(vanishing-error) broadcast rate R can be strictly smaller than Our main result is the following theorem.
its zero-error counterpart ρ. Theorem 1: For the vanishing-error leakage rate L, we have
Leakage to a guessing adversary: 1
ρ(Q) − |Q| + log P ≤ L ≤ R(Q). (7)
We assume the adversary knows messages XP and tries to max PX[n] (x[n] )
xQ
guess the remaining messages XQ , where Q = [n] \ P , via xP

maximum likelihood estimation within a number of trials. In For the zero-error leakage rate λ, we have
other words, the adversary generates a list of certain size of 1
guesses, and is satisfied iff the true message sequence is in ρ(Q) − |Q| + log P ≤ λ ≤ ρ(Q). (8)
max PX[n] (x[n] )
the list. We characterize the number of guesses the adversary xP xQ

can make by a function of sequence length, c : Z+ → Z+ , In the following, we prove the lower and upper bounds in
namely, the guessing capability function. We assume c(t) to (7). As for (8), the lower bound follows directly from the
be non-decreasing and upper-bounded2 by α(Γt (Q)), where lower bound in (7) and the fact that L ≤ λ, and the upper
α(·) denotes the independence number of a graph. bound can be shown using similar techniques to the proof of
Consider any valid (t, M, f, g) index code. Before eaves- the upper bound in (7).
dropping the codeword y, the expected probability of the Proof of the lower bound in (7): Consider any  > 0
2 It can be verified that if for some t we have c(t) > α(Γ (Q)), then the and any valid (t, M, f, g) index code for which Pe ≤ .
t
probability of the adversary successfully guessing xtQ after observing y is Consider any codeword y ∈ Y and any realization xtP ∈
at least 1 − Pe , which tends to 1 as  tends to 0, making the problem trivial. XP . Let GXQt (y, xtP ) denote the collection of realizations
t
xtQ such that (y, xtP , xtQ ) has nonzero probability, and for the where c(t)− = min{c(t), |GXQt (y, xtP )|}, and
event that xt[n] = (xtP , xtQ ) is the true message sequence tuple • (a) follows from the fact that each xtQ ∈ GXQt (y, xtP )
realization and y is the codeword realization, every receiver |GX t (y,xtP )−1|
can correctly decode its requested message. That is, appears in exactly Q

subsets of
c(t) −1
GXQt (y, xtP ) = {xtQ ∈ XQt : gi (y, xtAi ) = xti , ∀i ∈ [n]} GXQt (y, xtP ) of size c(t)− ,
• (b) follows from (9), (10), and that if c(t) ≤
Then, we have −
X X |GXQt (y, xtP )|, then |G c(t) t
t (y,x )|
= |G tc(t)
(y,xt )|

t t X P X P
t (y, x , x )
PY,X[n] Q Q
P Q c(t)
y,xtP xtQ ∈GX t (y,xtP ) α(Γt (Q)) , otherwise we have c(t) > |GXQt (y, xtP )| and
Q −
thus |G c(t) t )| = 1≥ c(t)
α(Γt (Q)) , where the inequality
= 1 − Pe ≥ 1 − . (9) X t
Q
(y,xP

is due to the assumption that c(t) ≤ α(Γt (Q)).


We also have
For bounding Ps (XPt ), consider any two disjoint subsets
|GXQt (y, xtP )| ≤ α(Γt (Q)), (10) A, B ⊆ [n]. Note that any realization xtA can be explicitly
which can be shown by contradiction as follows. Assume denoted as (xA,1 , xA,2 , . . . , xA,t ). We have
there exists two different xt[n] , z[n]
t
∈ X[n] t
, such that xtP =
X
max PXA∪B t (xtA , xtB )
zP , xQ ∈ GXQt (y, xP ), zQ ∈ GXQt (y, xtP ), and xtQ and
t t t t
xtA
t xB
t
zQ are adjacent (i.e., confusable) in Γt (Q). Hence, there X Y
exists some receiver i ∈ Q such that xti 6= zit and = max
t
PXA∪B (xA,j , xB,j )
xB
xtA j∈[t]
xtAi ∩Q = zA t
i ∩Q
. Then considering xt[n] and z[n] t
, since X t
they have the same realizations for messages in P , we have = max PXA∪B (xA , xB ) , (12)
xB
xtAi = (xtAi ∩P , xtAi ∩Q ) = (zAt
i ∩P
, zAt
i ∩Q
) = zAt
i
. From the xA
perspective of receiver i, upon receiving codeword y and where the last equality can be shown via induction.
observing side information xtAi = zA t
i
, it cannot tell whether Based on (11) and (12), we have
the true sequence for message i is xti or zit . Therefore, with c(t)(1−)
the transmitted codeword being y, either xt[n] or z[n] t
being 1 α(Γt (Q))
L ≥ lim lim log P P t )
the true realization will lead to an erroneous decoding at →0 t→∞ t
xt max
t
t (x
PX[n] [n]
P K⊆XQ :|K|≤c(t) xt ∈K
receiver i, which contradicts the assumption that both xtQ Q
t
and zQ belong to GXQt (y, xtP ). Therefore, any realizations 1
c(t)(1−)
α(Γt (Q))
xQ , zQ ∈ GXQt (y, xtP ) must be not confusable and thus not
t t ≥ lim lim log P t )
→0 t→∞ t c(t) · xt maxxtQ PX[n]
t (x
[n]
adjacent to each other in Γt (Q). In other words, the vertex P
(1−)|V (Γt (Q))|
subset GXQt (y, xtP ) ⊆ V (Γt (Q)) must be an independent set 1 · |V (Γt1(Q))|
α(Γt (Q))
in Γt (Q) and thus its cardinality is upper bounded by the = lim lim log P
→0 t→∞ t ( xP maxxQ PX[n] (x[n] ))t
independence number of Γt (Q).
(c) 1 |X |−|Q|
We lower bound Ps (XPt , Y ), i.e., the adversary’s expected = lim log χf (Γt (Q)) + log P
xP maxxQ PX[n] (x[n] )
t→∞ t
successful guessing probability after observing Y , as
(d) 1
= ρ(Q) − |Q| + log P
X X
max PY,X[n]t (y, x
t
[n] )
,
t
K⊆XQ :|K|≤c(t) xP maxxQ PX[n] (x[n] )
y,xtP xtQ ∈K
X X where (c) follows from the fact that for any vertex-transitive
t
≥ max
t
t (y, x
PY,X[n] [n] ) graph G, χf (G) = |V (G)|/α(G) [18, Proposition 3.1.1], and
K⊆GX t (y,xP ):|K|≤c(t)
y,xtP Q xtQ ∈K that any confusion graph for index coding is vertex-transitive
X 1 [3, Section 11.4], and (d) follows from (3).
≥ t −
|K⊆GX t (y,xP ):|K|=c(t) | Proof of the upper bound in (7): Consider any decoding
y,xtP Q
X X error  > 0. Construct a deterministic encoding function
t t
t (y, x
PY,X[n] [n] ) f that maps messages X[n] to codeword Y = (Y1 , Y2 )
K⊆GX t (y,xtP ):|K|=c(t)− xtQ ∈K according to the following rules.
Q

|GX t (y,xtP )−1|


1) Codeword Y1 is generated from XPt according to
t
P
t (y, x
PY,X[n] [n] ) some deterministic encoding function f1 : X t|P | →
Q
c(t)− −1
xtQ ∈GX t (y,xtP )
(a) X Q {1, 2, . . . , |Y1 |} such that there exist some decoding
= |GX t (y,xtP )| functions gi , i ∈ P allowing zero-error decoding for
Q
y,xtP
c(t)− all receivers i ∈ P and that t−1 log |Y1 | = ρ(P ).
t
X c(t)− X
t 2) Codeword Y2 is generated from XQ according to
= t (y, x
PY,X[n] [n] )
|GXQt (y, xtP )| some deterministic encoding function f2 : X t|Q| →
y,xtP xtQ ∈GX t (y,xtP )
Q {1, 2, . . . , |Y2 |} such that there exist some decoding
(b) c(t)(1 − ) functions gi , i ∈ Q allowing -error decoding for all
≥ , (11) receivers i ∈ Q and that t−1 log |Y2 | = R(Q).
α(Γt (Q))
Such encoding functions f1 and f2 exist for sufficiently in general, as shown in the following example.
large t. We further verify that the coding scheme de- Consider the 4-message index coding problem
scribed above leads to an average probability of er- (1|4), (2|3), (3|2), (4|1), where the messages are binary
ror Pe no more than  and thus is valid. Note and independent of each other with PX1 (0) = 1/4 and
that f1 and f2 are all deterministic. Define BXQt = PX1 (1) = 3/4, and X2 , X3 , and X4 all follow a uniform
{xtQ : there exists some i ∈ Q such that gi (f2 (xtQ )) 6= xti }. distribution. Consider an adversary knowing XP = X4 as
That is, BXQt denotes the set of xtQ for which there is at side information, and thus Q = {1, 2, 3}. The broadcast rate
least one receiver i ∈ Q that decodes erroneously. We have for the subproblem induced by Q has been previously found
X [20] to be R(Q) = 3 − 43 log 3. By Theorem 1, the leakage
≥ PXQt (xtQ ). (13)
rate L is upper bounded by R(Q), and lower bounded as
xtQ ∈BX t
Q 1
L ≥ρ(Q) − |Q| + log P
Hence, we have xP max xQ PX[n] (x[n] )
X X
Pe = t (x
PX[n] t
[n] )
1
= 2 − 3 + log = 3 − log 3.
xtP xtQ ∈BX t 3/32 + 3/32
Q
X X Note that ρ(Q) = 2 can be easily verified (for example, see
= PXQt (xtQ ) PXPt |XQt (xtP |xtQ )
[3, Section 8.6]). For the zero-error leakage rate λ, by (8) in
xtQ ∈BX t xtP
Q Theorem 1, we have
X
= PXQt (xtQ ) · 1 ≤ , 3 − log 3 ≤ λ ≤ 2.
xtQ ∈BX t
Q

where the last inequality follows from (13). Now we have B. Leakage Under A Uniform Message Distribution
shown that the proposed coding scheme is valid. In most existing works for index coding, the messages X[n]
The optimal leakage rate is upper bounded by the rate are assumed to be uniformly distributed and thus independent
of the information leakage of the proposed coding scheme of each other. In such cases, Theorem 1 simplifies to the
as  goes to 0. Let PY,X[n]
t denote the joint distribution of following corollary.
t
Y = (Y1 , Y2 ) and X[n] according to the proposed coding Corollary 1: If PX[n] follows a uniform distribution, then
scheme. For any xtP ∈ XPt and y2 ∈ Y2 , define L = λ = R(Q) = ρ(Q). (14)
XQt (xtP , y2 ) = {xtQ ∈ XQt : t t
t (y2 , x , x )
PY2 ,X[n] P Q > 0}, Proof: We have
Then we have 1
ρ(Q) − |Q| + log P
t max xQ PX[n] (x[n] )
P P
t
max t (y1 , y2 , x
PY,X[n] [n] ) xP
xtP , K⊆XQ : xtQ ∈K 1
1 y1 ,y2 |K|≤c(t) = ρ(Q) − |Q| + log
L ≤ lim lim log P P t ) |X |t|P | · (1/|X |tn )
→0 t→∞ t maxt xtQ ∈K t (x
PX[n] [n]
xtP K⊆XQ : = ρ(Q) = R(Q),
|K|≤c(t)
P
max
P
t (x
t where the last equality follows from the fact that the
PX[n] [n] )
t t
xtP ,y2 K⊆XQ (xP ,y2 ): xtQ ∈K vanishing-error and zero-error broadcast rates are equal when
(a) 1 |K|≤c(t) messages are uniformly distributed [19]. Combining Theorem
= lim lim log P P t )
→0 t→∞ t maxt
t (x
PX[n] [n] 1 and the above result yields (14).
xtP K⊆XQ : xt ∈K
|K|≤c(t)
Q Remark 3: Even though we have established the equiva-
|Y2 | · (
P
maxt
P
t (x
t lence between the leakage and broadcast rates under uniform
PX[n] [n] ))
xtP K⊆XQ : xt ∈K
Q
message distribution, a computable single-letter characteri-
(b) 1 |K|≤c(t) zation of the value in (14) is unknown. Nevertheless, the
≤ lim lim log P P t )
→0 t→∞ t maxt t (x
PX[n] [n] equivalence between the leakage and broadcast rates means
xtP K⊆XQ : xtQ ∈K that the extensive results on the broadcast rate of index coding
|K|≤c(t)
1 established in the literature (such as single-letter lower and
= lim lim log |Y2 | = R(Q), upper bounds, explicit characterization for special cases, and
→0 t→∞ t
structural properties) can be directly used to determine or
where (a) follows from the definition of XQt (xtP , y2 ) and bound the leakage rate.
the fact that Y1 is a deterministic function of XPt and Y2
t Remark 4: As the leakage rate in (14) can be achieved
is a deterministic function of XQ , and (b) follows from
t t t by the proposed coding scheme in the achievability proof
XQ (xP , y2 ) ⊆ XQ .
of Theorem 1, for any index coding instance with uniform
Remark 2: An interesting observation is that the bounds in message distribution satisfying R = R(P ) + R(Q) (or
Theorem 1 is independent of the guessing capability function equivalently, ρ = ρ(P ) + ρ(Q)), we know that the broadcast
c(t). Whether L and λ depend on c(t) remains unclear. rate and leakage rate can be simultaneously achieved by some
The upper and lower bounds in Theorem 1 do not match deterministic index code.
R EFERENCES
[1] Y. Birk and T. Kol, “Informed-source coding-on-demand (ISCOD) over
broadcast channels,” in IEEE INFOCOM, Mar. 1998, pp. 1257–1264.
[2] Z. Bar-Yossef, Y. Birk, T. Jayram, and T. Kol, “Index coding with side
information,” IEEE Trans. Inf. Theory, vol. 57, pp. 1479–1494, 2011.
[3] F. Arbabjolfaei and Y.-H. Kim, “Fundamentals of index coding,”
Foundations and Trends® in Communications and Information Theory,
vol. 14, no. 3-4, pp. 163–346, 2018.
[4] S. H. Dau, V. Skachek, and Y. M. Chee, “On the security of index
coding with side information,” IEEE Trans. Inf. Theory, vol. 58, no. 6,
pp. 3975–3988, 2012.
[5] L. Ong, B. N. Vellambi, P. L. Yeoh, J. Kliewer, and J. Yuan, “Secure
index coding: Existence and construction,” in Proc. IEEE Int. Symp.
on Information Theory (ISIT), Barcelona, Spain, 2016, pp. 2834–2838.
[6] Y. Liu, Y.-H. Kim, B. Vellambi, and P. Sadeghi, “On the capacity
region for secure index coding,” in Proc. IEEE Information Theory
Workshop (ITW), Guanzhou, China, Nov. 2018.
[7] L. Ong, B. N. Vellambi, J. Kliewer, and P. L. Yeoh, “A code and rate
equivalence between secure network and index coding,” IEEE J. Sel.
Areas Inf. Theory, vol. 2, no. 1, pp. 106–120, 2021.
[8] Y. Liu, P. Sadeghi, N. Aboutorab, and A. Sharififar, “Secure index
coding with security constraints on receivers,” in Proc. IEEE Int. Symp.
on Information Theory and its Applications (ISITA), Oct. 2020.
[9] V. Narayanan, J. Ravi, V. K. Mishra, B. K. Dey, N. Karamchan-
dani, and V. M. Prabhakaran, “Private index coding,” arXiv preprint
arXiv:2006.00257, 2020.
[10] M. Karmoose, L. Song, M. Cardone, and C. Fragouli, “Privacy in index
coding: k-limited-access schemes,” IEEE Trans. Inf. Theory, vol. 66,
no. 5, pp. 2625–2641, 2019.
[11] Y. Liu, N. Ding, P. Sadeghi, and T. Rakotoarivelo, “Privacy-utility
tradeoff in a guessing framework inspired by index coding,” in Proc.
IEEE Int. Symp. on Information Theory (ISIT), Jun. 2020, pp. 926–931.
[12] G. Smith, “On the foundations of quantitative information flow,” in
International Conference on Foundations of Software Science and
Computational Structures. Springer, 2009, pp. 288–302.
[13] C. Braun, K. Chatzikokolakis, and C. Palamidessi, “Quantitative no-
tions of leakage for one-try attacks,” Electronic Notes in Theoretical
Computer Science, vol. 249, pp. 75–91, 2009.
[14] I. Issa, A. B. Wagner, and S. Kamath, “An operational approach to
information leakage,” IEEE Trans. Inf. Theory, vol. 66, no. 3, pp. 1625–
1657, 2019.
[15] Y. Liu, L. Ong, S. Johnson, J. Kliewer, P. Sadeghi, and P. L. Yeoh,
“Information leakage in zero-error source coding: A graph-theoretic
perspective,” in Proc. IEEE Int. Symp. on Information Theory (ISIT),
Melbourne, Australia, 2021.
[16] J. Körner, “Coding of an information source having ambiguous al-
phabet and the entropy of graphs,” in 6th Prague Conference on
Information Theory, 1973, pp. 411–425.
[17] N. Alon, E. Lubetzky, U. Stav, A. Weinstein, and A. Hassidim,
“Broadcasting with side information,” in 49th Annu. IEEE Symp. on
Foundations of Computer Science (FOCS), Oct. 2008, pp. 823–832.
[18] E. R. Scheinerman and D. H. Ullman, Fractional Graph Theory: A
Rational Approach to the Theory of Graphs. Courier Corporation,
2011.
[19] M. Langberg and M. Effros, “Network coding: Is zero error always
possible?” in Proc. 49th Ann. Allerton Conf. Comm. Control Comput.,
2011, pp. 1478–1485.
[20] S. Miyake and J. Muramatsu, “Index coding over correlated sources,”
in Proc. IEEE Int. Symp. on Network Coding (NetCod), Sydney,
Australia, 2015, pp. 36–40.

You might also like