Next Article in Journal
Anticipation Next: System-Sensitive Technology Development and Integration in Work Contexts
Previous Article in Journal
The Human Digitalisation Journey: Technology First at the Expense of Humans?
Previous Article in Special Issue
On Two-Stage Guessing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Distributed Hypothesis Testing over Noisy Broadcast Channels

by
Sadaf Salehkalaibar
1 and
Michèle Wigger
2,*,†
1
Department of Electrical and Computer Engineering, College of Engineering, University of Tehran, Tehran 1433957131, Iran
2
LTCI, Telecom Paris, IP Paris, 91120 Paris, France
*
Author to whom correspondence should be addressed.
M. Wigger has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 programme, grant agreement No 715111.
Submission received: 21 May 2021 / Revised: 8 June 2021 / Accepted: 8 June 2021 / Published: 29 June 2021
(This article belongs to the Special Issue Statistical Communication and Information Theory)

Abstract

:
This paper studies binary hypothesis testing with a single sensor that communicates with two decision centers over a memoryless broadcast channel. The main focus lies on the tradeoff between the two type-II error exponents achievable at the two decision centers. In our proposed scheme, we can partially mitigate this tradeoff when the transmitter has a probability larger than 1 / 2 to distinguish the alternate hypotheses at the decision centers, i.e., the hypotheses under which the decision centers wish to maximize their error exponents. In the cases where these hypotheses cannot be distinguished at the transmitter (because both decision centers have the same alternative hypothesis or because the transmitter’s observations have the same marginal distribution under both hypotheses), our scheme shows an important tradeoff between the two exponents. The results in this paper thus reinforce the previous conclusions drawn for a setup where communication is over a common noiseless link. Compared to such a noiseless scenario, here, however, we observe that even when the transmitter can distinguish the two hypotheses, a small exponent tradeoff can persist, simply because the noise in the channel prevents the transmitter to perfectly describe its guess of the hypothesis to the two decision centers.

1. Introduction

In Internet of Things (IoT) networks, data are collected at sensors and transmitted over a wireless channel to remote decision centers, which decide on one or multiple hypotheses based on the collected information. In this paper, we study simple binary hypothesis testing with a single sensor but two decision centers. The results can be combined with previous studies focusing on multiple sensors and a single decision center to tackle the practically relevant case of multiple sensors and multiple decision centers. We consider a single sensor for simplicity and because our main focus is on studying the tradeoff between the performances at the two decision centers that can arise because the single sensor has to send information over the channel that can be used by both decision centers. A simple, but highly suboptimal, approach would be to time-share communication and serve each of the two decision centers only during a part of the transmission. As we will see, better schemes are possible, and, in some cases, it is even possible to serve each of the two decision centers as if the other center was not present in the system.
In this paper, we follow the information-theoretic framework introduced in [1,2]. That means each terminal observes a memoryless sequence, and depending on the underlying hypothesis H { 0 , 1 } , all sequences follow one of two possible joint distributions, which are known to all involved terminals. A priori, the transmitter, however, ignores the correct hypothesis and has to compute its transmit signal as a function of the observed source symbols only. Decision centers observe outputs of the channel, and, combined with their local observations, they have to make a decision on whether H = 0 or H = 1 . The performance of the decision center is measured by its type-II error exponent, i.e., the expontial decay in the length of the observations of the probability of deciding on H = 0 when the true hypothesis is H = 1 . As a constraint on the decision center, we impose that the type-I error probability, i.e., the probability of deciding H = 1 when the true hypothesis is H = 0 , vanishes (at any desired speed) with increasing observation lengths. The motivation for studying such asymmetric requirements on the two error probabilities stems, for example, from alert systems, where the miss-detection event is much more harmful than the false-alarm event, and, as a consequence, in our systems we require the miss-detection probability to decay much faster than the false-alarm probability.
This problem setting has first been considered for the setup with a single sensor and a single decision center when communication is over a noiseless link of given capacity [1,2]. For this canonical problem, the optimal error exponent has been identified in the special cases of testing against independence [1] and testing against conditional independence [3,4].
The scheme proposed by Shimokawa–Han–Amari in [3,4] yields an achievable error exponent for all distributed hypothesis testing problems (not only testing against conditional independence) [3,4], but it might not be optimal in general [5]. The Shimokawa–Han–Amari (SHA) scheme has been extended to various more involved setups such as noiseless networks with multiple sensors and a single decision center [2,6,7]; networks where the sensor and the decision center can communicate interactively [8,9]; multi-hop networks [10]; networks with multiple decision centers [10,11,12,13].
The works most closely related to the current paper are [10,12,13,14,15,16]. Specifically, Refs [10,12,13,14] consider a single-sensor multi-detector system where communication is over a common noiseless link from the sensor to all decision centers. Focusing on two decision centers, two scenarios can be encountered here: (1) the two decision centers have the same null and alternate hypotheses, and as a consequence, both aim at maximizing the error exponent under the same hypothesis H ; or (2) the two decision centers have opposite null and alternate hypotheses and thus one decision center wishes to maximize the error exponent under hypothesis H = 0 and the other under hypothesis H = 1 . The second scenario is motivated by applications where the decision centers have different goals. Hypothesis testing for scenario 1) was studied in [10,12,13,14], and the results showed a tradeoff between the exponents achieved at the two decision centers. Intuitively, the tradeoff comes from the fact that communication from the sensor is serving both decision centers at the same time. Scenario 2) was considered in [12,13]. In this case, a tradeoff only occurs when the sensor’s observation alone provides no advantage in guessing the hypothesis. Otherwise, a tradeoff-free exponent region can be achieved by the following simple scheme: the sensor takes a tentative guess on the hypothesis based only on its local observations. It communicates this tentative guess to both decision centers using a single bit and then dedicates the rest of the communication using a dedicated SHA scheme only to the decision center that wishes to maximize the error exponent under the hypothesis that does not correspond to its guess. The other decision center simply keeps the transmitter’s tentative guess and ignores the rest of the communication.
In this paper, we extend these previous works to memoryless broadcast channels (BC). Hypothesis testing over BCs was already considered in [10], however, only for above scenario 1 where both decision centers have the same null and alternate hypothesis and in the special case of testing against conditional independence, in which case the derived error exponents were proved to be optimal. Interestingly, when testing against conditional independence over noisy channels, only the capacity of the channel matters but not the properties, see [15,16]. General hypothesis testing over noisy channels is much more challenging and requires additional tools, such as joint source-channel coding and unequal error protection (UEP) coding [17]. The latter can, in particular, be used to specially protect the communication of the sensor’s tentative guess, which allows one to avoid a degradation of the performance of classical hypothesis testing schemes.
We present general distributed hypothesis testing schemes over memoryless BCs, and we analyze the performances of these schemes with a special focus on the tradeoff in exponents they achieve for the two decision centers. We propose two different schemes, depending on whether the sensor can distinguish with error probability 1 / 2 the two null hypotheses at the two decision centers. If a distinction is possible (because the decision centers have different null hypothesis and the sensor’s observations follow different marginal distributions under the two hypotheses), then we employ a similar scheme as proposed in [12,13] over a common noiseless link, but where the SHA scheme is replaced by the UEP-based scheme for DMCs in [15]. That means, the sensor makes a tentative guess about the hypothesis and conveys this guess to both decision centers using an UEP mechanism. Moreover, the joint source-channel coding scheme in [15] with dedicated codebooks is used to communicate to the decision center that aims to maximize the error exponent under the hypothesis that does not correspond to the sensor’s tentative guess. This scheme shows no tradeoff between the exponents achieved at the two decision centers in various interesting cases. Sometimes, however, a tradeoff arises because even under UEP the specially protected messages can be in error and because the decision centers can confuse the codewords of the two different sets of codebooks. For the case where the sensor cannot reasonably distinguish the alternate hypotheses at the two decision centers (because both decision centers have the same alternate hypotheses or the sensor’s observations have the same marginal observations under both hypotheses), we present a scheme similar to [10] but again including UEP. In this scheme, a tradeoff between the exponents achieved at the two decision centers naturally arises and mostly stems from the inherent tradeoff in distributed lossy compression systems with multiple decoders having different side informations.

Notation

We mostly follow the notation in [18]. Random variables are denoted by capital letters, e.g., X , Y , and their realizations by lower-case letters, e.g., x , y. Script symbols such as X and Y stand for alphabets of random variables, and X n and Y n for the corresponding n-fold Cartesian products. Sequences of random variables ( X i , , X j ) and realizations ( x i , , x j ) are abbreviated by X i j and x i j . When i = 1 , then we also use X j and x j instead of X 1 j and x 1 j .
We write the probability mass function (pmf) of a discrete random variable X as P X ; to indicate the pmf under hypothesis H = 1 , we also use Q X . The conditional pmf of X given Y is written as P X | Y , or as Q X | Y when H = 1 . The term D ( P Q ) stands for the Kullback–Leibler (KL) divergence between two pmfs P and Q over the same alphabet. We use t p ( a n , b n ) to denote the joint type of the pair of sequences ( a n , b n ) , and cond_tp ( a n | b n ) for the conditional type of a n given b n . For a joint type π A B C over alphabet A × B × C , we denote by I π A B C ( A ; B | C ) the conditional ßmutual information assuming that the random triple ( A , B , C ) has pmf π A B C ; similarly for the entropy H π A B C ( A ) and the conditional entropy H π A B C ( A | B ) . Sometimes we abbreviate π A B C by π . In addition, when π A B C has been defined and is clear from the context, we write π A or π A B for the corresponding subtypes. When the type π A B C coincides with the actual pmf of a triple ( A , B , C ) , we omit the subscript and simply write H ( A ) , H ( A | B ) , and I ( A ; B | C ) .
For a given P X and a constant μ > 0 , let T μ n ( P X ) be the set of μ-typical sequences in X n as defined in [8] (Section 2.4). Similarly, T μ n ( P X Y ) stands for the set of jointly μ-typical sequences. The expectation operator is written as E [ · ] . We abbreviate independent and identically distributed by i.i.d. The log function is taken with base 2. Finally, in our justifications, we use (DP) and (CR) for “data processing inequality” and “chain rule”.

2. System Model

Consider the distributed hypothesis testing problem in Figure 1, where a transmitter observes sequence X n , Receiver 1 sequence Y 1 n , and Receiver 2 sequence Y 2 n . Under the null hypothesis:
H = 0 : ( X n , Y 1 n , Y 2 n ) i . i . d . P X Y 1 Y 2 ,
and under the alternative hypothesis:
H = 1 : ( X n , Y 1 n , Y 2 n ) i . i . d . Q X Y 1 Y 2 ,
for two given pmfs P X Y 1 Y 2 and Q X Y 1 Y 2 . The transmitter can communicate with the receivers over n uses of a discrete memoryless broadcast channel ( W , V 1 × V 2 , P V 1 V 2 | W ) where W denotes the finite channel input alphabet and V 1 and V 2 the finite channel output alphabets. Specifically, the transmitter feeds inputs
W n = f ( n ) ( X n ) ,
to the channel, where f ( n ) denotes the chosen (possibly stochastic) encoding function
f ( n ) : X n W n .
Each Receiver i { 1 , 2 } observes the BC ouputs V i n , where for a given input W t = w t ,
( V 1 , t , V 2 , t ) Γ V 1 V 2 | W ( · , · | w t ) , t { 1 , , n } .
Based on the sequence of channel outputs V i n and the source sequence Y i n , Receiver i decides on the hypothesis H . That means it produces the guess
H ^ i = g ( n ) ( V i n , Y i n ) ,
for a chosen decoding function
g i ( n ) : V i n × Y i n { 0 , 1 } .
There are different possible scenarios regarding the requirements on error probabilities. We assume that each receiver is interested in only one of the two exponents. For each i { 1 , 2 } , let h i { 0 , 1 } be the hypothesis whose error exponent Receiver i wishes to maximize, and h ¯ i the other hypothesis, i.e., h ¯ i { 0 , 1 } and h i h ¯ i . (The values of h 1 and h 2 are fixed and part of the problem statement.) We then have:
Definition 1.
An exponent pair ( θ 1 , θ 2 ) is said to be achievable over a BC, if for each ϵ 1 , ϵ 2 ( 0 , 1 ) and sufficiently large blocklengths n, there exist encoding and decoding functions ( f ( n ) , g 1 ( n ) , g 2 ( n ) ) such that:
α 1 , n = Δ Pr [ H ^ 1 = h 1 | H = h ¯ 1 ] , α 2 , n = Δ Pr [ H ^ 2 = h 2 | H = h ¯ 2 ] ,
β 1 , n = Δ Pr [ H ^ 1 = h ¯ 1 | H = h 1 ] , β 2 , n = Δ Pr [ H ^ 2 = h ¯ 2 | H = h 2 ] ,
satisfy
α i , n ϵ i , i { 1 , 2 } ,
and
lim ¯ n 1 n log β i , n θ i , i { 1 , 2 } .
Definition 2.
The fundamental exponents region E is the set of all exponent pairs ( θ 1 , θ 2 ) that are achievable.
Remark 1.
Notice that both α 1 , n and β 1 , n depend on the BC law Γ V 1 V 2 | W only through the conditional marginal distribution Γ V 1 | W . Similarly, α 2 , n and β 2 , n only depend on Γ V 2 | W . As a consequence, also the fundamental exponents region E depends on the joint laws P X Y 1 Y 2 and Q X Y 1 Y 2 only through their marginal laws P X Y 1 , P X Y 2 , Q X Y 1 , and Q X Y 2 .
Remark 2.
As a consequence to the preceding Remark 1, when P X = Q X , one can restrict attention to a scenario where both receivers aim at maximizing the error exponent under hypothesis H = 1 , i.e., h 1 = h 2 = 1 . In fact, under P X = Q X , the fundamental exponents region E for arbitrary h 1 and h 2 coincides with the fundamental exponents region E for h 1 = 1 and h 2 = 1 if one exchanges pmfs P X Y 1 and Q X Y 1 in case h 1 = 0 and one exchanges pmfs P X Y 2 and Q X Y 2 in case h 2 = 0 .
To simplify the notation in the sequel, we use the following shorthand notations for the pmfs P X Y 1 Y 2 and Q X Y 1 Y 2 .
For each i { 1 , 2 } :
if h ¯ i = 0 p X Y 1 Y 2 i : = P X Y 1 Y 2 and q X Y 1 Y 2 i : = Q X Y 1 Y 2
and
if h ¯ i = 1 p X Y 1 Y 2 i : = Q X Y 1 Y 2 and q X Y 1 Y 2 i : = P X Y 1 Y 2 .
We propose two coding schemes yielding two different exponent regions, depending on whether
x X : p X 1 ( x ) = p X 2 ( x ) ,
or
x X : p X 1 ( x ) p X 2 ( x ) .
Notice that (13) always holds when h 1 = h 2 . In contrast, given (14), then obviously h 1 h 2 .

3. Results on Exponents Region

Before presenting our main results, we recall the achievable error exponent over a discrete memoryless channel reported in [15] (Theorem 1).

3.1. Achievable Exponent for Point-to-Point Channels

Consider a single-receiver setup with only Receiver 1 that wishes to maximize the error exponent under hypothesis h 1 = 1 . For simplicity then, we drop the user index 1 and simply call the receiver’s source observation Y n and its channel outputs V n .
Theorem 1
(Theorem 1 in [15]). Any exponent θ satisfying the following condition is achievable:
θ max min { θ standard ( P S | X ) , θ dec ( P S | X , P T , P W | T ) , θ miss ( P S | X , P T , P W | T ) } ,
where the maximization is over pmfs P S | X , P T , and P W | T such that the joint law P S T W V X Y : = P X Y P S | X P T P W | T P V | W satisfies
I ( S ; X | Y ) I ( W ; V | T ) ,
and where the exponents in (15) are defined as:
θ standard ( P S | X ) : = min P ˜ S X Y : P ˜ S X = P S X P ˜ S Y = P S Y D ( P ˜ S X Y P S | X Q X Y ) ,
θ dec ( P S | X , P T , P W | T ) : = min P ˜ S X Y : P ˜ S X = P S X P ˜ Y = P Y H P ( S | Y ) H P ˜ ( S | Y ) D ( P ˜ S X Y P S | X Q X Y ) I ( S ; X | Y ) + I ( W ; V | T ) ,
θ miss ( P S | X , P T , P W | T ) : = D ( P Y Q Y ) + E P T D P V | T Γ V | W = T I ( S ; X | Y ) + I ( W ; V | T ) .
Here, all mutual information terms are calculated with respect to the joint pmf P S T W V X Y defined above.
The exponent in Theorem 1 is obtained by the following scheme, which is also depicted in Figure 2.
The transmitter attempts to quantize the source sequence X n using a random codebook consisting of codewords { S n ( m , ) } . If the quantization fails because no codeword is jointly typical with the source sequence, then the transmitter applies the UEP mechanism in [17] by sending an IID P T -sequence T n over the channel. Otherwise, it sends the codeword W n ( m ) for m indicating the first index of the S n ( m , ) codeword that is jointly typical with its source observation X n . The receiver jointly decodes the channel and source codeword by verifying the existence of indices ( m , ) such that W n = W n ( m ) is jointly typical with its channel outputs V n and there is no other codeword S n ( m , ˜ ) with smaller conditional empirical entropy given Y n than S n ( m , ) . If the decoded codeword S n ( m , ) is jointly typical with the receiver’s observation Y n , then it produces H ^ = 0 , and otherwise H ^ = 1 .
The three competing type-II error exponents in Theorem 1 can be understood in view of this coding scheme as follows. Exponent θ standard indicates the event that a random codeword S n ( m , ) is jointly typical with the transmitter’s observation X n and with the receiver’s observation Y n while being under H = 1 . This is also the error exponent in Han’s scheme [2] over a noiseless communication link and does not depend on the channel law Γ V | W . Exponent θ dec is related to the joint decoding that checks the joint typicality of the source codeword, as well as of the channel codeword, and applies a conditional minimum entropy decoder. A similar error exponent is observed in the SHA scheme [3,4] over a noiseless link if the mutual information I ( W ; V | T ) is replaced by the rate of the link. The third exponent θ m i s s finally indicates an event where the transmitter sends T n (so as to indicate the receiver to decide for H ^ = 1 ) but the receiver detects a channel codeword W n ( m ) and a corresponding source codeword S n ( m , ) . This exponent is directly related to the channel transition law Γ V | W and not only to the mutual information of the channel and does not occur when transmission is over a noiseless link. Interestingly, it is redundant in view of exponent θ d e c whenever Q X Y = P X Q Y because in this case the minimization in (18) evaluates to D ( P Y Q Y ) . In this special case, the exponent can also be shown to be optimal, see [15].
We now present our achievable exponents region, where we distinguish the two cases (1) h 1 h 2 and P X Q X ; and (2) ( h 1 = h 2 ) or P X = Q X .

3.2. Achievable Exponents Region When h 1 h 2 and P X Q X

Theorem 2.
If h 1 h 2 and P X Q X , i.e., (14) holds, then all error exponent pairs ( θ 1 , θ 2 ) satisfying the following condition are achievable:
θ i min { θ standard , i θ dec , i , θ cross , i , θ miss , i } , i { 1 , 2 } ,
where the union is over pmfs p S | X i , p T , p T i | T i , and p W | T i i , for i { 1 , 2 } , so that the joint pmfs p 1 , p 2 defined through (12) and
p S X Y 1 Y 2 T T i W V 1 V 2 i : = p S | X i · p X Y 1 Y 2 i · p T · p T i | T i · p W | T T i i · Γ V 1 V 2 | W , i { 1 , 2 } ,
satisfy constraints
I p i ( S ; X | Y i ) < I p i ( W ; V i | T , T i ) , i { 1 , 2 } ,
and where the exponents in (20) are defined as follows, where we set q 1 = p 2 and q 2 = p 1 :
θ standard , i : = min P ˜ S X Y i : P ˜ S X = p S X i P ˜ S Y i = p S Y i i D ( P ˜ S X Y i p S | X i q X Y i i ) ,
θ dec , i : = min P ˜ S X Y i : P ˜ S X = p S X i P ˜ Y i = p Y i i H p i ( S | Y i ) H P ˜ ( S | Y i ) D ( P ˜ S X Y i p S | X i q X Y i i ) I p i ( S ; X | Y i ) + I p i ( W ; V i | T , T i ) ,
θ miss , i : = D ( p Y i i q Y i i ) + E p T D p V i | T i Γ V i | W = T I p i ( S ; X | Y i ) + I p i ( W ; V i | T , T i ) ,
θ cross , i : = min P ˜ S X Y i : P ˜ Y i = p Y i i H p i ( S | Y i ) H P ˜ ( S | Y i ) E P ˜ S X D P ˜ Y i | X S q Y i | X i + min P ˜ T T i W : P ˜ T W = q T W i P ˜ T T i = p T T i i E P ˜ T T i W D ( p V i | T T i i Γ V i | W ) I p i ( S ; X | Y i ) + I p i ( W ; V i | T , T i ) .
Proof. 
See Appendix A. □
In Theorem 2, the exponent triple θ standard , 1 , θ dec , 1 , θ miss , 1 can be optimized over the pmfs p S | X 1 , p T 1 | T 1 and p W | T T 1 1 and independently thereof the exponent triple θ standard , 2 , θ dec , 2 , θ miss , 2 can be optimized over the pmfs p S | X 2 , p T 2 | T 2 and p W | T , T 2 2 . The pmf p T is common to both optimizations. However, whenever the exponents θ cross , 1 and θ cross , 2 are not active, Theorem 2 depends only on p S | X i , p T i i and p W | T i i , for i = 1 , 2 , and there is thus no tradeoff between the two exponents θ 1 and θ 2 . In other words, the same exponents θ 1 and θ 2 can be attained as in a system where the transmitter communicates over two individual DMCs Γ V 1 | W and Γ V 2 | W to the two receivers, or equivalently each receiver achieves the same exponent as if the other receiver was not present in the system.
The scheme achieving the exponents region in Theorem 2 is described in detail in Section 4 and analyzed in Appendix A. The main feature is that the sensor makes a tentative decision on H and conveys this decision to both receivers through its choice of the codebooks and a special coded time-sharing sequence indicating this choice. The receiver that wishes to maximize the error exponent corresponding to the hypothesis guessed at the sensor directly decides on this hypothesis. The other receiver should compare its own observation to a quantized version of the source sequence observed at the sensor. The sensor uses the quantization and binning scheme presented in [15] tailored to this latter receiver using either coded time-sharing sequence T 1 n and codebooks { S n ( 1 ; m , ) } and { W n ( 1 ; m ) } or coded time-sharing sequence T 2 n and codebooks { S n ( 2 ; m , ) } and { W n ( 2 ; m ) } , respectively. The overall scheme is illustrated in Figure 3.
Exponents θ standard , i , θ dec , i , and θ miss , i have similar explanations as in the single-user case. Exponent θ cross , i corresponds to the event that the transmitter sends a codeword from { W ( j ; m ) } , for j = 3 i , but Receiver i decides that a codeword from { W ( i ; m ) } was sent and a source codeword S ( i ; m , ) satisfies the minimum conditional entropy condition and the typicality check with the observed source sequence Y i n . Notice that setting T i as a constant decreases the error exponent θ cross , i .
For the special case where the BC consists of a common noiseless link, Theorem 2 has been proved in [12,13]. (More precisely, [12] considers the more general case with K 2 receivers and M K hypotheses.) In this case, the exponents ( θ miss , 1 , θ cross , 1 ) and ( θ miss , 2 , θ cross , 2 ) are not active and there is no tradeoff between θ 1 and θ 2 .

3.3. Achievable Exponents Region for h 1 = h 2 or P X = Q X

Define for any pmfs P T , P S U 1 U 2 | X T and function
f : S × U 1 × U 2 × X W
the joint pmfs
p S U 1 U 2 X Y 1 Y 2 T V 1 V 2 i : = P S U 1 U 2 | X T · p X Y 1 Y 2 i · P T · Γ V 1 V 2 | S U 1 U 2 X , i { 1 , 2 } ,
and
Γ V 1 V 2 | S U 1 U 2 X : = Γ V 1 V 2 | W ( v 1 , v 2 | f ( s , u 1 , u 2 , x ) ) , s S , u 1 U 1 , u 2 U 2 , x X ,
and for each i { 1 , 2 } , the four exponents
θ standard , i : = min P ˜ S U i X Y i T V i : P ˜ S U i X T = p S U i X T i P ˜ S U i Y i T V i = p S U i Y i T V i i D P ˜ S U i X Y i T V i p S U i | X i q X Y i i P T Γ V i | S U 1 U 2 X ,
θ dec , i a : = min P ˜ S U i X Y i T V i : P ˜ S U i X T = p S U i X T i P ˜ Y i T V i = p Y i T V i i H p i ( S , U i | Y i , T , V i ) H P ˜ ( S , U i | Y i , T , V i ) D P ˜ S U i X Y i T V i p S U i | X i q X Y i i P T Γ V i | S U 1 U 2 X I p i ( S , U i ; X | T ) + I p i ( S , U i ; Y i , V i | T ) ,
θ dec , i b : = min P ˜ S U i X Y i T V i : P ˜ S U i X T = p S U i X T i P ˜ S Y i T V i = p S Y i T V i i H p i ( U i | S , Y i , T , V i ) H P ˜ ( U i | S , Y i , T , V i ) D P ˜ S U i X Y i T V i p S U i | X i q X Y i i P T Γ V i | S U 1 U 2 X I p i ( U i ; X | S , T ) + I p i ( U i ; Y i , V i | S , T ) ,
θ miss , i : = E P T D p Y i V i | T i q Y i i Γ V i | W = T I p i ( S , U i ; X | T ) + I p i ( S , U i ; Y i , V i | T ) .
Theorem 3.
If h 1 = h 2 or P X = Q X , i.e., (13) holds, then the union of all nonnegative error exponent pairs ( θ 1 , θ 2 ) satisfying the following conditions is achievable:
θ i min θ standard , i , θ dec , i a , θ dec , i b , θ miss , i , i { 1 , 2 } , θ 1 + θ 2 min { θ standard , 1 + θ standard , 2 , θ standard , 1 + θ dec , 2 a , θ standard , 1 + θ dec , 2 b , θ standard , 2 + θ dec , 1 a , θ standard , 2 + θ dec , 1 b , θ miss , 1 + θ miss , 2 }
I p 1 ( U 1 ; U 2 | S , T ) ,
θ 1 + θ 2 min θ dec , 1 a , θ dec , 1 b + min θ dec , 2 a , θ dec , 2 b 2 I p 1 ( U 1 ; U 2 | S , T ) ,
where the union is over pmfs P T , P S U 1 U 2 | X T and functions f as in (27) so that the pmfs (28) and (29) satisfy for i { 1 , 2 } :
I p i ( S , U i ; X | T ) I p i ( S , U i ; Y i , V i | T ) ,
I p i ( U i ; X | S , T ) I p i ( U i ; Y i , V i | S , T ) ,
I p 1 ( S , U 1 ; X | T ) + I p 1 ( S , U 2 ; X | T ) + I p 1 ( U 1 ; U 2 | S , T )              I p 1 ( S , U 1 ; Y 1 , V 1 | T ) + I p 2 ( S , U 2 ; Y 2 , V 2 | T ) ,
I p 1 ( U 1 ; X | S , T ) + I p 1 ( U 2 ; X | S , T ) + I p 1 ( U 1 ; U 2 | S , T )              I p 1 ( U 1 ; Y 1 , V 1 | S , T ) + I p 2 ( U 2 ; Y 2 , V 2 | S , T ) ,
I p 1 ( U 1 ; X | S , T ) + I p 1 ( S , U 2 ; X | T ) + I p 1 ( U 1 ; U 2 | S , T )              I p 1 ( U 1 ; Y 1 , V 1 | S , T ) + I p 2 ( S , U 2 ; Y 2 , V 2 | T ) ,
I p 1 ( S , U 1 ; X | T ) + I p 1 ( U 2 ; X | S , T ) + I p 1 ( U 1 ; U 2 | S , T )              I p 1 ( S , U 1 ; Y 1 , V 1 | T ) + I p 2 ( U 2 ; Y 2 , V 2 | S , T ) ,
Proof. 
The coding and testing scheme achieving these exponents is described in Section 5. The analysis of the scheme is similar to the proof of [15] (Theorem 4) and omitted for brevity. In particular, error exponent θ standard , i corresponds with the event that Receiver i decodes the correct cloud and satellite codewords but wrongly decides on H ^ i = 0 . In contrast, error exponents θ dec , i a and θ dec , i b correspond to the events that Receiver i wrongly decides on H ^ i = 0 after wrongly decoding both the cloud center and the satellite or only the satellite. Error exponent θ miss , i corresponds to the miss-detection event. Due to the implicit rate constraints in (46), the final constraints in (31) are obtained by eliminating the rates R 0 , R 1 , R 2 by means of Fourier–Motzkin elimination. Notice that in constraint (31c), the mutual information I p 1 ( U 1 ; U 2 | S , T ) is multiplied by a factor 2, whereas in (31b), it appears without a factor. The reason is that the error analysis includes union bounds over the codewords in a bin and when wrongly decoding the satellite codewords (which is the case of exponents θ dec , i a and θ dec , i b ) then the union bound is over pairs of codewords, whereas under correct decoding, it is over single codewords. In the former case, we have the factor 2 2 n R i in the error probability, and, in the latter case, the factor 2 n R i . The auxiliary rates R 1 and R 2 are then eliminated using the Fourier–Motzkin elimination algorithm. □
For each i { 1 , 2 } , exponents θ standard , i , θ dec , i a , θ dec , i b , and θ miss , i have the same form as the three exponents in [15] (Theorem 1) for the DMC. There is, however, a tradeoff between the two exponents θ 1 and θ 2 in the above theorem because they share the same choice of the auxiliary pmfs P T and P S U 1 U 2 | X T and the function f. In [10], the above setup is studied in the special case of testing against conditional independence, and the mentioned tradeoff is illustrated through a Gaussian example.

4. Coding and Testing Scheme When p X 1 p X 2

Fix μ > 0 , a sufficiently large blocklength n, auxiliary distributions p T , p T 1 | T 1 and p T 2 | T 2 over W , conditional channel input distributions p W | T T 1 1 and p W | T T 2 2 , and conditional pmfs p S | X 1 and p S | X 2 over a finite auxiliary alphabet S such that for each i { 1 , 2 } :
I p i ( S ; X | Y i ) < I p i ( W ; V i | T , T i ) .
The mutual information in (33) is calculated according to the joint distribution:
p S X Y 1 Y 2 T T i W V 1 V 2 i = p S | X i · p X Y 1 Y 2 i · p T · p T i | T i · p W | T T i i · Γ V 1 V 2 | W .
For each i { 1 , 2 } , if I p i ( S ; X ) < I p i ( W ; V i | T , T i ) , choose rates
R i : = I p i ( S ; X ) + μ ,
R i : = 0 .
If I p i ( S ; X ) I p i ( W ; V i | T , T i ) , then choose rates
R i : = I p i ( W ; V i | T , T i ) μ ,
R i : = I p i ( S ; X ) I p i ( W ; V i | T , T i ) + 2 μ .
Again, all mutual informations in (35)–(38) are calculated with respect to the pmf in (34).
Code Construction: Generate a sequence T n = ( T 1 , , T n ) by independently drawing each component T k according to p T . For each i { 1 , 2 } , generate a sequence T i n = ( T i , 1 , , T i , n ) by independently drawing each T i , k according to p T i | T i ( . | t ) when T k = t . In addition, construct a random codebook
C W i = W n ( i ; m ) : m { 1 , . . . , 2 n R i }
superpositioned on ( T n , T i n ) where the k-th symbol W k ( i ; m ) of codeword W n ( i ; m ) is drawn independently of all codeword symbols according to p W | T T i i ( · | t , t i ) when T k = t and T i , k = t i . Finally, construct a random codebook
C S i = { S n ( i ; m , ) : m { 1 , , 2 n R i } , { 1 , , 2 n R i } } , i { 1 , 2 } ,
by independently drawing the k-th component S k ( i ; m , ) of codeword S n ( i ; m , ) according to the marginal pmf p S i .
Reveal all codebooks and the realizations t n , t 1 n , t 2 n of the sequences T n , T 1 n , T 2 n to all terminals.
Transmitter: Given source sequence X n = x n , the transmitter looks for indices ( i , m , ) { 1 , 2 } × { 1 , , 2 n R 1 } × { 1 , , 2 n R i } such that codeword s n ( i ; m , ) from codebook C S i satisfies
( s n ( i ; m , ) , x n ) T μ / 2 n ( p S X i ) .
and the corresponding codeword w n ( i ; m ) from codebook C W i satisfies
( t n , t i n , w n ( i ; m ) ) T μ / 2 n ( p T T i W i ) .
(Notice that when μ is sufficiently small, then Condition (41) can be satisfied for at most one value i { 1 , 2 } , because p X 1 p X 2 .) If successful, the transmitter picks uniformly at random one of the triples ( i , m , ) that satisfy (41), and it sends the sequence w n ( i ; m ) over the channel. If no triple satisfies Condition (41), then the transmitter sends the sequence t n over the channel.
Receiver i { 1 , 2 } : Receives v i n and checks whether there exist indices ( m , ) such that the following three conditions are satisfied:
  • ( t n , t i n , w n ( i ; m ) , v i n ) T μ n ( p T T i W V i i ) ,
  • H tp ( s n ( i ; m , ) , y i n ) ( S | Y i ) = min ˜ H tp ( s n ( i ; m , ˜ ) , y i n ) ( S | Y i ) ,
  • ( s n ( i ; m , ) , y i n ) T μ n ( p S Y i i ) .
If successful, it declares H ^ i = h ¯ i . Otherwise, it declares H ^ i = h i .
Analysis: See Appendix A.

5. Coding and Testing Scheme When p X 1 = p X 2

In this case, the scheme is based on hybrid source-channel coding. Choose a large positive integer n, auxiliary alphabets S , U 1 , and U 2 , and a function f as in (27).
Choose an auxiliary distribution P T over W , and a conditional distribution P S U 1 U 2 | X T over S × U 1 × U 2 so that for i { 1 , 2 } inequalities (32) are satisfied with strict inequality.
Then, choose a positive μ and rates R 0 , R 1 , R 2 so that
R 0 = I p 1 ( S ; X | T ) + μ ,
R i > I p 1 ( U i ; X | S , T ) , i { 1 , 2 } ,
R 1 + R 2 > I p 1 ( U 1 ; X | S , T ) + I p 1 ( U 2 ; X | S , T ) + I p 1 ( U 1 ; U 2 | S , T ) ,
and
R 0 + R i < I p i ( S , U i ; Y i , V i | T ) ,
R i < I p i ( U i ; Y i , V i | S , T ) .
Generate a sequence T n i.i.d. according to P T and construct a random codebook
C S = S n ( m 0 ) : m 0 { 1 , . . . , 2 n R 0 }
superpositioned on T n where each codeword is drawn independently according to p S | T 1 conditioned on T n . Then, for each index m 0 and i { 1 , 2 } , randomly generate a codebook
C U i ( m 0 ) = U i n ( m 0 , m i ) : m i { 1 , . . . , 2 n R i }
superpositioned on ( T n , S n ( m 0 ) ) by drawing each entry of the n-length codeword U i n ( m 0 , m i ) i.i.d. according to the conditional pmf p U i | S T 1 ( · | S k ( m 0 ) , T ) where S k ( m 0 ) denotes the k-th symbol of S n ( m 0 ) . Reveal the realizations of the codebooks and the sequence T n to all terminals.
Transmitter: Given that it observes the source sequence X n = x n , the transmitter looks for indices ( m 0 , m 1 , m 2 ) that satisfy
s n ( m 0 ) , u 1 n ( m 0 , m 1 ) , u 2 n ( m 0 , m 2 ) , x n , t n T μ / 2 n p S U 1 U 2 X T 1 .
If successful, it picks one of these indices uniformly at random and sends the codeword w n over the channel, where
w k = f s k ( m 0 ) , u 1 , k ( m 0 , m 1 ) , u 2 , k ( m 0 , m 2 ) , x k , k { 1 , , n } ,
and where ( s k ( m 0 ) , u 1 , k ( m 0 , m 1 ) , u 2 , k ( m 0 , m 2 ) ) denote the k-th components of codewords ( s n ( m 0 ) , u 1 n ( m 0 , m 1 ) , u 2 n ( m 0 , m 2 ) ) . Otherwise, it sends the sequence of inputs t n over the channel.
Receiver i { 1 , 2 } : After observing V i n = v i n and Y i n = y i n , Receiver i { 1 , 2 } looks for indices m 0 { 1 , , 2 n R 0 } and m i { 1 , , 2 n R i } that satisfy the following conditions:
  • ( s n ( m 0 ) , u i n ( m 0 , m i ) , y i n , t n , v i n ) T μ n ( p S U i Y i T V i i ) .
  • H tp s n ( m 0 ) , u i n ( m 0 , m i ) , y i n , t n , v i n ( S , U i | Y i , T , V i )         = min m ˜ 0 , m ˜ i H tp s n ( m ˜ 0 ) , u i n ( m ˜ 0 , m ˜ i ) , y i n , t n , v i n ( S , U i | Y i , T , V i ) ,
If successful, Receiver i declares H ^ i = h ¯ i . Otherwise, it declares H ^ i = h i .
Analysis: Similar to [15] (Appendix D) and omitted.

6. Summary and Conclusions

The paper proposed and analyzed general distributed hypothesis testing schemes both for the case where the sensor can distinguish the two null hypotheses and where it cannot. Our general schemes recover all previously studied special cases. Moreover, our schemes illustrate a similar phenomenon for setups with common noisefree communication links from the sensor to all decision centers: while a tradeoff arises when the transmitter cannot distinguish the alternate hypotheses at the two decision centers, such a tradeoff can almost completely be mitigated when the transmitter can distinguish the alternate hypotheses. In contrast to the noise-free link scenario, under a noisy broadcast channel model, a tradeoff can still arise in this case because decision centers can confuse the decision taken at the transmitter, and thus misinterpret to whom the communication is dedicated.
Interesting directions for future research include information-theoretic converse results and extensions to multiple sensors or more than two decision centers.

Author Contributions

Author Formal analysis, S.S. and M.W.; writing—original draft preparation, S.S.; writing—review and editing, M.W. All authors have read and agreed to the published version of the manuscript.

Funding

Author has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 programme, grant agreement No 715111.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Proof of Theorem 2

The proof is based on the scheme of Section 4. Fix a choice of blocklength n, the small positive μ and the (conditional) pmfs p T , p T 1 | T 1 , p T 2 | T 2 , p W | T T 1 1 , p W | T T 2 2 , p S | X 1 and p S | X 2 so that (22) holds. Assume that I p 1 ( S ; X ) I p 1 ( W ; V 1 | T , T 1 ) and I p 2 ( S ; X ) I p 2 ( W ; V 2 | T , T 2 ) , in which case R 1 , R 2 , R 1 , R 2 are given by (37) and (38). Additionally, set for convenience of notation:
p S i ( s ) = p S i ( s ) , s S ,
p W | T T i i ( w | t , t i ) = p W | T T i i ( w | t , t i ) , t , t i , w W .
The analysis of type-I error probability is similar as in [15] (Appendix A). The main novelty is that because p X 1 ( x ) p X 2 ( x ) for some x X , for sufficiently small values of μ > 0 , the source sequence cannot lie in both T μ / 2 ( p X 1 ) and T μ / 2 ( p X 2 ) . Details are omitted.
Consider the type-II error probability at Receiver 1 averaged over all random codebooks. Define the following events for i { 1 , 2 } :
E Tx , i ( m , ) : { ( S n ( i ; m , ) , X n ) T μ / 2 n ( p S X i ) , ( T n , T i n , W n ( i ; m ) ) T μ / 2 n ( p T T i W i ) , W n ( i ; m ) ) is   sent } ,
E Rx , i ( m , ) : { ( S n ( i ; m , ) , Y i n ) T μ n ( p S Y i i ) , ( T n , T i n , W n ( i ; m ) , V i n ) T μ n ( p T T i W V i i ) , H tp ( S n ( i ; m , ) , Y i n ) ( S | Y i ) = min l ˜ H tp ( S n ( i ; m , ˜ ) , Y i n ) ( S | Y i ) } .
Notice that
E C [ β 1 , n ] = Pr [ H ^ 1 = 0 | H = h 1 ] Pr m , E Rx , 1 ( m , ) | H = h 1 .
Above probability is upper bounded as:
Pr m , E Rx , 1 ( m , ) | H = h 1 Pr m , E Rx , 1 ( m , ) m , E Tx , 1 ( m , ) | H = h 1 + Pr m , E Rx , 1 ( m , ) m , E Tx , 1 c ( m , ) m , E Tx , 2 ( m , ) | H = h 1 + Pr m , E Rx , 1 ( m , ) m , E Tx , 1 c ( m , ) m , E Tx , 2 c ( m , ) | H = h 1 .
The sum of above probabilities can be upper bounded by the sum of the probabilities of the following events:
B 1 : ( m , ) s . t . E Tx , 1 ( m , ) and E Rx , 1 ( m , ) ,
B 2 : ( m , , ) with s . t . E Tx , 1 ( m , ) and E Rx , 1 ( m , ) ,
B 3 : { ( m , m , , ) with and m m s . t . E Tx , 1 ( m , ) and E Rx , 1 ( m , ) } ,
B 4 : ( m , ) E Tx , 1 c ( m , ) ( m , m , , ) s . t . E Tx , 2 ( m , ) E Rx , 1 ( m , ) ,
B 5 : ( m , ) E Tx , 1 c ( m , ) and E Tx , 2 c ( m , ) ( m , ) s . t . E Rx , 1 ( m , ) .
Thus, we have
E C β 1 , n i = 1 5 Pr B i | H = h 1 .
The probabilities of events B 1 , B 2 , B 3 and B 5 can be bounded following similar steps to [15] (Appendix A). This yields:
Pr B 1 | H = h 1 2 n θ μ , standard , 1 δ 1 ( μ ) ,
Pr B 2 | H = h 1 2 n θ μ , dec , 1 δ 2 ( μ ) ,
Pr B 3 | H = h 1 2 n θ μ , dec , 1 δ 3 ( μ ) ,
Pr B 5 | H = h 1 2 n θ μ , miss , 1 δ 5 ( μ ) ,
for some functions δ 1 ( μ ) , δ 2 ( μ ) , δ 3 ( μ ) and δ 5 ( μ ) that go to zero as n goes to infinity and μ 0 , and where we define:
θ standard , i : = min P ˜ S X Y i : | π S X p S X i | < μ / 2 | π S Y i p S Y i i | < μ D ( π S X Y i p S | X i q X Y i i ) ,
θ dec , i : = min P ˜ S X Y i : | π S X p S X i | < μ / 2 | π Y i p Y i i | < μ H p i ( S | Y i ) H π ( S | Y i ) D ( π S X Y i p S | X i q X Y i i ) I p i ( S ; X | Y i ) + I p i ( W ; V i | T , T i ) ,
θ miss , i : = D ( p Y i i q Y i i ) + E p T D p V i | T i Γ V i | W = T I p i ( S ; X | Y i ) + I p i ( W ; V i | T , T i ) .
Consider event B 4 :
Pr B 4 | H = h 1 m , m , Pr [ ( S n ( 2 ; m , ) , X n ) T μ / 2 n ( p S X 2 ) , ( T n , W n ( 2 ; m ) ) T μ / 2 n ( p T W 2 ) , W n ( 2 ; m ) is   sent , ( S n ( 1 ; m , ) , Y 1 n ) T μ n ( p S Y 1 1 ) , ( T n , T 1 n , W n ( 1 ; m ) , V 1 n ) T μ n ( p T T 1 W V 1 1 ) H tp ( S n ( 1 ; m , ) , Y 1 n ) ( S | Y 1 ) = min ˜ H tp ( S n ( 1 ; m , ˜ ) , Y 1 n ) ( S | Y 1 ) | H = h 1 ] ( a ) m , m , Pr [ ( S n ( 2 ; m , ) , X n ) T μ / 2 n ( p S X 2 ) , ( S n ( 1 ; m , ) , Y 1 n ) T μ n ( p S Y 1 1 ) , H tp ( S n ( 1 ; m , ) , Y 1 n ) ( S | Y 1 ) = min ˜ H tp ( S n ( 1 ; m , ˜ ) , Y 1 n ) ( S | Y 1 ) | H = h 1 ] · Pr [ ( T n , T 1 n , W n ( 1 ; m ) , V 1 n ) T μ n ( p T T 1 W V 1 1 ) , ( T n , W n ( 2 ; m ) ) T μ / 2 n ( p T W 2 ) | W n ( 2 ; m ) is   sent , H = h 1 ] ( b ) 2 n ( R 1 + R 1 + R 2 + R 2 ) · max π S S X Y 1 : | π S X p S X 2 | < μ / 2 | π S Y 1 p S Y 1 1 | < μ H π ( S | Y 1 ) H π ( S | Y 1 ) 2 n D π S S X Y 1 p S 2 p S 1 q X Y 1 1 μ · max π T T 1 W W V 1 : | π T W p T W 2 | < μ / 2 | π T T 1 W V 1 p T T 1 W V 1 1 | < μ 2 n D π T T 1 W W V 1 p T p T 1 | T 1 p W | T T 1 1 p W | T 2 Γ V 1 | W μ ,
where ( a ) holds because the channel code is drawn independently of the source code and ( b ) holds by Sanov’s theorem.
Define
θ ˜ μ , cross , 1 : = min π S S X Y 1 : | π S X p S X 2 | < μ / 2 | π S Y 1 p S Y 1 1 | < μ H π ( S | Y 1 ) H π ( S | Y 1 ) D π S S X Y 1 p S 2 p S 1 q X Y 1 1 + min π T T 1 W W V 1 : | π T W p T W 2 | < μ / 2 | π T T 1 W V 1 p T T 1 W V 1 1 | < μ D π T T 1 W W V 1 p T p T 1 | T 1 p W | T T 1 1 p W | T 2 Γ V 1 | W R 1 R 2 R 1 R 2 2 μ ,
and notice that
θ ˜ μ , cross , 1 = ( ( 37 ) & ( 38 ) ) min π S S X Y 1 : | π S X p S X 2 | < μ / 2 | π S Y 1 p S Y 1 1 | < μ H π ( S | Y 1 ) H π ( S | Y 1 ) D π S S X Y 1 p S 2 p S 1 q X Y 1 1 + min π T T 1 W W V 1 : | π T W p T W 2 | < μ | π T T 1 W V 1 p T T 1 W V 1 1 | < μ D π T T 1 W W V 1 p T p T 1 | T 1 p W | T T 1 1 p W | T 2 Γ V 1 | W I p 1 ( S ; X ) I p 2 ( S ; X ) 4 μ = ( c ) min π S S X Y 1 : | π S X q S X 1 | < μ / 2 | π S Y 1 p S Y 1 1 | < μ H π ( S | Y 1 ) H π ( S | Y 1 ) D π S S X Y 1 q S 1 p S 1 q X Y 1 1 + min π T T 1 W W V 1 : | π T W q T W 1 | < μ | π T T 1 W V 1 p T T 1 W V 1 1 | < μ D π T T 1 W W V 1 p T p T 1 | T 1 p W | T T 1 1 q W | T 1 Γ V 1 | W I p 1 ( S ; X ) I q 1 ( S ; X ) 4 μ = ( CR ) min π S S X Y 1 : | π S X q S X 1 | < μ / 2 | π S Y 1 p S Y 1 1 | < μ H π ( S | Y 1 ) H π ( S | Y 1 ) D π S X Y 1 q S | X 1 q X Y 1 1 + E π S X Y 1 D ( π S | S X Y 1 p S 1 ) I p 1 ( S ; X ) + min π T T 1 W W V 1 : | π T W q T W 1 | < μ | π T T 1 W V 1 p T T 1 W V 1 1 | < μ [ D ( π T T 1 W W p T T 1 1 p W | T T 1 1 q W | T 1 ) + E T T 1 W W D ( π V 1 | T T 1 W W π V 1 | T T 1 ) + D ( π V 1 | T T 1 Γ V 1 | W ) ] 4 μ ( DP ) min π S S X Y 1 : | π S X q S X 1 | < μ / 2 | π S Y 1 p S Y 1 1 | < μ H π ( S | Y 1 ) H π ( S | Y 1 ) D π S X Y 1 q S | X 1 q X Y 1 1 + E π Y 1 D ( π S | Y 1 p S 1 ) I p 1 ( S ; X ) + min π T T 1 W W V 1 : | π T W q T W 1 | < μ | π T T 1 W V 1 p T T 1 W V 1 1 | < μ [ E π T T 1 W D ( π V 1 | T T 1 W π V 1 | T T 1 ) + D ( π V 1 | T T 1 Γ V 1 | W ) 4 μ = ( d ) min π S X Y 1 : | π Y 1 p Y 1 1 | < μ H p 1 ( S | Y 1 ) H π ( S | Y 1 ) E q X S 1 D π Y 1 | X S q Y 1 | X 1 + I p 1 ( S ; Y 1 ) I p 1 ( S ; X ) + I p 1 ( V 1 ; W | T , T 1 ) + min π T T 1 W V 1 : | π T W q T W 1 | < μ | π T T 1 V 1 p T T 1 V 1 1 | < μ E π T T 1 W D ( p V 1 | T T 1 1 Γ V 1 | W ) δ 3 ( μ ) = θ μ , cross , 1 δ 4 ( μ )
for a function δ 4 ( μ ) that goes to zero as μ 0 and
θ μ , cross , 1 : = min π S X Y 1 : | π Y 1 p Y 1 1 | < μ H p 1 ( S | Y 1 ) H π ( S | Y 1 ) E q X S 1 D π Y 1 | X S q Y 1 | X 1 + I p 1 ( S ; Y 1 ) I p 1 ( S ; X ) + I p 1 ( V 1 ; W | T , T 1 ) + min π T T 1 W V 1 : | π T W q T W 1 | < μ | π T T 1 V 1 p T T 1 V 1 1 | < μ E π T T 1 W D ( p V 1 | T T 1 1 Γ V 1 | W ) .
Here, ( c ) holds because the condition p X 1 p X 2 implies that h 1 = h ¯ 2 and thus p 2 = q 1 , and ( d ) holds by the constraints in the minimizations.
Combining (A20), (A21) and (A22) establishes:
Pr B 4 | H = h 1 2 n θ μ , cross , 1 δ 3 ( μ ) .
Considering (A12)–(A16) and (A24), we get:
E C β 1 , n max { 2 n θ μ , standard , 1 δ 1 ( μ ) , 2 n θ μ , dec , 1 δ 2 ( μ ) , 2 n θ μ , dec , 1 δ 2 ( μ ) , 2 n θ μ , cross , 1 δ 3 ( μ ) , 2 n θ μ , miss , 1 δ 4 ( μ ) } .
By standard arguments and successively eliminating the worst half of the codewords with respect to α 1 , n and the exponents θ μ , standard , 1 , θ μ , dec , 1 , θ μ , cross , 1 and θ μ , miss , 1 , it can be shown that there exists at least one codebook for which
α 1 , n < ϵ ,
β 1 , n 32 · max { 2 n θ μ , standard , 1 δ 1 ( μ ) , 2 n θ μ , dec , 1 δ 2 ( μ ) , 2 n θ μ , dec , 1 δ 2 ( μ ) , 2 n θ μ , cross , 1 δ 3 ( μ ) , 2 n θ μ , miss , 1 δ 4 ( μ ) } .
Letting μ 0 and n , we get θ μ , standard , 1 θ standard , 1 , θ μ , dec , 1 θ dec , 1 , θ μ , cross , 1 θ cross , 1 and θ μ , miss , 1 θ miss , 1 . A similar bound can be found for θ 2 . This concludes the proof.

References

  1. Ahlswede, R.; Csiszár, I. Hypothesis testing with communication constraints. IEEE Trans. Inf. Theory 1986, 32, 533–542. [Google Scholar] [CrossRef] [Green Version]
  2. Han, T.S. Hypothesis testing with multiterminal data compression. IEEE Trans. Inf. Theory 1987, 33, 759–772. [Google Scholar] [CrossRef]
  3. Shimokawa, H.; Han, T.; Amari, S.I. Error bound for hypothesis testing with data compression. In Proceedings of the 1994 IEEE International Symposium on Information Theory, Trondheim, Norway, 27 June–1 July 1994; p. 114. [Google Scholar]
  4. Shimokawa, H. Hypothesis Testing with Multiterminal Data Compression. Master’s Thesis, University of Tokyo, Tokyo, Janpan, 1994. [Google Scholar]
  5. Weinberger, N.; Kochman, Y. On the reliability function of distributed hypothesis testing under optimal detection. IEEE Trans. Inf. Theory 2019, 65, 4940–4965. [Google Scholar] [CrossRef]
  6. Rahman, M.S.; Wagner, A.B. On the optimality of binning for distributed hypothesis testing. IEEE Trans. Inf. Theory 2012, 58, 6282–6303. [Google Scholar] [CrossRef] [Green Version]
  7. Zhao, W.; Lai, L. Distributed testing against independence with multiple terminals. In Proceedings of the 2014 52nd Annual Allerton Conference on Communication, Control, and Computing (Allerton), Monticello, IL, USA, 30 September–3 October 2014; pp. 1246–1251. [Google Scholar]
  8. Xiang, Y.; Kim, Y.H. Interactive hypothesis testing against independence. In Proceedings of the 2013 IEEE International Symposium on Information Theory, Istanbul, Turkey, 7–12 July 2013; pp. 2840–2844. [Google Scholar]
  9. Katz, G.; Piantanida, P.; Debbah, M. Collaborative distributed hypothesis testing with general hypotheses. In Proceedings of the 2016 IEEE International Symposium on Information Theory (ISIT), Barcelona, Spain, 10–15 July 2016; pp. 1705–1709. [Google Scholar]
  10. Salehkalaibar, S.; Wigger, M.; Timo, R. On hypothesis testing against independence with multiple decision centers. IEEE Trans. Commun. 2018, 66, 2409–2420. [Google Scholar] [CrossRef] [Green Version]
  11. Salehkalaibar, S.; Wigger, M.; Wang, L. Hypothesis testing over the two-hop relay network. IEEE Trans. Inf. Theory 2019, 65, 4411–4433. [Google Scholar] [CrossRef]
  12. Escamilla, P.; Wigger, M.; Zaidi, A. Distributed hypothesis testing with concurrent detection. In Proceedings of the 2018 IEEE International Symposium on Information Theory (ISIT), Vail, CO, USA, 17–22 June 2018. [Google Scholar]
  13. Escamilla, P.; Wigger, M.; Zaidi, A. Distributed hypothesis testing: Cooperation and concurrent detection. IEEE Trans. Inf. Theory 2020, 66, 7550–7564. [Google Scholar] [CrossRef]
  14. Tian, C.; Chen, J. Successive refinement for hypothesis testing and lossless one-helper problem. IEEE Trans. Inf. Theory 2008, 54, 4666–4681. [Google Scholar] [CrossRef] [Green Version]
  15. Salehkalaibar, S.; Wigger, M. Distributed hypothesis testing based on unequal error protection codes. IEEE Trans. Inf. Theory 2020, 66, 4150–4182. [Google Scholar] [CrossRef]
  16. Sreekumar, S.; Gündüz, D. Distributed hypothesis testing over discrete memoryless channels. IEEE Trans. Inf. Theory 2020, 66, 2044–2066. [Google Scholar] [CrossRef]
  17. Borade, S.; Nakiboglu, B.; Zheng, L. Unequal error protection: An information-theoretic perspective. IEEE Trans. Inf. Theory 2009, 55, 5511–5539. [Google Scholar] [CrossRef]
  18. El Gamal, A.; Kim, Y.H. Network Information Theory; Cambridge University Press: Cambridge, MA, USA, 2011. [Google Scholar]
Figure 1. Hypothesis testing over a noisy BC.
Figure 1. Hypothesis testing over a noisy BC.
Information 12 00268 g001
Figure 2. Coding and testing scheme for hypothesis testing over a DMC.
Figure 2. Coding and testing scheme for hypothesis testing over a DMC.
Information 12 00268 g002
Figure 3. Coding and testing scheme for hypothesis testing over a BC.
Figure 3. Coding and testing scheme for hypothesis testing over a BC.
Information 12 00268 g003
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Salehkalaibar, S.; Wigger, M. Distributed Hypothesis Testing over Noisy Broadcast Channels. Information 2021, 12, 268. https://0-doi-org.brum.beds.ac.uk/10.3390/info12070268

AMA Style

Salehkalaibar S, Wigger M. Distributed Hypothesis Testing over Noisy Broadcast Channels. Information. 2021; 12(7):268. https://0-doi-org.brum.beds.ac.uk/10.3390/info12070268

Chicago/Turabian Style

Salehkalaibar, Sadaf, and Michèle Wigger. 2021. "Distributed Hypothesis Testing over Noisy Broadcast Channels" Information 12, no. 7: 268. https://0-doi-org.brum.beds.ac.uk/10.3390/info12070268

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop