Next Article in Journal
Quantum Trajectories: Real or Surreal?
Next Article in Special Issue
Rate-Distortion Function Upper Bounds for Gaussian Vectors and Their Applications in Coding AR Sources
Previous Article in Journal
Numerical Study on Entropy Generation in Thermal Convection with Differentially Discrete Heat Boundary Conditions
Previous Article in Special Issue
Content Adaptive Lagrange Multiplier Selection for Rate-Distortion Optimization in 3-D Wavelet-Based Scalable Video Coding
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Exponential Strong Converse for Source Coding with Side Information at the Decoder †

Department of Communication Engineering and Informatics, University of Electro-Communications, Tokyo 182-8585, Japan
This paper is an extended version of our paper published in 2016 International Symposium on Information Theory and Its Applications, Monterey, CA, USA, 6–9 November 2016; pp. 171–175.
Submission received: 31 January 2018 / Revised: 18 April 2018 / Accepted: 20 April 2018 / Published: 8 May 2018
(This article belongs to the Special Issue Rate-Distortion Theory and Information Theory)

Abstract

:
We consider the rate distortion problem with side information at the decoder posed and investigated by Wyner and Ziv. Using side information and encoded original data, the decoder must reconstruct the original data with an arbitrary prescribed distortion level. The rate distortion region indicating the trade-off between a data compression rate R and a prescribed distortion level Δ was determined by Wyner and Ziv. In this paper, we study the error probability of decoding for pairs of ( R , Δ ) outside the rate distortion region. We evaluate the probability of decoding such that the estimation of source outputs by the decoder has a distortion not exceeding a prescribed distortion level Δ . We prove that, when ( R , Δ ) is outside the rate distortion region, this probability goes to zero exponentially and derive an explicit lower bound of this exponent function. On the Wyner–Ziv source coding problem the strong converse coding theorem has not been established yet. We prove this as a simple corollary of our result.

1. Introduction

For single or multi terminal source coding systems, the converse coding theorems state that at any data compression rates below the fundamental theoretical limit of the system the error probability of decoding cannot go to zero when the block length n of the codes tends to infinity. On the other hand, the strong converse theorems state that, at any transmission rates exceeding the fundamental theoretical limit, the error probability of decoding must go to one when n tends to infinity. The former converse theorems are sometimes called the weak converse theorems to distinguish them with the strong converse theorems.
In this paper, we study the strong converse theorem for the rate distortion problem with side information at the decoder posed and investigated by Wyner and Ziv [1]. We call the above source coding system the Wyner and Ziv source coding system (the WZ system). The WZ system is shown in Figure 1. In this figure, the WZ system corresponds to the case where the switch is close. In Figure 1, the sequence ( X n , Y n ) represents independent copies of a pair of dependent random variables ( X , Y ) which take values in the finite sets X and Y , respectively. We assume that ( X , Y ) has a probability distribution denoted by p X Y . The encoder φ ( n ) outputs a binary sequence which appears at a rate R bits per input symbol. The decoder function ψ ( n ) observes φ ( n ) ( X n ) and Y n to output a sequence Z n . The t-th component Z t of Z n for t = 1 , 2 , , n take values in the finite reproduction alphabet Z . Let d : X × Z [ 0 , ) be an arbitrary distortion measure on X × Z . The distortion between x n X n and z n Z n is defined by
d ( x n , z n ) : = t = 1 n d ( x t , z t ) .
In general, we have two criteria on d ( X n , Z n ) . One is the excess-distortion probability of decoding defined by
P e ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) : = Pr 1 n d ( X n , Z n ) Δ .
The other is the average distortion defined by
Δ n : = E 1 n d ( X , n Z n ) : = ( x n , z n ) X n × Z n 1 n k = 1 n d ( x k , z k ) Pr { X n = x n , Z n = z n } = 1 n k = 1 n ( x k , z k ) X × Z d ( x k , z k ) Pr X k = x k , Z k = z k .
A pair ( R , Δ ) is ε -achievable for p X Y if there exist a sequence of pairs { ( φ ( n ) , ψ ( n ) ) } n 1 such that for any δ > 0 and any n with n n 0 = n 0 ( ε , δ )
1 n log φ ( n ) R + δ , P e ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) ε ,
where φ ( n ) stands for the range of cardinality of φ ( n ) . The rate distortion region R WZ ( ε | p X Y ) is defined by
R WZ ( ε | p X Y ) = ( R , Δ ) : ( R , Δ ) is   ε - achievable   for   p X Y .
Furthermore, set
R WZ ( p X Y ) : = ε > 0 R WZ ( ε | p X Y ) .
On the other hand, we can define a rate distortion region based on the average distortion criterion, a formal definition of which is the following. A pair ( R , Δ ) is achievable for p X Y if there exist a sequence of pairs { ( φ ( n ) , ψ ( n ) ) } n 1 such that for any δ > 0 and any n with n n 0 = n 0 ( δ ) ,
1 n log φ ( n ) R + δ , Δ ( n ) Δ + δ .
The rate distortion region R ˜ WZ ( p X Y ) is defined by
R ˜ WZ ( p X Y ) : = ( R , Δ ) : ( R , Δ ) is   achievable   for   p X Y .
If the switch is open, then the side information is not available to the decoder. In this case the communication system corresponds to the source coding for the discrete memoryless source (DMS) specified with p X . We define the rate distortion region R ˜ DMS ( p X ) in a similar manner to the definition of R ˜ WZ ( p X Y ) . We further define the region R DMS ( ε | p X ) , ε ( 0 , 1 ) and R DMS ( p X ) , respectively in a similar manner to the definition of R WZ ( ε | p X Y ) and R WZ ( p X Y ) .
Previous works on the characterizations of R ˜ DMS ( p X ) , R DMS ( ε | p X ) , ε ( 0 , 1 ) , and R DMS ( p X ) are shown in Table 1. Shannon [2] determined R ˜ DMS ( p X ) . Subsequently, Wolfowiz [3] proved that R ˜ DMS ( p X ) = R DMS ( p X ) . Furthermore, he proved the strong converse theorem. That is, if ( R , Δ ) R ˜ DMS ( p X ) , then for any sequence { ( φ ( n ) , ψ ( n ) ) } n = 1 of encoder and decoder functions satisfying the condition
lim sup n 1 n log | | φ ( n ) | | R ,
we have
lim n P e ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) = lim n Pr 1 n k d ( X n , Z n ) Δ = 1 .
The above strong converse theorem implies that, for any ε ( 0 , 1 ) ,
R ˜ DMS ( p X ) = R DMS ( p X ) = R DMS ( ε | p X ) .
Csiszár and Körner proved that in Equation (3), the probability P e ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) converges to one exponentially and determined the optimal exponent as a function of ( R , Δ ) .
The previous works on the coding theorems for the WZ system are summarized in Table 1. The rate distortion region R ˜ WZ ( p X Y ) was determined by Wyner and Ziv [1]. Csiszár and Körner [4] proved that R ˜ WZ ( p X Y ) = R WZ ( p X Y ) . On the other hand, we have had no result on the strong converse theorem for the WZ system.
Main results of this paper are summarized in Table 1. For the WZ system, we prove that if ( R , Δ ) is out side the rate distortion region R WZ ( p X Y ) , then we have that for any sequence { ( φ ( n ) , ψ ( n ) ) } n = 1 of encoder and decoder functions satisfying the conditionin Equation (2), the quantity P e ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) goes to zero exponentially and derive an explicit lower bound of this exponent function. This result corresponds to Theorem 3 in Table 1. As a corollary from this theorem, we obtain the strong converse result, which is stated in Corollary 2 in Table 1. This results states that we have an outer bound with O ( 1 n ) gap from the rate distortion region R WZ ( p X Y ) .
To derive our result, we use a new method called the recursive method. This method is a general powerful tool to prove strong converse theorems for several coding problems in information theory. In fact, the recursive method plays important roles in deriving exponential strong converse exponent for communication systems treated in [5,6,7,8].

2. Source Coding with Side Information at the Decoder

In the following argument, the operations E p [ · ] and Var p [ · ] , respectively, stand for the expectation and the variance with respect to a probability distribution p. When the value of p is obvious from the context, we omit the suffix p in those operations to simply write E [ · ] and Var [ · ] . Let X and Y be finite sets and ( X t , Y t ) t = 1 be a stationary discrete memoryless source. For each t = 1 , 2 , , the random pair ( X t , Y t ) takes values in X × Y , and has a probability distribution
p X Y = p X Y ( x , y ) ( x , y ) X × Y .
We write n independent copies of X t t = 1 and Y t t = 1 , respectively, as
X n = X 1 , X 2 , , X n   and   Y n = Y 1 , Y 2 , , Y n .
We consider a communication system depicted in Figure 2. Data sequences X n is separately encoded to φ ( n ) ( X n ) and is sent to the information processing center. At the centerm the decoder function ψ ( n ) observes φ ( n ) ( X n ) and Y n to output the estimation Z n of X n . The encoder function φ ( n ) is defined by
φ ( n ) : X n M n = 1 , 2 , , M n .
Let Z be a reproduction alphabet. The decoder function ψ ( n ) is defined by
ψ ( n ) : M n × Y n Z n .
Let d : X × Z [ 0 , ) be an arbitrary distortion measure on X × Z . The distortion between x n X n and z n Z n is defined by
d ( x n , z n ) : = t = 1 n d ( x t , z t ) .
The excess-distortion probability of decoding is
P e ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) = Pr 1 n d ( X n , Z n ) Δ ,
where Z n = ψ ( n ) ( φ ( n ) ( X n ) , Y n ) . The average distortion Δ ( n ) between X n and Z n is defined by
Δ ( n ) : = 1 n E d ( X n , Z n ) : = 1 n t = 1 n E d ( X t , Z t ) .
In the previous section, we gave the formal definitions of R WZ ( ε | p X Y ) , ε ( 0 , 1 ) , R WZ ( p X Y ) , and R ˜ WZ ( p X Y ) . We can show that the above three rate distortion regions satisfy the following property.
Property 1.
(a) 
The regions R WZ ( ε | p X Y ) , ε ( 0 , 1 ) , R WZ ( p X Y ) , and R ˜ WZ ( p X Y ) are closed convex sets of R + 2 , where
R + 2 : = { ( R , Δ ) : R 0 , Δ 0 } .
(b) 
R WZ ( ε | p X Y ) has another form using ( n , ε ) -rate distortion region, the definition of which is as follows. We set
R WZ ( n , ε | p X Y ) = { ( R , Δ ) : T h e r e   e x i s t s ( φ ( n ) , ψ ( n ) ) s u c h   t h a t 1 n log | | φ ( n ) | | R , P e ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) ε } ,
which is called the ( n , ε ) -rate distortion region. Using R WZ ( n , ε | p X Y ) , R WZ ( ε | p X Y ) can be expressed as
R WZ ( ε | p X Y ) = cl m 1 n m R WZ ( n , ε | p X Y ) ,
where cl ( · ) stands for the closure operation.
Proof of this property is given in Appendix A.
It is well known that R ˜ WZ ( p X Y ) was determined by Wyner and Ziv [1]. To describe their result we introduce auxiliary random variables U and Z, respectively, taking values in finite sets U and Z . We assume that the joint distribution of ( U , X , Y , Z ) is
p U X Y Z ( u , x , y , z ) = p U ( u ) p X | U ( x | u ) p Y | X ( y | x ) p Z | U Y ( z | u , y ) .
The above condition is equivalent to
U X Y , X ( U , Y ) Z .
Define the set of probability distribution p = p U X Y Z by
P ( p X Y ) : = { p = p U X Y Z : | U | | X | + 1 , U X Y , X ( U , Y ) Z } , P * ( p X Y ) : = { p = p U X Y Z : | U | | X | + 1 , U X Y , Z = ϕ ( U , Y ) for   some ϕ : U × Y Z } .
By definitions, it is obvious that P * ( p X Y ) P ( p X Y ) . Set
R ( p ) : = { ( R , Δ ) : R , Δ 0 , R I p ( X ; U | Y ) , Δ E p d ( X , Z ) } , R ( p X Y ) : = p P ( p X Y ) R ( p ) , R * ( p X Y ) : = p P * ( p X Y ) R ( p ) .
We can show that the above functions and sets satisfy the following property:
Property 2.
(a) 
The region R ( p X Y ) is a closed convex set of R + 2 .
(b) 
For any p X Y , we have
R ( p X Y ) = R * ( p X Y ) .
Proof of Property 2 is given in Appendix C. In Property 2 Part (b), R ( p X Y ) is regarded as another expression of R * ( p X Y ) . This expression is useful for deriving our main result. The rate region R WZ ( p X Y ) was determined by Wyner and Ziv [1]. Their result is the following:
Theorem 1 
(Wyner and Ziv [1]).
R ˜ WZ ( p X Y ) = R * ( p X Y ) = R ( p X Y ) .
On R WZ ( p X Y ) , Csiszár and Körner [4] obtained the following result.
Theorem 2 
(Csiszár and Körner [4]).
R WZ ( p X Y ) = R ˜ WZ ( p X Y ) = R * ( p X Y ) = R ( p X Y ) .
We are interested in an asymptotic behavior of the error probability of decoding to tend to one as n for ( R , Δ ) R WZ ( p X Y ) . To examine the rate of convergence, we define the following quantity. Set
P c ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) : = 1 P e ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) , G ( n ) ( R , Δ | p X Y ) : = min ( φ ( n ) , ψ ( n ) ) : ( 1 / n ) log φ ( n ) R 1 n log P c ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) .
By time sharing, we have that
G ( n + m ) n R + m R n + m , n Δ + m Δ n + m p X Y n G ( n ) ( R , Δ | p X Y ) + m G ( m ) ( R , Δ | p X Y ) n + m .
Choosing R = R and Δ = Δ in Equation (7), we obtain the following subadditivity property on { G ( n ) ( R , Δ | p X Y ) } n 1 :
G ( n + m ) ( R , Δ | p X Y ) n G ( n ) ( R , Δ | p X Y ) + m G ( m ) ( R , Δ | p X Y ) n + m ,
which together with Fekete’s lemma yields that G ( n ) ( R , Δ | p X Y ) exists and satisfies the following:
lim n G ( n ) ( R , Δ | p X Y ) = inf n 1 G ( n ) ( R , Δ | p X Y ) .
Set
G ( R , Δ | p X Y ) : = lim n G ( n ) ( R , Δ | p X Y ) , G ( p X Y ) : = { ( R , Δ , G ) : G G ( R , Δ | p X Y ) } .
The exponent function G ( R , Δ | p X Y ) is a convex function of ( R , Δ ) . In fact, from Equation (7), we have that for any α [ 0 , 1 ]
G ( α R + α ¯ R , α Δ + α ¯ Δ | p X Y ) α G ( R , Δ | p X Y ) + α ¯ G ( R , Δ | p X Y ) ,
where α ¯ = 1 α . The region G ( p X Y ) is also a closed convex set. Our main aim is to find an explicit characterization of G ( p X Y ) . In this paper, we derive an explicit outer bound of G ( p X Y ) whose section by the plane G = 0 coincides with R WZ ( p X Y ) .

3. Main Results

In this section, we state our main results. We first explain that the rate distortion region R ( p X Y ) can be expressed with two families of supporting hyperplanes. To describe this result, we define two sets of probability distributions on U × X × Y × Z by
P sh ( p X Y ) : = { p U X Y Z : | U | | X | , U X Y , X ( U , Y ) Z } . Q : = { q = q U X Y Z : | U | | X | } .
Let μ ¯ = 1 μ . We set
R ( μ ) ( p X Y ) : = min p P sh ( p X Y ) μ ¯ I p ( X ; U | Y ) + μ E p d ( X ; Z ) , R sh ( p X Y ) : = μ [ 0 , 1 ] { ( R , Δ ) : μ ¯ R + μ Δ R ( μ ) ( p X Y ) } .
Then, we have the following property:
Property 3.
For any p X Y , we have
R sh ( p X Y ) = R ( p X Y ) .
Proof of Property 3 is given in Appendix D. For μ [ 0 , 1 ] and λ , α 0 , define
ω q | | p ( μ , λ ) ( x , y , z | u ) : = log q X ( x ) q Y | X U ( y | x , u ) q Z | U Y X ( z | u , y , x ) p X ( x ) p Y | X ( y | x ) q Z | U Y ( z | u , y ) + λ μ ¯ log q X | Y U ( x | y , u ) p X | Y ( x | y ) + μ d ( x , z ) , Ω ( μ , λ , α ) ( q | p X Y ) : = log E q exp α ω q | | p ( μ , λ ) ( X , Y , Z | U ) , Ω ( μ , λ , α ) ( p X Y ) : = min q Q Ω ( μ , λ , α ) ( q | p X Y ) , F ( μ , λ , α ) ( μ ¯ R + μ Δ | p X Y ) : = Ω ( μ , λ , α ) ( p X Y ) λ α ( μ ¯ R + μ Δ ) 1 + ( 4 + λ μ ¯ ) α .
Furthermore, set
F ( R , Δ | p X Y ) : = sup μ [ 0 , 1 ] , λ , α 0 F ( μ , λ , α ) ( μ ¯ R + μ Δ | p X Y ) , G ¯ ( p X Y ) : = ( R , Δ , G ) : G F ( R , Δ | p X Y ) .
We next define a functions serving as a lower bound of F ( R , Δ | p X Y ) . For each p = p U X Y Z P sh ( p X Y ) , define
ω ˜ p ( μ ) ( x , y , z | u ) : = μ ¯ log p X | Y U ( x | y , u ) p X | Y ( x | y ) + μ d ( x , z ) , Ω ˜ ( μ , λ ) ( p ) : = log E p exp λ ω p ( μ ) ( X , Y , Z | U ) .
Furthermore, set
Ω ˜ ( μ , λ ) ( p X Y ) : = min p P sh ( p X Y ) Ω ˜ ( μ , λ ) ( p ) , F ˜ ( μ , λ ) ( μ R + μ ¯ Δ | p X Y ) : = Ω ˜ ( μ , λ ) ( p X Y ) λ ( μ ¯ R + μ Δ ) 5 + λ ( 1 + μ ¯ ) ,
F ˜ ( R , Δ | p X Y ) : = sup λ 0 , μ [ 0 , 1 ] F ˜ ( μ , λ ) ( μ ¯ R + μ Δ | p X Y ) .
We can show that the above functions satisfies the following properties:
Property 4.
(a) 
The cardinality bound | U | | X | appearing in Q is sufficient to describe the quantity Ω ( μ , λ , α ) ( p X Y ) . Furthermore, the cardinality bound | U | | X | in P sh ( p X Y ) is sufficient to describe the quantity Ω ˜ ( μ , λ ) ( p X Y ) .
(b) 
For any R , Δ 0 , we have
F ( R , Δ | p X Y ) F ˜ ( R , Δ | p X Y ) .
(c) 
Fix any p = p U X Y P sh ( p X Y ) and μ [ 0 , 1 ] . For λ [ 0 , 1 ] , Ω ˜ ( μ , λ ) ( p ) exists and is nonnegative. For p = p U X Y Z P sh ( p X Y ) , define a probability distribution p ( λ ) = p U X Y Z ( λ ) by
p ( λ ) ( u , x , y , z ) : = p ( u , x , y , z ) exp λ ω ˜ p ( μ ) ( x , y , z | u ) E p exp λ ω ˜ p ( μ ) ( X , Y , Z | U ) .
Then, for λ [ 0 , 1 / 2 ] , Ω ˜ ( μ , λ ) ( p ) is twice differentiable. Furthermore, for λ [ 0 , 1 / 2 ] , we have
d d λ Ω ˜ ( μ , λ ) ( p ) = E p ( λ ) ω p ( μ ) ( X , Y , Z | U ) , d 2 d λ 2 Ω ˜ ( μ , λ ) ( p ) = Var p ( λ ) ω ˜ p ( μ ) ( X , Y , Z | U ) .
The second equality implies that Ω ˜ ( μ , λ ) ( p ) is a concave function of λ [ 0 , 1 / 2 ] .
(d) 
For ( μ , λ ) [ 0 , 1 ] × [ 0 , 1 / 2 ] , define
ρ ( μ , λ ) ( p X Y ) : = max ( ν , p ) [ 0 , λ ] × P sh ( p X Y ) : Ω ˜ ( μ , λ ) ( p ) = Ω ˜ ( μ , λ ) ( p X Y ) Var p ( ν ) ω ˜ p ( μ ) ( X , Y , Z | U ) ,
and set
ρ = ρ ( p X Y ) : = max ( μ , λ ) [ 0 , 1 ] × [ 0 , 1 / 2 ] ρ ( μ , λ ) ( p X Y ) .
Then, we have ρ ( p X Y ) < . Furthermore, for any ( μ , λ ) [ 0 , 1 ] × [ 0 , 1 / 2 ] ,
Ω ˜ ( μ , λ ) ( p X Y ) λ R ( μ ) ( p X Y ) λ 2 2 ρ ( p X Y ) .
(e) 
For every τ ( 0 , ( 1 / 2 ) ρ ( p X Y ) ) , the condition ( R + τ , Δ + τ ) R ( p X Y ) implies
F ˜ ( R , Δ | p X Y ) > ρ ( p X Y ) 10 · g 2 τ ρ ( p X Y ) > 0 ,
where g is the inverse function of ϑ ( a ) : = a + ( 1 / 5 ) a 2 , a 0 .
Proof of Property 4 Part (a) is given in Appendix B. Proof of Property 4 Part (b) is given in Appendix E. Proofs of Property 4 Parts (c), (d), and (e) are given in Appendix F.
Our main result is the following:
Theorem 3.
For any R , Δ 0 , any p X Y , and for any ( φ ( n ) , ψ ( n ) ) satisfying ( 1 / n ) log | | φ ( n ) | | R , we have
P c ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) 5 exp n F ( R , Δ | p X Y ) .
It follows from Theorem 3 and Property 4 Part (d) that if ( R , Δ ) is outside the rate distortion region, then the error probability of decoding goes to one exponentially and its exponent is not below F ( R , Δ | p X Y ) .
It immediately follows from Theorem 3 that we have the following corollary.
Corollary 1.
For any R , Δ 0 and any p X Y , we have
G ( R , Δ | p X Y ) F ( R , Δ | p X Y ) .
Furthermore, for any p X Y , we have
G ( p X Y ) G ¯ ( p X Y ) : = ( R , Δ , G ) : G F ( R , Δ | p X Y ) .
Proof of Theorem 3 will be given in the next section. The exponent function in the case of Δ = 0 can be obtained as a corollary of the result of Oohama and Han [9] for the separate source coding problem of correlated sources [10]. The techniques used by them is a method of types [4], which is not useful for proving Theorem 3. In fact, when we use this method, it is very hard to extract a condition related to the Markov chain condition U X Y , which the auxiliary random variable U U must satisfy when ( R , Δ ) is on the boundary of the set R ( p X Y ) . Some novel techniques based on the information spectrum method introduced by Han [11] are necessary to prove this theorem.
From Theorem 3 and Property 4 Part (e), we can obtain an explicit outer bound of R WZ ( ε | p X Y ) with an asymptotically vanishing deviation from R WZ ( p X Y ) = R ( p X Y ) . The strong converse theorem immediately follows from this corollary. To describe this outer bound, for κ > 0 , we set
R ( p X Y ) κ ( 1 , 1 ) : = { ( R κ , Δ κ ) : ( R , Δ ) R ( p X Y ) } ,
which serves as an outer bound of R ( p X Y ) . For each fixed ε ( 0 , 1 ) , we define κ n = κ n ( ε , ρ ( p X Y ) ) by
κ n : = ρ ( p X Y ) ϑ 10 n ρ ( p X Y ) log 5 1 ε = ( a ) 10 ρ ( p X Y ) n log 5 1 ε + 2 n log 5 1 ε .
Step (a) follows from ϑ ( a ) = a + ( 1 / 5 ) a 2 . Since κ n 0 as n , we have the smallest positive integer n 0 = n 0 ( ε , ρ ( p X Y ) ) such that κ n ( 1 / 2 ) ρ ( p X Y ) for n n 0 . From Theorem 3 and Property 4 Part (e), we have the following corollary.
Corollary 2.
For each fixed ε ( 0 , 1 ) , we choose the above positive integer n 0 = n 0 ( ε , ρ ( p X Y ) ) Then, for any n n 0 , we have
R WZ ( n , ε | p X Y ) R ( p X Y ) κ n ( 1 , 1 ) .
The above result together with
R WZ ( ε | p X Y ) = cl m 1 n m R WZ ( n , ε | p X Y ) ,
yields that for each fixed ε ( 0 , 1 ) , we have
R WZ ( ε | p X Y ) = R WZ ( p X Y ) = R ( p X Y ) .
Proof of this corollary will be given in the next section.
The direct part of coding theorem, i.e., the inclusion of R ( p X Y ) R WZ ( ε | p X Y ) was established by Csiszár and Körner [4]. They proved a weak converse theorem to obtain the inclusion R WZ ( p X Y ) R ( p X Y ) . Until now, we have had no result on the strong converse theorem. The above corollary stating the strong converse theorem for the Wyner–Ziv source coding problem implies that a long standing open problem since Csiszár and Körner [4] has been resolved.

4. Proof of the Main Results

In this section, we prove Theorem 3 and Corollary 2. We first present a lemma which upper bounds the correct probability of decoding by the information spectrum quantities. We set
S n : = φ ( n ) ( X n ) , Z n : = ψ n ( φ ( n ) ( X n ) , Y n ) .
It is obvious that
S n X n Y n , X n ( S n , Y n ) Z n .
Then, we have the following:
Lemma 1.
For any η > 0 and for any ( φ ( n ) , ψ ( n ) ) satisfying ( 1 / n ) log | | φ ( n ) | | R , we have
P c ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) p S n X n Y n Z n {
η 1 n log Q X n ( i ) ( X n ) p X n ( X n ) ,
η 1 n log Q Y n | S n X n ( ii ) ( Y n | S n , X n ) p Y n | X n ( Y n | X n ) ,
η 1 n log Q X n | S n Y n Z n ( iii ) ( X n | S n , Y n , Z n ) p X n | S n Y n ( X n | S n , Y n ) ,
R + η 1 n log Q X n | S n Y n ( iv ) ( X n | S n , Y n ) p X n | Y n ( X n | Y n ) ,
Δ 1 n log exp d ( X n , Z n ) + 4 e n η .
The probability distribution and stochastic matrices appearing in the right members of Equation (18) have a property that we can select them arbitrary. In Equation (14), we can choose any probability distribution Q X n ( i ) on X n . In Equation (15), we can choose any stochastic matrix Q Y n | S n X n ( ii ) : M n × X n Y n . In Equation (16), we can choose any stochastic matrix Q X n | S n Y n Z n ( iii ) : M n × Y n × Z n X n . In Equation (17), we can choose any stochastic matrix Q X n | S n Y n ( iv ) : M n × Y n X n .
Proof of this lemma is given in Appendix G.
Lemma 2.
Suppose that, for each t = 1 , 2 , , n , the joint distribution p S n X t Y n of the random vector S n X t Y n is a marginal distribution of p S n X n Y n . Then, for t = 1 , 2 , , n , we have the following Markov chain:
X t S n X t 1 Y t n Y t 1
or equivalently that I ( X t ; Y t 1 | S n X t 1 Y t n ) = 0 .
Proof of this lemma is given in Appendix H. For t = 1 , 2 , , n , set u t : = ( s , x t 1 , y t + 1 n ) . Let U t : = ( S n , X t 1 , Y t + 1 n ) be a random vector taking values in M n × X t 1 × Y t + 1 n . From Lemmas 1 and 2, we have the following:
Lemma 3.
For any η > 0 and for any ( φ ( n ) , ψ ( n ) ) satisfying ( 1 / n ) log | | φ ( n ) | | R , we have the following:
P c ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) p S n X n Y n Z n { η 1 n t = 1 n log Q X t ( i ) ( X t ) p X t ( X t ) , η 1 n t = 1 n log Q Y t | U t X t ( ii ) ( Y t | U t , X t ) p Y t | X t ( Y t | X t ) , η 1 n t = 1 n log Q X t | U t Y t Z t ( iii ) ( X t | U t , Y t , Z t ) p X t | U t Y t ( X t | U t , Y t ) , R + η 1 n t = 1 n log Q X t | U t Y t ( iv ) ( X t | U t , Y t ) p X t | Y t ( X t | Y t ) , Δ 1 n t = 1 n log e d ( X t , Z t ) + 4 e n η ,
where for each t = 1 , 2 , , n , the following probability distribution and stochastic matrices:
Q X t ( i ) , Q Y t | U t X t ( ii ) , Q X t | U t Y t Z t ( iii ) ,   a n d   Q X t | U t Y t ( iv )
appearing in the first term in the right members of Equation (21) have a property that we can choose their values arbitrary.
Proof. 
On the probability distributions appearing in the right members of Equation (18), we take the following choices. In Equation (14), we choose Q X n ( i ) so that
Q X n ( i ) ( X n ) = t = 1 n Q X t ( i ) ( X t ) .
In Equation (15), we choose Q Y n | S n X n ( ii ) so that
Q Y n | S n X n ( ii ) ( Y n | S n , X n ) = t = 1 n Q Y t | S n X t Y t + 1 n ( ii ) ( Y t | S n , X t , Y t + 1 n ) = t = 1 n Q Y t | X t U t ( ii ) ( Y t | U t X t ) .
In Equation (16), we choose Q X n | S n Y n Z n ( iii ) so that
Q X n | S n Y n Z n ( iii ) ( X n | S n , Y n , Z n ) = t = 1 n Q X t | S n X t 1 Y t n Z t ( iii ) ( X t | S n X t 1 , Y t n , Z t ) = t = 1 n Q X t | U t Y t Z t ( iii ) ( X t | U t Y t Z t ) .
In Equation (16), we note that
p X n | S n Y n ( X n | S n , Y n ) = t = 1 n p X t | S n X t 1 Y n ( X t | S n , X t 1 , Y n ) = ( a ) t = 1 n p X t | S n X t 1 Y t n ( X t | S n , X t 1 , Y t n ) = t = 1 n p X t | U t Y t ( X t | U t , Y t ) .
Step (a) follows from Lemma 2. In Equation (17), we choose Q X n | S n Y n ( iv ) so that
Q X n | S n Y n ( iv ) ( X n | S n , Y n ) = t = 1 n Q X t | S n X t 1 Y t n ( iv ) ( X t | S n , X t 1 , Y t n ) = t = 1 n Q X t | U t Y t ( iv ) ( X t | U t , Y t ) .
From Lemma 1 and Equations (21)–(25), we have the bound of Equation (21) in Lemma 3. ☐
To evaluate an upper btound of Equation (21) in Lemma 3, we use the following lemma, which is well known as the Cramér’s bound in the large deviation principle.
Lemma 4.
For any real valued random variable A and any θ 0 , we have
Pr { A a } exp θ a + log E [ exp ( θ A ) ] .
Here, we define a quantity which serves as an exponential upper bound of P c ( n ) ( φ ( n ) , ψ ( n ) ) . For each t = 1 , 2 , , n , let Q ̲ t be a set of all
Q ̲ t = ( Q X t ( i ) , Q Y t | U t X t ( ii ) , Q X t | U t Y t Z t ( iii ) , Q X t | U t Y t ( iv ) ) .
Set
Q ̲ n : = t = 1 n Q ̲ t , Q ̲ n : = Q ̲ t t = 1 n Q ̲ n .
Let P ( n ) ( p X Y ) be a set of all probability distributions p S n X n Y n Z n on M n × X n × Y n × Z n having the form:
p S n X n Y n Z n ( s , x n , y n , z n ) = p S n | X n ( s | x n ) t = 1 n p X t Y t ( x t , y t ) p Z n | Y n S n ( z n | y n , s ) .
For simplicity of notation, we use the notation p ( n ) for p S n X n Y n Z n P ( n ) ( p X Y ) . We assume that p U t X t Y t Z t = p S n X t Y t n Z t is a marginal distribution of p ( n ) . For t = 1 , 2 , , n , we simply write p t = p U t X t Y t Z t . For p ( n ) P ( n ) ( p X Y ) and Q ̲ n Q ̲ n , we define
Ω ( μ , λ , θ ) ( p ( n ) , Q ̲ n | p X Y ) : = log E p ( n ) t = 1 n p X t ( X t ) Q X t ( i ) ( X t ) p Y t | X t ( Y t | X t ) Q Y t | X t U t ( ii ) ( Y t | X t , U t ) θ × t = 1 n p X t | U t Y t ( X t | U t , Y t ) Q X t | U t Y t Z t ( iii ) ( X t | U t , Y t , Z t ) θ t = 1 n p X t | Y t ( X t | Y t ) Q X t | Y t U t ( iv ) ( X t | U t , Y t ) μ ¯ e μ d ( X t , Z t ) λ θ ,
where, for each t = 1 , 2 , , n , the following probability distribution and stochastic matrices:
Q X t ( i ) , Q X t | U t Y t ( ii ) , Q X t | U t Y t Z t ( iii ) , Q Y t | X t U t ( iv )
appearing in the definition of Ω ( μ , λ , θ ) ( p ( n ) , Q ̲ n | p X Y ) are chosen so that they are induced by the joint distribution q t = q U t X t Y t Z t Q t .
By Lemmas 3 and 4, we have the following proposition:
Proposition 1.
For any μ [ 0 , 1 ] , λ , θ 0 , any q n Q n , and any ( φ ( n ) , ψ ( n ) ) satisfying ( 1 / n ) log | | φ ( n ) | | R , we have
P c ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) 5 exp { n 1 + ( 3 + λ μ ¯ ) θ 1 1 n Ω ( μ , λ , θ ) ( p ( n ) , Q ̲ n | p X Y ) λ θ ( μ ¯ R + μ Δ ) .
Proof. 
When Ω ( μ , λ , θ ) ( p ( n ) , Q ̲ n | p X Y ) n λ θ ( μ ¯ R + μ Δ ) , the bound we wish to prove is obvious. In the following argument, we assume that Ω ( μ , λ , θ ) ( p ( n ) , Q ̲ n | p X Y ) > n λ θ ( μ ¯ R + μ Δ ) . We define five random variables A i , i = 1 , 2 , , 5 by
A 1 = 1 n t = 1 n log Q X t ( i ) ( X t ) p X t ( X t ) , A 2 = 1 n t = 1 n log Q Y t | X t U t ( ii ) ( Y t | X t , U t ) p Y t | X t ( Y t | X t ) , A 3 = 1 n t = 1 n log Q X t | U t Y t Z t ( iii ) ( X t | U t , Y t , Z t ) p X t | U t Y t ( X t | U t , Y t ) , A 4 = 1 n t = 1 n log Q X t | U t Y t ( iv ) ( X t | U t , Y t ) p X t | Y t ( X t | Y t ) , A 5 = 1 n t = 1 n log e d ( X t , Z t ) .
By Lemma 3, for any ( φ ( n ) , ψ ( n ) ) satisfying ( 1 / n ) log | | φ ( n ) | | R , we have
P c ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) p S n X n Y n Z n { A i η f o r i = 1 , 2 , 3 , A 4 R + η , A 5 Δ } + 4 e n η p S n X n Y n Z n { A 1 + A 2 + A 3 + λ ( μ ¯ A 4 + μ A 5 ) λ ( μ ¯ R + μ Δ ) + ( 3 + λ μ ¯ ) η } + 4 e n η = p S n X n Y n Z n { A a } + 4 e n η ,
where we set
A : = A 1 + A 2 + A 3 + λ ( μ ¯ A 4 + μ A 5 ) , a : = λ ( μ ¯ R + μ Δ ) + ( 3 + λ μ ¯ ) η .
Applying Lemma 4 to the first term in the right member of Equation (26), we have
P c ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) exp θ a + log E p ( n ) [ exp ( θ A ) ] + 4 e n η = exp [ n { λ θ ( μ ¯ R + μ Δ ) + ( 3 + λ μ ¯ ) θ η 1 n Ω ( μ , λ , θ ) ( p ( n ) , Q ̲ n | p X Y ) + 4 e n η .
We choose η so that
η = λ θ ( μ ¯ R + μ Δ ) + θ ( 3 + λ μ ¯ ) η 1 n Ω ( μ , λ , θ ) ( p ( n ) , Q ̲ n | p X Y ) .
Solving Equation (28) with respect to η , we have
η = 1 n Ω ( μ , λ , θ ) ( p ( n ) , Q ̲ n | p X Y ) λ θ ( μ ¯ R + μ Δ ) 1 + ( 3 + λ μ ¯ ) θ .
For this choice of η and Equation (27), we have
P c ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) 5 e n η = 5 exp { n 1 + ( 3 + λ μ ¯ ) θ 1 1 n Ω ( μ , λ , θ ) ( p ( n ) , Q ̲ n | p X Y ) λ θ ( μ ¯ R + μ Δ ) ,
completing the proof. ☐
Set
Ω ̲ ( μ , λ , θ ) ( p X Y ) : = inf n 1 min p ( n ) P ( n ) ( p X Y ) max Q ̲ n Q ̲ n 1 n Ω ( μ , λ , θ ) ( p ( n ) , Q ̲ n | p X Y ) .
By Proposition 1, we have the following corollary.
Corollary 3.
For any μ [ 0 , 1 ] , λ 0 , for any θ 0 , and for any ( φ ( n ) , ψ ( n ) ) satisfying ( 1 / n ) log | | φ ( n ) | | R , we have
P c ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) 5 exp n Ω ̲ ( μ , λ , θ ) ( p X Y ) λ θ ( μ ¯ R + μ Δ ) 1 + ( 3 + λ μ ¯ ) θ .
We shall call Ω ̲ ( μ , λ , θ ) ( p X Y ) the communication potential. The above corollary implies that the analysis of Ω ̲ ( μ , λ , θ ) ( p X Y ) leads to an establishment of a strong converse theorem for Wyner–Ziv source coding problem. In the following argument, we drive an explicit lower bound of Ω ̲ ( μ , λ , θ ) ( p X Y ) . We use a new technique we call the recursive method. The recursive method is a powerful tool to drive a single letterized exponent function for rates below the rate distortion function. This method is also applicable to prove the exponential strong converse theorems for other network information theory problems [5,6,7]. Set
F t : = ( p X t | U t Y t , Q ̲ t ) , F t : = { F i } i = 1 t .
For each t = 1 , 2 , , n , define a function of ( u t , x t , y t , z t ) U t × X × Y × Z by
f F t ( μ , λ , θ ) ( x t , y t , z t | u t ) : = p X t ( x t ) Q X t ( i ) ( x t ) p Y t | X t ( y t | x t ) Q Y t | X t U t ( ii ) ( y t | x t , u t ) p X t | U t Y t ( x t | u t , y t ) Q X t | U t Y t Z t ( iii ) ( x t | u t , y t , z t ) p X t | Y t ( x t | y t ) Q X t | Y t U t ( iv ) ( x t | u t , y t ) λ μ ¯ e λ μ d ( x t , z t ) θ .
By definition, we have
exp Ω ( μ , λ , θ ) ( p ( n ) , Q ̲ n | p X Y ) = s , y n p S n Y n ( s , y n ) x n , z n p X n Z n | S n Y n ( x n , z n | s , y n ) t = 1 n f F t ( μ , λ , θ ) ( x t , y t , z t | u t ) .
For each t = 1 , 2 , , n , we define the conditional probability distribution
p X t Z t | S n Y n ; F t ( μ , λ , θ ) : = p X t Z t | S n Y n ; F t ( μ , λ , θ ) ( x t , z t | s , y n ) ( x t , z t , s , y n ) X t × Z t × M n × Y n
by
p X t Z t | S n Y n ; F t ( μ , λ , θ ) ( x t , z t | s , y n ) : = C t 1 ( s , y n ) p X t Z t | S n Y n ( x t , z t | s , y n ) i = 1 t f F i ( μ , λ , θ ) ( x i , y i , z i | u i )
where
C t ( s , y n ) : = x t , z t p X t Z t | S n Y n ( x t , z t | s , y n ) i = 1 t f F i ( μ , λ , θ ) ( x i , y i , z i | u i )
are constants for normalization. For t = 1 , 2 , , n , define
Φ t , F t ( μ , λ , θ ) ( s , y n ) : = C t ( s , y n ) C t 1 1 ( s , y n ) ,
where we define C 0 ( s , y n ) = 1 for ( s , y n ) M n × Y n . Then, we have the following lemma:
Lemma 5.
For each t = 1 , 2 , , n , and for any ( s , y n x t , z t ) M n × Y n × X t × Z t , we have
p X t Z t | S n Y n ; F t ( μ , λ , θ ) ( x t , z t | s , y n ) = ( Φ t , F t ( μ , λ , θ ) ( s , y n ) ) 1 p X t 1 Y t 1 | S n Y n ; F t 1 ( μ , λ , θ ) ( x t 1 , z t 1 | s , y n )
× p X t Z t | S n X t 1 Y n ( x t , z t | s , x t 1 , z t 1 , y n ) f F t ( μ , λ , θ ) ( x t , y t , z t | u t ) . Φ t , F t ( μ , λ , θ ) ( s , y n ) = x t , z t p X t 1 Z t 1 | S n Y n ; F t 1 ( μ , λ , θ ) ( x t 1 , z t 1 | s , y n )
× p X t Z t | S n X t 1 Y n ( x t , z t | s , x t 1 , z t 1 , y n ) f F t ( μ , λ , θ ) ( x t , y t , z t | u t ) .
Furthermore, we have
exp Ω ( μ , λ , θ ) ( p ( n ) , Q ̲ n | p X Y ) = s , y n p S n Y n ( s , y n ) t = 1 n Φ t , F t ( μ , λ , θ ) ( s , y n ) .
The equality in Equation (34) in Lemma 5 is obvious from Equations (29)–(31). Proofs of Equations (32) and (33) in this lemma are given in Appendix I. Next, we define a probability distribution of the random pair ( S n , Y n ) taking values in M n × Y n by
p S n Y n ; F t ( μ , λ , θ ) ( s , y n ) = C ˜ t 1 p S n Y n ( s , y n ) i = 1 t Φ i , F i ( μ , λ , θ ) ( s , y n ) ,
where C ˜ t is a constant for normalization given by
C ˜ t = s , y n p S n Y n ( s , y n ) i = 1 t Φ i , F i ( μ , λ , θ ) ( s , y n ) .
For t = 1 , 2 , , n , define
Λ t , F t ( μ , λ , θ ) : = C t ˜ C ˜ t 1 1 ,
where we define C ˜ 0 = 1 . Set
p S n X t Y t n Z t ; F t 1 ( μ , λ , θ ) ( s , x t , y t n , z t ) = p U t X t Y t Z t ; F t 1 ( μ , λ , θ ) ( u t , x t , y t , z t ) : = y t 1 , z t 1 p S n Y n ; F t 1 ( μ , λ , θ ) ( s , y n ) p X t 1 Z t 1 | S n Y n ; F t 1 ( μ , λ , θ ) ( x t 1 , z t 1 | s , y n ) × p X t Z t | X t 1 Z t 1 S n Y n ( x t , z t | x t 1 , z t 1 , s , y n ) .
Then, we have the following:
Lemma 6.
exp Ω ( μ , λ , θ ) ( p ( n ) , Q ̲ n | p X Y ) = t = 1 n Λ t , F t ( μ , λ , θ ) ,
Λ t , F t ( μ , λ , θ ) = u t , x t , y t , z t p U t X t Y t Z t ; F t 1 ( μ , λ , θ ) ( u t , x t , y t , z t ) f F t ( μ , λ , θ ) ( x t , y t , z t | u t ) .
Proof. 
By the equality Equation (34) in Lemma 5, we have
exp Ω ( μ , λ , θ ) ( p ( n ) , Q ̲ n | p X Y ) = C ˜ n = t = 1 n C ˜ t C ˜ t 1 1 = ( a ) t = 1 n Λ t , F t ( μ , λ , θ ) .
Step (a) follows from the definition in Equation (36) of Λ t , F t ( μ , λ , ν , θ ) . We next prove Equation (39) in Lemma 6. Multiplying Λ t , F t ( μ , λ , θ ) = C ˜ t / C ˜ t 1 to both sides of Equation (35), we have
Λ t , F t ( μ , λ , θ ) p S n Y n ; F t ( μ , λ , θ ) ( s , y n ) = C ˜ t 1 1 p S n Y n ( s , y n ) i = 1 t Φ i , F i ( μ , λ , θ ) ( s , y n )
= p S n Y n ; F t 1 ( μ , λ , θ ) ( s , y n ) Φ t , F t ( μ , λ , θ ) ( s , y n ) .
Taking summations of Equations (41) and (42) with respect to ( s , y n ) , we have
Λ t , F t ( μ , λ , θ ) = s , y n p S n Y n ; F t 1 ( μ , λ , θ ) ( s , y n ) Φ t , F t ( μ , λ , θ ) ( s , y n ) = ( a ) s , y n p S n Y n ; F t 1 ( μ , λ , θ ) ( s , y n ) x t , z t p X t 1 Z t 1 | S n Y n ; F t 1 ( μ , λ , θ ) ( x t 1 , z t 1 | s , y n ) × p X t Z t | X t 1 Z t 1 S n Y n ( x t , z t | x t 1 , z t 1 , s , y n ) f F t ( μ , λ , θ ) ( x t , y t , z t | u t ) = s , x t , y t n , z t y t 1 , z t 1 p S n Y n ; F t 1 ( μ , λ , θ ) ( s , y n ) p X t 1 Z t 1 | S n Y n ; F t 1 ( μ , λ , θ ) ( x t 1 , z t 1 | s , y n ) × p X t Z t | X t 1 Z t 1 S n Y n ( x t , z t | x t 1 , z t 1 , s , y n ) f F t ( μ , λ , θ ) ( x t , y t , z t | u t ) . = ( b ) u t , x t , y t , z t p U t X t Y t Z t ; F t 1 ( μ , λ , θ ) ( u t , x t , y t , z t ) f F t ( μ , λ , θ ) ( x t , y t , z t | u t ) .
Step (a) follows from Equation (33) in Lemma 5. Step (b) follows from the definition in Equation (37) of p U t X t Y t Z t ; F t 1 ( μ , λ , θ ) . ☐
The following proposition is a mathematical core to prove our main result.
Proposition 2.
For θ [ 0 , 1 ) , we choose the parameter α such that
α = θ 1 θ θ = α 1 + α .
Then, for any λ 0 , μ [ 0 , 1 ] and for any θ [ 0 , 1 ) , we have
Ω ̲ ( μ , λ , θ ) ( p X Y ) Ω ( μ , λ , α ) ( p X Y ) 1 + α .
Proof. 
Set
Q ^ n : = { q = q U X Y Z : | U | | M n | | X n 1 | | Y n 1 | } , Ω ^ n ( μ , λ , α ) ( p X Y ) : = min q Q ^ n Ω ( μ , λ , α ) ( q | p X Y ) .
Then, by Lemma 6, we have
Λ t , F t ( μ , λ , θ ) = u t , x t , y t , z t p U t X t Y t Z t ; F t 1 ( μ , λ , θ ) ( u t , x t , y t , z t ) f F t ( μ , λ , θ ) ( x t , y t , z t | u t ) .
For each t = 1 , 2 , , n , we recursively choose q t = q U t X t Y t Z t so that q U t X t Y t Z t = p U t X t Y t Z t ; F t 1 ( μ , λ , θ ) and choose Q X t ( i ) , Q Y t | X t U t ( ii ) , Q X t | U t Y t Z t ( iii ) , and Q X t | Y t U t ( iv ) appearing in
f F t ( μ , λ , θ ) ( x t , y t , z t | u t ) : = p X t ( x t ) Q X t ( i ) ( x t ) p Y t | X t ( y t | x t ) Q Y t | X t U t ( ii ) ( y t | x t , u t ) p X t | U t Y t ( x t | u t , y t ) Q X t | U t Y t Z t ( iii ) ( x t | u t , y t , z t ) p X t | Y t ( x t | y t ) Q X t | Y t U t ( iv ) ( x t | u t , y t ) λ μ ¯ e λ μ d ( x t , z t ) θ
such that they are the distributions induced by q U t X t Y t Z t . Then, for each t = 1 , 2 , ⋯, n, we have the following chain of inequalities:
Λ t , F t ( μ , λ , θ ) = E q t p X t ( X t ) q X t ( X t ) p Y t | X t ( Y t | X t ) q Y t | X t U t ( Y t | X t , U t ) p X t | U t Y t ( X t | U t , Y t ) q X t | U t Y t Z t ( X t | U t , Y t , Z t ) p X t | Y t λ μ ¯ ( X t | Y t ) e λ μ d ( X t , Z t ) q X t | Y t U t λ μ ¯ ( X t | U t , Y t ) θ = E q t p X t ( X t ) q X t ( X t ) p Y t | X t ( Y t | X t ) q Y t | X t U t ( Y t | X t , U t ) q X t | U t Y t ( X t | U t , Y t ) q X t | U t Y t Z t ( X t | U t , Y t , Z t ) p X t | Y t λ μ ¯ ( X t | Y t ) e λ μ d ( X t , Z t ) q X t | Y t U t λ μ ¯ ( X t | U t , Y t ) θ × p X t | U t Y t ( X t | U t , Y t ) q X t | U t Y t ( X t | U t , Y t ) θ ( a ) E q t p X t ( X t ) q X t ( X t ) p Y t | X t ( Y t | X t ) q Y t | X t U t ( Y t | X t , U t ) q Z t | U t Y t ( Z t | U t , Y t ) q Z t | U t X t Y t ( Z t | U t , X t , Y t ) p X t | Y t λ μ ¯ ( X t | Y t ) e λ μ d ( X t , Z t ) q X t | Y t U t λ μ ¯ ( X t | U t , Y t ) θ 1 θ 1 θ × E q t p X t | U t Y t ( X t | U t , Y t ) q X t | U t Y t ( X t | U t , Y t ) θ = exp ( 1 θ ) Ω ( μ , λ , θ 1 θ ) ( q t | p X Y ) = ( b ) exp Ω ( μ , λ , α ) ( q t | p X Y ) 1 + α ( c ) exp Ω ^ n ( μ , λ , α ) ( p X Y ) 1 + α = ( d ) exp Ω ( μ , λ , α ) ( p X Y ) 1 + α .
Step (a) follows from Hölder’s inequality and the following identity:
q X t | U t Y t ( X t | U t , Y t ) q X t | U t Y t Z t ( X t | U t , Y t , Z t ) = q Z t | U t Y t ( Z t | U t , Y t ) q Z t | U t X t Y t ( Z t | U t , X t , Y t ) for   t = 1 , 2 , , n .
Step (b) follows from Equation (43). Step (c) follows from the definition of Ω ^ n ( μ , λ , α ) ( p X Y ) . Step (d) follows from that by Property 4 Part (a), the bound | U | | X | , is sufficient to describe Ω ^ n ( μ , λ , α ) ( p X Y ) . Hence, we have the following:
max q n Q n 1 n Ω ( μ , λ , θ ) ( p ( n ) , Q ̲ n | p X Y ) 1 n Ω ( μ , λ , θ ) ( p ( n ) , Q ̲ n | p X Y ) = ( a ) 1 n t = 1 n log Λ t , F t ( μ , λ , θ ) ( b ) Ω ( μ , λ , α ) ( p X Y ) 1 + α .
Step (a) follows from Equation (38) in Lemma 6. Step (b) follows from Equation (45). Since Equation (46) holds for any n 1 and any p ( n ) P ( n ) ( p X Y ) , we have
Ω ̲ ( μ , λ , θ ) ( p X Y ) Ω ( μ , λ , α ) ( p X Y ) 1 + α .
Thus, we have Equation (44) in Proposition 2. ☐
Proof of Theorem 3:
For θ [ 0 , 1 ) , set
α = θ 1 θ θ = α 1 + α .
Then, we have the following:
1 n log 5 P c ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) ( a ) Ω ̲ ( μ , λ , θ ) ( p X Y ) λ θ ( μ ¯ R + μ Δ ) 1 + θ ( 3 + λ μ ¯ ) ( b ) 1 1 + α Ω ( μ , λ , α ) ( p X Y ) λ α 1 + α ( μ ¯ R + μ Δ ) 1 + α 1 + α ( 3 + λ μ ¯ ) = Ω ( μ , λ , α ) ( p X Y ) λ α ( μ ¯ R + μ Δ ) 1 + α + α ( 3 + λ μ ¯ ) = F ( μ , λ , α ) ( μ ¯ R + μ Δ | p X Y ) .
Step (a) follows from Corollary 3. Step (b) follows from Proposition 2 and Equation (47). Since the above bound holds for any positive λ 0 , μ [ 0 , 1 ] , and α 0 , we have
1 n log 5 P c ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) F ( R , Δ | p X Y ) .
Thus, Equation (10) in Theorem 3 is proved. ☐
Proof of Corollary 2:
Since g is an inverse function of ϑ , the definition in Equation (13) of κ n is equivalent to
g κ n ρ ( p X Y ) = 10 n ρ ( p X Y ) log 5 1 ε .
By the definition of n 0 = n 0 ( ε , ρ ( p X Y ) ) , we have that κ n ( 1 / 2 ) ρ ( p X Y ) for n n 0 . We assume that for n n 0 , ( R , Δ ) R WZ ( n , ε | p X Y ) . Then, there exists a sequence { ( φ ( n ) , ψ ( n ) ) } n n 0 such that for n n 0 , we have
1 n log | | φ ( n ) | | R , P e ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) ε .
Then, by Theorem 3, we have
1 ε P c ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) 5 exp n F ( R , Δ | p X Y )
for any n n 0 . We claim that for n n 0 , we have ( R + κ n , Δ + κ n ) R ( p X Y ) . To prove this claim, we suppose that ( R + κ n * , Δ + κ n * ) does not belong to R ( p X Y ) for some n * n 0 . Then, we have the following chain of inequalities:
5 exp n * F ( R , Δ | p X Y ) < ( a ) 5 exp n * ρ ( p X Y ) 10 · g 2 κ n * ρ ( p X Y ) = ( b ) 5 exp n * ρ ( p X Y ) 10 10 n * ρ ( p X Y ) log 5 1 ε = 5 exp log 1 ε 5 = 1 ε .
Step (a) follows from κ n * ( 1 / 2 ) ρ ( p X Y ) and Property 4 Part (e). Step (b) follows from Equation (48). The bound of Equation (50) contradicts Equation (49). Hence, we have ( R + κ n , Δ + κ n ) R ( p X Y ) or equivalent to
( R , Δ ) R ( p X Y ) κ n ( 1 , 1 )
for n n 0 , which implies that for n n 0 ,
R WZ ( n , ε | p X Y ) R ( p X Y ) κ n ( 1 , 1 ) ,
completing the proof. ☐

5. Conclusions

For the WZ system, we have derived an explicit lower bound of the optimal exponent function on the correct probability of decoding for for ( R , Δ ) R WZ ( p X Y ) . We have described this result in Theorem 3. The determination problem of the optimal exponent remains to be resolved. This problem is our future work.
In this paper, we have treated the case where X and Y are finite sets. Extension of Theorem 3 to the case where X and Y are arbitrary sets is also our future work. Wyner [12] investigated the characterization of the rate distortion region in the case where X and Y are general sets and { ( X t , Y t ) } t = 1 is a correlated stationary memoryless source. This work may provide a good tool to investigate the second future work.

Acknowledgments

I am very grateful to Shun Watanabe for his helpful comments.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Properties of the Rate Distortion Regions

In this Appendix, we prove Property 1. Property 1 Part (a) can easily be proved by the definitions of the rate distortion regions. We omit the proofs of this part. In the following argument, we prove Part (b).
Proof of Property 1 Part (b): 
We set
R ̲ WZ ( m , ε | p X Y ) = n m R WZ ( n , ε | p X Y ) .
By the definitions of R ̲ WZ ( m , ε | p X Y ) and R WZ ( ε | p X Y ) , we have that R ̲ WZ ( m , ε | p X Y ) R WZ ( ε | p X Y ) for m 1 . Hence, we have that
m 1 R ̲ WZ ( m , ε | p X Y ) R WZ ( ε | p X Y ) .
We next assume that ( R , Δ ) R WZ ( ε | p X Y ) . Set
R WZ ( δ ) ( ε | p X Y ) : = { ( R + δ , Δ ) : ( R , Δ ) R WZ ( ε | p X Y ) } .
Then, by the definitions of R WZ ( n , ε | p X Y ) and R WZ ( ε | p X Y ) , we have that, for any δ > 0 , there exists n 0 ( ε , δ ) such that for any n n 0 ( ε , δ ) , ( R + δ , Δ ) R WZ ( n , ε | p X Y ) , which implies that
R WZ ( δ ) ( ε | p X Y ) n n 0 ( ε , δ ) R WZ ( n , ε | p X Y ) = R ̲ WZ ( n 0 ( ε , δ ) , ε | p X Y ) cl m 1 R ̲ WZ ( m , ε | p X Y ) .
Here, we assume that there exists a pair ( R , Δ ) belonging to R WZ ( ε | p X Y ) such that
( R , Δ ) cl m 1 R ̲ WZ ( m , ε | p X Y ) .
Since the set in the right hand side of Equation (A3) is a closed set, we have
( R + δ , Δ ) cl m 1 R ̲ WZ ( m , ε | p X Y )
for some small δ > 0 . Note that ( R + δ , Δ ) R WZ ( δ ) ( ε | p X Y ) . Then, Equation (A4) contradicts Equation (A2). Thus, we have
m 1 R ̲ WZ ( m , ε | p X Y ) R WZ ( ε | p X Y ) cl m 1 R ̲ WZ ( m , ε | p X Y ) .
Note here that R WZ ( ε | p X Y ) is a closed set. Then, from Equation (A5), we conclude that
R WZ ( ε | p X Y ) = cl m 1 R ̲ WZ ( m , ε | p X Y ) = cl m 1 n m R WZ ( n , ε | p X Y ) ,
completing the proof. ☐

Appendix B. Cardinality Bound on Auxiliary Random Variables

Set
R ( μ ) ( p X Y ) : = min q P sh ( p X Y ) μ ¯ I q ( X ; U | Y ) + μ E q d ( X , Z ) , R ̲ ( μ ) ( p X Y ) : = min q P ( p X Y ) μ ¯ I q ( X ; U | Y ) + μ E q d ( X , Z ) ,
Since P sh ( p X Y ) P ( p X Y ) , it is obvious that
R ( μ ) ( p X Y ) R ̲ ( μ ) ( p X Y ) .
We first prove the following lemma.
Lemma A1.
R ( μ ) ( p X Y ) = R ̲ ( μ ) ( p X Y ) .
To prove Lemma A1, we set
P 1 ( p X Y ) : = { q 1 = q U X Y : | U | | X | , q X Y = p X Y , U X Y } , Q 1 ( p X Y ) : = { q 1 = q U X Y : | U | | X | + 1 , q X Y = p X Y , U X Y } , Q 2 ( q U X Y ) : = { q 2 = q Z | U X Y : q U X Y Z = ( q U X Y , q 2 ) , X ( U , Y ) Z } .
By definition, it is obvious that
P sh ( p X Y ) = { q = ( q 1 , q 2 ) : q 1 P 1 ( p X Y ) , q 2 Q 2 ( q 1 ) } ,
P ( p X Y ) = { q = ( q 1 , q 2 ) : q 1 Q 1 ( p X Y ) , q 2 Q 2 ( q 1 ) } .
Proof of Lemma A1: 
Since we have Equation (A6), it suffices to show R ( μ ) ( p X Y ) R ̲ ( μ ) ( p X Y ) to prove Lemma A1. We bound the cardinality | U | of U to show that the bound | U | | X | is sufficient to describe R ̲ ( μ ) ( p X Y ) . We first observe that, by Equation (A8), we have
R ̲ ( μ ) ( p X Y ) = min q 1 Q 1 ( p X Y ) min q 2 Q 2 ( q 1 ) μ ¯ I q 1 ( X ; U | Y ) + μ E ( q 1 , q 2 ) d ( X , Z ) = min q 1 Q 1 ( p X Y ) { μ ¯ I q 1 ( X ; U | Y ) + μ min q 2 Q 2 ( q 1 ) E ( q 1 , q 2 ) d ( X , Z ) = min q 1 Q 1 ( p X Y ) μ ¯ I q 1 ( X ; U | Y ) + μ E ( q 1 , q 2 * ( q 1 ) ) d ( X , Z ) ,
where
q 2 * = q 2 * ( q 1 ) = q Z | U Y * = { q Z | U Y * ( z | u , y ) } ( u , y , z ) U × Y × Z
is a conditional probability distribution that attains the following optimization problem:
min q 2 Q 2 ( q 1 ) E ( q 1 , q 2 ) d ( X , Z ) .
Observe that
p X ( x ) = u U q U ( u ) q X | U ( x | u ) ,
μ ¯ I q 1 ( X ; U | Y ) + μ E ( q 1 , q 2 * ) ( X , Z ) = u U p U ( u ) π ( q X | U ( · | u ) ) ,
where
π ( q X | U ( · | u ) ) : = ( x , y , z ) X × Y × Z q X | U ( x | u ) p Y | X ( y | x ) q Z | U Y * ( z | u , y ) × log q X | U μ ¯ ( x | u ) p X μ ¯ ( x ) p Y μ ¯ ( y ) e μ d ( x , z ) x ˜ X p Y | X ( y | x ˜ ) q X | U ( x ˜ | u ) μ ¯ .
For each u U , π ( q X | U ( · | u ) ) is a continuous function of q X | U ( · | u ) . Then, by the support lemma,
| U | | X | 1 + 1 = | X |
is sufficient to express | X | 1 values of Equation (A9) and one value of Equation (A10). ☐
Next, we give a proof of Property 4 Part (a).
Proof of Property 4 Part (a): 
We first bound the cardinality | U | of U in Q to show that the bound | U | | X | is sufficient to describe Ω ( μ , λ , α ) ( p X Y ) . Observe that
q X ( x ) = u U q U ( u ) q X | U ( x | u ) ,
exp Ω ( μ , λ , α ) ( q | p X Y ) = u U q U ( u ) Π ( μ , λ , α ) ( q X , q X Y Z | U ( · , · , · | u ) ) ,
where
Π ( μ , λ , α ) ( q X , q X Y Z | U ( · , · , · | u ) ) : = ( x , y , z ) X × Y × Z q X Y Z | U ( x , y , z | u ) exp α ω q | | p ( μ , λ ) ( x , y , z | u ) .
The value of q X included in Π ( μ , λ , α ) ( q X , q X Y Z | U ( · , · , · | u ) ) must be preserved under the reduction of U . For each u U , Π ( μ , λ ) ( q X , q X Y Z | U ( · , · , · | u ) ) is a continuous function of q X Y Z | U ( · , · , · | u ) . Then, by the support lemma,
| U | | X | 1 + 1 = | X |
is sufficient to express | X | 1 values of Equation (A11) and one value of Equation (A12). We next bound the cardinality | U | of U in P sh ( p X Y ) to show that the bound | U | | X | is sufficient to describe Ω ˜ ( μ , λ ) ( p X Y ) . Observe that
p X ( x ) = u U p U ( u ) p X | U ( x | u ) ,
exp Ω ˜ ( μ , λ ) ( p ) = u U p U ( u ) Π ˜ ( μ , λ ) ( p X , p X Y Z | U ( · , · , · | u ) ) ,
where
Π ˜ ( μ , λ ) ( p X , p X Y Z | U ( · , · , · | u ) ) : = ( x , y , z ) X × Y × Z p X Y Z | U ( x , y , z | u ) exp λ ω p ( μ ) ( x , y , z | u ) .
The value of p X included in Π ˜ ( μ , λ ) ( p X , p X Y Z | U ( · , · , · | u ) ) must be preserved under the reduction of U . For each u U , Π ( μ , λ ) ( p X , p X Y Z | U ( · , · , · | u ) ) is a continuous function of p X Y Z | U ( · , · , · | u ) . Then, by the support lemma,
| U | | X | 1 + 1 = | X |
is sufficient to express | X | 1 values of Equation (A13) and one value of Equation (A14). ☐

Appendix C. Proof of Property 2

In this Appendix, we prove Property 2. Property 2 Part (a) is a well known property. Proof of this property is omitted here. We only prove Property 2 Part (b).
Proof of Property 2 Part (b): 
Since P * ( p X Y ) P ( p X Y ) , it is obvious that R * ( p X Y ) R ( p X Y ) . Hence it suffices to prove that R ( p X Y ) R * ( p X Y ) . We assume that ( R , Δ ) R ( p X Y ) . Then, there exists p P ( p X Y ) such that
R I p ( U ; X | Y ) a n d Δ E p d ( X , Z ) .
On the second inequality in Equation (A15), we have the following:
Δ E p d ( X , Z ) = ( u , y ) U × Y p U Y ( u , y ) z Z p Z | U Y ( z | u , y ) x X d ( x , z ) p X | U Y ( x | u , y ) ( u , y ) U × Y p U Y ( u , y ) min z Z x X d ( x , z ) p X | U Y ( x | u , y ) = ( u , y ) U × Y p U Y ( u , y ) x X d ( x , z * ) p X | U Y ( x | u , y ) ,
where z * = z * ( u , y ) is one of the minimizers of the function
x X d ( x , z ) p X | U Y ( x | u , y ) .
Define ϕ : U × Y Z by ϕ ( u , y ) = z * . We further define q = q U X Y Z by q U X Y = p U X Y , q Z = q ϕ ( U , Y ) . It is obvious that
q P * ( p X Y ) and R I p ( X ; U | Y ) = I q ( X ; U | Y ) .
Furthermore, from Equation (A16), we have
Δ E p d ( X , Z ) E q d ( X , ϕ ( U , Y ) ) .
From Equations (A17) and (A18), we have ( R , Δ ) R * ( p X Y ) . Thus R ( p X Y ) R * ( p X Y ) is proved.

Appendix D. Proof of Property 3

In this Appendix, we prove Property 3. From Property 2 Part (a), we have the following lemma.
Lemma A2.
Suppose that ( R ^ , Δ ^ ) does not belong to R ( p X Y ) . Then, there exist ϵ > 0 and μ * [ 0 , 1 ] such that for any ( R , Δ ) R ( p X Y ) we have
μ * ¯ ( R R ^ ) + μ * ( Δ Δ ^ ) ϵ 0 .
Proof of this lemma is omitted here. Lemma A2 is equivalent to the fact that if the region R ( p X Y ) is a convex set, then for any point ( R ^ , Δ ^ ) outside the region R ( p X Y ) , there exits a line which separates the point ( R ^ , Δ ^ ) from the region R ( p X Y ) . Lemma A2 will be used to prove Equation (8) in Property 3.
Proof of Equation (8) in Property 3:
We first recall the following definitions of P ( p X Y ) and P sh ( p X Y ) :
P ( p X Y ) : = { p U X Y Z : | U | | X | + 1 , U X Y , X ( U , Y ) Z } , P sh ( p X Y ) : = { p U X Y Z : | U | | X | , U X Y , X ( U , Y ) Z } .
We prove R sh ( p X Y ) R ( p X Y ) . We assume that ( R ^ , Δ ^ ) R ( p X Y ) . Then, by Lemma A2, there exist ϵ > 0 and μ * [ 0 , 1 ] such that for any ( R , Δ ) R ( p X Y ) , we have
μ ¯ * R ^ + μ * Δ ^ μ ¯ * R + μ * Δ ϵ .
Hence, we have
μ ¯ * R ^ + μ * Δ ^ min ( R , Δ ) R ( p X Y ) μ ¯ * R + μ * Δ ϵ = ( a ) min p P ( p X Y ) μ ¯ * I p ( U ; X | Y ) + μ * E p d ( X , Z ) ϵ min p P sh ( p X Y ) μ ¯ * I p ( U ; X | Y ) + μ * E p d ( X , Z ) ϵ = R ( μ * ) ( p X Y ) ϵ .
Step (a) follows from the definition of R ( p X Y ) . The inequality in Equation (A19) implies that ( R ^ , Δ ^ ) R sh ( p X Y ) . Thus, R sh ( p X Y ) R ( p X Y ) is concluded. We next prove R ( p X Y ) R sh ( p X Y ) . We assume that ( R , Δ ) R ( p X Y ) . Then, there exists q P ( p X Y ) such that
R I q ( X ; U | Y ) , Δ E q d ( X , Z ) .
Then, for each μ [ 0 , 1 ] and for ( R , Δ ) R ( p X Y ) , we have the following chain of inequalities:
μ ¯ R + μ Δ ( a ) μ ¯ I q ( X ; U | Y ) + μ E q d ( X , Z ) min q P ( p X Y ) μ ¯ I q ( X ; U | Y ) + μ E q d ( X , Z ) = R ̲ ( μ ) ( p X Y ) = ( b ) R ( μ ) ( p X Y ) .
Step (a) follows from Equation (A20). Step (b) follows from Lemma A1. Hence, we have R ( p X Y ) R sh ( p X Y ) . ☐

Appendix E. Proof of Property 4 Part (b)

In this Appendix, we prove Property 4 Part (b). We have the following lemma.
Lemma A3.
For any μ [ 0 , 1 ] , λ 0 , α [ 0 , 1 λ + 1 ] and any q = q U X Y Q ( p Y | X ) , there exists p = p U X Y Z P sh ( p X Y ) such that
Ω ( μ , γ , α ) ( q | p X Y ) α Ω ˜ ( μ , λ ) ( p ) .
This implies that for any μ [ 0 , 1 ] , λ 0 , α [ 0 , 1 λ + 1 ] , we have
Ω ( μ , γ , α ) ( p X Y ) α Ω ˜ ( μ , λ ) ( p X Y ) .
Proof. 
Since Equation (A22) is obvious from Equation (A21), we only prove Equation (A21). We consider the case where ( μ , λ , α ) satisfies ( μ , λ ) [ 0 , 1 ] , λ 0 , and α [ 0 , 1 1 + λ ] . In this case, we have
λ μ ¯ α α ¯ λ α α ¯ λ 1 + λ 1 1 1 + λ = 1 .
For each q = q U X Y Z Q , we choose p = p U X Y Z P sh ( p X Y ) so that p U | X = q U | X and p Z | U Y = q Z | U Y . Then, for any ( μ , λ , α ) satisfying μ [ 0 , 1 ] , λ 0 , and α [ 0 , 1 1 + λ ] , we have the following chain of inequalities:
exp Ω ( μ , λ , α ) ( q | p X Y ) = E q p X ( X ) q X ( X ) p Y | X ( Y | X ) q Y | X U ( Y | X , U ) q Z | U Y ( Z | U , Y ) q Z | U Y X ( Z | U , Y , X ) p X | Y λ μ ¯ ( X | Y ) e λ μ d ( X , Z ) q X | U Y λ μ ¯ ( X | U , Y ) α = E q p U X Y Z ( X , Y , Z , U ) q U X Y Z ( X , Y , Z , U ) p X | Y λ μ ¯ ( X | Y ) e λ μ d ( X , Z ) p X | U Y λ μ ¯ ( X | U , Y ) α p X | U Y ( X | U , Y ) q X | U Y ( X | U , Y ) λ μ ¯ α ( a ) E q p U X Y Z ( X , Y , Z , U ) q U X Y Z ( X , Y , Z , U ) p X | Y λ μ ¯ ( X | Y ) e λ μ d ( X , Z ) p X | U Y λ μ ¯ ( X | U , Y ) α E q p X | U Y ( X | U , Y ) q X | U Y ( X | U , Y ) λ μ ¯ α α ¯ α ¯ = exp α Ω ˜ ( μ , λ ) ( p ) E q p X | U Y ( X | U , Y ) q X | U Y ( X | U , Y ) λ μ ¯ α α ¯ α ¯ ( b ) exp α Ω ˜ ( μ , λ ) ( p ) .
Step (a) follows from Hölder’s inequality. Step (b) follows from Equation (A23) and Hölder’s inequality. ☐
Proof of Property 4 Part (b): 
We have the following chain of inequalities:
F ( R , Δ | p X Y ) = sup μ [ 0 , 1 ] , λ , α 0 Ω ( μ , λ , α ) ( p X Y ) λ α ( μ ¯ R + μ Δ ) 1 + ( 4 + μ ¯ λ ) α sup μ [ 0 , 1 ] , λ 0 sup α 0 , 1 1 + λ Ω ( μ , λ , α ) ( p X Y ) λ α ( μ ¯ R + μ Δ ) 1 + ( 4 + μ ¯ λ ) α ( a ) sup μ [ 0 , 1 ] , λ 0 sup α 0 , 1 1 + λ α [ Ω ˜ ( μ , λ ) ( p X Y ) λ ( μ ¯ R + μ Δ ) ] 1 + ( 4 + μ ¯ λ ) α = ( b ) sup μ [ 0 , 1 ] , λ 0 Ω ˜ ( μ , λ ) ( p X Y ) λ ( μ ¯ R + μ Δ ) 5 + λ ( 1 + μ ¯ ) = F ˜ ( R , Δ | p X Y ) .
Step (a) follows from Lemma A3. Step (b) follows from
sup α 0 , 1 1 + λ α 1 + ( 4 + μ ¯ λ ) α = 1 5 + λ ( 1 + μ ¯ ) ,
completing the proof. ☐

Appendix F. Proof of Property 4 Parts (c), (d), and (e)

In this Appendix, we prove Property 4 Parts (c), (d), and (e). We first prove Parts (c) and (d).
Proof of Property 4 Parts (c) and (d): 
We first prove Part (c). We first show that for each p P sh ( p X Y ) and μ [ 0 , 1 ] , Ω ˜ ( μ , λ ) ( p ) exists for λ [ 0 , 1 ] . On a lower bound of exp [ Ω ˜ ( μ , λ ) ( p ) ] , for λ [ 0 , 1 ] , we have the following:
exp [ Ω ˜ ( μ , λ ) ( p ) ] = ( u , x , y , z ) U × X × Y × Z p U X Y Z ( u , x , y , z ) p X | Y ( x | y ) p X | U Y ( x | u , y ) μ ¯ λ e μ λ d ( x , z )
( a ) ( u , x , y , z ) U × X × Y × Y p U X Y Z ( u , x , y , z ) p X | Y ( x | y ) e d ( x , z ) .
Step (a) follows from that, for μ , λ [ 0 , 1 ] ,
p X | Y ( x | y ) p X | U Y ( x | u , y ) μ ¯ λ e μ λ d ( x , z ) p X | Y ( x | y ) e d ( x , z ) .
It is obvious that the lower bound of Equation (A26) of exp [ Ω ˜ ( μ , λ ) ( p ) ] takes some positive value. This implies that Ω ˜ ( μ , λ ) ( p ) exists for λ [ 0 , 1 ] . We next show that Ω ˜ ( μ , λ ) ( p ) 0 for λ [ 0 , 1 ] . On upper bounds of exp [ Ω ˜ ( μ , λ ) ( p ) ] for λ [ 0 , 1 ] , we have the following chain of inequalities:
exp [ Ω ˜ ( μ , λ ) ( p ) ] ( a ) ( u , x , y , z ) U × X × Y × Z p U X Y Z ( u , x , y , z ) p X | Y ( x | y ) p X | U Y ( x | u , y ) μ ¯ λ = ( u , x , y ) U × X × Y p U Y ( u , y ) p X | U Y 1 μ ¯ λ ( x | u , y ) p X | Y μ ¯ λ ( x | y ) ( b ) ( u , y ) U × Y p U Y ( u , y ) x X p X | U Y ( x | u , y ) 1 μ ¯ λ x X p X | Y ( x | y ) μ ¯ λ = ( u , y ) U × Y p U Y ( u , y ) = 1 .
Step (a) follows from Equation (A25) and e μ λ d ( x , z ) 1 . Step (b) follows from μ ¯ λ [ 0 , 1 ] and Hölder’s inequality. We next prove that that, for each p P sh ( p X Y ) and μ [ 0 , 1 ] , Ω ˜ ( μ , λ ) ( p ) is twice differentiable for λ [ 0 , 1 / 2 ] . For simplicity of notations, set
a ̲ : = ( u , x , y , z ) , A ̲ : = ( U , X , Y , Z ) , A ̲ : = U × X × Y × Z , ω ˜ p ( μ ) ( x , y , z | u ) : = ς ( a ̲ ) , Ω ˜ ( μ , λ ) ( p ) : = ξ ( λ ) .
Then, we have
Ω ˜ ( μ , λ ) ( p ) = ξ ( λ ) = log a ̲ A ̲ p A ̲ ( a ̲ ) e λ ς ( a ̲ ) .
The quantity p ( λ ) ( a ̲ ) = p A ̲ ( λ ) ( a ̲ ) , a ̲ A has the following form:
p ( λ ) ( a ̲ ) = e ξ ( λ ) p ( a ̲ ) e λ ς ( a ̲ ) .
By simple computations, we have
ξ ( λ ) = e ξ ( λ ) a ̲ A ̲ p ( a ̲ ) ς ( a ̲ ) e λ ς ( a ̲ ) = a ̲ A ̲ p ( λ ) ( a ̲ ) ς ( a ̲ ) , ξ ( λ ) = e 2 ξ ( λ ) a ̲ , b ̲ A ̲ p ( a ̲ ) p ( b ̲ ) ς ( a ̲ ) ς ( b ̲ ) 2 2 e λ ς ( a ̲ ) + ς ( b ̲ ) = a ̲ , b ̲ A ̲ p ( λ ) ( a ̲ ) p ( λ ) ( b ̲ ) ς ( a ̲ ) ς ( b ̲ ) 2 2 = a ̲ A ̲ p ( λ ) ( a ̲ ) ς 2 ( a ̲ ) + a ̲ A ̲ p ( λ ) ( a ̲ ) ς ( a ̲ ) 2 0 .
On upper bound of ξ ( λ ) 0 for λ [ 0 , 1 / 2 ] , we have the following chain of inequalities:
ξ ( λ ) ( a ) a ̲ A ̲ p ( λ ) ( a ̲ ) ς 2 ( a ̲ ) = ( b ) a ̲ A ̲ p ( a ̲ ) ς 2 ( a ̲ ) e λ ς ( a ̲ ) + ξ ( λ ) = e ξ ( λ ) a ̲ A ̲ p ( a ̲ ) e 2 λ ς ( a ̲ ) ς 4 ( a ̲ ) ( c ) e 2 ξ ( λ ) ξ ( 2 λ ) a ̲ A ̲ p ( a ̲ ) ς 4 ( a ̲ ) ( d ) e 2 ξ ( λ ) a ̲ A ̲ p ( a ̲ ) ς 4 ( a ̲ ) .
Step (a) follows from Equation (A30). Step (b) follows from Equation (A29). Step (c) follows from Cauchy–Schwarz inequality and Equation (A28). Step (d) follows from that ξ ( 2 λ ) 0 for 2 λ [ 0 , 1 ] . Note that ξ ( λ ) exists for λ [ 0 , 1 / 2 ] . Furthermore, we have the following:
a ̲ A ̲ p ( a ̲ ) ς 4 ( a ̲ ) < .
Hence, by Equation (A31), ξ ( λ ) exists for λ [ 0 , 1 / 2 ] . We next prove Part (d). We derive the lower bound Equation (9) of Ω ˜ ( μ , λ ) ( p X Y ) . Fix any ( μ , λ ) [ 0 , 1 ] × [ 0 , 1 / 2 ] and any p P sh ( p X Y ) . By the Taylor expansion of ξ ( λ ) = Ω ˜ ( μ , λ ) ( p ) with respect to λ around λ = 0 , we have that, for any p P sh ( p X Y ) and for some ν [ 0 , λ ] ,
Ω ˜ ( μ , λ ) ( p ) = ξ ( 0 ) + ξ ( 0 ) λ + 1 2 ξ ( ν ) λ 2 = λ E p ω ˜ p ( μ ) ( X , Y , Z | U ) λ 2 2 Var p ( ν ) ω ˜ p ( μ ) ( X , Y , Z | U ) ( a ) λ R ( μ ) ( p X Y ) λ 2 2 Var p ( ν ) ω ˜ p ( μ ) ( X , Y , Z | U ) .
Step (a) follows from p P sh ( p X Y ) ,
E p ω ˜ p ( μ ) ( X , Y , Z | U ) = μ ¯ I p ( X ; U | Y ) + μ E p d ( X , Z ) ,
and the definition of R ( μ ) ( p X Y ) . Let ( ν opt , p opt ) [ 0 , λ ] × P sh ( p X Y ) be a pair which attains ρ ( μ , λ ) ( p X Y ) .
By this definition, we have that
Ω ˜ ( μ , λ ) ( p opt ) = Ω ˜ ( μ , λ ) ( p X Y )
and that, for any ν [ 0 , λ ] ,
Var p opt ( ν ) ω p opt ( μ ) ( X , Y , Z | U ) Var p opt ( ν opt ) ω p opt ( μ ) ( X , Y , Z | U ) = ρ ( μ , λ ) ( p X Y ) .
On lower bounds of Ω ( μ , λ ) ( p X Y ) , we have the following chain of inequalities:
Ω ˜ ( μ , λ ) ( p X Y ) = ( a ) Ω ˜ ( μ , λ ) ( p opt ) ( b ) λ R ( μ ) ( p X Y ) λ 2 2 Var p opt ( ν ) ω ˜ p opt ( μ ) ( X , Y , Z | U ) ( c ) λ R ( μ ) ( p X Y ) λ 2 2 ρ ( μ , λ ) ( p X Y ) ( d ) λ R ( μ ) ( p X Y ) λ 2 2 ρ ( p X Y ) .
Step (a) follows from Equation (A33). Step (b) follows from Equation (A32). Step (c) follows from Equation (A34). Step (d) follows from the definition of ρ ( p X Y ) . ☐
We next prove Part (e). For the proof we use the following lemma.
Lemma A4.
When τ ( 0 , ( 1 / 2 ) ρ ] , the maximum of
1 5 + 2 λ 1 2 ρ λ 2 + τ λ
for λ ( 0 , 1 / 2 ] is attained by the positive λ 0 satisfying
ϑ ( λ 0 ) : = 1 5 λ 0 2 + λ 0 = τ ρ .
Let g ( a ) be the inverse function of ϑ ( a ) for a 0 . Then, the condition of Equation (A35) is equivalent to λ 0 = g ( τ ρ ) . The maximum is given by
1 5 + 2 λ 0 1 2 ρ λ 0 2 + τ λ 0 = ρ 10 λ 0 2 = ρ 10 g 2 τ ρ .
By an elementary computation we can prove this lemma. We omit the detail.
Proof of Property 4 Part (e): 
By the hyperplane expression R sh ( p X Y ) of R ( p X Y ) stated Property 3 we have that when ( R + τ , Δ + τ ) R ( p X Y ) , we have
μ ¯ * ( R + τ ) + μ * ( Δ + τ ) < R ( μ * ) ( p X Y )
or equivalent to
R ( μ * ) ( p X Y ) ( μ ¯ * R + μ * Δ ) > τ
for some μ * [ 0 , 1 ] . Then, for each positive τ , we have the following chain of inequalities:
F ˜ ( R , Δ | p X Y ) sup λ ( 0 , 1 / 2 ] F ˜ ( μ * , λ ) ( μ ¯ * R + μ * Δ | p X Y ) = sup λ ( 0 , 1 / 2 ] Ω ˜ ( μ * , λ ) ( p X Y ) λ ( μ ¯ * R + μ * Δ ) 5 + λ ( 1 + μ ¯ * ) ( a ) sup λ ( 0 , 1 / 2 ] 1 5 + 2 λ 1 2 ρ λ 2 + λ R ( μ * ) ( p X Y ) λ ( μ ¯ * R + μ * Δ ) } > ( b ) sup λ ( 0 , 1 / 2 ] 1 5 + 2 λ 1 2 ρ λ 2 + τ λ = ( c ) ρ 10 g 2 τ ρ .
Step (a) follows from Property 4 Part (d). Step (b) follows from Equation (A36). Step (c) follows from Lemma A4. ☐

Appendix G. Proof of Lemma 1

To prove Lemma 1, we prepare a lemma. Set
A ˜ n : = x n : 1 n log p X n ( x n ) Q X n ( i ) ( x n ) η , A n : = A ˜ n × M n × Y n × Z n , A n c : = A ˜ n c × M n × Y n × Z n , B ˜ n : = ( s , x n , y n ) : 1 n log p Y n | X n ( y n | x n ) Q Y n | X n S n ( ii ) ( y n | x n , s ) η , B n : = B ˜ n × Z n , B n c : = B ˜ n c × Z n , C n : = ( s , x n , y n , z n ) : 1 n log p X n | S n Y n ( x n | s , y n ) Q X n | S n Y n Z n ( iii ) ( x n | s , y n , z n ) η , D ˜ n : = { ( s , x n , y n ) : s = φ ( n ) ( x n ) , Q X n | S n Y n ( iv ) ( x n | s , y n ) M n e n η p X n | Y n ( x n | y n ) } , D n : = D ˜ n × Z n , D n c : = D ˜ n c × Z n .
Then, we have the following lemma:
Lemma A5.
p S n X n Y n Z n A n c e n η , p S n X n Y n Z n B n c e n η , p S n X n Y n Z n C n c e n η , p S n X n Y n Z n D n c e n η .
Proof. 
We first prove the first inequality. We have the following chain of inequalities:
p S n X n Y n Z n ( A n c ) = p X n ( A ˜ n c ) = x n A ˜ n c p X n ( x n ) ( a ) x n A ˜ n c e n η Q X n ( i ) ( x n ) e n η x n Q X n ( i ) ( x n ) = e n η .
Step (a) follows from the definition of A n . We next prove the second inequality. We have the following chain of inequalities:
p S n X n Y n Z n ( B n c ) = p S n X n Y n ( B ˜ n c ) = ( a ) ( s , x n , y n ) B ˜ n c p S n X n ( s , x n ) p Y n | X n ( y n | x n ) ( b ) ( s , x n , y n ) B ˜ n c e n η p S n X n ( s , x n ) Q Y n | S n X n ( ii ) ( y n | s , x n ) e n η s , x n , y n p S n X n ( s , x n ) Q Y n | S n X n ( ii ) ( y n | s , x n ) = e n η .
Step (a) follows from the Markov chain S n X n Y n . Step (b) follows from the definition of B n . On the third inequality, we have the following chain of inequalities:
p S n X n Y n Z n ( C n c ) = ( a ) ( s , x n , y n , z n ) C n c p X n | S n Y n ( x n | s , y n ) p S n Y n Z n ( s , y n , z n ) ( b ) ( s , x n , y n , z n ) C n c e n η Q X n | S n Y n Z n ( iii ) ( x n | s , y n , z n ) p S n Y n Z n ( s , y n , z n ) e n η s , x n , y n , z n Q X n | S n Y n Z n ( iii ) ( x n | s , y n , z n ) p S n Y n Z n ( s , y n , z n ) = e n η .
Step (a) follows from the Markov chain X n S n Y n Z n . Step (b) follows from the definition of C n . We finally prove the fourth inequality. We have the following chain of inequalities:
p S n X n Y n Z n ( D n c ) = p S n X n Y n ( D ˜ n c ) = s M n ( x n , y n ) : φ ( n ) ( x n ) = s , p X n | Y n ( x n | y n ) ( 1 / M n ) e n η × Q X n | S n , Y n ( iv ) ( x n | s , y n ) p X n | Y n ( x n | y n ) p Y n ( y n ) e n η M n s M n ( x n , y n ) : φ ( n ) ( x n ) = s , p X n | Y n ( x n | y n ) ( 1 / M n ) e n η × Q X n | S n Y n ( iv ) ( x n | s , y n ) Q X n | S n Y n ( iv ) ( x n | s , y n ) p Y n ( y n ) e n η M n s M n x n , y n Q X n | S n Y n ( iv ) ( x n | s , y n ) p Y n ( y n ) = e n η .
Proof of Lemma 1:
We set
E n : = ( s , x n , y n , z n ) : 1 n d ( X n , Z n ) Δ .
Set R ( n ) : = ( 1 / n ) log M n . By definition, we have
p S n X n Y n Z n A n B n C n D n E n = p S n X n Y n Z n { η 1 n log Q X n ( i ) ( X n ) p X n ( X n ) , η 1 n log Q Y n | X n S ( ii ) ( Y n | X n S ) p Y n | X n ( Y n | X n ) , η 1 n log Q X n | S n Y n Z n ( iii ) ( X n | S n Y n Z n ) p X n | S n Y n ( X n | S n Y n ) , R ( n ) + η 1 n log Q X n | S n Y n ( iv ) ( X n | S n Y n ) p X n | Y n ( X n | Y n ) , Δ 1 n log exp d ( X n , Z n ) .
Then, for any ( φ ( n ) , ψ ( n ) ) satisfying
R ( n ) = 1 n log M n R ,
we have
p S n X n Y n Z n A n B n C n D n E n p S n X n Y n Z n { η 1 n log Q X n ( i ) ( X n ) p X n ( X n ) , η 1 n log Q Y n | X n S ( ii ) ( Y n | X n S ) p Y n | X n ( Y n | X n ) , η 1 n log Q X n | S n Y n Z n ( iii ) ( X n | S n Y n Z n ) p X n | S n Y n ( X n | S n Y n ) , R + η 1 n log Q X n | S n Y n ( iv ) ( X n | S n Y n ) p X n | Y n ( X n | Y n ) , Δ 1 n log exp d ( X n , Z n ) .
Hence, it suffices to show
P c ( n ) ( φ 1 ( n ) , φ 2 ( n ) , ψ ( n ) ; Δ ) p S n X n Y n A n B n C n D n E n + 4 e n η
to prove Lemma 1. By definition, we have
P c ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) = p S n X n Y n Z n E n .
Then, we have the following:
P c ( n ) ( φ ( n ) , ψ ( n ) ; Δ ) = p S n X n Y n Z n E n = p S n X n Y n Z n A n B n C n D n E n + p S n X n Y n Z n A n B n C n D n c E n p S n X n Y n Z n A n B n C n D n E n + p S n X n Y n Z n A n c + p S n X n Y n Z n B n c + p S n X n Y n Z n C n c + p S n X n Y n Z n D n c ( a ) p S n X n Y n Z n A n B n C n D n E n + 4 e n η .
Step (a) follows from Lemma A5. ☐

Appendix H. Proof of Lemma 2

In this Appendix, we prove Lemma 2.
Proof of Lemma 2:
We have the following chain of inequalities:
I ( X t ; Y t 1 | S n X t 1 Y t n ) = H ( Y t 1 | S n X t 1 Y t n ) H ( Y t 1 | S n X t Y t n ) H ( Y t 1 | X t 1 ) H ( Y t 1 | S n X n Y t n ) = ( a ) H ( Y t 1 | X t 1 ) H ( Y t 1 | X n Y t n ) = ( b ) H ( Y t 1 | X t 1 ) H ( Y t 1 | X t 1 ) = 0 .
Step (a) follows from that S n = φ ( n ) ( X n ) is a function of X n . Step (b) follows from the memoryless property of the information source { ( X t , Y t ) } t = 1 . ☐

Appendix I. Proof of Lemma 5

In this Appendix, we prove Equations (32) and (33) in Lemma 5.
Proofs Equations (32) and (33) in Lemma 5:
By the definition of p X t Z t | S n Y n , q t ( μ , λ , θ ) ( x t , z t | s , y n ) , for t = 1 , 2 , , n , we have
p X t Z t | S n Y n ; F t ( μ , λ , θ ) ( x t , z t | s , y n ) = C t 1 ( s , y n ) p X t Z t | S n Y n ( x t , z t | s , y n ) i = 1 t f F i ( μ , λ , θ ) ( x i , y i , z i | u i ) .
Then, we have the following chain of equalities:
p X t Z t | S n Y n ; F t ( μ , λ , θ ) ( x t , z t | s , y n ) = ( a ) C t 1 ( s , y n ) p X t Z t | S n Y n ( x t , z t | s , y n ) i = 1 t f F i ( μ , λ , θ ) ( x i , y i , z i | u i ) = C t 1 ( s , y n ) p X t 1 Z t 1 | S n Y n ( x t 1 , z t 1 | s , y n ) i = 1 t 1 f F i ( μ , λ , θ ) ( x i , y i , z i | u i ) × p X t | Z t | X t 1 Z t 1 S Y n ( x t , z t | x t 1 , z t 1 , s , y n ) f F t ( μ , λ , θ ) ( x t , y t | u t ) = ( b ) C t 1 ( s , y n ) C t ( s , y n ) p X t 1 Z t 1 | S n Y n ; F t 1 ( μ , λ , θ ) ( x t 1 , z t 1 | s , y n ) × p X t | Z t | X t 1 Z t 1 S n Y n ( x t , z t | x t 1 , z t 1 , s , y n ) f F t ( μ , λ , θ ) ( x t , y t , z t | u t ) = ( Φ t ( μ , λ , θ ) ( s , y n ) ) 1 p X t 1 Z t 1 | S n Y n ; F t 1 ( μ , λ , θ ) ( x t , y t , z t | u t ) × p X t | Z t | X t 1 Z t 1 S n Y n ( x t , z t | x t 1 , z t 1 , s , y n ) f F t ( μ , λ , θ ) ( x t , y t , z t | u t ) .
Steps (a) and (b) follow from Equation (A39). From Equation (A40), we have
Φ t , q t ( μ , λ , θ ) ( s , y n ) p X t Z t | S n Y n ( μ , λ , θ ) ( x t , z t | s , y n ) = p X t 1 Z t 1 | S n Y n ; F t 1 ( μ , λ , θ ) ( x t 1 , z t 1 | s , y n )
× p X t Z t | X t 1 Z t 1 S n Y n ( x t , z t | x t 1 , z t 1 , s , y n ) f F t ( μ , λ , θ ) ( x t , y t , z t | u t ) .
Taking summations of Equation (A41) and (A42) with respect to x t , z t , we obtain
Φ t , F t ( μ , λ , θ ) ( s , y n ) = x t , z t p X t 1 Z t 1 | S n Y n ; F t 1 ( μ , λ , θ ) ( x t 1 , z t 1 | s , y n ) × p X t Z t | X t 1 Z t 1 S n Y n ( x t , z t | x t 1 , z t 1 , s , y n ) f F t ( μ , λ , θ ) ( x t , y t , z t | u t ) ,
completing the proof. ☐

References

  1. Wyner, A.D.; Ziv, J. The rate-distortion function for source coding with side information at the decoder. IEEE Trans. Inf. Theory 1976, 22, 1–10. [Google Scholar] [CrossRef]
  2. Shannon, C.E. Coding theorems for a discrete source with a fidelity criterion. IRE Nat. Conv. Rec. 1959, 7, 142–163. [Google Scholar]
  3. Wolfowiz, J. Approximation with a fidelity criterion. In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability; University of California Press: Berkeley, CA, USA, 1966; Volume 1, pp. 566–573. [Google Scholar]
  4. Csiszár, I.; Körner, J. Information Theory: Coding Theorems for Discrete Memoryless Systems; Academic: London, UK, 1981. [Google Scholar]
  5. Oohama, Y. Exponent function for one helper source coding problem at rates outside the rate region. In Proceedings of the 2015 IEEE International Symposium on Information Theory, Hong Kong, China, 14–19 June 2015; pp. 1575–1579. [Google Scholar]
  6. Oohama, Y. Strong converse exponent for degraded broadcast channels at rates outside the capacity region. In Proceedings of the 2015 IEEE International Symposium on Information Theory, Hong Kong, China, 14–19 June 2015; pp. 939–943. [Google Scholar]
  7. Oohama, Y. Strong converse theorems for degraded broadcast channels with feedback. In Proceedings of the 2015 IEEE International Symposium on Information Theory, Hong Kong, China, 14–19 June 2015; pp. 2510–2514. [Google Scholar]
  8. Oohama, Y. Exponent function for asymmetric broadcast channels at rates outside the capacity region. In Proceedings of the 2016 IEEE InternationalSymposium on Information Theory and Its Applications, Monterey, CA, USA, 30 October–2 November 2016; pp. 568–572. [Google Scholar]
  9. Oohama, Y.; Han, T.S. Universal coding for the Slepian-wolf data compression system and the strong converse theorem. IEEE Trans. Inf. Theory 1994, 40, 1908–1919. [Google Scholar] [CrossRef]
  10. Slepian, D.; Wolf, J.K. Noiseless coding of correlated information sources. IEEE Trans. Inf. Theory 1973, 19, 471–480. [Google Scholar] [CrossRef]
  11. Han, T.S. Information-Spectrum Methods in Information Theory; Springer: Berlin, Germany, 2002; The Japanese edition was published by Baifukan-Publisher, Tokyo, 1998. [Google Scholar]
  12. Wyner, A.D. The rate-distortion function for source coding with side information at the decoder-II: General sources. Inf. Control 1978, 38, 60–80. [Google Scholar] [CrossRef]
Figure 1. Source encoding with a fidelity criterion with or without side inforamion at the decoder.
Figure 1. Source encoding with a fidelity criterion with or without side inforamion at the decoder.
Entropy 20 00352 g001
Figure 2. Wyner–Ziv source coding system.
Figure 2. Wyner–Ziv source coding system.
Entropy 20 00352 g002
Table 1. Previous results on the converse coding theorems for DMS, WZ. Main results in the present paper on WZ are also included.
Table 1. Previous results on the converse coding theorems for DMS, WZ. Main results in the present paper on WZ are also included.
Characterization of
the Rate Distortion Region
Strong ConverseStrong Converse Exponent
DMSShannon [2] (1959)
(Explicit form of R ˜ DMS ( p X ) )
Wolfowitz [3] (1966)
( R ˜ DMS ( p X ) = R DMS ( p X ) )
Wolfowitz [3] (1966)
( R DMS ( ε | p X ) = R DMS ( p X )
for any ε ( 0 , 1 ) )
Csiszár and Körner [4] (1981)
(The optimal exponent)
WZWyner and Ziv [1] (1976)
(Explicit form of R ˜ WZ ( p X Y ) )
Csiszár and Körner [4] (1981)
( R ˜ WZ ( p X Y ) = R WZ ( p X Y ) )
Corollary 2
(Outer bound with
O 1 / n gap from
the rate distortion region,
R WZ ( ε | p X Y ) = R WZ ( p X Y )
for any ε ( 0 , 1 ) )
Theorem 3
(Lower bound F of
the opt. exp. G)

Share and Cite

MDPI and ACS Style

Oohama, Y. Exponential Strong Converse for Source Coding with Side Information at the Decoder. Entropy 2018, 20, 352. https://0-doi-org.brum.beds.ac.uk/10.3390/e20050352

AMA Style

Oohama Y. Exponential Strong Converse for Source Coding with Side Information at the Decoder. Entropy. 2018; 20(5):352. https://0-doi-org.brum.beds.ac.uk/10.3390/e20050352

Chicago/Turabian Style

Oohama, Yasutada. 2018. "Exponential Strong Converse for Source Coding with Side Information at the Decoder" Entropy 20, no. 5: 352. https://0-doi-org.brum.beds.ac.uk/10.3390/e20050352

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop