Next Article in Journal
The General Solution of Singular Fractional-Order Linear Time-Invariant Continuous Systems with Regular Pencils
Next Article in Special Issue
Fixed-Rate Universal Lossy Source Coding and Model Identification: Connection with Zero-Rate Density Estimation and the Skeleton Estimator
Previous Article in Journal
Spatial Heterogeneity Analysis: Introducing a New Form of Spatial Entropy
Previous Article in Special Issue
Exponential Strong Converse for Source Coding with Side Information at the Decoder
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Rate-Distortion Function Upper Bounds for Gaussian Vectors and Their Applications in Coding AR Sources

by
Jesús Gutiérrez-Gutiérrez
*,
Marta Zárraga-Rodríguez
,
Fernando M. Villar-Rosety
and
Xabier Insausti
Tecnun, University of Navarra, Paseo de Manuel Lardizábal 13, 20018 San Sebastián, Spain
*
Author to whom correspondence should be addressed.
Submission received: 16 April 2018 / Revised: 14 May 2018 / Accepted: 20 May 2018 / Published: 23 May 2018
(This article belongs to the Special Issue Rate-Distortion Theory and Information Theory)

Abstract

:
In this paper, we give upper bounds for the rate-distortion function (RDF) of any Gaussian vector, and we propose coding strategies to achieve such bounds. We use these strategies to reduce the computational complexity of coding Gaussian asymptotically wide sense stationary (AWSS) autoregressive (AR) sources. Furthermore, we also give sufficient conditions for AR processes to be AWSS.

1. Introduction

In 1956, Kolmogorov [1] gave a formula for the rate-distortion function (RDF) of Gaussian vectors and the RDF of Gaussian wide sense stationary (WSS) sources. Later, in 1970 Gray [2] obtained a formula for the RDF of Gaussian autoregressive (AR) sources.
In 1973, Pearl [3] gave an upper bound for the RDF of finite-length data blocks of Gaussian WSS sources, but he did not propose a coding strategy to achieve his bound for a given block length. In [4], we presented two tighter upper bounds for the RDF of finite-length data blocks of Gaussian WSS sources, and we proposed low-complexity coding strategies, based on the discrete Fourier transform (DFT), to achieve such bounds. Moreover, we proved that those two upper bounds tend to the RDF of the WSS source (computed by Kolmogorov in [1]) when the size of the data block grows.
In the present paper, we generalize the upper bounds and the two low-complexity coding strategies presented in [4] to any Gaussian vector. Therefore, in contrast to [4], here no assumption about the structure of the correlation matrix of the Gaussian vector has been made (observe that since the sources in [4] were WSS the correlation matrix of the Gaussian vectors there considered was Toeplitz). To obtain such generalization we start our analysis by first proving several new results on the DFT of random vectors. Although in [4] (Theorem 1) another new result on the DFT was presented, it cannot be used here, because such result and its proof rely on the power spectral density (PSD) of a WSS process and its properties.
The two low-complexity strategies here presented are applied in coding finite-length data blocks of Gaussian AR sources. Specifically, we prove that the rates (upper bounds) corresponding to these two strategies tend to the RDF of the AR source (computed by Gray in [2]) when the size of the data block grows and the AR source is asymptotically WSS (AWSS).
The definition of AWSS process was introduced by Gray in [5] (Chapter 6) and it is based on his concept of asymptotically equivalent sequences of matrices [6]. Sufficient conditions for AR processes to be AWSS can be found in [5] (Theorem 6.2) and [7] (Theorem 7). In this paper we present other sufficient conditions which make easier to check in practice whether an AR process is AWSS.
The paper is organized as follows. In Section 2 we obtain several new results on the DFT of random vectors which are used in Section 3. In Section 3 we give upper bounds for the RDF of Gaussian vectors, and we propose coding strategies to achieve such bounds. In Section 4 we apply the strategies proposed in Section 3 to reduce the computational complexity of coding Gaussian AWSS AR sources. In Section 5 we give sufficient conditions for AR processes to be AWSS. We finish the paper with a numerical example and conclusions.

2. Several New Results on the DFT of Random Vectors

We begin by introducing some notation. C denotes the set of (finite) complex numbers, i is the imaginary unit, Re and Im denote real and imaginary parts, respectively. * stands for conjugate transpose, ⊤ denotes transpose, and λ k ( A ) , k { 1 , , n } , are the eigenvalues of an n × n Hermitian matrix A arranged in decreasing order. E stands for expectation, and V n is the n × n Fourier unitary matrix, i.e.,
[ V n ] j , k = 1 n e 2 π ( j 1 ) ( k 1 ) n i , j , k { 1 , , n } .
If z C then z ^ denotes the real (column) vector
z ^ = Re ( z ) Im ( z ) .
If z k C for all k { 1 , , n } then z n : 1 is the n-dimensional vector given by
z n : 1 = z n z n 1 z n 2 z 1 .
In this section, we give several new results on the DFT of random vectors in two theorems and one lemma.
Theorem 1.
Let y n : 1 be the DFT of an n-dimensional random vector x n : 1 , that is, y n : 1 = V n * x n : 1 .
  • If k { 1 , , n } then
    λ n ( E x n : 1 x n : 1 * ) E x k 2 λ 1 ( E x n : 1 x n : 1 * )
    and
    λ n ( E x n : 1 x n : 1 * ) E y k 2 λ 1 ( E x n : 1 x n : 1 * ) .
  • If the random vector x n : 1 is real and k { 1 , , n 1 } { n 2 } then
    λ n ( E x n : 1 x n : 1 ) 2 E Re ( y k ) 2 λ 1 ( E x n : 1 x n : 1 ) 2 ,
    and
    λ n ( E x n : 1 x n : 1 ) 2 E Im ( y k ) 2 λ 1 ( E x n : 1 x n : 1 ) 2 .
Proof. 
(1) We first prove that if W n is an n × n unitary matrix then
λ n ( E x n : 1 x n : 1 * ) W n diag 1 j n λ j E x n : 1 x n : 1 * W n * n k + 1 , n k + 1 λ 1 ( E x n : 1 x n : 1 * ) .
We have
W n diag 1 j n λ j E x n : 1 x n : 1 * W n * k 1 , k 2 = h = 1 n [ W n ] k 1 , h diag 1 j n λ j E x n : 1 x n : 1 * W n * h , k 2 = h = 1 n [ W n ] k 1 , h l = 1 n diag 1 j n λ j E x n : 1 x n : 1 * h , l W n * l , k 2 = h = 1 n [ W n ] k 1 , h λ h ( E x n : 1 x n : 1 * ) [ W n ] k 2 , h ¯
for all k 1 , k 2 { 1 , , n } , and hence,
W n diag 1 j n λ j E x n : 1 x n : 1 * W n * n k + 1 , n k + 1 = h = 1 n λ h ( E x n : 1 x n : 1 * ) | [ W n ] n k + 1 , h | 2 .
Consequently,
λ n ( E x n : 1 x n : 1 * ) h = 1 n | [ W n ] n k + 1 , h | 2 W n diag 1 j n λ j E x n : 1 x n : 1 * W n * n k + 1 , n k + 1 λ 1 ( E x n : 1 x n : 1 * ) h = 1 n | [ W n ] n k + 1 , h | 2 ,
and applying
h = 1 n | [ W n ] n k + 1 , h | 2 = h = 1 n [ W n ] n k + 1 , h [ W n * ] h , n k + 1 = [ W n W n * ] n k + 1 , n k + 1 = [ I n ] n k + 1 , n k + 1 = 1 ,
where I n denotes the n × n identity matrix, we obtain Equation (5).
Let E x n : 1 x n : 1 * = U n diag 1 j n λ j E x n : 1 x n : 1 * U n 1 be a diagonalization of E x n : 1 x n : 1 * where the eigenvector matrix U n is unitary. As
E x k 2 = E x n : 1 x n : 1 * n k + 1 , n k + 1 = U n diag 1 j n λ j E x n : 1 x n : 1 * U n * n k + 1 , n k + 1 ,
Equation (1) follows directly by taking W n = U n in Equation (5).
Since
E y k 2 = E y n : 1 y n : 1 * n k + 1 , n k + 1 = E V n * x n : 1 x n : 1 * V n * * n k + 1 , n k + 1 = V n * E x n : 1 x n : 1 * V n * * n k + 1 , n k + 1 = V n * U n diag 1 j n λ j E x n : 1 x n : 1 * U n * V n * * n k + 1 , n k + 1 = V n * U n diag 1 j n λ j E x n : 1 x n : 1 * V n * U n * n k + 1 , n k + 1 ,
taking W n = V n * U n in Equation (5) we obtain Equation (2).
(2) Applying [4] (Equation (10)) and taking W n = U n in Equation (6) yields
E Re ( y k ) 2 = 1 n k 1 , k 2 = 1 n cos 2 π ( 1 k 1 ) k n cos 2 π ( 1 k 2 ) k n E x n k 1 + 1 x n k 2 + 1 = 1 n k 1 , k 2 = 1 n cos 2 π ( 1 k 1 ) k n cos 2 π ( 1 k 2 ) k n E x n : 1 x n : 1 k 1 , k 2 = 1 n k 1 , k 2 = 1 n cos 2 π ( 1 k 1 ) k n cos 2 π ( 1 k 2 ) k n U n diag 1 j n λ j E x n : 1 x n : 1 U n * k 1 , k 2 = 1 n k 1 , k 2 = 1 n cos 2 π ( 1 k 1 ) k n cos 2 π ( 1 k 2 ) k n h = 1 n [ U n ] k 1 , h λ h E x n : 1 x n : 1 [ U n ] k 2 , h ¯ = 1 n h = 1 n λ h E x n : 1 x n : 1 k 1 = 1 n cos 2 π ( 1 k 1 ) k n [ U n ] k 1 , h k 2 = 1 n cos 2 π ( 1 k 2 ) k n [ U n ] k 2 , h ¯ = 1 n h = 1 n λ h E x n : 1 x n : 1 l = 1 n cos 2 π ( 1 l ) k n [ U n ] l , h 2 ,
and therefore,
λ n E x n : 1 x n : 1 1 n h = 1 n l = 1 n cos 2 π ( 1 l ) k n [ U n ] l , h 2 E Re ( y k ) 2 λ 1 E x n : 1 x n : 1 1 n h = 1 n l = 1 n cos 2 π ( 1 l ) k n [ U n ] l , h 2 .
Analogously, it can be proved that
λ n E x n : 1 x n : 1 1 n h = 1 n l = 1 n sin 2 π ( 1 l ) k n [ U n ] l , h 2 E Im ( y k ) 2 λ 1 E x n : 1 x n : 1 1 n h = 1 n l = 1 n sin 2 π ( 1 l ) k n [ U n ] l , h 2 .
To finish the proof we only need to show that
1 n h = 1 n l = 1 n cos 2 π ( 1 l ) k n [ U n ] l , h 2 = 1 n h = 1 n l = 1 n sin 2 π ( 1 l ) k n [ U n ] l , h 2 = 1 2 .
If b 1 , , b n are n real numbers then
1 n h = 1 n l = 1 n b l [ U n ] l , h 2 = 1 n h = 1 n k 1 = 1 n b k 1 [ U n ] k 1 , h k 2 = 1 n b k 2 [ U n ] k 2 , h ¯ = 1 n k 1 , k 2 = 1 n b k 1 b k 2 h = 1 n [ U n ] k 1 , h [ U n * ] h , k 2 = 1 n k 1 , k 2 = 1 n b k 1 b k 2 U n U n * k 1 , k 2 = 1 n k 1 , k 2 = 1 n b k 1 b k 2 I n k 1 , k 2 = 1 n l = 1 n b l 2 ,
and thus,
1 n h = 1 n l = 1 n sin 2 π ( 1 l ) k n [ U n ] l , h 2 = 1 n l = 1 n sin 2 π ( 1 l ) k n 2 = 1 n l = 1 n 1 cos 2 π ( 1 l ) k n 2 = 1 1 n l = 1 n cos 2 π ( 1 l ) k n 2 = 1 1 n h = 1 n l = 1 n cos 2 π ( 1 l ) k n [ U n ] l , h 2 .
Equation (8) now follows directly from [4] (Equation (15)). ☐
Lemma 1.
Let y n : 1 be the DFT of an n-dimensional random vector x n : 1 . If k { 1 , , n } then
  • E y k 2 = V n * E x n : 1 x n : 1 * V n n k + 1 , n k + 1 .
  • E y k 2 = V n * E x n : 1 x n : 1 V n ¯ n k + 1 , n k + 1 .
  • E Re y k Im y k = 1 2 Im E y k 2 .
  • E Re ( y k ) 2 = E y k 2 + Re E y k 2 2 .
  • E Im ( y k ) 2 = E y k 2 Re E y k 2 2 .
Proof. 
(1) It is a direct consequence of Equation (7).
(2) We have
E y k 2 = E y n : 1 y n : 1 n k + 1 , n k + 1 = E V n * x n : 1 x n : 1 V n * n k + 1 , n k + 1 = E V n * x n : 1 x n : 1 V n ¯ n k + 1 , n k + 1 = V n * E x n : 1 x n : 1 V n ¯ n k + 1 , n k + 1 .
(3) Observe that
E y k 2 = E Re ( y k ) 2 Im ( y k ) 2 + 2 Re ( y k ) Im ( y k ) i = E Re ( y k ) 2 E Im ( y k ) 2 + 2 E Re ( y k ) Im ( y k ) i ,
and hence,
Im E y k 2 = 2 E Re ( y k ) Im ( y k ) .
(4) and (5) From Equation (10) we obtain
Re E y k 2 = E Re ( y k ) 2 E Im ( y k ) 2 .
Furthermore,
E y k 2 = E Re ( y k ) 2 + Im ( y k ) 2 = E Re ( y k ) 2 + E Im ( y k ) 2 .
(4) and (5) follow directly from Equations (11) and (12). ☐
Theorem 2.
Let y n : 1 be the DFT of a real n-dimensional random vector x n : 1 . If k { 1 , , n 1 } { n 2 } then
λ n ( E x n : 1 x n : 1 ) 2 λ 2 E y k ^ y k ^ λ 1 E y k ^ y k ^ λ 1 ( E x n : 1 x n : 1 ) 2 .
Proof. 
Fix r { 1 , 2 } and consider a real unit eigenvector v = ( v 1 , v 2 ) corresponding to λ r E y k ^ y k ^ . We have
λ r E y k ^ y k ^ = λ r E y k ^ y k ^ v v = v λ r E y k ^ y k ^ v = v E y k ^ y k ^ v .
From [4] (Equation (10)) we obtain
E y k ^ y k ^ = 1 n k 1 , k 2 = 1 n cos 2 π ( 1 k 1 ) k n cos 2 π ( 1 k 2 ) k n E x n k 1 + 1 x n k 2 + 1 cos 2 π ( 1 k 1 ) k n sin 2 π ( 1 k 2 ) k n E x n k 1 + 1 x n k 2 + 1 sin 2 π ( 1 k 1 ) k n cos 2 π ( 1 k 2 ) k n E x n k 1 + 1 x n k 2 + 1 sin 2 π ( 1 k 1 ) k n sin 2 π ( 1 k 2 ) k n E x n k 1 + 1 x n k 2 + 1 = 1 n k 1 , k 2 = 1 n E x n : 1 x n : 1 k 1 , k 2 w k 1 w k 2
with
w l = cos 2 π ( 1 l ) k n sin 2 π ( 1 l ) k n , l { 1 , , n } ,
and consequently,
λ r E y k ^ y k ^ = 1 n k 1 , k 2 = 1 n E x n : 1 x n : 1 k 1 , k 2 v w k 1 w k 2 v = 1 n k 1 , k 2 = 1 n h = 1 n [ U n ] k 1 , h λ h E x n : 1 x n : 1 [ U n ] k 2 , h ¯ v w k 1 w k 2 v = 1 n k 1 , k 2 = 1 n w k 1 v h = 1 n [ U n ] k 1 , h λ h E x n : 1 x n : 1 [ U n ] k 2 , h ¯ w k 2 v = 1 n h = 1 n λ h E x n : 1 x n : 1 k 1 = 1 n w k 1 v [ U n ] k 1 , h k 2 = 1 n w k 2 v [ U n ] k 2 , h ¯ = 1 n h = 1 n λ h E x n : 1 x n : 1 l = 1 n w l v [ U n ] l , h 2
with E x n : 1 x n : 1 = U n diag 1 j n λ j E x n : 1 x n : 1 U n 1 being a diagonalization of E x n : 1 x n : 1 where the eigenvector matrix U n is unitary. Therefore,
λ n E x n : 1 x n : 1 1 n h = 1 n l = 1 n w l v [ U n ] l , h 2 λ r E y k ^ y k ^ λ 1 E x n : 1 x n : 1 1 n h = 1 n l = 1 n w l v [ U n ] l , h 2 .
To finish the proof we only need to show that
1 n h = 1 n l = 1 n w l v [ U n ] l , h 2 = 1 2 .
Applying Equation (9) and [4] (Equations (14) and (15)) yields
1 n h = 1 n l = 1 n w l v [ U n ] l , h 2 = 1 n l = 1 n w l v 2 = 1 n l = 1 n cos 2 π ( 1 l ) k n v 1 + sin 2 π ( 1 l ) k n v 2 2 = v 1 2 1 n l = 1 n cos 2 π ( 1 l ) k n 2 + v 2 2 1 n l = 1 n sin 2 π ( 1 l ) k n 2 + 2 v 1 v 2 1 n l = 1 n cos 2 π ( 1 l ) k n sin 2 π ( 1 l ) k n = v 1 2 1 n l = 1 n cos 2 π ( 1 l ) k n 2 + v 2 2 2 + v 1 v 2 1 n l = 1 n sin 4 π ( 1 l ) k n = v 1 2 1 n l = 1 n 1 sin 2 π ( 1 l ) k n 2 + v 2 2 2 v 1 v 2 1 n l = 1 n sin 4 π ( l 1 ) k n = v 1 2 1 1 n l = 1 n sin 2 π ( 1 l ) k n 2 + v 2 2 2 v 1 v 2 1 n l = 1 n Im e 4 π ( l 1 ) k n i = v 1 2 2 + v 2 2 2 v 1 v 2 1 n Im l = 1 n e 4 π ( l 1 ) k n i = 1 2 v v = 1 2 .

3. RDF Upper Bounds for Real Gaussian Vectors

We first review the formula for the RDF of a real Gaussian vector given by Kolmogorov in [1].
Theorem 3.
If x n : 1 is a real zero-mean Gaussian n-dimensional vector with positive definite correlation matrix, its RDF is given by
R x n : 1 ( D ) = 1 n k = 1 n max 0 , 1 2 ln λ k E x n : 1 x n : 1 θ D 0 , tr E x n : 1 x n : 1 n ,
where tr denotes trace and θ is a real number satisfying
D = 1 n k = 1 n min θ , λ k E x n : 1 x n : 1 .
We recall that R x n : 1 ( D ) can be thought of as the minimum rate (measured in nats) at which one must encode (compress) x n : 1 in order to be able to recover it with a mean square error (MSE) per dimension not larger than D, that is:
E x n : 1 x n : 1 ˜ 2 2 n D ,
where x n : 1 ˜ denotes the estimation of x n : 1 and · 2 is the spectral norm.
The following result provides an optimal coding strategy for x n : 1 in order to achieve R x n : 1 ( D ) whenever D λ n E x n : 1 x n : 1 . Observe that if D λ n E x n : 1 x n : 1 then
R x n : 1 ( D ) = 1 2 n k = 1 n ln λ k E x n : 1 x n : 1 D = 1 2 n ln det E x n : 1 x n : 1 D n .
Corollary 1.
Suppose that x n : 1 is as in Theorem 3. Let E x n : 1 x n : 1 = U n diag 1 k n λ k E x n : 1 x n : 1 U n 1 be a diagonalization of E x n : 1 x n : 1 where the eigenvector matrix U n is real and orthogonal. If D 0 , λ n E x n : 1 x n : 1 then
R x n : 1 ( D ) = 1 n k = 1 n R z k ( D ) = 1 2 n k = 1 n ln E z k 2 D
with z n : 1 = U n x n : 1 .
Proof. 
We encode z 1 , , z n separately with E z k z k ˜ 2 2 D for all k { 1 , , n } . Let x n : 1 ˜ : = U n z n : 1 ˜ , where
z n : 1 ˜ : = z n ˜ z 1 ˜ .
As U n is unitary (in fact, it is a real orthogonal matrix) and the spectral norm is unitarily invariant, we have
E x n : 1 x n : 1 ˜ 2 2 n = E U n x n : 1 U n x n : 1 ˜ 2 2 n = E z n : 1 z n : 1 ˜ 2 2 n = E k = 1 n z k z k ˜ 2 n = k = 1 n E z k z k ˜ 2 n = k = 1 n E z k z k ˜ 2 2 n D ,
and thus,
R x n : 1 ( D ) 1 n k = 1 n R z k ( D ) .
To finish the proof we show Equation (14). Since
E z n : 1 z n : 1 = E U n x n : 1 x n : 1 U n = U n E x n : 1 x n : 1 U n = diag 1 k n λ k E x n : 1 x n : 1 ,
we obtain
E z k 2 = E z n : 1 z n : 1 n k + 1 , n k + 1 = λ n k + 1 E x n : 1 x n : 1 λ n E x n : 1 x n : 1 D > 0 .
Hence, applying Equation (13) yields
1 n k = 1 n R z k ( D ) = 1 n k = 1 n 1 2 ln E z k 2 D = 1 2 n k = 1 n ln λ n k + 1 E x n : 1 x n : 1 D = 1 2 n k = 1 n ln λ k E x n : 1 x n : 1 D = R x n : 1 ( D ) .
Corollary 1 shows that an optimal coding strategy for x n : 1 is to encode z 1 , , z n separately.
We now give two coding strategies for x n : 1 based on the DFT whose computational complexity is lower than the computational complexity of the optimal coding strategy provided in Corollary 1.
Theorem 4.
Let x n : 1 be as in Theorem 3. Suppose that y n : 1 is the DFT of x n : 1 and D 0 , λ n E x n : 1 x n : 1 . Then
R x n : 1 ( D ) R ˜ x n : 1 ( D ) R ˘ x n : 1 ( D ) 1 2 n k = 1 n ln E ( | y k | 2 ) D
R x n : 1 ( D ) + 1 2 ln 1 + E x n : 1 x n : 1 V n diag 1 k n V n * E x n : 1 x n : 1 V n k , k V n * F n λ n E x n : 1 x n : 1 ,
where · F is the Frobenius norm,
R ˜ x n : 1 ( D ) : = R y n 2 D + 2 k = n 2 + 1 n 1 R y k ^ D 2 + R y n ( D ) n i f n i s e v e n , 2 k = n + 1 2 n 1 R y k ^ D 2 + R y n ( D ) n i f n i s o d d ,
and
R ˘ x n : 1 ( D ) : = R y n 2 D + k = n 2 + 1 n 1 R Re y k D 2 + R Im y k D 2 + R y n ( D ) n i f n i s e v e n , k = n + 1 2 n 1 R Re y k D 2 + R Im y k D 2 + R y n ( D ) n i f n i s o d d .
Proof. 
Equations (15) and (16) were presented in [4] (Equations (16) and (20)) for the case where the correlation matrix E x n : 1 x n : 1 is Toeplitz. They were proved by using a result on the DFT of random vectors with Toeplitz correlation matrix, namely, ref. [4] (Theorem 1). The proof of Theorem 4 is similar to the proof of [4] (Equations (16) and (20)) but using Theorem 1 instead of [4] (Theorem 1). Observe that in Theorems 1 and 4 no assumption about the structure of E x n : 1 x n : 1 has been made. ☐
Theorem 4 shows that a coding strategy for x n : 1 is to encode y n 2 , , y n separately, where n 2 denotes the smallest integer higher than or equal to n 2 . Theorem 4 also shows that another coding strategy for x n : 1 is to encode separately the real part and the imaginary part of y k instead of encoding y k when k { n 2 , , n 1 } { n 2 } . The computational complexity of these two coding strategies based on the DFT is lower than the computational complexity of the optimal coding strategy provided in Corollary 1. Specifically, the complexity of computing the DFT ( y n : 1 = V n * x n : 1 ) is O ( n log n ) whenever the fast Fourier transform (FFT) algorithm is used, while the complexity of computing z n : 1 = U n x n : 1 is O ( n 2 ) . Moreover, when the coding strategies based on the DFT are used, we do not need to compute a real orthogonal eigenvector matrix U n of E x n : 1 x n : 1 . It should also be mentioned that for these coding strategies based on the DFT the knowledge of E x n : 1 x n : 1 is not even required, in fact, for them we only need to know E y k ^ y k ^ with k { n 2 , , n } .
The rates corresponding to the two coding strategies given in Theorem 4, R ˜ x n : 1 ( D ) and R ˘ x n : 1 ( D ) , can be written in terms of E x n : 1 x n : 1 and V n by using Lemma 1 and the following lemma.
Lemma 2.
Let y n : 1 and D be as in Theorem 4. Then
  • R y k ( D ) = 1 2 ln E y k 2 D for all k { 1 , , n } { n 2 , n } .
  • R y k ^ D 2 = 1 4 ln E Re y k 2 E Im y k 2 E Re y k Im y k 2 D 2 2 for all k { 1 , , n 1 } { n 2 } .
  • R Re y k D 2 = 1 2 ln E Re y k 2 D 2 for all k { 1 , , n 1 } { n 2 } .
  • R Im y k D 2 = 1 2 ln E Im y k 2 D 2 for all k { 1 , , n 1 } { n 2 } .
Proof. 
(1) Applying Equation (2) and [4] (Lemma 1) yields
0 < D λ n E x n : 1 x n : 1 E y k 2 = E y k 2 .
Assertion (1) now follows directly from Equation (13).
(2) Applying Theorem 2 we have
0 < D 2 λ n ( E x n : 1 x n : 1 ) 2 λ 2 E y k ^ y k ^ .
Consequently, from Equation (13) we obtain
R y k ^ D 2 = 1 4 ln det E y k ^ y k ^ D 2 2 = 1 4 ln det E Re y k 2 E Re y k Im y k E Im y k Re y k E Im y k 2 D 2 2 .
Assertions (3) and (4) Applying Equations (3) and (4) yields
0 < D 2 λ n E x n : 1 x n : 1 2 E Re y k 2 .
and
0 < D 2 λ n E x n : 1 x n : 1 2 E Im y k 2 .
Assertions (3) and (4) now follow directly from Equation (13). ☐
We end this section with a result that is a direct consequence of Lemma 2. This result shows when the rates corresponding to the two coding strategies given in Theorem 4, R ˜ x n : 1 ( D ) and R ˘ x n : 1 ( D ) , are equal.
Lemma 3.
Let x n : 1 , y n : 1 , and D be as in Theorem 4. Then the two following assertions are equivalent:
  • R ˜ x n : 1 ( D ) = R ˘ x n : 1 ( D ) .
  • E Re y k Im y k = 0 for all k { n 2 , , n 1 } { n 2 } .
Proof. 
Fix k { n 2 , , n 1 } { n 2 } . From Lemma 2 we have
2 R y k ^ D 2 = 1 2 ln E Re y k 2 E Im y k 2 E Re y k Im y k 2 D 2 2 1 2 ln E Re y k 2 E Im y k 2 D 2 2 = 1 2 ln E Re y k 2 D 2 + 1 2 ln E Im y k 2 D 2 = R Re y k D 2 + R Im y k D 2 .

4. Low-Complexity Coding Strategies for Gaussian AWSS AR Sources

We begin by introducing some notation. The symbols N , Z , and R denote the set of positive integers, integers, and (finite) real numbers, respectively. If f : R C is continuous and 2 π -periodic, we denote by T n ( f ) the n × n Toeplitz matrix given by
[ T n ( f ) ] j , k = t j k ,
where { t k } k Z is the sequence of Fourier coefficients of f, i.e.,
t k = 1 2 π 0 2 π f ( ω ) e k ω i d ω k Z .
If A n and B n are n × n matrices for all n N , we write { A n } { B n } if the sequences { A n } and { B n } are asymptotically equivalent, that is, { A n 2 } and { B n 2 } are bounded and lim n A n B n F n = 0 (see [5] (Section 2.3) and [6]).
We now review the definitions of AWSS processes and AR processes.
Definition 1.
A random process { x n } is said to be AWSS if it has constant mean (i.e., E ( x j ) = E ( x k ) for all j , k N ) and there exists a continuous 2 π -periodic function f : R C such that { E x n : 1 x n : 1 * } { T n ( f ) } . The function f is called (asymptotic) PSD of { x n } .
Definition 2.
A real zero-mean random process { x n } is said to be AR if
x n = w n k = 1 n 1 a k x n k n N ,
or equivalently,
k = 0 n 1 a k x n k = w n n N ,
where a 0 = 1 , a k R for all k N , and { w n } is a real zero-mean random process satisfying that E w j w k = δ j , k σ 2 for all j , k N with σ 2 > 0 and δ j , k being the Kronecker delta (i.e., δ j , k = 1 if j = k , and it is zero otherwise).
The AR process { x n } in Equation (17) is of finite order if there exists p N such that a k = 0 for all k > p . In this case, { x n } is called an AR ( p ) process.
The following theorem shows that if x n : 1 is a large enough data block of a Gaussian AWSS AR source, the rate does not increase whenever we encode it using the two coding strategies based on the DFT presented in Section 3, instead of encoding x n : 1 using an eigenvector matrix of its correlation matrix.
Theorem 5.
Let { x n } be as in Definition 2. Suppose that { a k } k Z , with a k = 0 for all k N , is the sequence of Fourier coefficients of a function a : R C which is continuous and 2 π -periodic. Then
  • inf n N λ n E x n : 1 x n : 1 σ 2 max ω [ 0 , 2 π ] | a ( ω ) | 2 > 0 .
  • Consider D 0 , inf n N λ n E x n : 1 x n : 1 .
    (a)
    If { x n } is Gaussian,
    1 2 ln σ 2 D = R x n : 1 ( D ) R ˜ x n : 1 ( D ) R ˘ x n : 1 ( D ) K 1 ( n , D ) K 2 ( n , D ) K 3 ( n , D ) n N ,
    where K 1 ( n , D ) is given by Equation (16), and K 2 ( n , D ) and K 3 ( n , D ) are obtained by replacing λ n E x n : 1 x n : 1 in Equation (16) by inf n N λ n E x n : 1 x n : 1 and σ 2 max ω [ 0 , 2 π ] | a ( ω ) | 2 , respectively.
    (b)
    If { x n } is Gaussian and AWSS,
    lim n R x n : 1 ( D ) = lim n R ˜ x n : 1 ( D ) = lim n R ˘ x n : 1 ( D ) = lim n K 3 ( n , D ) .
Proof. (1) Equation (17) can be rewritten as
T n ( a ) x n : 1 = w n : 1 n N .
Consequently,
T n ( a ) E x n : 1 x n : 1 T n ( a ) = E T n ( a ) x n : 1 T n ( a ) x n : 1 = E w n : 1 w n : 1 = σ 2 I n n N .
As det ( T n ( a ) ) = 1 , T n ( a ) is invertible, and therefore,
E x n : 1 x n : 1 = σ 2 T n ( a ) 1 T n ( a ) 1 = σ 2 T n ( a ) T n ( a ) 1 = σ 2 T n ( a ) * T n ( a ) 1 = σ 2 N n diag 1 k n σ k T n ( a ) 2 N n * 1 = N n diag 1 k n σ 2 σ k T n ( a ) 2 N n *
for all n N , where T n ( a ) = M n diag 1 k n σ k T n ( a ) N n * is a singular value decomposition of T n ( a ) . Thus, applying [8] (Theorem 4.3) yields
λ n E x n : 1 x n : 1 = σ 2 σ 1 T n ( a ) 2 σ 2 max ω [ 0 , 2 π ] | a ( ω ) | 2 > 0 n N .
(2a) From Equation (13) we have
R x n : 1 ( D ) = 1 2 n ln det E x n : 1 x n : 1 D n = 1 2 n ln det σ 2 T n ( a ) 1 T n ( a ) 1 D n = 1 2 n ln σ 2 n D n det T n ( a ) det T n ( a ) = 1 2 n ln σ 2 n D n = 1 2 ln σ 2 D n N .
Assertion (2a) now follows from Theorem 4 and Assertion (1).
(2b) From Assertion (2a) we only need to show that
lim n E x n : 1 x n : 1 V n diag 1 k n V n * E x n : 1 x n : 1 V n k , k V n * F n = 0 .
As the Frobenius norm is unitarily invariant we obtain
0 E x n : 1 x n : 1 V n diag 1 k n V n * E x n : 1 x n : 1 V n k , k V n * F n E x n : 1 x n : 1 T n ( f ) F n + T n ( f ) C ^ n ( f ) F n + V n diag 1 k n V n * E x n : 1 x n : 1 V n k , k V n * C ^ n ( f ) F n = E x n : 1 x n : 1 T n ( f ) F n + T n ( f ) C ^ n ( f ) F n + V n diag 1 k n V n * E x n : 1 x n : 1 T n ( f ) V n k , k V n * F n = E x n : 1 x n : 1 T n ( f ) F n + T n ( f ) C ^ n ( f ) F n + diag 1 k n V n * E x n : 1 x n : 1 T n ( f ) V n k , k F n E x n : 1 x n : 1 T n ( f ) F n + T n ( f ) C ^ n ( f ) F n + V n * E x n : 1 x n : 1 T n ( f ) V n F n = 2 E x n : 1 x n : 1 T n ( f ) F n + T n ( f ) C ^ n ( f ) F n ,
where f is (asymptotic) PSD of { x n } and C ^ n ( f ) = V n diag 1 k n [ V n * T n ( f ) V n ] k , k V n * . Assertion (2b) now follows from { E x n : 1 x n : 1 } { T n ( f ) } and [9] (Lemma 4.2). ☐
If k = 0 | a k | < , there always exists such function a and it is given by a ( ω ) = k = 0 a k e k ω i for all ω R (see, e.g., [8] (Appendix B)). In particular, if { x n } is an AR ( p ) process, a ( ω ) = k = p 0 a k e k ω i for all ω R .

5. Sufficient Conditions for AR Processes to be AWSS

In the following two results we give sufficient conditions for AR processes to be AWSS.
Theorem 6.
Let { x n } be as in Definition 2. Suppose that { a k } k Z , with a k = 0 for all k N , is the sequence of Fourier coefficients of a function a : R C which is continuous and 2 π -periodic. Then the following assertions are equivalent:
  • { x n } is AWSS.
  • { E x n : 1 x n : 1 2 } is bounded.
  • { T n ( a ) } is stable (that is, { ( T n ( a ) ) 1 2 } is bounded).
  • a ( ω ) 0 for all ω R and { x n } is AWSS with (asymptotic) PSD σ 2 | a | 2 .
Proof. 
(1)⇒(2) This is a direct consequence of the definition of AWSS process, i.e., of Definition 1.
(2)⇔(3) From Equation (20) we have
E x n : 1 x n : 1 2 = σ 2 σ n T n ( a ) 2 = σ 2 N n diag 1 k n 1 σ k T n ( a ) M n * 2 2 = σ 2 T n ( a ) 1 2 2
for all n N .
(3)⇒(4) It is well known that if f : R C is continuous and 2 π -periodic, and { T n ( f ) } is stable then f ( ω ) 0 for all ω R . Hence, a ( ω ) 0 for all ω R .
Applying [8] (Lemma 4.2.1) yields T n ( a ) = T n ( a ) * = T n ( a ¯ ) . Consequently, from [7] (Theorem 3) we obtain
T n ( a ) T n ( a ) = T n ( a ¯ ) T n ( a ) T n ( a ¯ a ) = T n | a | 2 .
Observe that the sequence
T n ( a ) T n ( a ) 1 2 = 1 σ 2 E x n : 1 x n : 1 2 = 1 σ 2 E x n : 1 x n : 1 2
is bounded. As the function | a | 2 is real, applying [8] (Theorem 4.4) we have that T n | a | 2 is Hermitian and 0 < min ω [ 0 , 2 π ] | a ( ω ) | 2 λ n ( T n | a | 2 ) for all n N , and therefore,
T n | a | 2 1 2 = 1 λ n T n | a | 2 1 min ω [ 0 , 2 π ] | a ( ω ) | 2 n N .
Thus, from [5] (Theorem 1.4) we obtain
1 σ 2 E x n : 1 x n : 1 = T n ( a ) T n ( a ) 1 T n | a | 2 1 .
Hence, applying [10] (Theorem 4.2) and [5] (Theorem 1.2) yields
1 σ 2 E x n : 1 x n : 1 T n 1 | a | 2 .
Consequently, from [8] (Lemma 3.1.3) and [8] (Lemma 4.2.3) we have
E x n : 1 x n : 1 σ 2 T n 1 | a | 2 = T n σ 2 | a | 2 .
(4)⇒(1) It is obvious.
Corollary 2.
Let { x n } be as in Definition 2 with k = 0 | a k | < . If k = 0 a k z k 0 for all | z | 1 then { x n } is AWSS.
Proof. 
It is well known that if a sequence of complex numbers { t k } k Z satisfies that k = | t k | < and that k = t k z k 0 for all | z | 1 then { T n ( f ) } is stable with f ( ω ) = k = t k e k ω i for all ω R . Therefore, { T n ( b ) } is stable with b ( ω ) = k = 0 a k e k ω i for all ω R . Thus,
( T n ( a ) ) 1 2 = ( T n ( a ) ) 1 2 = ( T n ( a ) ) 1 2 = ( T n ( b ) ) 1 2
is bounded with a ( ω ) = k = 0 a k e k ω i for all ω R . As { T n ( a ) } is stable, from Theorem 6 we conclude that { x n } is AWSS. ☐

6. Numerical Example and Conclusions

6.1. Example

Let { x n } be as in Definition 2 with a k = 0 for all k > 1 . Observe that σ 2 max ω [ 0 , 2 π ] | a ( ω ) | 2 = σ 2 ( 1 + | a 1 | ) 2 . If | a 1 | < 1 from Corollary 2 we obtain that the A R ( 1 ) process { x n } is AWSS. Figure 1 shows R x n : 1 ( D ) , R ˜ x n : 1 ( D ) , and R ˘ x n : 1 ( D ) by assuming that { x n } is Gaussian, a 1 = 1 2 , σ 2 = 1 , D = σ 2 ( 1 + | a 1 | ) 2 = 4 9 , and n 100 . Figure 1 also shows the highest upper bound of R x n : 1 ( D ) presented in Theorem 5, namely, K 3 ( n , D ) . Observe that the figure bears evidence of the equalities and inequalities given in Equations (18) and (19).

6.2. Conclusions

The computational complexity of coding finite-length data blocks of Gaussian sources can be reduced by using any of the two low-complexity coding strategies here presented instead of the optimal coding strategy. Moreover, the rate does not increase if we use those strategies instead of the optimal one whenever the Gaussian source is AWSS and AR, and the considered data block is large enough.

Author Contributions

Authors are listed in order of their degree of involvement in the work, with the most active contributors listed first. All authors have read and approved the final manuscript.

Funding

This work was supported in part by the Spanish Ministry of Economy and Competitiveness through the CARMEN project (TEC2016-75067-C4-3-R).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kolmogorov, A.N. On the Shannon theory of information transmission in the case of continuous signals. IRE Trans. Inf. Theory 1956, 2, 102–108. [Google Scholar] [CrossRef]
  2. Gray, R.M. Information rates of autoregressive processes. IEEE Trans. Inf. Theory 1970, 16, 412–421. [Google Scholar] [CrossRef]
  3. Pearl, J. On coding and filtering stationary signals by discrete Fourier transforms. IEEE Trans. Inf. Theory 1973, 19, 229–232. [Google Scholar] [CrossRef]
  4. Gutiérrez-Gutiérrez, J.; Zárraga-Rodríguez, M.; Insausti, X. Upper bounds for the rate distortion function of finite-length data blocks of Gaussian WSS sources. Entropy 2017, 19, 554. [Google Scholar] [CrossRef]
  5. Gray, R.M. Toeplitz and circulant matrices: A review. Found. Trends Commun. Inf. Theory 2006, 2, 155–239. [Google Scholar] [CrossRef]
  6. Gray, R.M. On the asymptotic eigenvalue distribution of Toeplitz matrices. IEEE Trans. Inf. Theory 1972, 18, 725–730. [Google Scholar] [CrossRef]
  7. Gutiérrez-Gutiérrez, J.; Crespo, P.M. Asymptotically equivalent sequences of matrices and multivariate ARMA processes. IEEE Trans. Inf. Theory 2011, 57, 5444–5454. [Google Scholar] [CrossRef]
  8. Gutiérrez-Gutiérrez, J.; Crespo, P.M. Block Toeplitz matrices: Asymptotic results and applications. Found. Trends Commun. Inf. Theory 2011, 8, 179–257. [Google Scholar] [CrossRef]
  9. Gutiérrez-Gutiérrez, J.; Zárraga-Rodríguez, M.; Insausti, X.; Hogstad, B.O. On the complexity reduction of coding WSS vector processes by using a sequence of block circulant matrices. Entropy 2017, 19, 95. [Google Scholar] [CrossRef]
  10. Gutiérrez-Gutiérrez, J.; Crespo, P.M. Asymptotically equivalent sequences of matrices and Hermitian block Toeplitz matrices with continuous symbols: Applications to MIMO systems. IEEE Trans. Inf. Theory 2008, 54, 5671–5680. [Google Scholar] [CrossRef]
Figure 1. Considered rates for a Gaussian AWSS AR(1) source.
Figure 1. Considered rates for a Gaussian AWSS AR(1) source.
Entropy 20 00399 g001

Share and Cite

MDPI and ACS Style

Gutiérrez-Gutiérrez, J.; Zárraga-Rodríguez, M.; Villar-Rosety, F.M.; Insausti, X. Rate-Distortion Function Upper Bounds for Gaussian Vectors and Their Applications in Coding AR Sources. Entropy 2018, 20, 399. https://0-doi-org.brum.beds.ac.uk/10.3390/e20060399

AMA Style

Gutiérrez-Gutiérrez J, Zárraga-Rodríguez M, Villar-Rosety FM, Insausti X. Rate-Distortion Function Upper Bounds for Gaussian Vectors and Their Applications in Coding AR Sources. Entropy. 2018; 20(6):399. https://0-doi-org.brum.beds.ac.uk/10.3390/e20060399

Chicago/Turabian Style

Gutiérrez-Gutiérrez, Jesús, Marta Zárraga-Rodríguez, Fernando M. Villar-Rosety, and Xabier Insausti. 2018. "Rate-Distortion Function Upper Bounds for Gaussian Vectors and Their Applications in Coding AR Sources" Entropy 20, no. 6: 399. https://0-doi-org.brum.beds.ac.uk/10.3390/e20060399

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop