Chapter III: Causal rate-constrained sampling
3.4 Optimal causal rate-constrained deterministic sampling
Second, the SOI code can be implemented as the symmetric threshold sampling policy with sampling frequencyRsamples per sec followed by an SOI compressor withRs= 1bit per sample. Thus, it holds that
D(R) =D(R,1). (3.21)
From (3.20)–(3.21), we conclude (3.19).
Corollary 5.1illuminates the working principle of the optimal causal code for the stochastic processes considered in Section2.2: the optimal encoding policy trans- mits1-bit codewords, representing the signs of process innovations, as frequently as possible. In other words, the optimal causal code uses the minimum compression rate (1 bit per sample) in exchange for the maximum average sampling frequencyR (samples per sec).
which can be decomposed as (AppendixB.3) DopDET(R) = lim sup
T→∞
inf
πT∈ΠDETT
1 T
( E
" N X
i=0
Z τi+1
τi
(Xt−X¯t)2dt
#
(3.23a)
+ inf
{ft}Tt=0∈CT: (3.3a)
E
" N X
i=0
Z τi+1
τi
( ¯Xt−Xˆt)2dt
# )
, (3.23b)
whereX¯t is defined in (2.10) andXˆt is defined in (3.15). For the Wiener process, X¯t=Wτi,Xˆt= ˆWτi. The expectation on the right side of (3.23a) is the distortion due to causally estimating the source process from its samples. The expectation in (3.23b) is the distortion due to quantization.
The informational counterpart ofDDETop (R)for the Wiener process is defined below.
Definition 13 (Informational distortion-rate function (IDRF)). The IDRF for the Wiener process under deterministic sampling policies is defined as
DDET(R)≜lim sup
T→∞
inf
πT∈ΠDETT
1 T
( E
N X
i=0
Z τi+1
τi
(Wt−Wτi)2dt
(3.24a)
+ inf
NN i=1Pˆ
Wτi|W τi ,Wˆτi−1:
I(WτN→WˆτN)
T ≤R
E N
X
i=1
(τi+1−τi)(Wτi −Wˆτi)2 )
. (3.24b)
The minimization problem (3.24b) inDDET(R)is the causal IDRF for the discrete- time stochastic process formed by the samples. Note that (3.24b) is minimized over the directed information rate, which gives an information-theoretic lower bound to the rate in (3.3a). According to [77, Sec. II-C], we have
DDETop (R)≥DDET(R). (3.25)
To gain insight into the tradeoffs between the sampling frequencyF at the sampler and the rate per sampleRsat the compressor, we introduce informational distortion- frequency-rate function below.
Definition 14(Informational distortion-frequency-rate function (IDFRF)). The ID- FRF for the Wiener process under deterministic sampling policies is defined as
DDET(F, Rs)≜lim sup
T→∞
inf
πT∈ΠDETT : (2.3a)
1 T
( E
N X
i=0
Z τi+1
τi
(Wt−Wτi)2dt
(3.26a)
+ inf
NN i=1Pˆ
Wτi|W τi ,Wˆτi−1:
I(WτN→WˆτN)
N ≤Rs
E N
X
i=1
(τi+1−τi)(Wτi −Wˆτi)2 )
. (3.26b)
Similar to (3.24b), the optimization problem in (3.26b) is the causal IDRF for the Guass-Markov (GM) process formed by the samples, but the rate in (3.26b) is the rate per sampleRs rather than the rate per secondRin (3.24b).
We show the optimal deterministic sampling policy that achieves the IDRF.
Theorem 6. In causal coding of the Wiener process, the uniform sampling policy with the sampling interval equal to
τi+1−τi = 1
R, i= 0,1,2, . . . , (3.27) achieves
DDET(R) = min
f >0,Rs≥1 :f Rs≤RDDET(F, Rs) (3.28)
=DDET(R,1) (3.29)
= 5
6R. (3.30)
Proof sketch. See details in AppendixB.4. In Lemma11, we writeDDET(F, Rs)in (3.26) aslim supN→∞DN(F, Rs)and writeDN(F, Rs)as a minimization problem building on existing results on the causal IDRF (3.26b) of discrete-time GM pro- cesses. In Lemma12, we provide a lower bound onDN(F, Rs). In Lemma13, we provide an upper bound onDN(F, Rs)achieved by uniform sampling. In Lemma14, we show that the lower bound and the upper bound coincide asN → ∞and obtain
DDET(F, Rs) = 1
2F + 1
F(22Rs −1). (3.31) In Lemma 15, we prove (3.28) by showing that the minimization in (3.28) can be interchanged with the limit inDDET(F, Rs). To prove (3.29), it remains to minimize
DDET(F, Rs)in (3.28) over feasibleF andRs: min
F >0,Rs≥1 :F Rs≤RDDET(F, Rs) = min
Rs≥1DDET R
Rs, Rs
(3.32a)
=DDET(R,1) (3.32b)
= 1 2R + 1
3R = 5
6R, (3.32c)
where (3.32a) holds because DDET(F, Rs) in (3.31) decreases monotonically in F for any given Rs ≥ 1, and (3.32b) holds because DDET
R Rs, Rs
increases monotonically asRsincreases in the rangeRs ≥1. Thus, the minimum is achieved atF =R,Rs= 1. Note that 2R1 in (3.32c) comes from the sampling distortion and
1
3R comes from the causal IDRF for the discrete-time samples.
Theorem6shows that the uniform sampling policy (3.27) operates at the maximum sampling frequency R. Proposition 5.1 and Theorem 6 indicate that the working principleof the optimal encoding policy is to transmit1-bit codewords as frequently as possible.
In the setting of Theorem 6, although evaluating DDET(R) does not give us an operational compressing policy, we know that the stochastic kernel that achieves the causal IDRF for discrete-time GM processes formed by the samples under uniform sampling policies has the formN∞
i=1PWˆ
τi|Wτi−Wˆτi−1,Wˆτi−1 [23, Eq. (5.12)], suggesting that at the encoder, it is sufficient to compress the quantization innovation Wτi−Wˆτi−1only. The decoder computes the estimateWˆτiasWˆτi = ˆWτi−1+qi(Wτi− Wˆτi−1), where qi = gi ◦fi, fi
Wτi −Wˆτi−1
is thei-th binary codeword Ui, and gi(·)∈ Ris the quantization representation point of its argument. In practice, one can use thegreedy Lloyd-Max quantizer[28] that runs the Lloyd-Max algorithm for the quantization innovation in each step based on its prior pdf. Specifically, the prior pdf for the(i+ 1)-th step quantization innovationWτi+1−Wˆτican be computed by convolving the pdfs of the quantization errorWτi−Wˆτi and the process increment Wτi+1 −Wτi. The globally optimal scheme has a negligible gain over the greedy Lloyd-Max algorithm even in the finite horizon [28].
Fig.3.3displays distortion-rate tradeoffs obtained in Theorems5and6for the Wiener process, as well as a numerical simulation of the uniform sampler in Theorem 6 with the greedy Lloyd-Max quantizer. The symmetric threshold sampling policy followed by the 1-bit SOI compressor leads to a much lower MSE than uniform sampling. Indeed, according to Theorems5and6, DDETD(R)(R) = 5, andDopDET(R)for
0 1 2 3 4 5 6 7 8 9 10 0
0.5 1 1.5 2 2.5
Figure 3.3: MSE versus rate.
the uniform sampling is even higher thanDDET(R)by (3.25). Note that the greedy Lloyd-Max curve is rather close to the DDET(R)curve, indicating that the IDRF is a meaningful gauge of what is attainable in zero-delay continuous-time causal compression.