• Tidak ada hasil yang ditemukan

An Inequality for Logarithms and its Application in Coding Theory

N/A
N/A
Protected

Academic year: 2025

Membagikan "An Inequality for Logarithms and its Application in Coding Theory"

Copied!
7
0
0

Teks penuh

(1)

An Inequality for Logarithms and its Application in Coding Theory

This is the Published version of the following publication

Dragomir, Nicoleta and Dragomir, Sever S (1998) An Inequality for Logarithms and its Application in Coding Theory. RGMIA research report collection, 1 (1).

The publisher’s official version can be found at

Note that access to this version may require subscription.

Downloaded from VU Research Repository https://vuir.vu.edu.au/17110/

(2)

AN INEQUALITY FOR LOGARITHMS AND ITS APPLICATION IN CODING THEORY

N.M. DRAGOMIR AND S.S. DRAGOMIR

Abstract. In this paper we prove a new analytic inequality for logarithms and apply it for the Noiseless Coding Theorem.

1 Introduction

The following analytic inequality for logarithms is well known in the literature (see for example [1, Lemma 1.2.2, p. 22] ):

Lemma 1.1. LetP = (p1, ..., pn)be a probability distribution, that is,0≤pi 1 and Pn

i=1pi = 1. Let Q= (q1, ..., qn)have the property that 0 ≤qi 1 and Pn

i=1qi1 (note the inequality here). Then Xn

i=1

pilogb 1

pi

Xn

i=1

pilogb 1

qi

(1.1)

where b >1, 0·logb(1/0) = 0and logb(1/0) = +∞. Furthermore, equality holds if and only if qi=pi for alli∈ {1, ..., n}.

Note that the proof of this fact uses the elementary inequality for logarithms (see [1, p. 22])

lnx≤x−1 for allx >0.

(1.2)

Also, we would like to remark that the inequality (1.1) was used to obtain many important results from the foundations of Information Theory such as:

the range of the entropy mapping, the Noiseless Coding Theorem, etc. For some recent results which provide similar inequalities see the papers [2-6].

The main aim of this paper is to point out a counterpart inequality for (1.1) and to use it in connection with theNoiseless Coding Theorem.

2 The Results

We shall start with the following inequality.

Date.November, 1998

1991 Mathematics Subject Classification. Primary 26D15; Secondary 94Xxx Key words and phrases.Analytic Inequalities, Noiseless Coding Theorem

(3)

84 Dragomir and Dragomir

Lemma 2.1. Let pi, qi be strictly positive real numbers for i = 1, ..., n. Then we have the double inequality:

1 lnr

Xn

i=1

(pi−qi) (2.1)

Xn

i=1

logr 1

qi logr 1 pi

pi 1 lnr

Xn

i=1

pi

qi 1

pi

wherer >1, r∈R.The equality holds in both inequalities iff pi=qi for alli.

Proof. The mapping f(x) = logrx is a concave mapping on (0,∞) and thus satisfies the double inequality

f0(y)(x−y)≥f(x)−f(y)≥f0(x)(x−y) for allx, y >0,and as

f0(x) = 1 ln 1

x we get

1

lnr·x−y

y logrx−logry 1

ln x−y

x for allx, y >0.

(2.2)

Let us choose x=q1

i, y= p1

i in (2.2) to get 1

lnr ·(pi−qi)

qi logr 1

qi logr 1 pi 1

ln(pi−qi) pi

(2.3)

for alli∈ {1, ..., n}.

Now, if we multiply this inequality bypi>0 (i= 1, ..., n) we get:

1 lnr

pi

pi

qi 1

≥pilogr 1

qi −pilogr 1 pi 1

ln(pi−qi) (2.4)

for alli∈ {1, ...n}.

Now, summing over ifrom 1 ton, we obtain the desired inequality (2.1).

The statement on equality holds by the strict concavity of the mapping logr(.) . We shall omit the details.

Corollary 2.2. Let P = (p1, ..., pn)be a probability distribution, that is, pi [0,1]andPn

i=1pi = 1. LetQ= (q1, ..., qn)have the property thatqi[0,1]and Pn

i=1qi1( note the inequality here). Then we have:

0 1 lnr 1

Xn

i=1

qi

! (2.5)

Xn

i=1

pilogr 1 qi

Xn

i=1

pilogr 1 pi 1

lnr Xn

i=1

p2i qi 1

!

wherer >1, r∈R. The Equality holds iffpi=qi (i= 1, ...n).

RGMIA Research Report Collection, Vol. 1, No. 1, 1998

(4)

The proof is obvious by Lemma 2.1 taking into account that Pn

i=1pi = 1 and 1Pn

i=1qi.

Remark 2.1. Note that the above corollary is a worth-while improvement of Lemma 1.2.2 from the book [1] which plays there a very important role in ob- taining the basically inequalities for entropy, conditional entropy, mutual infor- mation, etc.

Now, consider an encoding scheme (c1, ..., cn) for a probability distribution (p1, ..., pn). Recall that the average codeword length of an encoding scheme (c1, ..., cn) for (p1, ..., pn) is

AveLen(c1, ..., cn) = Xn

i=1

pilen(ci). We denote the length len(ci) byli .

Recall also that the r−ary entropy of a probability distribution (or of a source) is given by:

Hr(p1, ..., pn) = Xn

i=1

pilogr 1 pi

.

The following theorem is well known in the literature (see for example [1, Theorem 2.3.1, p. 62] ):

Theorem 2.3. LetC= (c1, ..., cn) be an instantaneous (decipherable) encoding scheme for P= (p1, ..., pn). Then we have the inequality:

Hr(p1, ..., pn)≤AveLen(c1, ..., cn), (2.6)

with equality if and only ifli= logr(p1

i)for alli= 1, ..., n.

We shall give now the following sharpening of (2.6) which has important consequences in connection with Noiseless Coding Theorem as follows.

Theorem 2.4. Let C and P be as in the above theorem. Then we have the inequality:

0 1 lnr 1

Xn

i=1

1 rli

! (2.7)

≤AveLen(c1, ..., cn)−Hr(p1, ..., pn) 1 lnr

Xn

i=1

pi pirli1 .

The Equality holds iff li= logr

1 pi

.

Proof. Defineqi:=r1li (i= 1, ...n) . Thenqi[0,1] andPn

i=1qi=Pn i=1

1 rli 1 by Kraft’s theorem (see for example [1, Theorem 2.1.2, p. 44]) and by a simple computation (as in [1, p. 62] ) we have :

Xn

i=1

pilogr 1 qi =

Xn

i=1

pilogr rli

= Xn

i=1

pili=AveLen(c1, ...cn). Also

(5)

86 Dragomir and Dragomir

1 lnr

Xn

i=1

p2i qi 1

!

= 1 lnr

Xn

i=1

p rli1 .

Thus inequality (2.5) yields (2.7).

The following theorem also holds.

Theorem 2.5. Let P = (p1, ..., pn) be a given probability distribution and r∈ N,r≥2. Ifε >0 is given and there exists natural numbersl1, ..., ln such that

logr 1

pi

≤lilogr

1 +εlnr pi

for all i∈ {1, ..., n}, (2.8)

then there exists an instantaneous r-ary code C = (c1, ..., cn) with codeword lengthlen(ci) =li such that:

Hr(p1, ..., pn) AveLen(c1, ..., cn)≤Hr(p1, ..., pn) +ε.

(2.9)

Proof. First of all, let us observe that (2.8) is equivalent to 1

pi ≤rli 1 +εlnr

pi , for alli∈ {1, ..., n}. (2.10)

Now, as r1li ≤pi,we deduce that Xn

i=1

1 rli

Xn

i=1

pi = 1

and by Kraft’s theorem, there exists an instantaneousr−ary code C= (c1, ..., cn) so thatlen(ci) =li.Obviously, by the Theorem 2.3, the first inequality in (2.9) holds.

We prove the second inequality.

By Theorem 2.4 we have the estimate

AveLen(c1, ..., cn)−Hr(p1, ..., pn) (2.11)

1 lnr

Xn

i=1

p pirli1

1 lnr

Xn

i=1

pi

pirli1 max

i=1,...,npirli1 1 lnr

Xn

i=1

pi

= 1 lnr max

i=1,...,npirli1 .

Now, we observe that (2.10) implies 1−εlnr

pi 1

pi ≤rli 1 +εlnr

pi , i∈ {1, ..., n}, i.e. ,

1−εlnr≤pirli1 +εlnr, i∈ {1, ..., n},

RGMIA Research Report Collection, Vol. 1, No. 1, 1998

(6)

which is equivalent to

pirli1≤εlnr for alli∈ {1, ..., n} and then, by (2.11),we deduce the second part of (2.9).

Remark 2.2. Since for ε∈(0,1),we have for all r >0, logr

1 +εlnr pi

logr 1

pi

= logr(1 +εlnr)<logrr= 1, (because1 +εlnr < rfor allr for a givenε∈(0,1)) we are not sure always we can find a natural numberli so that inequality(2.8) holds.

Before giving some sufficient conditions for the probability P = (p1, ..., pn) so that we can find natural numbers li satisfying the inequalities (2.8), let us recall the Noiseless Coding Theorem.

We shall use the notation

M inAveLenr(p1, ..., pn)

to denote the minimum average codeword length among allr−aryinstantaneous encoding scheme for the probability distributionP = (p1, ..., pn).

The following Noiseless Coding Theorem is well known in the literature ( see for example [1, Theorem 2.3.2, p. 64] ) :

Theorem 2.6. For any probability distribution P= (p1, ..., pn)we have Hr(p1, ..., pn) ˙≤M inAveLenr(p1, ..., pn)< Hr(p1, ..., pn) + 1.

(2.12)

The following question arises naturally:

Question: Is it possible to replace the constant 1 on (2.12) by a smaller constant ε∈ (0,1) under some conditions on the probability distribution P = (p1, ..., pn)?

We are able to give the following (partial) answer to this question.

Theorem 2.7. Let r be a given natural number andε∈(0,1). If a probability distributionP = (p1, ..., pn)satisfies the condition that every closed interval

Ii =

logr 1

pi

,logr

1 +εlnr pi

, i∈ {1, ..., n}

contains at least one natural numberli, then for that probability distributionP we have

Hr(p1, ..., pn)≤M inAveLenr(p1, ..., pn)≤Hr(p1, ..., pn) +ε.

(2.13)

Proof. Under the hypotheses Xn

i=1

1 rli

Xn

i=1

pi= 1

and by Kraft’s theorem, there exists an instantaneous code C = (c1, ..., cn) so that len(ci) = li. For that code we have the condition (2.8) and then, by Theorem 2.5, we have the inequality (2.9) . Taking the infimum in that inequality over allr−ary instantaneous codes, we get (2.13).

(7)

88 Dragomir and Dragomir

The following theorem could be useful for applications.

Theorem 2.8. Letai (i= 1, ..., n)bennatural numbers. Ifpi(i= 1, ..., n)are such that

1

rai ≤pi 1 +εlnr

rai fori= 1, ..., n;

(2.14) and Pn

i=1pi = 1, then there exists an instantaneous code C = (c1, ..., cn) with len(ci) =ai so that (2.9) holds for the probability distribution P = (p1, ..., pn).

Furthermore, for that distribution, we have the inequality (2.13). Proof. The condition (2.14) is equivalent to

1

pi ≤rai and 1 +εlnr

pi ≥rai, i= 1, ..., n;

which implies logr

1 pi

≤ailogr

1 +εlnr pi

, i= 1, ..., n;

and thenai∈Ii,i= 1, ..., n.

Applying the above results, we get the desired conclusion.

References

[1] S. ROMAN, Coding and Information Theory, Springer-Verlag, New York, Berlin, Heidelberg, 1992.

[2] S.S. DRAGOMIR and C.J. GOH, A counterpart of Jensen’s discrete inequal- ity for differentiable convex mapping and applications in information theory, Math. Comput. Modeling,24(2) (1996), 1 - 11 .

[3] S.S. DRAGOMIR and C.J. GOH, Further counterparts of some inequalities in information theory, submitted.

[4] S.S. DRAGOMIR and C.J. GOH, A counterpart of Jensen’s continuous in- equality and applications in information theory, submitted.

[5] S.S. DRAGOMIR and C.J. GOH, Some monotonicity and superadditivity properties of the entropy function,submitted.

[6] S.S. DRAGOMIR and C.J. GOH, Some bound on entropy measures in in- formation theory, Appl. Math. Lett.10(3)(1997), 23-28.

School of Communications and Informatics, Victoria University of Technology, PO Box 14428, Melbourne City MC, Victoria 8001, Australia.

E-mail address: {nico, sever}@matilda.vut.edu.au

RGMIA Research Report Collection, Vol. 1, No. 1, 1998

Referensi

Dokumen terkait

For the last conditions, sludge water, in order to obtain clear image for this particular type of water, it is important to ensure that the distance of the

By using Dreams of Trespass novel, Fatima Mernissi wants to criticize the social life in Morocco and others country that still obtain gender inequality.. There are some

to b Large m~xed species communal dung prles of large vertebrates like these white rh~no latr~nes In South Afr~ca wth seedlings of the rh~no-d~spersed Datura $p The env~ronmental

The following were the objectives of the study: To Investigate the perceptions of the rural women in Mungwi district regarding poverty and inequality; to examine the challenges faced by

Basing on Taylor’s formula with an integral remaider, an integral is estimated inLp norm of the n+ 1-th derivative of its integrand, and the Iyengar’s inequality and many other useful

Aims To determine the factors related to the incidence of RAS and examine the social class inequality in RAS among the Indonesian population using data from the fifth wave of the

Darma Persada University | 63 Gender Inequality and Feminism Seen in Characters of "Battle of the Sexes" Movie Script APPENDIX 1 SCHEME OF THE RESEARCH Gender Inequality and

General hypothesis that progestin treatment prior to the GnRH-PG estrus synchronization protocol would successfully: 1 induce ovulation in anestrous postpartum beef cows and per