Cryptography: Theory and Practice:Shannon s Theory

cryptography: theory and practice Cryptography: Theory and Practice
by Douglas Stinson
CRC Press, CRC Press LLC
ISBN: 0849385210   Pub Date: 03/17/95
  

Previous Table of Contents Next


2.3 Properties of Entropy

In this section, we prove some fundamental results concerning entropy. First, we state a fundamental result, known as Jensen’s Inequality, that will be very useful to us. Jensen’s Inequality involves concave functions, which we now define.

DEFINITION 2.4  A real-valued function f is concave on an interval I if

for all x, y ∈ I. f is strictly concave on an interval I if

for all x, y ∈ I, x ≠ y.

Here is Jensen’s Inequality, which we state without proof.

THEOREM 2.5  (Jensen’s Inequality)

Suppose f is a continuous strictly concave function on the interval I,

and ai > 0, 1 ≤ in. Then

where xi ∈ I, 1 ≤ in. Further, equality occurs if and only if x1 = ... = xn.

We now proceed to derive several results on entropy. In the next theorem, we make use of the fact that the function log2 x is strictly concave on the interval (0, ∞). (In fact, this follows easily from elementary calculus since the second deriviative of the logarithm function is negative on the interval (0, ∞).)

THEOREM 2.6

Suppose X is a random variable having probability distribution p1, p2, pn, where pi > 0, 1 ≤ in. Then H(X) ≤ log2 n, with equality if and only if pi = 1/n, 1 ≤ in.

PROOF  Applying Jensen’s Inequality, we have the following:

Further, equality occurs if and only if pi = 1/n, 1 ≤ in.

THEOREM 2.7

H (X, Y) ≤ H(X) + H(Y), with equality if and only if X and Y are independent events.

PROOF  Suppose X takes on values xi, 1 ≤ im, and Y takes on values yj, 1 ≤ jn. Denote pi = p(X = xi), 1 ≤ im, and qj = p(Y = yj), 1 ≤ jn. Denote rij = p(X = xi, Y = yj), 1 ≤ im, 1 ≤ jn (this is the joint probability distribution).

Observe that

(1 ≤ im) and

(1 ≤ jn). We compute as follows:

On the other hand,

Combining, we obtain the following:

(Here, we apply Jensen’s Inequality, using the fact that the rij’s form a probability distribution.)

We can also say when equality occurs: it must be the case that there is a constant c such that piqj/rij = c for all i, j. Using the fact that

it follows that c = 1. Hence, equality occurs if and only if rij = piqj, i.e., if and only if

1 ≤ im, 1 ≤ jn. But this says that X and Y are independent.

We next define the idea of conditional entropy.

DEFINITION 2.5  Suppose X and Y are two random variables. Then for any fixed value y of Y, we get a (conditional) probability distribution p(X|y). Clearly,

We define the conditional entropy H(X|Y) to be the weighted average (with respect to the probabilities p(y)) of the entropies H(X|y) over all possible values y. It is computed to be

The conditional entropy measures the average amount of information about X that is revealed by Y.

The next two results are straightforward; we leave the proofs as exercises.

THEOREM 2.8

H(X, Y) = H(Y) + H(X|Y).

COROLLARY 2.9

H(X|Y) ≤ H(X), with equality if and only if X and Y are independent.


Previous Table of Contents Next

Copyright © CRC Press LLC



Cryptography. Theory and Practice
Modern Cryptography: Theory and Practice
ISBN: 0130669431
EAN: 2147483647
Year: 1995
Pages: 133
Authors: Wenbo Mao

flylib.com © 2008-2017.
If you may any questions please contact us: flylib@qtcs.net