256 Channel capacity and coding theorem
CCT can be stated in a number of different and equivalent ways. A possible definition,
which reflects the original one from Shannon,
13
is:
Given a noisy communication channel with capacity C, and a symbol source X with entropy
H(X ), which is used at an information rate R ≤ C, there exists a code for which message symbols
can be transmitted through the channel with an arbitrary small error ε.
The demonstration of the CCT rests upon the subtle notion of “typical sets,” as
analyzed in the previous section. It proceeds according to the following steps:
r
Assume the input source messages x to have a length (number of symbols) n.
With independent symbol outcomes, the corresponding extended-source entropy is
H(X
n
) = nH(X). The typical set of X
n
roughly contains 2
nH(X)
possible sequences,
which represent the most probable input messages.
r
Call y the output message sequences received after transmission through the noisy
channel. The set of y sequences corresponds to a random source Y
n
of entropy
H(Y
n
) = nH(Y ).
r
The channel capacity C is the maximum of the mutual information H (X; Y ) =
H(X ) − H(X |Y ). Assume that the source X corresponds (or nearly corresponds)
to this optimal condition.
Refer now to Fig. 13.5 and observe that:
r
The typical set of Y
n
roughly contains 2
nH(Y )
possible sequences, which represent the
most probable output sequences y, other outputs having a comparatively small total
probability.
r
Given an output sequence y
j
, there exist 2
nH(X|Y )
most likely and equiprobable input
sequences x (also called “reasonable causes”), other inputs having comparatively
small total probability.
r
Given an input sequence x
i
, there exist 2
nH(Y |X)
most likely and equiprobable output
sequences y (also called “reasonable effects”), other outputs having comparatively
small total probability.
Assume next that the originator is using the channel at an information (or code) rate
R < C per unit time, i.e., R payload bits are generated per second, but the rate is strictly
less than C bits per second. Thus nR payload bits are generated in the duration of each
message sequence of length n bits. To encode the payload information into message
sequences, the originator chooses to use only the sequences belonging to the typical set
of X. Therefore, 2
nR
coded message sequences (or codewords) are randomly chosen
from the set of 2
nH(X)
typical sequences, and with a uniform probability. Accordingly,
the probability that a given typical sequence x
i
will be selected for transmitting the
coded message is:
p(x
i
) =
2
nR
2
nH(X)
= 2
n[R−H (X)]
. (13.28)
13
C. E. Shannon, A mathematical theory of communication. Bell Syst. Tech. J., 27 (1948), 379–423, 623–56,
http://cm.bell-labs.com/cm/ms/what/shannonday/shannon1948.pdf.