His ground-breaking approach introduced a simple abstraction of human
communication, called the channel. Shannon's communication channel consisted
of a sender (a source of information), a transmission medium (with noise
and distortion), and a receiver (whose goal is to reconstruct the sender's
messages).
In order to quantitatively analyze transmission through the channel
he also introduced a measure of the amount of information in a message.
To Shannon the amount of information is a measure of surprise and is closely
related to the chance of one of several messages being transmitted. For
Shannon a message is very informative if the chance of its occurrence is
small. If, in contrast, a message is very predictable, then it has a small
amount of information---one is not surprised to receive it.
To complete his quantiative analysis of the communication channel,
Shannon introduced the entropy rate, a quantity that measured a source's
information production rate and also a measure of the information carrying
capacity, called the communication channel capacity.
He showed that if the entropy rate, the amount of information you wish to transmit, excceds the channel capacity, then there were unavoidable and uncorrectable errors in the transmission. This is intuitive enough. What was truly surprising, though, is that he also showed that if the sender's entropy rate is below the channel capacity, then there is a way to encode the information so that it can be received without errors. This is true even if the channel distorts the message during transmission.
Shannon adapted his theory to analyze ordinary human (written) language.
He showed that it is quite redundant, using more symbols and words than
necessary to convey messages. Presumably, this redundancy is used by us
to improve our ability to recognize messages reliably and to communicate
different types of information.
In the study of complex systems today, we view deterministic chaotic
processes as information sources and use Shannon's entropy rate, as adapted
by Kolmogorov
and his student Y. Sinai in the late 1950s, to measure how
random a chaotic system is.
0 komentar:
Posting Komentar