Professional Documents
Culture Documents
theory
The first one who brought up the idea of measuring the amount of
information, was “Claude Shannon”
Shannon published his landmark paper
“A Mathematical Theory of Communication” in 1948
and created a turning point in the field of digital
communications and Information theory was first
established on a firm basis
Significant aspects of Information Theory
It sets bounds on what can be done but does little to aid in design of a
particular system
Despite most bodies of knowledge, Information theory gives a central role to
errors (noise) and consider it as an issue that already exists and we have to
deal with that
Presents the ideal model of a communication system with considering the
limits we’ll face in practice and recognizes the maximum efficiency of the
channel in the presence of noise
Clarifies the concept of information and particularly discriminates theses
two different concepts: amount & meaning of information
Provides mathematical tools for measuring the amount of information
Represents Information in terms of bits and established the basis of digital
communications and more generally, digitized all electronic devices
The conventional signaling system is modeled as bellow:
Noise 4
1 2 3 5 6
1.An information source 2.An encoding of this source
3.A channel (the medium through which the signal is transmitted)
4.A noise (error) source that is added to the signal in the channel
5.A decoding of the original information from the contaminated received signal
6.A sink for the information
Elaboration of the blocks of the typical signaling system:
1. Information source
Any arbitrary information source is represented as a sequence of
symbols in a source alphabet s1 , s2 , . . . , sq having q symbols.
The power of both coding and information theory is that they do not
define what information is; but try to measure the amount of it instead
Continuous forms of information (analog signals) are mostly sampled
and digitized due to 2 main reasons:
1. storage, manipulation, transmission and generally processing of
digital signals is much more reliable that analog signals
2. Integrated circuits and specially computers are all digital devices;
therefore working with digital signals would be much easier
2&5. Encoding & Decoding
Both encoding and decoding blocks implicitly should be divided into
two parts.
Encoding block consists of two main stages:
1. Source encoding
2. Channel encoding
1. Source encoding
This stage of encoding basically deals with how to encode the
mentioned source symbols more efficiently; The more complex the
structure of the encoding method is, the more efficient the channel
capacity would be utilized.
Therefore “Data Compression” is highly influenced and controlled by
the method we use to encode our information source
2. Channel encoding
Strategies and methods in channel encoding, mainly are devised in
order to compensate for the alterations and damages that may be
applied to the original signal by the inevitable noise of the channel
through which the signal is transmitted;
Channel encoding methods on the other hand, expands the length of
the sent message but don’t contain any information;
In conclusion, at the “Encode” block there will be two stages as
following:
1. Source encoding
2. Channel encoding
At the “Decode” block, of course, these two stages will be reversed as
following:
1. Channel decoding
2. Source decoding
Coding Theory entry
From the moment we start dealing with strategies and ideas of
encoding the information source, “coding theory” is being
utilized and implemented.
Information theory and Coding theory are greatly related to each
other; hence, the boundaries of these two are not very clear most
of the times. Coding theory tries to bring Information theory
into application using efficient methods of encoding.
Coding theory leads to information theory, and information
theory provides bounds on what can be done by suitable encoding
of the information.
Channel Coding
Methods of channel coding are generally
divided into two parts:
1. Error-Detecting codes
2. Error-Correcting codes
also/ Escape characters
Why Error-Detecting codes?
We require a highly reliable transmission whether
through a certain distance (transmission) or through
time (storage); Therefore a reliable transmission
definitely requires error-detection.
Note/ It is obviously not possible to detect an error if
every possible symbol, or set of symbols, that can be
received is a legitimate message.
It is possible to catch errors only if there are some
restrictions on what is a proper message
An example of Error-Detecting codes
Simple Parity Checks
At the receiving end how do you recognize each symbol of the code?
How do you recognize the end of one code word and the beginning of
the next?
Note/ The message may have other larger structures (plus its source
symbols having different probabilities) which could be used to make
the code more efficient.
Example: in English the letter Q is usually followed by the letter U
Unique Decoding
The received message must have a single, unique possible interpretation
Consider a code in which the source alphabet S has four symbols and they are to be
encoded in binary as follows:
s1 = 0
s2 = 01
s3 = 11
s4 = 00
The particular received message 0011 could be one of these two:
0011 = s4 , s3 or 0011 = s1 , s1 , s3
Thus the code is not uniquely decodable.
Condition/ Only if every distinct sequence of source symbols has a corresponding
encoded sequence that is unique, can we have a uniquely decodable signaling
system.
Instantaneous Codes
Decision Tree:
Consider the following code:
s1 = 0
s2 = 10
s3 = 110
s4 = 111
The two least frequent source symbols of an efficient code have the same
encoded lengths(why?).
Huffman encoding includes two stages:
1. Reduction process
2. Splitting process
Limitation on the efficiency of different codes
An important question/ How efficient could be the best encoding method?
Obviously the most efficient coding method is the one which compress
the information as great as possible.
Therefore the problem is to find a limitation on the amount of efficiency
and data compression.
Information
The concept of information in the field of communication and
information theory is different from the common expression of
information between people and these two separate meanings should not
be confused; since they are not equivalent.
Definition of information in mathematics
A signaling system is set up so that it could transmit a message to the
receiver where they already don’t know what the content of the message
is. Therefore, if you already know what you’re gonna receive at the
receiving end, the whole process is a waste and you won’t get any
information.(Example: p1 = 1 and all other pi = 0)
On the other hand, if the probabilities are all very different, then when a
symbol with a low probability arrives, you feel more surprised, get more
information than when a symbol with a higher probability arrives.
Thus information is somewhat inversely related to the probability of
occurrence.
I(p) is the mathematical function which will measure the amount of
information, surprise and uncertainty in the occurrence of an event of
probability p.
I(p) is represented as bellow:
With the base 2 for the logarithm, the unit of information is called “bit”.
If we use base e then the unit would be called “nat”.
For base 10 the unit would be called “hartley”.
The are three things assumed about I(p):
Which will tell us that on the average, over the whole alphabet of symbols
si , how much information we will get. (information per symbol of the
alphabet S)
How to maximize entropy?
If you have no idea which one of the source symbols are going to arrive,
then you’ll become most surprise and you’ll get the most amount of
information; hence, The entropy will be maximum. Therefore:
Whenever all source symbols are equally probable, the entropy of the
distribution would become the maximum of all other ones.
There also seem to be another factor affecting entropy.
Example:
For a die you’ll have : { , , , , , }
Where
C is the capacity of the channel (transmission rate) which is represented in “bits
per second”. It shows the amount of discrete information in bits that the medium
(channel) can hold.
W is the bandwidth of the channel in hertz.
S/N is the signal-to-noise ratio of the communication channel
Note/ SNR is expressed as a straight power ratio (not as decibels)
Some examples
Example 1:
Assume that:
SNR = 20 dB ; W = 4 kHz
Hence
=> 26.63 kbit/sec
Note/ the value of 100 is appropriate for an SNR of 20 dB
Meaning that the transmission rate could not exceeds over
26.63 bit/sec
Example 2:
Assume that :
C = 50 kbit/sec ; W = 1 MHz
Hence
=>
This example shows that it is possible to transmit the signal while the
original signal (carrying information) is much weaker than the
background noise. However the Information theory won’t tell us how? It
only convinces us that it is possible.
Conclusion
Information theory recognizes the limitations of communication
systems in practice and sets bounds on ideal transmission in the
presence of noise; However it does not tell us how to achieve them.
Coding theory generally, control the reliability of transmission
(channel encoding) and affects the efficiency and data compression of
the transmitted signal.
In fact, “Coding theory” will help us to approach the bounds
determined by “Information theory” in practice.
It should be emphasized that it is hard to specify clear and distinct
boundaries for these two theories;
Your idea ?
References
W.Hamming, R. (1986). Coding and Infomation theory. Prentice-Hall.
Forchhammer, J. J. (2010). Two-Dimentional Information and Coding
Theory. Cambridge University Press.
Tilborg, H. C. (1993). Coding Theory a first course. Eindhoven.
(n.d.). Retrieved from wikipedia: http://www.wikipedia.org
Questions & ambiguities ?!