Information coding theory Chandramohan - Enathur

38
Informa DEPARTMENT OF E SRI CHANDR VIS (Deemed to be Univ ENA A Course Material on ation Theory and Coding By S.CHANDRAMOHAN Assistant Professor ELECTRONICS AND COMMUN ENGINEERING RASEKHARENDRA SARASWATH SWA MAHAVIDYALAYA versity established under section 3 of UGC act 19 ATHUR, KANCHIPURAM – 631 561 NICATION HI 956)

Transcript of Information coding theory Chandramohan - Enathur

Page 1: Information coding theory Chandramohan - Enathur

Information T

DEPARTMENT OF ELECTRONICS AND COMMUNICATION

SRI CHANDRASEKHARENDRA SARASWATHI

VISWA MAHAVIDYALAYA

(Deemed to be University established under section 3 of UGC act 1956)

ENATHUR, KANCHIPURAM

A Course Material on

Information Theory and Coding

By

S.CHANDRAMOHAN Assistant Professor

DEPARTMENT OF ELECTRONICS AND COMMUNICATION

ENGINEERING

SRI CHANDRASEKHARENDRA SARASWATHI

VISWA MAHAVIDYALAYA

(Deemed to be University established under section 3 of UGC act 1956)

ENATHUR, KANCHIPURAM – 631 561

DEPARTMENT OF ELECTRONICS AND COMMUNICATION

SRI CHANDRASEKHARENDRA SARASWATHI

(Deemed to be University established under section 3 of UGC act 1956)

Page 2: Information coding theory Chandramohan - Enathur

INFORMATION THEORY AND CODING

Pre-requisite: Basic knowledge of Digital Communication

OBJECTIVE: To get exposed to information and entropy, compression technique, audio & video

UNIT I: INFORMATION THEORY

Information – Entropy, Information rate, classification of codes, Kraft McMillan inequality, Source coding theorem, Shannon-Fano coding, Huffman coding, Extended Huffman coding - Joint and conditional entropies, Mutual information - Discrete memory less channels – BSC, BEC – Channel capacity, Shannon limit

UNIT II: ERROR CONTROL CODING:

BLOCK CODES Definitions and Principles: Hamming weight, Hamming distance, Minimum distance decoding - Single parity codes, Hamming codes, Repetition codes - Linear block codes, Cyclic codes - Syndrome calculation, Encoder and decoder - CRC

UNIT III: ERROR CONTROL CODING: CONVOLUTIONAL CODES

Convolution codes – code tree, trellis, state diagram - Encoding – Decoding: Sequential search and Viterbi algorithm – Principle of Turbo coding

UNIT IV: SOURCE CODING:

TEXT, AUDIO AND SPEECH Text: Adaptive Huffman Coding, Arithmetic Coding, LZW algorithm – Audio: Perceptual coding, Masking techniques, Psychoacoustic model, MEG Audio layers I,II,III, Dolby AC3 - Speech: Channel Vocoder, Linear Predictive Coding

UNIT V: SOURCE CODING:

IMAGE AND VIDEO Image and Video Formats – GIF, TIFF, SIF, CIF, QCIF – Image compression: READ, JPEG – Video Compression: Principles-I, B, P frames, Motion estimation, Motion compensation, H.261, MPEG standard

Text Books: 1. Ranjan Bose, Information Theory, Coding and Cryptography, Publication,2005.

2. Cover, Thomas, and Joy Thomas. Elements of Information Theory. 2nd ed. New York, NY: Wiley-Interscience, 2006. ISBN: 9780471241959

E books and online learning materials:

1. http://www-public.tem-tsp.eu/~uro/cours-pdf/poly.pdf

2. http://www.cl.cam.ac.uk/teaching/0910/InfoTheory/InfoTheoryLectures.pdf

Page 3: Information coding theory Chandramohan - Enathur

INFORMATION THEORY:– Entropy, Information rate, classification of codes, Kraft McMillan inequality, Source coding theorem, Shannon-Fano coding, Huffman coding, Extended Huffman coding - Joint and conditional entropies, Mutual information - Discrete memory less channels – BSC, BEC – Channel capacity, Shannon limit

***************************************************************************************************

Information is the source of a communication system, whether it is analog or digital.

Information theory is a mathematical approach to the study of coding of information along with the quantification, storage, and communication of information.

Conditions of Occurrence of Events

If we consider an event, there are three conditions of occurrence.

If the event has not occurred, there is a condition of uncertainty.

If the event has just occurred, there is a condition of surprise.

If the event has occurred, a time back, there is a condition of having some information.

These three events occur at different times. The differences in these conditions help us gain knowledge on the probabilities of the occurrence of events.

Page 4: Information coding theory Chandramohan - Enathur
Page 5: Information coding theory Chandramohan - Enathur

Definition of Information:

Page 6: Information coding theory Chandramohan - Enathur
Page 7: Information coding theory Chandramohan - Enathur
Page 8: Information coding theory Chandramohan - Enathur

ENTROPY:

Page 9: Information coding theory Chandramohan - Enathur
Page 10: Information coding theory Chandramohan - Enathur
Page 11: Information coding theory Chandramohan - Enathur

Entropy:When we observe the possibilities of the occurrence of an event, how surprising or

uncertain it would be, it means that we are trying to have an idea on the average content of the information from the source of the event.

Entropy can be defined as a measure of the average information content per source symbol.

Where pi is the probability of the occurrence of character number i from a given stream of characters and b is the base of the algorithm used. Hence, this is also called as Shannon’s Entropy.

Conditional Entropy: The amount of uncertainty remaining about the channel input after observing the channel output, is called as Conditional Entropy.

It is denoted by

Page 12: Information coding theory Chandramohan - Enathur
Page 13: Information coding theory Chandramohan - Enathur

Mutual Information :

Let us consider a channel whose output is Y and input is X

Let the entropy for prior uncertainty be X = Hx Thisisassumedbeforetheinputisapplied

To know about the uncertainty of the output, after the input is applied, let us consider Conditional Entropy, given that Y = yk

This is a random variable for

With probabilities p (yo)….p(yk-1)

respectively

Page 14: Information coding theory Chandramohan - Enathur

Now, considering both the uncertainty conditions beforeandafterapplyingtheinputsbeforeandafterapplyingtheinputs,

we come to know that the difference, i.e. H(x)−H(x∣y)must represent the uncertainty about the channel

input that is resolved by observing the channel output.

This is called as the Mutual Information of the channel.

Denoting the Mutual Information as I(x;y),we can write the whole thing in an equation, as follows

Hence, this is the equational representation of Mutual Information.

Page 15: Information coding theory Chandramohan - Enathur

Discrete Memoryless Channel (DMC):

Page 16: Information coding theory Chandramohan - Enathur

Fig1: Noiseless Channel

Page 17: Information coding theory Chandramohan - Enathur

Fig2: Deterministic Channel

Page 18: Information coding theory Chandramohan - Enathur

Fig2: Noiseless Channel

Page 19: Information coding theory Chandramohan - Enathur
Page 20: Information coding theory Chandramohan - Enathur
Page 21: Information coding theory Chandramohan - Enathur
Page 22: Information coding theory Chandramohan - Enathur
Page 23: Information coding theory Chandramohan - Enathur
Page 24: Information coding theory Chandramohan - Enathur
Page 25: Information coding theory Chandramohan - Enathur
Page 26: Information coding theory Chandramohan - Enathur
Page 27: Information coding theory Chandramohan - Enathur
Page 28: Information coding theory Chandramohan - Enathur

Next Page 72 ?????

SHANNON- FANO CODING:

Page 29: Information coding theory Chandramohan - Enathur
Page 30: Information coding theory Chandramohan - Enathur
Page 31: Information coding theory Chandramohan - Enathur
Page 32: Information coding theory Chandramohan - Enathur
Page 33: Information coding theory Chandramohan - Enathur
Page 34: Information coding theory Chandramohan - Enathur
Page 35: Information coding theory Chandramohan - Enathur

MCQ Test

1. Self information should be

a) Negative b) Positive c) Positive & Negative d) None of the mentioned Ans: b

2. The unit of average mutual information is

a) Bytes per symbol b) Bytes c) Bits per symbol d) Bits Ans: d

3. In discrete memoryless source, the current letter produced by a source is

statistically independent of _____

a. Past output

b. Future output

c. Both a and b

d. None of the aboveAns: c 4. When the base of the logarithm is 2, then the unit of measure of information is

a) Bits b) Bytes c) Nats d) None of the mentioned Ans: a

5. The self information of random variable is a) 0 b) 1 c) Infinite d) Cannot be determined Ans: c

6. Entropy of a random variable is a) 0 b) 1 c) Infinite d) Cannot be determined Ans: c

7. Which is more efficient method? a) Encoding each symbol of a block b) Encoding block of symbols

Page 36: Information coding theory Chandramohan - Enathur

c) Encoding each symbol of a block & Encoding block of symbols d) None of the mentioned Ans: b

8. The mutual information between a pair of events is a) Positive b) Negative c) Zero d) All of the mentionedAns: d

9. When the base of the logarithm is e, the unit of measure of information is a) Bits b) Bytes c) Nats d) None of the mentionedAns: c

10. When probability of error during transmission is 0.5, it indicates that a) Channel is very noisy b) No information is received c) Channel is very noisy & No information is received d) None of the mentioned Ans: c

11.Types of compression

a. Lossless b. Lossy. c. both a and b

d. None of the aboveAns: C

12. What is significance of D- frames in video coding

a. They generate low resolution picture. b. highly compressed technique. c. They generate high resolution picture.

d. None of the aboveAns: A

13. MPEG coders are used for

a. compression of audio b. compression of text c. compression of audio and video d. None of the aboveAns: A

14. B-frame is also known as

Page 37: Information coding theory Chandramohan - Enathur

a. unidirectional. b. B- De-compression technique c. B- compression technique

d. bidirectional frame. Ans: D

15. I-frame is

a. It basically searches the frames. b. It basically predicts the movement of objects. c. It basically compress the movement of objects

d. None of the above.Ans: B

74. Video Coding consists of two process

a. Processing for reducing Temporal Redundancy.

b. Processing for reducing Spatial Redundancy

c. Both a and b

d. None of the above.Ans: C

16. H.261 is

a. compression of audio b. De-compression of audio c. Video Compression standard

d. None of the aboveAns: A

Page 38: Information coding theory Chandramohan - Enathur

Assignment

(1)Explain the terms (i) Self information (ii) Average information (iii) Mutual Information. (2)Discuss the reason for using logarithmic measure for measuring the amount of information. (3)Explain the concept of amount of information associated with message. (4) A binary source emitting an independent sequence of 0’s and 1’s with probabilities p and (1p) respectively. Plottheentropyofthesource. (5) Explain the concept of information, average information, information rate and redundancy as referred to information transmission. (6)Let X represents the outcome of a single roll of a fair dice. What is the entropy of X? (7)A code is composed of dots and dashes. Assume that the dash is 3 times as long as the dot and has one-third the probability of occurrence. (i) Calculate the information in dot and that in a dash; (ii) Calculate the average information in dot-dash code; and (iii) Assume that a dot lasts for 10 ms and this same time interval is allowed between symbols. Calculate the average rate of information transmission. (8)What do you understand by the term extension of a discrete memory less source? Show that the entropy of the nth extension of a DMS is n times the entropy of the original source. (9)A card is drawn from a deck of playing cards. A) You are informed that the card you draw is spade. How much information did you receive in bits? B) How much information did you receive if you are told that the card you drew is an ace? C) How much information did you receive if you are told that the card you drew is an ace of spades? Is the information content of the message “ace of spades” the sum of the information contents of the messages ”spade” and “ace”? (10) The output of an information source consists OF 128 symbols, 16 of which occurs with probability of 1/32 and remaining 112 occur with a probability of 1/224. The source emits 1000 symbols/sec. assuming that the symbols are chosen independently; find the rate of information of the source.