Chapter 1 Introduction to Digital and Analog Communication Systems (Sections 1 – 9)
-
Upload
adele-norton -
Category
Documents
-
view
218 -
download
4
Transcript of Chapter 1 Introduction to Digital and Analog Communication Systems (Sections 1 – 9)
Chapter 1
Introduction to Digital and Analog Communication Systems
(Sections 1 – 9)
Communication Model
Source Transmitter Channel Receiver Destination
SignalProcessing
Communication(Transmitting)
SignalProcessing
Communication(Receiving)
ComputerTelephoneHumanCameraServer
ModemLaserMicrowave TransmitterMicrophone
WireAirOptical fiberWater
ModemPhotodetectorMicrowave DishMicrophoneRadio
ComputerHumanSpeakerVideo screenServer
Definition: A communication system is a system which transmits information from a source to a destination.
Goals of ESE 471
• Understand how communication systems work.
• Develop mathematical models for methods and components of communication systems.
• Analyze performance of communication systems and methods.
• Understand practical systems in use.
• A digital information source produces a finite set of possible messages
• An analog information source produces messages that are defined on a continuum.
• A digital communication system transfers information from a digital source to the intended receiver.
• An analog communication system transfers information from an analog source to the intended receiver.
• A signal is a measurable quantity (e.g., voltage) which bears information.
• Noise is a measurable quantity which carries undesired interference.
Definitions
Why Digital?
• Less expensive circuits
• Privacy and security
• Small signals (less power)
• Converged multimedia
• Error correction and reduction
Why Not Digital?
• More bandwidth
• Synchronization in electrical circuits
• Approximated information
Couch, Digital and Analog Communication Systems, Seventh Edition ©2007 Pearson Education, Inc. All rights reserved. 0-13-142492-0
Realistic Communication Model
Noise and errors make communication inaccurate.The goal is to achieve , or make them as close as possible.m(t) is the original signal at baseband.s(t) is called the transmitted signal at carrier frequency.
)( )( ~ tmtm
Wireless Frequency Allocations in the U.S.A.
Modes of Wireless Wave Propagation
Information
• Electronic communication systems are inherently analog.
• The process of communication is inherently discrete.
• Limitations– Noise or uncertainty– Changes in system characteristics
Measure of Information
• The amount of information sent from a digital source when the jth message is transmitted is given by:
where Pj is the probability of transmitting the jth message. (0 Pj 1, Ij 0)
• Average information measure of a digital source is given by:
where m is the number of possible different source messages. H is called the entropy. H is maximum when all messages are equally likely.
bits 1
log2
jj P
I
bits 1
log 211
j
m
jj
m
jjj P
PIPH
• The redundancy of a source is H / Hmax.e.g., parity bit adds redundancy to source with 8 bit output.
• Deterministic signal processing never creates information.• If the source generates independent messages once every T
seconds, then the source rate R is:
R can be viewed as the average information per second.second / bits T
HR
Example 1-1
Digital words consisting of 12 digits and each digits can be one of 4 equally likely possibilities.
Each word, j, has probability of occurrence, Pj:
Then:
12
12 4
1
4
1
jP
bits 2424 4
1
1
12
m
j
H
bits 24bits
41
1log 122
jI
Another Example: 2 Dice
Suppose we are throwing 2 dice simultaneously and sum the two numbers. The outcome, x, can be 2 x 12.
The possible scenarios for these outcomes are:x = 2: (1, 1) Pr (x =2) = 1/36x = 3: (1, 2), (2, 1) Pr (x =3) = 2/36x = 4: (1, 3), (2, 2), (3, 1) Pr (x =4) = 3/36x = 5: (1, 4), (2, 3), (3, 2), (4, 1) Pr (x =5) = 4/36x = 6: (1, 5), (2, 4), (3, 3), (4, 2), (5, 1) Pr (x =6) = 5/36x = 7: (1, 6), (2, 5), (3, 4), (4, 3), (5, 2), (6, 1) Pr (x =7) = 6/36x = 8: (2, 6), (3, 5), (4, 4), (5, 3), (6, 2) Pr (x =8) = 5/36x = 9: (3, 6), (4, 5), (5, 4), (6, 3) Pr (x =9) = 4/36x = 10: (4, 6), (5, 5), (6, 4) Pr (x =10) = 3/36x = 11: (5, 6), (6, 5) Pr (x =11) = 2/36x = 12: (6, 6) Pr (x =12) = 1/36
bits 27.2
3.58 36
12.89
36
22.48
36
32.20
36
4
1.97 36
51.79
36
61.97
36
52.20
36
42.48
36
32.89
36
2 3.58
36
1
bits 58.3 36log
3611
log bits 89.2 18log
3621
log
bits 48.2 12log
3631
log bits 20.2 9log
3641
log bits 97.1 2.7log
3651
log
bits 79.1 6log
3661
log bits 97.1 2.7log
3651
log bits 20.2 9log
3641
log
bits 48.2 12log
3631
log bits 89.2 18log
3621
log bits 3.58 36log
3611
log
22122211
2210229228
227226225
224223222
H
H
II
III
III
III
Dice Example Continued
• H = 2.27 bits implies that in order to communicate the outcome of throwing two dice, it requires a minimum of 2.27 bits. How can we achieve this?
• For example, when we get a certain x, send a binary sequence corresponding to it as shown below.x = 2: 11010 5 bits with Pr (x =2) = 1/36 x = 3: 11011 5 bits with Pr (x =3) = 2/36x = 4: 1100 4 bits with Pr (x =4) = 3/36x = 5: 000 3 bits with Pr (x =5) = 4/36x = 6: 001 3 bits with Pr (x =6) = 5/36x = 7: 10 2 bits with Pr (x =7) = 6/36x = 8: 010 3 bits with Pr (x =8) = 5/36x = 9: 011 3 bits with Pr (x =9) = 4/36x = 10: 1110 4 bits with Pr (x =10) = 3/36x = 11: 11110 5 bits with Pr (x =11) = 2/36x = 12: 11111 5 bits with Pr (x =12) = 1/36
Average bits needed per message = 3.33 bits. Can you think of any other scheme?
Channel Capacity
In 1948, Claude Shannon developed a mathematical model for channel such that, insofar as the model is realistic, there exists an upper limit on the rate at which information can be transmitted from source to user error-free. This upper limit is called the channel capacity, C. It is a function of bandwidth (B) and signal-to-noise ration (S/N or SNR).
N
SBC 1log2
Coding
Coding Theory is the study of preprocessing the source information such that the channel capacity can be achieved (or get as close to it as possible).
coding