0% found this document useful (0 votes)
248 views82 pages

AWGN Channel Signal Transmission

The document discusses the geometric representation of signals transmitted over an additive white Gaussian noise (AWGN) channel. It describes how any signal can be represented as a linear combination of orthogonal basis functions, represented by a vector in N-dimensional space. The length and angle between these signal vectors relates to the energy and inner product of the original signals. Representing signals geometrically allows analysis of their performance as modulation signals in a simplified way. It also describes how the continuous AWGN channel can be converted into a vector channel by correlating the received signal with the basis functions.

Uploaded by

anu
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
248 views82 pages

AWGN Channel Signal Transmission

The document discusses the geometric representation of signals transmitted over an additive white Gaussian noise (AWGN) channel. It describes how any signal can be represented as a linear combination of orthogonal basis functions, represented by a vector in N-dimensional space. The length and angle between these signal vectors relates to the energy and inner product of the original signals. Representing signals geometrically allows analysis of their performance as modulation signals in a simplified way. It also describes how the continuous AWGN channel can be converted into a vector channel by correlating the received signal with the basis functions.

Uploaded by

anu
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 82

ECT305 Analog and Digital Communication

MODULE-4 , PART-1

Dr. Susan Dominic


As s i stant Profess or
De p a rtm ent O f E l e ct ronic s a n d C o m m u n i cat ion E n g i n eeri ng
R a ja giri S c h o ol Of E n gi neering A n d Te c hno logy

RAJAGIRI SCHOOL OF ENGINEERING AND TECHNOLOGY, KOCHI


In this module we will learn about some basic issues pertaining to the transmission of
signals over an AWGN channel

Consider a most basic form of a digital communication system as above


• A message source emits one symbol every T seconds with the symbols belonging to
an alphabet of M symbols denoted by 𝑚1, 𝑚2 , … . . , 𝑚𝑀
• In general each symbol is encoded into a distinct signal 𝑠𝑖 (𝑡) which is suitable for
transmission over the channel
• The signal 𝑠𝑖 (𝑡) occupies the full duration T allotted for symbol 𝑚𝑖
• 𝑠𝑖 (𝑡) is a real-value energy signal (i.e., with finite energy), as shown by

• The channel is represented by an AWGN model with the received signal as


• The receiver observes the received signal for a duration of T seconds and then makes an
estimate of the transmitted signal 𝑠𝑖 (𝑡) or equivalently the 𝑖𝑡ℎ symbol 𝑚𝑖
• Due to the presence of noise the receiver will occasionally make erroneous decisions
• The requirement is to design the receiver so as to minimize the average probability of
symbol error
which is defined as Conditional probability of making
𝑀 an error( conditional error
𝑃𝑒 = ෍ 𝑝𝑖 𝑃(𝑚
ෝ ≠ 𝑚𝑖 |𝑚𝑖 ) probability) given that 𝑖𝑡ℎ symbol
𝑖=1 was sent

Prior probability
of transmitting 𝑚𝑖
• The aim is to make the system as reliable as possible by minimizing the average
probability of symbol error between the transmitted symbol by the source and the
receiver output
• For this, we have to tackle two basic issues
1. How to optimize the design of the receiver to minimize the average
probability of error
2. How to chose the set of signals 𝑠1 (𝑡), 𝑠2 (𝑡)…., 𝑠𝑀 (𝑡) for representing the
symbols 𝑚1, 𝑚2 , … . . , 𝑚𝑀 resp. since this affects the average probability of
symbol error
Both require an understanding of the geometric representation of signals
Geometric Representation of Signals

• The essence of geometric representation of signals is to represent any set of M


energy signals {𝑠𝑖 (𝑡) } as linear combinations of N orthonormal basis functions,
where N ≤ M,
• i.e given a set of real-valued energy signals 𝑠1 (𝑡), 𝑠2 (𝑡) ,…, 𝑠𝑀 (𝑡), each of duration T
seconds ,

1
where the coefficients 𝑠𝑖𝑗 are defined as

• The real-valued functions form an orthonormal set , i.e.

where is Kronecker delta


First condition of 3 ensures that each basis function is normalized to have unit energy
Second condition ensures that the basis functions are orthogonal wrt to each other within
the interval 0 ≤ t ≤ T
• For a given i, the coefficients can be viewed as an N-dimensional vector
denoted by
• The vector bears a one-to-one relationship with the transmitted signal 𝑠𝑖 (𝑡), i.e.

1. Given the N elements of the vector (i.e., 𝑠𝑖1 , 𝑠𝑖2 , … , 𝑠𝑖𝑁 ) operating as input, we can
generate the signal 𝑠𝑖 (𝑡) from 1

We can use a bank of N multipliers with each multiplier having its own basis function
followed by a summer can be viewed as a synthesizer
2. Conversely, given the signals operating as inputs, the
coefficients can be calculated using 2

The scheme consists of a bank of N product integrators or correlators with a common


input and each one of them supplied with its own basis function can be viewed
as an analyzer
We may state that each signal 𝑠𝑖 (𝑡) can be completely determined by the vector
of its coefficients

The vector is called a signal vector


We may visualize the set of signal vectors as corresponding set
of M points in an N-dimensional Euclidean space, with N mutually perpendicular axes
labeled . This N-dimensional space is called as the signal space
Example: Consider the case of a two dimensional signal space ( N=2) with three signals (M=3)
• In the N-dimensional space, we may define lengths of vectors and angles between
vectors
• The length ( called absolute value or norm) of a signal vector is represented by the
symbol
• The squared length of any signal vector is defined as the inner product or dot
product of with itself , i.e.

𝑠𝑖𝑗 is the jth element


of and ‘T’ stands
for transpose
• In the N-dimensional space, we may define lengths of vectors and angles between
vectors
• The length ( called absolute value or norm) of a signal vector is represented by the
symbol
• The squared length of any signal vector is defined as the inner product or dot
product of with itself , i.e.

𝑠𝑖𝑗 is the jth element


of and ‘T’ stands
for transpose
Relationship between the energy of a signal and its representation as a vector

Energy of a signal 𝑠𝑖 (𝑡) of duration T seconds,

Substituting

We have
Interchanging the order of summation and integration and then rearranging the terms,
we have

Since { } form an orthonormal set,

Hence, the energy of a signal 𝑠𝑖 (𝑡) is equal to the squared length of the signal vector
representing it
Similarly, for a pair of signals 𝑠𝑖 (𝑡) and 𝑠𝑘 (𝑡) represented by vectors 𝐬𝑖 and 𝐬𝑘
respectively, we may show that

Shows that inner product of energy signals 𝑠𝑖 (𝑡) and 𝑠𝑘 (𝑡) over an interval [0, T ] using their
time-domain representations is equal to the inner product of their respective vector
representations 𝐬𝑖 and 𝐬𝑘
Another useful relation is

represents the Euclidean distance between the points represented by the


signal vectors 𝐬𝑖 and 𝐬𝑘
The angle subtended between two signal vectors 𝐬𝑖 and 𝐬𝑘 is given by

Hence, two vectors 𝐬𝑖 and 𝐬𝑘 are orthogonal if their inner product is equal to zero ,
i.e. equal to zero in which case, = 90 degrees
Therefore, geometric representation of signals can provide a compact characterization of
signals and can simplify analysis of their performance as modulation signals
How to choose the orthonormal basis functions ?

For a given set of M energy signals, , the orthonormal basis


functions can be extracted from the signals by following Gram-Schmidt orthogonalization
procedure
1
2

1 2

3
2

3
Geometric Representation of Signals
Any signal 𝑠𝑖 𝑡 of finite energy can be represented as a linear combination of N orthonormal
basis functions

Synthesizer Analyzer

44
Conversion of continuous AWGN channel into a vector channel
• The received signal in accordance with idealized AWGN channel is given by
0≤t≤T
x t = si t + w t ቊ
i = 1, 2, . . M
where 𝑤(𝑡) is a sample function of white gaussian noise process 𝑊(𝑡) of zero mean and
power spectral density 𝑁0 /2

• On giving 𝑥(𝑡) as input to a bank of 𝑁 product integrators or correlators, the output of


𝑗th correlator is given by
T

xj = න x t ϕj t dt In terms of RVs
0 𝑿𝒋 = 𝒔𝒊𝒋 + 𝑾𝒋
= 𝒔𝒊𝒋 + 𝒘𝒋 for 𝑗 = 1, 2, . . 𝑁
where 𝑠𝑖𝑗 is a deterministic quantity and 𝑤𝑗 is the sample value of a random variable 𝑊𝑗 due
to channel noise w(t)
45
• Consider next a random process 𝑋 ′ (𝑡) whose sample function 𝑥 ′ (𝑡) is related to received signal
𝑥(𝑡) as

𝑥 ′ 𝑡 = 𝑠𝑖 𝑡 + 𝑤 𝑡 − ෍(𝑠𝑖𝑗 +𝑤𝑗 )𝜙𝑗 𝑡


𝑗=1
𝑁 𝑁

= 𝑠𝑖 𝑡 + 𝑤 𝑡 − ෍ 𝑠𝑖𝑗 𝜙𝑗 𝑡 − ෍ 𝑤𝑗 𝜙𝑗 𝑡
𝑗=1 𝑗=1
𝑁

= 𝑤 𝑡 − ෍ 𝑤𝑗 𝜙𝑗 𝑡 = 𝑤′(𝑡)
𝑗=1
• The sample function 𝑥 ′ (𝑡) depends
solely on channel noise w(t)
• Thus the received signal is expressed as On comparing with expression of transmitted
signal, 𝑁

𝑠𝑖 (𝑡) = ෍ 𝑠𝑖𝑗 𝜙𝑗 (𝑡) Deterministic


𝑗=1

Stochastic / Random
46
Statistical Characterization of Correlator Output

• Let 𝑋𝑗 denote random variable (sample value is represented by correlator output 𝑥𝑗 ,


𝑗 = 1, 2 … 𝑁) in response to received signal 𝑥 𝑡
• To find mean and variance of observation vector, X 𝑥1

• The mean of 𝑋𝑗 is given by

𝝁𝑿𝒋 = E Xj x(t)
𝑥2

= E s𝑖j + 𝑊𝑗
= s𝑖j + E 𝑊𝑗
= s𝑖j [𝑾𝒋 has zero mean (AWGN)]
𝑥𝑁

𝝁𝑿𝒋 = 𝒔𝒊𝒋

47
• The variance of 𝑋𝑗 is given by
𝑤1
𝟐
𝑬 𝑿𝒋 − 𝝁𝑿𝒋

w(t)
𝑤2

• The random variable 𝑊𝑗 is defined as

Therefore 𝑤𝑁

48
• Since noise process W(𝑡) is stationary, 𝑹𝑾 𝒕, 𝒖 = 𝑹𝑾 (𝒕 − 𝒖)

Autocorrelation of noise process =

49
Covariance of observation vector:

• Since 𝜙𝑗 (𝑡) form an orthogonal set, 𝑋𝑗 are mutually uncorrelated

50
Hence 𝑋𝑗 and 𝑋𝑘 are uncorrelated

• The vector 𝑋 is defined as vector of 𝑁 random variables. Each element is an independent


Gaussian R.V. with mean 𝑠𝑖𝑗 and variance 𝑁0 /2

51
𝑁𝑜
Since each 𝑋𝑗 is a Gaussian RV with mean 𝑠𝑖𝑗 and variance , we can write the conditional PDF
2

as

𝟏 𝒙 − 𝝁𝑿 𝟐
𝒇𝑿 𝒙 = 𝐞𝐱𝐩
𝟐𝝅𝝈𝟐 𝟐𝝈𝟐

Since the elements of X are independent , the conditional probability density function of
vector X, given that signal 𝑠𝑖 (𝑡) or correspondingly the symbol 𝑚𝑖 was transmitted, can be
expressed as the product of the conditional PDFs of its individual elements as

52
Hence we have

Here vector x and 𝑥𝑗 are sample values of random vector X and random variable 𝑋𝑗 ,
resp.
• Vector x is called observation vector
• 𝑥𝑖 is called observable element
• A channel satisfying the above equation is called a memoryless channel

53
Now we have fully characterised the first part of the above equation. Now we have to
account for 𝑤 ′ (𝑡).
We can prove that 𝑤 ′ 𝑡 is orthogonal to σ𝑁
𝑗=1 𝑥𝑗 𝜙𝑗 (𝑡)

Hence we can conclude that


Insofar as signal detection in additive white Gaussian noise is concerned, only the
projections of the noise onto the basis functions of the signal set affects the
sufficient statistics of the detection problem; the remainder of the noise is
irrelevant
54
Putting the theorem into a mathematical context we can say that the AWGN channel
below

is equivalent to

55
Maximum Likelihood Decoding
• At the receiver, given the observation vector 𝐱, we need to estimate the message symbol
responsible for generating 𝐱
• The likelihood function is defined as
𝐿 𝑚𝑖 = 𝑓𝐗 (𝐱/𝑚𝑖 ) where 𝑖 = 1, 2, . . 𝑀
where 𝑓𝐗 (𝐱/𝑚𝑖 ) is the conditional probability density function (pdf) of the vector X given that
symbol 𝑚𝑖 was transmitted and is given by

• Each element 𝑋𝑗 is a gaussian R.V with mean 𝑠𝑖𝑗 and variance 𝑁0 /2

56
• In practice it is more convenient to use the log likelihood function for the AWGN channel is
defined as

𝑙 𝑚𝑖 = l𝑛(𝐿 𝑚𝑖 ) = l𝑛[𝑓𝑋 (x/𝑚𝑖 )]

𝑁 2
− 1
= 𝑙𝑛 𝜋𝑁0 2 𝑒𝑥𝑝 − σ𝑁
𝑗=1 𝑥𝑗 − 𝑠𝑖𝑗
𝑁0

𝑁 1 2
=− 𝑙𝑜𝑔 𝜋𝑁0 − σ𝑁
𝑗=1 𝑥𝑗 − 𝑠𝑖𝑗 ,
2 𝑁0

1 2
=− σ𝑁
𝑗=1 𝑥𝑗 − 𝑠𝑖𝑗 , 𝑖 = 1, 2, . . , 𝑀
𝑁0

𝑁
1 2
𝑙 𝑚𝑖 = − ෍ 𝑥𝑗 − 𝑠𝑖𝑗 , 𝑖 = 1, 2, . . , 𝑀
𝑁0
𝑗=1
57
Maximum Likelihood Detection
• Suppose one of M possible signals 𝑠1 𝑡 , 𝑠2 𝑡 … 𝑠𝑀 𝑡
transmitted over a duration of T seconds with equal
1
probability
𝑀

• The signal 𝑠𝑖 𝑡 (or correspondingly the signal vector 𝐬𝑖 ) can


be represented by a point in Euclidean space of dimension
N ≤ 𝑀.
• This point is known as transmitted signal / message point.
• The set of message points corresponding to the set of
transmitted signals {𝑠𝑖 (𝑡)} is called signal constellation
• The correlator outputs determined by received signal x(t)
define observation vector 𝐱
• The vector 𝐱 differs from signal vector 𝐬𝑖 by noise vector 𝐰

58
• The signal detection problem can be stated as follows:

Given the observation vector x, the mapping from x to an estimate m


ෝ of transmitted symbol 𝑚𝑖
must be done to minimize the probability of error in decision making process
• Suppose given an observation vector x, a decision is made, 𝑚
ෝ = 𝑚𝑖 . The probability of error in this
decision is defined as

• The optimum decision rule to minimize the probability of error is stated as

where k=1,2..M. This decision rule is known as Maximum A Posteriori Probability Rule (MAP)
59
The decoder which implements the MAP decision rule is called as a MAP decoder.

• Using Bayes’ probability rule,


𝑃 𝐴 𝑃(𝐵Τ𝐴)
𝑃(𝐴Τ𝐵) =
𝑃(𝐵)
• The MAP decision rule can be now rewritten as

where 𝑝𝑘 is the prior probability of message symbol 𝑚𝑘 .

60
• The denominator is independent of the transmitted symbol
• Since all message symbols are transmitted with equal probabilities, the prior
probabilities are equal
• The conditional probability density function bears a one-to-one relationship
with the log likelihood function
The decision rule can be hence restated as

Set m
ෝ = mi if l mk is maximum for k = i

This decision rule is called as maximum likelihood (ML) rule and the decoder that
implements this is called as maximum likelihood decoder (MAP rule becomes ML rule
when messgaes are assumed to be equiprobable)
61
• maximum likelihood decoder computes the log-likelihood functions for all M
message symbols, compares them and decides in favour of the maximum
• Thus the maximum likelihood decoder is simplified version of the MAP decoder where
the messages are assumed to be equally likely
1 2
• For AWGN channel, since 𝑙 𝑚𝑖 = − σ𝑁
𝑗=1 𝑥𝑗 − 𝑠𝑖𝑗 , 𝑖 = 1, 2, . . , 𝑀, maximum
𝑁0

2
value for log-likelihood function would mean, minimum value for σ𝑁
𝑗=1 𝑥𝑗 − 𝑠𝑖𝑗 ,
hence the decision rule can be restated as
2
Set m
ෝ = mi if σN
j=1 xj − skj is minimum for k = i

62
• Hence maximum likelihood decision rule is nothing but choosing the the message
point in an N-dimensional signal space that is closest to the received signal point
2
(observation vector) since σN
j=1 x j − skj determines the distance between the
observation vector x and the signal vector 𝒔𝒌

• Hence decision rule may be restated as


Set m
ෝ = mi if ||x-𝒔𝑘 || is minimum for k=i

63
Graphical Interpretation of Maximum Likelihood Rule

• Since ML rule is simply choosing the message point closest to the received signal
point, all points that are closest to a message point can be thought of as a region
around the message point called as the decision region
• Since there are M number of message points, we can picture the N-dimensional
space to be divided into M such decision regions surrounding the M message
points
• Hence by simply observing which decision region a received signal point falls in,
the estimate of the message point can be easily obtained

64
65
Hence we can restate the decision rule as
𝐎𝐛𝐬𝐞𝐫𝐯𝐚𝐭𝐢𝐨𝐧 𝐯𝐞𝐜𝐭𝐨𝐫 𝐱 𝐥𝐢𝐞𝐬 𝐢𝐧 𝐫𝐞𝐠𝐢𝐨𝐧 𝐙𝐢 𝐢𝐟
𝐭𝐡𝐞 𝐄𝐮𝐜𝐥𝐢𝐝𝐞𝐚𝐧 𝐝𝐢𝐬𝐭𝐚𝐧𝐜𝐞 𝐱 − 𝐬𝐤 𝐢𝐬 𝐦𝐢𝐧𝐢𝐦𝐮𝐦 𝐟𝐨𝐫 𝐤 = 𝐢

• In practice, the decision rule can be further simplified by expanding the


summation as
N N N N
2 2
෍ xj − skj = ෍ xj2 − ෍ 2xj skj + ෍ skj
j=1 j=1 j=1 j=1
• The first summation term is independent of k and can be ignored
• The second term is the inner product of observation vector and signal
vector
• The third term is the energy of the transmitted signal
2 1
σ𝑁
𝑗=1 𝑥𝑗 − 𝑠𝑘𝑗 ≅ −(σ𝑁
𝑗=1 𝑥𝑗 𝑠𝑘𝑗 − 𝐸𝑘 )
2

66
• The maximum likelihood decision rule for AWGN channel which was stated as

Observation vector 𝐱 lies in region Zi if


N
2
෍ xj − skj is minimum for k = i
j=1

can be restated as:


Observation vector 𝐱 lies in region Zi if
1
σN
j=1 x j skj − Ek is maximum for k = 𝑖
2

Next we look at the optimum receivers for an AWGN channel

67
Correlation Receiver

For an AWGN channel, for the case when the transmitted signals 𝑠1 (𝑡), 𝑠2 (𝑡), … , 𝑠𝑀 (𝑡) are
equally likely, the optimum receiver consists of two subsystems namely:

The detector/ Demodulator: It consists of a bank of product integrators/correlators each


supplied with orthonormal basis functions operating on the received signal x(t) to produce the
observation vector x

The signal transmission decoder: that operates on the observation vector x to produce an
estimate of transmitted symbol to minimize probability of error – implemented in the form of a
maximum likelihood decoder

68
Detector

69
Signal
Transmission
Decoder

Maximum Likelihood Decoding Principle:


𝑶𝒃𝒔𝒆𝒓𝒗𝒂𝒕𝒊𝒐𝒏 𝒗𝒆𝒄𝒕𝒐𝒓 𝐱 𝒍𝒊𝒆𝒔 𝒊𝒏 𝒓𝒆𝒈𝒊𝒐𝒏 𝒁𝒊 𝐢𝐟
𝟏
σ𝑵
𝒋=𝟏 𝒙𝒋 𝒔𝒌𝒋 − 𝟐 𝑬𝒌 𝐢𝐬 𝐦𝐚𝐱𝐢𝐦𝐮𝐦 𝐟𝐨𝐫 𝐤 = 𝐢

Inner product of x
and 𝒔𝑘

Detector + Signal transmission


decoder = Correlation receiver

70
71
Matched Filter Receiver

• The detector discussed has a set of correlators


• Alternatively, we may use an equivalent structure in place of correlators (each
correlator can be replaced by an LTI filter)
• Consider an LTI filter with IR ℎ𝑗 (𝑡) (j th arm of the bank of filters)
• With the received signal 𝑥(𝑡) as input, the output is defined by convolution integral

• Evaluating this integral over the duration of the transmitted signal 0 ≤ 𝑡 ≤ 𝑇 seconds
and observing the output at t=T, we have

72
73
• Consider next, a detector based on a bank of correlators
• The output of the j th correlator is

• For to equal , we choose

• Or equivalently , the IR of the desired filter can be expressed as

74
• An LTI filter with IR defined in such a way is called as a matched filter
• Hence we can have an equivalent receiver with matched filters in place of correlators
followed by a signal transmission decoder which implements the ML detection rule

75
76
Probability of Error for a correlation receiver or an equivalent matched-
filter receiver in the presence of AWGN (probability of error of maximum
likelihood detection)

• We need to evaluate the performance of the receivers in the presence of AWGN


𝑀
• Assume that the observation space Z is partitioned into a set of M regions, 𝑍𝑖 𝑖=1

, in accordance with the maximum likelihood decision rule


• Suppose also that symbol 𝑚𝑖 (or, equivalently, signal vector 𝒔𝑖 ) is transmitted and
an observation vector x is received
• Then, an error occurs whenever the received signal point represented by x does
not fall inside region 𝑍𝑖 associated with the message point 𝒔𝑖

77
• Averaging over all possible transmitted symbols assumed to be equiprobable, we
see that the average probability of symbol error is

An N-dimensional integration

78
Capacity of an AWGN channel
• The channel capacity, C, is defined to be the maximum rate at which information
can be transmitted through a channel.
• For an AWGN (zero mean) channel of bandwidth B Hz, and PSD of , the
channel capacity is given by
P- received signal power (at the
channel o/p)
𝜎 2 - variance of noise
P/ 𝜎 2 and P/𝑁0 𝐵 are the
expressions for SNR at the
channel output
79
• The channel capacity is an extremely important quantity, since it is possible to
transmit information through a channel at any rate less than the channel
capacity with smallest possible probability of error
• Completely reliable transmission is not possible if the information is sent at a
rate greater than the channel capacity

80
81
82

You might also like