Detection and Estimation
Detection and Estimation
Detection Theory:
Goal: Identify which symbol m (from M possible) was transmitted.
i
Message Source:
Emits one symbol every T seconds.
Symbols belong to a set {m 1
.
, m2 , … , mM }
Examples:
Binary communication: Alphabet = {0, 1}.
Quaternary signaling: Alphabet = {00, 01, 10, 11}.
Assumption: All M symbols are equally likely, simplifying analysis and system
design.
A Priori Probability of Message Symbols
Definition:
Assumption:
Modulator:
Signal s i (t)
One signal s
≤ M
p i = P (m i emitted)
Vector Transmitter:
Dimension: N
i
si =
i (t)
Ei = ∫
0
M
T
1
⎢⎥
⎡
2
f or all i
s i1
s i2
s iN
s i (t) dt < ∞
Channel Assumptions:
1. Linear with sufficient bandwidth ⇒ transmits s
2. Additive White Gaussian Noise (AWGN):
Additive: Noise W (t) adds to the signal.
White: Power is uniformly spread across all frequencies.
Gaussian: Noise samples are normally distributed.
⎤
i (t)
(3.1)
(3.2)
of duration T .
i (t)
(3.3)
without distortion.
Zero-mean, stationary: E[W (t)] = 0, statistical properties are time-invariant.
P e = P (m
^ ≠ mi )
Signal Representation
j=1
First, we check for linear independence. If the signals are not independent, then
there exists a set of constants a , a , . . . , a , not all zero, such that:
1 2 M
⇒ This shows s M
(t) is a linear combination of the others.
We continue checking the remaining signals recursively:
M −2
bi
s M −1 (t) = − ∑ s i (t) (3.12)
b M −1
i=1
This means:
Normalize g 2
(t) to get ϕ 2
(t) :
T
g 2 (t)
2
ϕ 2 (t) = where E 2 = ∫ g 2 (t) dt (3.17)
√E 2 0
Substitute to write:
s 2 (t) − s 21 ϕ 1 (t)
ϕ 2 (t) = (3.18)
√E 2
i−1
j=1
Where:
T
Then normalize:
T
g i (t) 2
ϕ i (t) = with E i = ∫ g i (t) dt (3.21)
√E i 0
Final Result
Now we have a set of orthonormal basis functions ϕ 1 (t), .
. . . , ϕ N (t)
j=1
Remarks
Example
7
2
E1 = ∫ s 1 (t)dt =
3
From Eq. 3.13 (Gram-Schmidt basis function formula), the first orthonormal basis
function is:
s 1 (t) 3
φ 1 (t) = = √ s 1 (t)
√E 1 7
3 7
√ s 1 (t) if 0 < t <
φ 1 (t) = { 7 3
φ 1 (t) = 0 elsewhere
Coefficient s : 21
T T /3
3 T
s 21 = ∫ s 2 (t)ϕ 1 (t) dt = ∫ (1) (√ )dt = √
0 0
T 3
Energy of s 2 (t) :
T 2T /3
2 2
2T
E2 = ∫ s 2 (t) dt = ∫ (1) dt =
0 0
3
3 T 2T
s 2 (t) − s 21 ϕ 1 (t) √ , ≤ t ≤
ϕ 2 (t) = = { T 3 3
√E 2 − s 2 0, elsewhere
21
Coefficient s : 31
s 31 = ∫ s 3 (t)ϕ 1 (t) dt = 0
0
Coefficient s : 32
T 2T /3
3 T
s 32 = ∫ s 3 (t)ϕ 2 (t) dt = ∫ (1) (√ )dt = √
0 T /3
T 3
3 2T
g 3 (t) √ , ≤ t ≤ T
ϕ 3 (t) = = { T 3 $$ ∗ ∗Conclusion : ∗∗
T
√∫ 2 0, elsewhere
g (t) dt
0 3
Signal in 2D:
S 11
S 1 (t) → S 1 = [ ]
S 12
2 2
∥S 1 ∥ = √ S + S
11 12
Generalized to N-Dimensions:
S 11
⎡ ⎤
S 12
S1 = ,
⋮
⎣ ⎦
S 1N
1/2
N 2
∥S 1 ∥ = (∑ S )
j=1 1j
Signals:
N N
j=1 j=1
Coordinates in 2D: (S 11
, S 12 ) and (S 21
, S 22 )
Energy:
2
T T N
2
Ei = ∫ S i (t)dt = ∫ (∑ S ij ϕ j (t)) dt
0 0
j=1
N N T
= ∑ ∑ S ij S ik ∫ ϕ j (t)ϕ k (t)dt
0
j=1 k=1
T N
1 j = k 2
∫ ϕ j (t)ϕ k (t)dt = { ⇒ E i = ∑ S ij
0 j ≠ k
0
j=1
Energy:
2 T
E i = ∥S i ∥ = S Si
i
Matrix form:
S i1
⎡ ⎤
N
S i2
T 2
S i S i = [S i1 S i2 … S iN ] = ∑ S ij
⋮ j=1
⎣ ⎦
S iN
Signals:
T T
S i = [S i1 , … , S iN ] , S k = [S k1 , … , S kN ]
Inner product:
T N
T
∫ S i (t)S k (t)dt = S i S k = ∑ S ij S kj
0
j=1
⮕ This measures similarity (projection) between signals.
Distance in 2D:
2 2
d = √ (x 1 − x 2 ) + (y 1 − y 2 )
Generalized to N-D:
N
2 2
d = ∑(S ij − S kj )
j=1
T N
2 2 2 2
∫ [S i (t) − S k (t)] dt = ∑(S ij − S kj ) = ∥S i − S k ∥ = d
0
j=1
In a digital communication system, in each time slot of duration T seconds, one out of
M possible signals, s (t), s (t), … , s (t), is transmitted. These signals are assumed to
1 2 M
When these signals are transmitted through an Additive White Gaussian Noise
(AWGN) channel, the received signal, x(t), is a sample function of a random process
X(t) . It can be described as:
Here:
w(t) is a sample function of the white Gaussian noise process W (t), which is
assumed to have:
Zero mean
Power spectral density N 0 /2
The primary task of the receiver is to observe the received signal x(t) and make
the "best estimate" of which signal s (t) (and thus which symbol m ) was originally
i i
transmitted.
vector w:
x = si + w f or i = 1, 2, … , M (Eq. 3.52)
Here, x and w are sample values of random vectors X and W, respectively. The
orientation of the noise vector w is completely random.
The observation vector x (representing the received signal x(t)) can also be
represented as a received signal point in the same Euclidean space as the
transmitted signals.
Given the observed vector x, the receiver must perform a mapping from x to an
estimate m
^ of the transmitted symbol m . The goal is to perform this mapping in a way
i
that minimizes the average probability of symbol error in the decision.
Maximum-Likelihood Detector
Decision Rule Goal
Given an observation vector x, choose the message m that was most likely i
transmitted.
1. Error Probability
Average probability of symbol error if decision M = mi is made:
p i f X|m i (x|m i )
P (m i sent |x) = (3.55)
f X (x)
p k f x (x|m k )
Decide M = m i if is maximum f or all k = i (3.55)
f x (x)
M
, then:
4. Decision Regions
Observation space Z is partitioned into regions Z 1, Z2 , . . . , ZM , where:
j=1
2 2
∑(x j − s kj ) = ∥x − s k ∥ (3.63)
j=1
x ∈ Z i if ∥x − s k ∥ is minimum f or k = i (3.64)
Interpretation:
The observation space is divided into Voronoi regions, where each region
contains all points closest to one s . i
{Z 1 , Z 2 , … , Z M }
occurs.
2. Average Symbol Error Probability
M
i=1
1
P (m i sent) =
M
So:
M M
1 1
Pe = ∑ P (x ∉ Z i ∣ m i sent) = 1 − ∑ P (x ∈ Z i ∣ m i sent)
M M
i=1 i=1
3. Likelihood Form
The conditional probability that x ∈ Z given m was sent is:
i i
P (x ∈ Z i ∣ m i sent) = ∫ f (x ∣ m i ) dx
Zi
Interpretation
Correct decision probability for m : integral of the likelihood over the correct
i
region Z .
i
⟨x, s k ⟩ = ∑ x j s kj
j=1
Conclusion:
This structure is called a correlation receiver.
It is optimum in the maximum-likelihood sense for AWGN channels.
Minimizes average probability of symbol error.
Let h j
(t) be the impulse response of a linear filter.
If input is received signal x(t), output is:
∞
h j (t) = ϕ j (T − t) (3.77)
Sampling at t = T
Sample filter output at t = T :
∞
Estimation error: ϵ = a − a
^
Average cost:
2
E = ∫ (a − a
^) f A|X (a|x) da
a
^ M L = arg max f X|A (x|a)
a
: Likelihood function of a
f X|A (x|a)
The Minimum Mean-Square Estimate (MMSE), also known as the Bayes' MMSE, aims
to estimate an unknown parameter α by minimizing the average squared error of the
estimate α
^.
1. Cost Function: The average cost C is defined as the expected value of the
1
squared error (α − α)
^ , conditioned on the observed data x:
2
2
^
C 1 = ∫ (α − α) f A (α|x)dα
where f A
(α|x) is the a posteriori PDF of α given x.
2. Minimization: The MMSE estimate is the α
^ that minimizes this average squared
observation X by minimizing the mean squared error (MSE) between the desired
n
1. Input Samples:
3. Filter Output:
Y n = ∑ h k X n−k
k=0
4. Estimation Error:
εn = Dn − Yn
Objective is to minimize:
2
ζ = E[ε n ]
Substituting Y :
n
M M M
2
ζ = σ D − 2 ∑ h k R DX (k) + ∑ ∑ h k h m R X (m − k)
Where:
σ
2
D
2
= E[D n ] : Power of desired signal
R DX (k) = E[D n X n−k ] : Cross-correlation
R X (m − k) = E[X n−k X n−m ] : Autocorrelation
6. Optimality Condition:
7. Wiener-Hopf Equations:
∑ h m R X (m − k) = R DX (k), k = 0, 1, … , M
m=0
8. Minimum MSE:
M
2
ζ min = σ D − ∑ h k R DX (k)
k=0
Questions
(b) Explain Wiener filter for waveform estimation. (4)
3(a) Draw block diagrams of the sub-systems of a matched filter receiver. (2)
(b) Derive the discrete-time version of Wiener-Hopf equations for a transversal Wiener
filter. (4)
(c) Write down the steps involved in the application of a Least-Mean-Square (LMS)
algorithm. (2.75)
3. (a) What are the criteria for estimation in digital communication? Discuss briefly.
(b) Draw and explain prediction-error process.
(c) Write down the characteristics of adaptive filter.
4. (a) Why are detection and estimation considered fundamental issues in
determining the optimum design of a receiver?
State and prove the Gram-Schmidt orthogonalization procedure.
How can you define a matched filter receiver? Write down the properties of a
matched filter.
5. (a) What is adaptive filter? Write down the characteristics of adaptive filter. 3
(b) With suitable illustration, explain a classical model for the speech production
process. 5
(c) Define matched filter receiver. Describe the detector part of matched filter
receiver with necessary figure(s). 3 2/3
Q3.
a) Why are detection and estimation considered fundamental issues in determining the
optimum design of a receiver? (2%)
b) How can you characterize a matched filter? Write down the properties of matched
filters. (3%)
c) Describe the subsystems of a typical optimum digital communication receiver with
suitable block diagrams. (4%)
d) What do you mean by the average probability of symbol error? (2%)
Answers
(c) Write down the steps involved in the application of a Least-Mean-Square (LMS)
algorithm.
1. Initialization:
Initialize the filter coefficients h to zero or small random values.
k
Y n = ∑ h k X n−k
k=0
3. Error Calculation:
Calculate the estimation error ε as the difference between the desired signal
n
εn = Dn − Yn
4. Coefficient Update:
Update the filter coefficients using the LMS update rule:
h k (n + 1) = h k (n) + με n X n−k , k = 0, 1, … , M
An adaptive filter is a filter whose parameters adjust automatically based on the input
signal to minimize an error criterion, such as mean square error.
(c) Define matched filter receiver. Describe the detector part of matched filter
receiver with necessary figure(s).
Q3.
(a) Why are detection and estimation considered fundamental issues in determining
the optimum design of a receiver?
Detection determines transmitted symbols (e.g., "0" or "1"), while estimation recovers
parameters (e.g., phase, amplitude). Both are fundamental because:
(b) How can you characterize a matched filter? Write down the properties of
matched filters.
A matched filter is a linear filter designed to maximize the signal-to-noise ratio (SNR)
at the output of the filter. It is matched to the shape of the transmitted signal.
2E
SNR max =
N0
8. Correlator Equivalence
Output at t=T equals correlation:
T
∗
y(T ) = ∫ x(τ )s (τ )dτ
0
The average probability of symbol error is the probability that the receiver makes an
incorrect decision about which symbol was transmitted. It is given by:
P e = P (m
^ ≠ mi )
where m
^ is the estimated symbol and m is the transmitted symbol.
i
^ 2
E[(θ − θ) ]
Optimal for Gaussian noise; equivalent to the conditional mean of the posterior
distribution.
2. Maximum Likelihood (ML)
Finds the parameter value that maximizes the likelihood function f (x|θ).
Asymptotically efficient (achieves Cramér-Rao lower bound for large data).
3. Maximum A Posteriori (MAP)
Incorporates prior knowledge of the parameter via f (θ).
Maximizes the posterior distribution f (θ|x).
4. Unbiasedness
Ensures E[θ]
^ = θ
(no systematic error).
5. Consistency
The estimate converges to the true value as sample size increases.
^
en = Xn − X n
M
^
X = ∑ a k X n−k
n k=1
2. Prediction-Error Filter:
Use the present and past samples of a stationary process X n, X n−1 , … , X n−M
This allows the computation of the present sample X using past samples and
n
process σ . 2
X
Detector Part
Source:
Voiced sounds: periodic pulses from vocal cords (glottal excitation).
Unvoiced sounds: random noise source.
Filter:
Vocal tract shapes the sound by acting as a time-varying linear filter
(resonator).
Output:
Speech = Source signal convolved with vocal tract filter response.
Example:
For a voiced sound, the source block generates a periodic pulse train, and the
filter block shapes the spectrum of the pulse train to produce the desired speech
sound.
For an unvoiced sound, the source block generates a random noise signal, and the
filter block shapes the spectrum of the noise signal to produce the desired speech
sound.