EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th , 2006
Jan 18, 2018
EE 551/451, Fall, 2006
Communication Systems
Zhu Han
Department of Electrical and Computer Engineering
Class 15
Oct. 10th, 2006
EE 541/451 Fall 2006
OutlineOutline Homework Exam format Second half schedule
– Chapter 7– Chapter 16– Chapter 8– Chapter 9– Standards
Estimation and detection this class: chapter 14, not required– Estimation theory, methods, and examples– Detection theory, methods, and examples
Information theory next Tuesday: chapter 15, not required
EE 541/451 Fall 2006
Estimation TheoryEstimation Theory Consider a linear process
y = H + ny = observed data
= sending informationn = additive noise
If is known, H is unknown. Then estimation is the problem of finding the statistically optimal , given y, and knowledge of noise properties.
If H is known, then detection is the problem of finding the most likely sending information , given y, and knowledge of noise properties.
In practical system, the above two steps are conducted iteratively to track the channel changes then transmit data.
EE 541/451 Fall 2006
Different Approaches for EstimationDifferent Approaches for Estimation
Minimum variance unbiased estimators Subspace estimators Least Squares Maximum-likelihood Maximum a posteriori
has no statistical
basis
uses knowledge of noise PDF
uses prior information
about
EE 541/451 Fall 2006
Least Squares EstimatorLeast Squares Estimator Least Squares:
LS = argmin ||y – H||2
Natural estimator– want solution to match observation Does not use any information about noise There is a simple solution (a.k.a. pseudo-inverse):
LS = (HTH)-1 HTy What if we know something about the noise? Say we know Pr(n)…
EE 541/451 Fall 2006
Maximum Likelihood EstimatorMaximum Likelihood Estimator Simple idea: want to maximize Pr(y|) Can write Pr(n) = e-L(n) , n = y – H, and
Pr(n) = Pr(y|) = e-L(y, )
if white Gaussian n, Pr(n) = e-||n||2/2 2 and
L(y, ) = ||y-H||2/22
ML = argmax Pr(y|) = argmin L(y, )– called the likelihood function
ML = argmin ||y-H||2/22
This is the same as Least Squares!
EE 541/451 Fall 2006
Maximum Likelihood EstimatorMaximum Likelihood Estimator But if noise is jointly Gaussian with cov. matrix C Recall C , E(nnT). Then
Pr(n) = e-½ nT C-1 n
L(y|) = ½ (y-H)T C-1 (y-H)
ML = argmin ½ (y-H)TC-1(y-H) This also has a closed form solution
ML = (HTC-1H)-1 HTC-1y If n is not Gaussian at all, ML estimators become complicated
and non-linear Fortunately, in most channel noise is usually Gaussian
EE 541/451 Fall 2006
Estimation example - DenoisingEstimation example - Denoising Suppose we have a noisy signal y, and wish to obtain the
noiseless image x, where
y = x + n Can we use Estimation theory to find x? Try: H = I, = x in the linear model Both LS and ML estimators simply give x = y! we need a more powerful model Suppose x can be approximated by a polynomial, i.e. a mixture
of 1st p powers of r:
x = i=0p ai ri
EE 541/451 Fall 2006
Example – DenoisingExample – Denoising
LS = (HTH)-1HTy
x = i=0p ai ri
H
y
Least Squares estimate:
y1
y2
yn
1 r11 r1
p
1 r21 r2
p
1 rn1 rn
p
=
a0
a1
ap
n1
n2
nn
+
EE 541/451 Fall 2006
Maximum a Posteriori (MAP) EstimateMaximum a Posteriori (MAP) Estimate
This is an example of using a signal prior information Priors are generally expressed in the form of a PDF Pr(x) Once the likelihood L(x) and prior are known, we have
complete statistical knowledge LS/ML are suboptimal in presence of prior MAP (aka Bayesian) estimates are optimal
Bayes Theorem:Pr(x|y) = Pr(y|x) Pr(x) Pr(y)
likelihood
priorposterior
EE 541/451 Fall 2006
Maximum a Posteriori (Bayesian) EstimateMaximum a Posteriori (Bayesian) Estimate Consider the class of linear systems y = Hx + n Bayesian methods maximize the posterior probability:
Pr(x|y) ∝ Pr(y|x) . Pr(x) Pr(y|x) (likelihood function) = exp(- ||y-Hx||2) Pr(x) (prior PDF) = exp(-G(x)) Non-Bayesian: maximize only likelihood
xest = arg min ||y-Hx||2
Bayesian:
xest = arg min ||y-Hx||2 + G(x) ,where G(x) is obtained from the prior distribution of x
If G(x) = ||Gx||2 Tikhonov Regularization
EE 541/451 Fall 2006
Expectation and Maximization (EM)Expectation and Maximization (EM) Expectation and Maximization (EM) algorithm alternates
between performing an expectation (E) step, which computes an expectation of the likelihood by including the latent variables as if they were observed, and a maximization (M) step, which computes the maximum likelihood estimates of the parameters by maximizing the expected likelihood found on the E step. The parameters found on the M step are then used to begin another E step, and the process is repeated.– E-step: Estimation for unobserved event (which Gaussian is
used), conditioned on the observation, using the values from the last maximization step.
– M-step: You want to maximize the expected log-likelihood of the joint event
EE 541/451 Fall 2006
Minimum-variance unbiased estimatorMinimum-variance unbiased estimator Biased and unbiased estimators An unbiased estimator of parameters, whose variance is
minimized for all values of the parameters. The Cramer-Rao Lower Bound (CRLB) sets a lower bound
on the variance of any unbiased estimator. Biased estimator might have better performances than unbiased
estimator in terms of variance. Subspace methods
– MUSIC– ESPRIT – Widely used in RADA– Helicopter, Weapon detection (from feature)
EE 541/451 Fall 2006
What is DetectionWhat is Detection Deciding whether, and when, an event occurs a.k.a. Decision Theory, Hypothesis testing Presence/absence of signal
– RADA– Received signal is 0 or 1– Stock goes high or not– Criminal is convicted or set free
Measures whether statistically significant change has occurred or not
EE 541/451 Fall 2006
DetectionDetection
“Spot the Money”
EE 541/451 Fall 2006
Hypothesis Testing with Matched FilterHypothesis Testing with Matched Filter Let the signal be y(t), model be h(t)
Hypothesis testing:
H0: y(t) = n(t) (no signal)
H1: y(t) = h(t) + n(t) (signal) The optimal decision is given by the Likelihood ratio test
(Nieman-Pearson Theorem)
Select H1 if L(y) = Pr(y|H1)/Pr(y|H0) > g
otherwise select H0
EE 541/451 Fall 2006
Signal detection paradigmSignal detection paradigm Signal trials Noise trials
EE 541/451 Fall 2006
Signal DetectionSignal Detection
EE 541/451 Fall 2006
Receiver operating characteristic (ROC) curveReceiver operating characteristic (ROC) curve
EE 541/451 Fall 2006
Matched FiltersMatched Filters Optimal linear filter for maximizing the signal to noise ratio (SNR) at the
sampling time in the presence of additive stochastic noise Given transmitter pulse shape g(t) of duration T, matched filter is given by hopt(t) = k g*(T-t) for all k
g(t)
Pulse signal
w(t)
x(t) h(t) y(t)
t = T
y(T)
Matched filter
EE 541/451 Fall 2006
Questions?Questions?