Top Banner
EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th , 2006
21

EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

Jan 18, 2018

Download

Documents

Abigail Rose

EE 541/451 Fall 2006 Estimation Theory Consider a linear process y = H  + n y = observed data  = sending information n = additive noise If  is known, H is unknown. Then estimation is the problem of finding the statistically optimal , given y,  and knowledge of noise properties. If H is known, then detection is the problem of finding the most likely sending information , given y,  and knowledge of noise properties. In practical system, the above two steps are conducted iteratively to track the channel changes then transmit data.
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 551/451, Fall, 2006

Communication Systems

Zhu Han

Department of Electrical and Computer Engineering

Class 15

Oct. 10th, 2006

                                                           

Page 2: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

OutlineOutline Homework Exam format Second half schedule

– Chapter 7– Chapter 16– Chapter 8– Chapter 9– Standards

Estimation and detection this class: chapter 14, not required– Estimation theory, methods, and examples– Detection theory, methods, and examples

Information theory next Tuesday: chapter 15, not required

Page 3: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

Estimation TheoryEstimation Theory Consider a linear process

y = H + ny = observed data

= sending informationn = additive noise

If is known, H is unknown. Then estimation is the problem of finding the statistically optimal , given y, and knowledge of noise properties.

If H is known, then detection is the problem of finding the most likely sending information , given y, and knowledge of noise properties.

In practical system, the above two steps are conducted iteratively to track the channel changes then transmit data.

Page 4: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

Different Approaches for EstimationDifferent Approaches for Estimation

Minimum variance unbiased estimators Subspace estimators Least Squares Maximum-likelihood Maximum a posteriori

has no statistical

basis

uses knowledge of noise PDF

uses prior information

about

Page 5: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

Least Squares EstimatorLeast Squares Estimator Least Squares:

LS = argmin ||y – H||2

Natural estimator– want solution to match observation Does not use any information about noise There is a simple solution (a.k.a. pseudo-inverse):

LS = (HTH)-1 HTy What if we know something about the noise? Say we know Pr(n)…

Page 6: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

Maximum Likelihood EstimatorMaximum Likelihood Estimator Simple idea: want to maximize Pr(y|) Can write Pr(n) = e-L(n) , n = y – H, and

Pr(n) = Pr(y|) = e-L(y, )

if white Gaussian n, Pr(n) = e-||n||2/2 2 and

L(y, ) = ||y-H||2/22

ML = argmax Pr(y|) = argmin L(y, )– called the likelihood function

ML = argmin ||y-H||2/22

This is the same as Least Squares!

Page 7: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

Maximum Likelihood EstimatorMaximum Likelihood Estimator But if noise is jointly Gaussian with cov. matrix C Recall C , E(nnT). Then

Pr(n) = e-½ nT C-1 n

L(y|) = ½ (y-H)T C-1 (y-H)

ML = argmin ½ (y-H)TC-1(y-H) This also has a closed form solution

ML = (HTC-1H)-1 HTC-1y If n is not Gaussian at all, ML estimators become complicated

and non-linear Fortunately, in most channel noise is usually Gaussian

Page 8: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

Estimation example - DenoisingEstimation example - Denoising Suppose we have a noisy signal y, and wish to obtain the

noiseless image x, where

y = x + n Can we use Estimation theory to find x? Try: H = I, = x in the linear model Both LS and ML estimators simply give x = y! we need a more powerful model Suppose x can be approximated by a polynomial, i.e. a mixture

of 1st p powers of r:

x = i=0p ai ri

Page 9: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

Example – DenoisingExample – Denoising

LS = (HTH)-1HTy

x = i=0p ai ri

H

y

Least Squares estimate:

y1

y2

yn

1 r11 r1

p

1 r21 r2

p

1 rn1 rn

p

=

a0

a1

ap

n1

n2

nn

+

Page 10: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

Maximum a Posteriori (MAP) EstimateMaximum a Posteriori (MAP) Estimate

This is an example of using a signal prior information Priors are generally expressed in the form of a PDF Pr(x) Once the likelihood L(x) and prior are known, we have

complete statistical knowledge LS/ML are suboptimal in presence of prior MAP (aka Bayesian) estimates are optimal

Bayes Theorem:Pr(x|y) = Pr(y|x) Pr(x) Pr(y)

likelihood

priorposterior

Page 11: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

Maximum a Posteriori (Bayesian) EstimateMaximum a Posteriori (Bayesian) Estimate Consider the class of linear systems y = Hx + n Bayesian methods maximize the posterior probability:

Pr(x|y) ∝ Pr(y|x) . Pr(x) Pr(y|x) (likelihood function) = exp(- ||y-Hx||2) Pr(x) (prior PDF) = exp(-G(x)) Non-Bayesian: maximize only likelihood

xest = arg min ||y-Hx||2

Bayesian:

xest = arg min ||y-Hx||2 + G(x) ,where G(x) is obtained from the prior distribution of x

If G(x) = ||Gx||2 Tikhonov Regularization

Page 12: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

Expectation and Maximization (EM)Expectation and Maximization (EM) Expectation and Maximization (EM) algorithm alternates

between performing an expectation (E) step, which computes an expectation of the likelihood by including the latent variables as if they were observed, and a maximization (M) step, which computes the maximum likelihood estimates of the parameters by maximizing the expected likelihood found on the E step. The parameters found on the M step are then used to begin another E step, and the process is repeated.– E-step: Estimation for unobserved event (which Gaussian is

used), conditioned on the observation, using the values from the last maximization step.

– M-step: You want to maximize the expected log-likelihood of the joint event

Page 13: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

Minimum-variance unbiased estimatorMinimum-variance unbiased estimator Biased and unbiased estimators An unbiased estimator of parameters, whose variance is

minimized for all values of the parameters. The Cramer-Rao Lower Bound (CRLB) sets a lower bound

on the variance of any unbiased estimator. Biased estimator might have better performances than unbiased

estimator in terms of variance. Subspace methods

– MUSIC– ESPRIT – Widely used in RADA– Helicopter, Weapon detection (from feature)

Page 14: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

What is DetectionWhat is Detection Deciding whether, and when, an event occurs a.k.a. Decision Theory, Hypothesis testing Presence/absence of signal

– RADA– Received signal is 0 or 1– Stock goes high or not– Criminal is convicted or set free

Measures whether statistically significant change has occurred or not

Page 15: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

DetectionDetection

“Spot the Money”

Page 16: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

Hypothesis Testing with Matched FilterHypothesis Testing with Matched Filter Let the signal be y(t), model be h(t)

Hypothesis testing:

H0: y(t) = n(t) (no signal)

H1: y(t) = h(t) + n(t) (signal) The optimal decision is given by the Likelihood ratio test

(Nieman-Pearson Theorem)

Select H1 if L(y) = Pr(y|H1)/Pr(y|H0) > g

otherwise select H0

Page 17: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

Signal detection paradigmSignal detection paradigm Signal trials Noise trials

Page 18: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

Signal DetectionSignal Detection

Page 19: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

Receiver operating characteristic (ROC) curveReceiver operating characteristic (ROC) curve

Page 20: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

Matched FiltersMatched Filters Optimal linear filter for maximizing the signal to noise ratio (SNR) at the

sampling time in the presence of additive stochastic noise Given transmitter pulse shape g(t) of duration T, matched filter is given by hopt(t) = k g*(T-t) for all k

g(t)

Pulse signal

w(t)

x(t) h(t) y(t)

t = T

y(T)

Matched filter

Page 21: EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.

EE 541/451 Fall 2006

                                                           

Questions?Questions?