+ All Categories
Home > Documents > Digital Communications Fredrik Rusek

Digital Communications Fredrik Rusek

Date post: 16-Oct-2021
Category:
Upload: others
View: 7 times
Download: 0 times
Share this document with a friend
88
Estimation Theory Fredrik Rusek Chapter 11
Transcript
Page 1: Digital Communications Fredrik Rusek

Estimation Theory Fredrik Rusek

Chapter 11

Page 2: Digital Communications Fredrik Rusek

Chapter 10 – Bayesian Estimation

Section 10.8 Bayesian estimators for deterministic parameters

Compute the MSE for a given value of A

If no MVU estimator exists, or is very hard to find, we can apply an MMSE estimator to deterministic parameters Recall the form of the Bayesian estimator for DC-levels in WGN

Page 3: Digital Communications Fredrik Rusek

Chapter 10 – Bayesian Estimation

Section 10.8 Bayesian estimators for deterministic parameters

Compute the MSE for a given value of A

If no MVU estimator exists, or is very hard to find, we can apply an MMSE estimator to deterministic parameters Recall the form of the Bayesian estimator for DC-levels in WGN

α<1

Page 4: Digital Communications Fredrik Rusek

Chapter 10 – Bayesian Estimation

Section 10.8 Bayesian estimators for deterministic parameters

Compute the MSE for a given value of A

If no MVU estimator exists, or is very hard to find, we can apply an MMSE estimator to deterministic parameters Recall the form of the Bayesian estimator for DC-levels in WGN

α<1

Variance smaller than classical estimator Large bias for large A

Page 5: Digital Communications Fredrik Rusek

Chapter 10 – Bayesian Estimation

Section 10.8 Bayesian estimators for deterministic parameters

Compute the MSE for a given value of A

If no MVU estimator exists, or is very hard to find, we can apply an MMSE estimator To deterministic parameters Recall the form of the Bayesian estimator for DC-levels in WGN

α<1

MSE for Bayesian is smaller for A close to the prior mean, but larger far away

Page 6: Digital Communications Fredrik Rusek

Chapter 10 – Bayesian Estimation

Section 10.8 Bayesian estimators for deterministic parameters

However, the BMSE is smaller To deterministic parameters

Page 7: Digital Communications Fredrik Rusek

Chapter 10 – Bayesian Estimation

Section 10.8 Bayesian estimators for deterministic parameters

However, the BMSE is smaller To deterministic parameters

Page 8: Digital Communications Fredrik Rusek

Chapter 10 – Bayesian Estimation

Section 10.8 Bayesian estimators for deterministic parameters

However, the BMSE is smaller To deterministic parameters

Page 9: Digital Communications Fredrik Rusek

Chapter 10 – Bayesian Estimation

Section 10.8 Bayesian estimators for deterministic parameters

However, the BMSE is smaller To deterministic parameters

Page 10: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Risk Functions

To deterministic parameters

p(θ) p(x|θ) θ x

Estimator θ

Error: ε = θ - θ

Page 11: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Risk Functions

To arameters

p(θ) p(x|θ) θ x

Estimator θ

Error: ε = θ - θ

The MMSE estimator minimizes Bayes Risk where the cost function is

Page 12: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Risk Functions

To arameters

p(θ) p(x|θ) θ x

Estimator θ

Error: ε = θ - θ

The MMSE estimator minimizes Bayes Risk where the cost function is

Page 13: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Risk Functions

To arameters

p(θ) p(x|θ) θ x

Estimator θ

Error: ε = θ - θ

The MMSE estimator minimizes Bayes Risk where the cost function is

An estimator that minimizez Bayes risk, for some cost, is termed a Bayes estimator

Page 14: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Let us now optimize for different

To arameters

For a quadratic cost, we already know that

Page 15: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Let us now optimize for different

Bayes risk equals

Page 16: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Let us now optimize for different

Bayes risk equals

Minimize this to minimize Bayes risk

Page 17: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Let us now optimize for different

Bayes risk equals

Page 18: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Let us now optimize for different

Bayes risk equals

Page 19: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Let us now optimize for different

Bayes risk equals

We need , but the limits of the integral depends on Not standard differential

θ

Page 20: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Interlude: Leibnitz’s rule (very useful)

Page 21: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Leibnitz’s rule (very useful)

We have:

Page 22: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Leibnitz’s rule (very useful)

Page 23: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Leibnitz’s rule (very useful)

u =

φ2(u)=

θ

θ

Page 24: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Leibnitz’s rule (very useful)

Page 25: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Leibnitz’s rule (very useful)

Page 26: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Leibnitz’s rule (very useful)

Page 27: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Leibnitz’s rule (very useful)

u =

θ

Lower limit does not depend on u:

Page 28: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Leibnitz’s rule (very useful)

Page 29: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Leibnitz’s rule (very useful)

Page 30: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Leibnitz’s rule (very useful)

Page 31: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Let us now optimize for different

Bayes risk equals

We need , but the limits of the integral depends on Not standard differential

θ

Page 32: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Let us now optimize for different

Bayes risk equals

Page 33: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Let us now optimize for different

Bayes risk equals

θ is the median of the posterior

Page 34: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Let us now optimize for different

Bayes risk equals

θ = median

Page 35: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Let us now optimize for different

Bayes risk equals

θ = median

Page 36: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Let us now optimize for different

Bayes risk equals

θ = median

θ = arg max

Page 37: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Let us now optimize for different

Bayes risk equals

θ = median

θ = arg max

θ = arg max Let δ->0: θ = arg max (maximum a posterori (MAP))

Page 38: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Gausian posterior

What is relation between mean, median and max ?

θ = median

θ = arg max

Page 39: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Gausian posterior

What is relation between mean, median and max ?

Gaussian posterior makes the three risk functions identical

θ = median

θ = arg max

Page 40: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Extension to vector parameter

Suppose we have a vector parameter of unknowns θ

Consider estimation of θ1. It still holds that the MAP estimator uses

Page 41: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Extension to vector parameter

Suppose we have a vector parameter of unknowns θ

Consider estimation of θ1. It still holds that the MAP estimator uses

The parameters θ2 …. θN are nuisance parameters, but we can integrate them away

Page 42: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Extension to vector parameter

Suppose we have a vector parameter of unknowns θ

Consider estimation of θ1. It still holds that the MAP estimator uses

The parameters θ2 …. θN are nuisance parameters, but we can integrate them away The estimator is

Page 43: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Extension to vector parameter

Suppose we have a vector parameter of unknowns θ

Consider estimation of θ1. It still holds that the MAP estimator uses

The parameters θ2 …. θN are nuisance parameters, but we can integrate them away The estimator is

Page 44: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Extension to vector parameter

Suppose we have a vector parameter of unknowns θ

Consider estimation of θ1. It still holds that the MAP estimator uses

The parameters θ2 …. θN are nuisance parameters, but we can integrate them away The estimator is

Page 45: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Extension to vector parameter

In vector form

Page 46: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Extension to vector parameter

Observations Classical approach (non-Bayesian): We must estimate all unknown paramters jointly, except if…..what holds????

Page 47: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Extension to vector parameter

Observations Classical approach (non-Bayesian): We must estimate all unknown paramters jointly, except if Fisher information is diagonal

Vector MMSE estimator minimizes the MSE for each component of the unknown vector parameter θ, i.e.,

Page 48: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Performance of MMSE estimator

Page 49: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Performance of MMSE estimator Function of x

Page 50: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Performance of MMSE estimator

Bayes rule

MMSE estimator

Page 51: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Performance of MMSE estimator

By definition

Page 52: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Performance of MMSE estimator

Page 53: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Performance of MMSE estimator

Element [1,1] of

Page 54: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Additive property Independent observations x1,x2 Estimate θ Assume that x1,x2, θ are jointly Gaussian Theorem 10.2

Page 55: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Additive property Independent observations x1,x2 Estimate θ Assume that x1,x2, θ are jointly Gaussian

Independent observations

Typo in book, should include means as well

Page 56: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Additive property Independent observations x1,x2 Estimate θ Assume that x1,x2, θ are jointly Gaussian

MMSE estimate can be updated sequentially !!!

Page 57: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

MAP estimator

n

Page 58: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

MAP estimator

n

Benefits compared with MMSE

Not needed (typically hard to find)

Optimization generally easier than finding the conditional expectation

Page 59: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

MAP vs ML estimator

Alexander Aljechin (1882-1946) became world chess champion 1927 (by defeating Capablanca) Aljechin defended his title twice, and regained it once

Page 60: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

MAP vs ML estimator

Alexander Aljechin (1882-1946) became world chess champion 1927 (by defeating Capablanca) Aljechin defended his title twice, and regained it once Magnus Calrsen became world champion 2013, and defended the title Once in 2014

Page 61: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

MAP vs ML estimator

Alexander Aljechin (1882-1946) became world chess champion 1927 (by defeating Capablanca) Aljechin defended his title twice, and regained it once Magnus Calrsen became world champion 2013, and defended the title Once in 2014 Now consider a title game in 2015. Observe Y=y1, where y1=win Two hypotheses: • H1: Aljechin defends title • H2: Carlsen defends title

Page 62: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

MAP vs ML estimator

Alexander Aljechin (1882-1946) became world chess champion 1927 (by defeating Capablanca) Aljechin defended his title twice, and regained it once Magnus Calrsen became world champion 2013, and defended the title Once in 2014 Now consider a title game in 2015. Observe Y=y1, where y1=win Two hypotheses: • H1: Aljechin defends title • H2: Carlsen defends title Given the above statistics f(y1|H1)>f(y1|H2)

Page 63: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

MAP vs ML estimator

Alexander Aljechin (1882-1946) became world chess champion 1927 (by defeating Capablanca) Aljechin defended his title twice, and regained it once Magnus Calrsen became world champion 2013, and defended the title Once in 2014 Now consider a title game in 2015. Observe Y=y1, where y1=win Two hypotheses: • H1: Aljechin defends title • H2: Carlsen defends title Given the above statistics f(y1|H1)>f(y1|H2)

ML rule: Aljechin takes title (although he died in 1946)

Page 64: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

MAP vs ML estimator

Alexander Aljechin (1882-1946) became world chess champion 1927 (by defeating Capablanca) Aljechin defended his title twice, and regained it once Magnus Calrsen became world champion 2013, and defended the title Once in 2014 Now consider a title game in 2015. Observe Y=y1, where y1=win Two hypotheses: • H1: Aljechin defends title • H2: Carlsen defends title Given the above statistics f(y1|H1)>f(y1|H2)

MAP rule: f(H1)=0, -> Carlsen defends title

Page 65: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Example DC-level in white noise, uniform prior U[-A0,A0]

The posterior is

We got stuck here: Cannot put the denominator in closed form Cannot integrate the nominator Lets try with the MAP estimator

Page 66: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Example DC-level in white noise, uniform prior U[-A0,A0]

The posterior is

Denominator: Does not depend on A -> irrelevant

Page 67: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Example DC-level in white noise, uniform prior U[-A0,A0]

The posterior is

Denominator: Does not depend on A -> irrelevant We need to maximize the nominator

Page 68: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Example DC-level in white noise, uniform prior U[-A0,A0]

Page 69: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Example DC-level in white noise, uniform prior U[-A0,A0]

Page 70: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Example DC-level in white noise, uniform prior U[-A0,A0]

MAP estimator can be found! Lesson learned (generally true) MAP is easier to find than MMSE

Page 71: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Element-wise MAP for vector-valued parameter

“No-integration-needed” benefit gone

Page 72: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Element-wise MAP for vector-valued parameter

“No-integration-needed” benefit gone

The estimator Minimizes the “hit-or-miss” risk for each I, where δ->0

Page 73: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Element-wise MAP for vector-valued parameter

“No-integration-needed” benefit gone

Let us now define another risk function Easy to prove that as δ->0, Bayes risk is minimized by the vector-MAP-estimator

Page 74: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Element-wise MAP and vector valued MAP are not the same

Vector-valued MAP solution Element-wise MAP solution

Page 75: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Two properties of vector-MAP • For jointly Gaussian x and θ, the conditional mean E(θ|x) coincides with

the peak of p(θ|x). Hence, the vector-MAP and the MMSE coincide.

Page 76: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Two properties of vector-MAP • For jointly Gaussian x and θ, the conditional mean E(θ|x) coincides with

the peak of p(θ|x). Hence, the vector-MAP and the MMSE coincide.

• Invariance does not hold for MAP (as opposed to MLE)

Page 77: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Invariance Why does invariance hold for MLE? With α=g(θ), it holds that p(x|α) = pθ(x|g-1(α))

Page 78: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Invariance Why does invariance hold for MLE? With α=g(θ), it holds that p(x|α) = pθ(x|g-1(α)) However, MAP involves the prior, and it doesn’t hold that pα(α)=pθ(g-1(α)), since the two distributions are related through the Jacobian

Page 79: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Example

Exponential Inverse gamma

Page 80: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Example

Exponential Inverse gamma

MAP

Page 81: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Example

Exponential Inverse gamma

MAP

Page 82: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Example

Exponential Inverse gamma

MAP

Page 83: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Example

Consider estimation of

? (holds for MLE)

Page 84: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Example

Consider estimation of

? (holds for MLE)

Page 85: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Example

Consider estimation of

? (holds for MLE)

Page 86: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Example

Consider estimation of

? (holds for MLE)

Page 87: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Example

Consider estimation of

? (holds for MLE)

Page 88: Digital Communications Fredrik Rusek

Chapter 11 – General Bayesian Estimators

Example

Consider estimation of

….


Recommended