Top Banner
E. Somersalo Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x 1 ,x 2 ,...,x N , x j R n . The problem is to infer on the underlying probability distribution that gives rise to the data S . Statistical modeling Statistical analysis. Computational Methods in Inverse Problems, Mat–1.3626 0-0
36

Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

Jul 22, 2020

Download

Documents

dariahiddleston
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Basic Problem of Statistical Inference

Assume that we have a set of observations

S =x1, x2, . . . , xN

, xj ∈ Rn.

The problem is to infer on the underlying probability distribution that givesrise to the data S.

• Statistical modeling

• Statistical analysis.

Computational Methods in Inverse Problems, Mat–1.3626 0-0

Page 2: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Parametric or non-parametric?

• Parametric problem: The underlying probability density has a specifiedform and depends on a number of parameters. The problem is to inferon those parameters.

• Non-parametric problem: No analytic expression for the probability den-sity is available. Description consists of defining the dependency/non-dependency of the data. Numerical exploration.

Typical situation for parametric model: The distribution is the probabilitydensity of a random variable X : Ω → Rn.

• Parametric problem suitable for inverse problems

• Model for a learning process

Computational Methods in Inverse Problems, Mat–1.3626 0-1

Page 3: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Law of Large Numbers

General result (“Statistical law of nature”):

Assume that X1, X2, . . . are independent and identically distributed randomvariables with finite mean µ and variance σ2. Then,

limn→∞

1n

(X1 + X2 + · · ·+ Xn

)= µ

almost certainly.

Almost certainly means that with probability one,

limn→∞

1n

(x1 + x2 + · · ·+ xn

)= µ,

xj being a realization of Xj .

Computational Methods in Inverse Problems, Mat–1.3626 0-2

Page 4: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Example

SampleS =

x1, x2, . . . , xN

, xj ∈ R2.

Parametric model: xj realizations of

X ∼ N (x0, Γ),

with unknown mean x0 ∈ R2 and covariance matrix Γ ∈ R2×2.

Probability density of X:

π(x | x0, Γ) =1

2πdet(Γ)1/2exp

(−1

2(x− x0)TΓ−1(x− x0)

).

Problem: Estimate the parameters x0 and Γ.

Computational Methods in Inverse Problems, Mat–1.3626 0-3

Page 5: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

The Law of Large Number suggests that we calculate

x0 = EX

≈ 1n

n∑

j=1

xj = x0. (1)

Covariance matrix: observe that if X1, X2, . . . are i.i.d, so are f(X1), f(X2), . . .for any function f : R2 7→ Rk.

Try

Γ = cov(X) = E(X − x0)(X − x0)T

≈ E(X − x0)(X − x0)T

(2)

≈ 1n

n∑

j=1

(xj − x0)(xj − x0)T = Γ.

Formulas (1) and (2) are known as empirical mean and covariance, respec-tively.

Computational Methods in Inverse Problems, Mat–1.3626 0-4

Page 6: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Case 1: Gaussian sample

0 1 2 3 4 50

0.5

1

1.5

2

2.5

3

3.5

4

4.5

5

0 1 2 3 4 50

0.5

1

1.5

2

2.5

3

3.5

4

4.5

5

Computational Methods in Inverse Problems, Mat–1.3626 0-5

Page 7: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Sample size N = 200.

Eigenvectors of the covariance matrix:

Γ = UDUT, (3)

where U ∈ R2×2 is an orthogonal matrix and D ∈ R2×2 is a diagonal,

UT = U−1.

U =[

v1 v2

], D =

[λ1

λ2

],

Γvj = λjv, , j = 1, 2.

Scaled eigenvectors,vj,scaled = 2

√λjvj ,

where√

λj =standard deviation (STD).

Computational Methods in Inverse Problems, Mat–1.3626 0-6

Page 8: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Case 2: Non-Gaussian Sample

−1.5 −1 −0.5 0 0.5 1 1.5−1.5

−1

−0.5

0

0.5

1

1.5

−1.5 −1 −0.5 0 0.5 1 1.5−1.5

−1

−0.5

0

0.5

1

1.5

Computational Methods in Inverse Problems, Mat–1.3626 0-7

Page 9: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Estimate of normality/non-normality

Consider the sets

Bα =x ∈ R2 | π(x) ≥ α

, α > 0.

If π is Gaussian, Bα is an ellipse or ∅.Calculate the integral

PX ∈ Bα

=

π(x)dx. (4)

We call Bα the credibility ellipse with credibility p, 0 < p < 1, if

PX ∈ Bα

= p, giving α = α(p). (5)

Computational Methods in Inverse Problems, Mat–1.3626 0-8

Page 10: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Assume that the Gaussian density π has the center of mass and covariancematrix x0 and Γ estimated from the sample S of size N .

If S is normally distributed,

#xj ∈ Bα(p)

≈ pN. (6)

Deviations due to non-normality.

Computational Methods in Inverse Problems, Mat–1.3626 0-9

Page 11: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

How do we calculate the quantity?

Eigenvalue decomposition:

(x− x0)TΓ−1(x− x0) = (x− x0)TUD−1UT(x− x0)

= ‖D−1/2UT(x− x0)‖2,

since U is orthogonal, i.e., U−1 = UT, and we wrote

D−1/2 =[

1/√

λ1

1/√

λ2

].

We introduce the change of variables,

w = f(x) = W (x− x0), W = D−1/2UT.

Computational Methods in Inverse Problems, Mat–1.3626 0-10

Page 12: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Write the the integral in terms of the new variable w,∫

π(x)dx =1

2π(det(Γ)

)1/2

exp(−1

2(x− x0)TΓ−1(x− x0)

)dx

=1

2π(det(Γ)

)1/2

exp(−1

2‖W (x− x0‖2

)dx

=12π

f(Bα)

exp(−1

2‖w‖2

)dw,

where we used the fact that

dw = det(W )dx =1√

λ1λ2

dx =1

det(Γ)1/2dx.

Note:det(Γ) = det(UDUT) = det(UTUD) = det(D) = λ1λ2.

Computational Methods in Inverse Problems, Mat–1.3626 0-11

Page 13: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

The equiprobability curves for the density for w are circles centered aroundthe origin, i.e.,

f(Bα) = Dδ =w ∈ R2 | ‖w‖ < δ

for some δ > 0.

Solve δ: Integrate in radial coordinates (r, θ),

12π

exp(−1

2‖w‖2

)dw =

∫ δ

0

exp(−1

2r2

)rdr

= 1− exp(−1

2δ2

)= p,

implying that

δ = δ(p) =

√2 log

(1

1− p

).

Computational Methods in Inverse Problems, Mat–1.3626 0-12

Page 14: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

To see if the sample points xj is within the confidence ellipse with confidencep, it is enough to check if the condition

‖wj‖ < δ(p), wj = W (xj − x0), 1 ≤ j ≤ N

is valid.

Plotp 7→ 1

N#

xj ∈ Bα(p)

Computational Methods in Inverse Problems, Mat–1.3626 0-13

Page 15: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Example

0 20 40 60 80 1000

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1GaussianNon−gaussian

Computational Methods in Inverse Problems, Mat–1.3626 0-14

Page 16: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Matlab code

N = length(S(1,:)); % Size of the samplexmean = (1/N)*(sum(S’)’); % Mean of the sampleCS = S - xmean*ones(1,N); % Centered sampleGamma = 1/N*CS*CS’; % Covariance matrix

% Whitening of the sample

[V,D] = eig(Gamma); % Eigenvalue decompositionW = diag([1/sqrt(D(1,1));1/sqrt(D(2,2))])*V’;WS = W*CS; % Whitened samplenormWS2 = sum(WS.^2);

Computational Methods in Inverse Problems, Mat–1.3626 0-15

Page 17: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

% Calculating percentual amount of scatter points that are% included in the confidence ellipses

rinside = zeros(11,1);rinside(11) = N;for j = 1:9

delta2 = 2*log(1/(1-j/10));rinside(j+1) = sum(normWS2<delta2);

endrinside = (1/N)*rinside;

plot([0:10:100],rinside,’k.-’,’MarkerSize’,12)

Computational Methods in Inverse Problems, Mat–1.3626 0-16

Page 18: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Which one of the following formulae?

Γ =1N

N∑

j=1

(xj − x0)(xj − x0)T,

or

Γ =1N

N∑

j=1

xjxTj − x0x

T0 .

The former, please.

Computational Methods in Inverse Problems, Mat–1.3626 0-17

Page 19: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Example

Calibration of a measurement instrument:

• Measure a dummy load whose output known

• Subtract from actual measurement

• Analyze the noise

Discrete sampling; Output is a vector of length n.

Noise vector x ∈ Rn is a realization of

X : Ω → Rn.

Estimate mean and variance

x0 =1n

n∑

j=1

xj (offset), σ2 =1n

n∑

j=1

(xj − x0)2.

Computational Methods in Inverse Problems, Mat–1.3626 0-18

Page 20: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Improving Signal-to-Noise Ratio (SNR):

• Repeat the measurement

• Average

• Hope that the target is stationary

Averaged noise:

x =1N

N∑

k=1

x(k) ∈ Rn.

How large must N be to reduce the noise enough?

Computational Methods in Inverse Problems, Mat–1.3626 0-19

Page 21: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Averaged noise x is a realization of a random variable

X =1N

N∑

k=1

X(k) ∈ Rn.

If X(1), X(2), . . . i.i.d., X is asymptotically Gaussian by Central Limit Theo-rem, and its variance is

var(X) =σ2

N.

Repeat until the variance is below a given threshold,

σ2

N< τ2.

Computational Methods in Inverse Problems, Mat–1.3626 0-20

Page 22: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

0 10 20 30 40 50−3

−2

−1

0

1

2

3Number of averaged signals = 1

0 10 20 30 40 50−3

−2

−1

0

1

2

3Number of averaged signals = 5

Computational Methods in Inverse Problems, Mat–1.3626 0-21

Page 23: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

0 10 20 30 40 50−3

−2

−1

0

1

2

3Number of averaged signals = 10

0 10 20 30 40 50−3

−2

−1

0

1

2

3Number of averaged signals = 25

Computational Methods in Inverse Problems, Mat–1.3626 0-22

Page 24: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Maximum Likelihood Estimator: frequentist’s approach

Parametric problem,

X ∼ πθ(x) = π(x | θ), θ ∈ Rk.

Independent realizations: Assume that the observations xj are obtained inde-pendently.

More precisely: X1, X2, . . . , XN i.i.d, xj is a realization of Xj .

Independency:

π(x1, x2, . . . , xN | θ) = π(x1 | θ)π(x2 | θ) · · ·π(xN | θ),

or, briefly,

π(S | θ) =N∏

j=1

π(xj | θ),

Computational Methods in Inverse Problems, Mat–1.3626 0-23

Page 25: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Maximum likelihood (ML) estimator of θ = parameter value that maximizesthe probability of the outcome:

θML = arg maxN∏

j=1

π(xj | θ).

DefineL(S | θ) = − log(π(S | θ)).

Minimizer of L(S | θ) = maximizer of π(S | θ).

Computational Methods in Inverse Problems, Mat–1.3626 0-24

Page 26: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Example

Gaussian model

π(x | x0, σ2) =

1√2πσ2

exp(

12σ2

(x− x0)2)

, θ =[

x0

σ2

]=

[θ1

θ2

].

Likelihood function is

N∏

j=1

π(xj | θ) =(

12πθ2

)N/2

exp

− 1

2θ2

N∑

j=1

(xj − θ1)2

= exp

− 1

2θ2

N∑

j=1

(xj − θ1)2 − N

2log

(2πθ2

)

= exp (−L(S | θ)) .

Computational Methods in Inverse Problems, Mat–1.3626 0-25

Page 27: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

We have

∇θL(S | θ) =

∂L

∂θ1

∂L

∂θ2

=

− 1θ22

N∑

j=1

xj +N

θ22

θ1

− 12θ2

2

N∑

j=1

(xj − θ1)2 +N

2θ2

.

Setting ∇θL(S | θ) = 0 gives

x0 = θML,1 =1N

N∑

j=1

xj ,

σ2 = θML,2 =1N

N∑

j=1

(xj − θML,1)2.

Computational Methods in Inverse Problems, Mat–1.3626 0-26

Page 28: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Example

Parametric modelπ(n | θ) =

θn

n!e−θ,

sample S = n1, · · · , nN, nk ∈ N, obtained by independent sampling.

The likelihood density is

π(S | θ) =N∏

k=1

π(nk) = e−NθN∏

k=1

θnk

nk!,

and its negative logarithm is

L(S | θ) = − log π(S | θ) =N∑

k=1

(θ − nk log θ + log nk!

).

Computational Methods in Inverse Problems, Mat–1.3626 0-27

Page 29: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Derivative with respect to θ to zero:

∂θL(S | θ) =

N∑

k=1

(1− nk

θ

)= 0, (7)

leading to

θML =1N

N∑

k=1

nk.

Warning:

var(N) ≈ 1N

N∑

k=1

nk − 1

N

N∑

j=1

nj

2

,

which is different from the estimate of θML obtained above.

Computational Methods in Inverse Problems, Mat–1.3626 0-28

Page 30: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Assume that θ is known a priori to be relatively large.

Use Gaussian approximation:

N∏

j=1

πPoisson(nj | θ) ≈(

12πθ

)N/2

exp

− 1

N∑

j=1

(nj − θ)2

=(

12π

)N/2

exp

−1

2

1

θ

N∑

j=1

(nj − θ)2 + N log θ

.

L(S | θ) =1θ

N∑

j=1

(nj − θ)2 + N log θ.

Computational Methods in Inverse Problems, Mat–1.3626 0-29

Page 31: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

An approximation for θML: Minimize

L(S | θ) =1θ

N∑

j=1

(nj − θ)2 + N log θ.

Write

∂θL(S | θ) = − 1

θ2

N∑

j=1

(nj − θ)2 − 2θ

N∑

j=1

(nj − θ) +N

θ= 0,

or

−N∑

j=1

(nj − θ)2 − 2N∑

j=1

θ(nj − θ) + Nθ = Nθ2 + Nθ −N∑

j=1

n2j = 0,

giving

θ =

1

4+

1N

N∑

j=1

n2j

1/2

− 12

6= 1

N

N∑

j=1

nj

.

Computational Methods in Inverse Problems, Mat–1.3626 0-30

Page 32: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Example

Multivariate Gaussian model,

X ∼ N (x0, Γ),

where x0 ∈ Rn is unknown, Γ ∈ Rn×n is symmetric positive definite (SPD)and known.

Model reduction: assume that x0 depends on hidden parameters z ∈ Rk

through a linear equation,

x0 = Az, A ∈ Rn×k, z ∈ Rk. (8)

Model for an inverse problem: z is the true physical quantity that in the idealcase is related to the observable x0 through the linear model (8).

Computational Methods in Inverse Problems, Mat–1.3626 0-31

Page 33: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Noisy observations:X = Az + E, E ∼ N (0, Γ).

Obviously,E

X

= Az + E

E

= Az = x0,

andcov(X) = E

(X −Az)(X −Az)T

= E

EET

= Γ.

The probability density of X, given z, is

π(x | z) =1

(2π)n/2det(Γ)1/2exp

(−1

2(x−Az)TΓ−1(x−Az)

).

Computational Methods in Inverse Problems, Mat–1.3626 0-32

Page 34: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Independent observations:

S =x1, . . . , xN

, xj ∈ Rn.

Likelihood function

N∏

j=1

π(xj | z) ∝ exp

−1

2

N∑

j=1

(xj −Az)TΓ−1(xj −Az)

is maximized by minimizing

L(S | z) =12

N∑

j=1

(xj −Az)TΓ−1(xj −Az)

=N

2zT

[ATΓ−1A

]z − zT

[ATΓ−1

N∑

j=1

xj

]+

12

N∑

j=1

xTj Γ−1xj .

Computational Methods in Inverse Problems, Mat–1.3626 0-33

Page 35: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Zeroing of the gradient gives

∇zL(S | z) = N[ATΓ−1A

]z −ATΓ−1

N∑

j=1

xj = 0,

i.e., the maximum likelihood estimator zML is the solution of the linear system

[ATΓ−1A

]z = ATΓ−1x, x =

1N

N∑

j=1

xj .

The solution may not exist; All depends on the properties of the model reduc-tion matrix A ∈ Rn×k.

Computational Methods in Inverse Problems, Mat–1.3626 0-34

Page 36: Basic Problem of Statistical Inference · Basic Problem of Statistical Inference Assume that we have a set of observations S = ' x1;x2;:::;xN “; xj 2 Rn: The problem is to infer

E. Somersalo

Particular case: one observation, S =x,

L(z | x) = (x−Az)TΓ−1(x−Az).

Eigenvalue decomposition of the covariance matrix,

Γ = UDUT,

or,Γ−1 = WTW, W = D−1/2UT,

we haveL(z | x) = ‖W (Az − x)‖2.

Hence, the problem reduces to a weighted least squares problem

Computational Methods in Inverse Problems, Mat–1.3626 0-35