Top Banner
Computer Vision Group Prof. Daniel Cremers 14. Sampling Methods
56

14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

Apr 01, 2018

Download

Documents

dangkhanh
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

Computer Vision Group Prof. Daniel Cremers

14. Sampling Methods

Page 2: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Sampling MethodsSampling Methods are widely used in Computer Science

• as an approximation of a deterministic algorithm

• to represent uncertainty without a parametric model

• to obtain higher computational efficiency with a small approximation error

Sampling Methods are also oftencalled Monte Carlo Methods

Example: Monte-Carlo Integration

• Sample in the bounding box

• Compute fraction of inliers

•Multiply fraction with box size

2

Page 3: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Non-Parametric Representation

Probability distributions (e.g. a robot‘s belief) can be represeted:

• Parametrically: e.g. using mean and covariance of a Gaussian

• Non-parametrically: using a set of hypotheses (samples) drawn from the distribution

Advantage of non-parametric representation:

• No restriction on the type of distribution (e.g. can be multi-modal, non- Gaussian, etc.)

3

Page 4: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Non-Parametric Representation

The more samples are in an interval, the higher the probability

of that interval

But:

How to draw samples from a function/distribution?

4

Page 5: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Sampling from a Distribution

There are several approaches:

• Probability transformation

• Uses inverse of the c.d.f h

• Rejection Sampling

• Importance Sampling

• MCMC

Probability transformation:

• Sample uniformly in [0,1]

• Transform using h-1

But:

• Requires calculation of h and its inverse

“Cumulative distribution Function”

5

Page 6: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

• If : keep the sampleotherwise: reject the sample

Rejection Sampling

1. Simplification:

• Assume for all z

• Sample z uniformly

• Sample c from

c

p(z)c’

z’

p(z’)

OKz

6

Page 7: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Rejection Sampling

2. General case:

Assume we can evaluate

• Find proposal distribution q

• Easy to sample from q

• Find k with

• Sample from q

• Sample uniformly from [0,kq(z0)]

• Reject if

But: Rejection sampling is inefficient.

(unnormalized)

Rejection

area

7

Page 8: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

•Idea: assign an importance weight w to each

sample

•With the importance weights, we can account for the

“differences between p and q ”

•p is called target

•q is called proposal (as before)

Importance Sampling

8

Page 9: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

•Explanation: The prob. of falling in an interval A is the area under p

•This is equal to the expectation of the indicator function

Importance Sampling

A

9

Page 10: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

•Explanation: The prob. of falling in an interval A is the area under p

•This is equal to the expectation of the indicator function

Approximation with samples drawn from q:

Importance Sampling

Requirement:

A

10

Page 11: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

•Non-parametric implementation of Bayes filter

•Represents the belief (posterior) by a set of

random state samples.

•This representation is approximate.

•Can represent distributions that are not Gaussian.

•Can model non-linear transformations.

Basic principle:

•Set of state hypotheses (“particles”)

•Survival-of-the-fittest

The Particle Filter

11

Page 12: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Algorithm Bayes_filter :

1. if is a sensor measurement then

2.

3. for all do

4.

5.

6. for all do

7. else if is an action then

8. for all do

9. return

Machine Learning for Computer Vision

The Bayes Filter Algorithm (Rep.)

12

Page 13: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Set of weighted samples:

Mathematical Description

The samples represent the probability distribution:

State hypotheses Importance weights

Point mass distribution (“Dirac” )

13

Page 14: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

The Particle Filter Algorithm

Algorithm Particle_filter :

1.

2. for to do

3.

4.

5.

6. for to do

7. return

Sample from proposal

Compute sample weights

Resampling

14

Page 15: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Localization with Particle Filters

•Each particle is a potential pose of the robot

•Proposal distribution is the motion model of the robot

(prediction step)

•The observation model is used to compute the

importance weight (correction step)

Randomized algorithms are usually called Monte Carlo

algorithms, therefore we call this:

Monte-Carlo Localization

15

Page 16: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

A Simple Example

• The initial belief is a uniform distribution (global localization).

• This is represented by an (approximately) uniform sampling of initial particles.

16

Page 17: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Sensor Information

The sensor model is used to compute the new importance weights:

17

Page 18: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Robot Motion

After resampling and applying the motion model the particles are distributed more densely at three locations.

18

Page 19: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Again, we set the new importance weights equal to the sensor model.

Sensor Information

19

Page 20: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Resampling and application of the motion model:

One location of dense particles is left.

The robot is localized.

Robot Motion

20

Page 21: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

A Closer Look at the Algorithm…

Algorithm Particle_filter :

1.

2. for to do

3.

4.

5.

6. for to do

7. return

Sample from proposal

Compute sample weights

Resampling

21

Page 22: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Sampling from Proposal

This can be done in the following ways:

• Adding the motion vector to each particle directly (this assumes perfect motion)

• Sampling from the motion model , e.g. for a 2D motion with translation velocity v and rotation velocity w we have:

Position

Orientation

22

Page 23: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Start

Motion Model Sampling (Example)

23

Page 24: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Computation of the sample weights:

• Proposal distribution:(we sample from that using the motion model)

• Target distribution (new belief):(we can not directly sample from that → importance sampling)

• Computation of importance weights:

Computation of Importance Weights

24

Page 25: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Proximity Sensor Models

• How can we obtain the sensor model ?

• Sensor Calibration:

Laser sensor Sonar sensor

25

Page 26: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

• Given: Set of weighted samples.

• Wanted : Random sample, where the probability of drawing xi is equal to wi.

• Typically done M times with replacement to generate new sample set .

Resampling

for to do

26

Page 27: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

w2

w3

w1wn

Wn-1

Resampling

w2

w3

w1wn

Wn-1

•Standard n-times sampling results in high variance

•This requires more particles

•O(nlog n) complexity

• Instead: low variance sampling only samples once

• Linear time complexity

• Easy to implement

27

Page 28: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Sample-based Localization (sonar)

28

Page 29: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Initial Distribution

29

Page 30: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

After Ten Ultrasound Scans

30

Page 31: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

After 65 Ultrasound Scans

31

Page 32: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Estimated Path

32

Page 33: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Kidnapped Robot Problem

The approach described so far is able to

• track the pose of a mobile robot and to

• globally localize the robot.

• How can we deal with localization errors (i.e., the kidnapped robot problem)?

Idea: Introduce uniform samples at every resampling step

• This adds new hypotheses

33

Page 34: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Summary

• There are mainly 4 different types of sampling methods: Transformation method, rejections sampling, importance sampling and MCMC

• Transformation only rarely applicable

• Rejection sampling is often very inefficient

• Importance sampling is used in the particle filter which can be used for robot localization

• An efficient implementation of the resampling step is the low variance sampling

34

Page 35: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

Computer Vision Group Prof. Daniel Cremers

Markov Chain Monte Carlo

Page 36: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Markov Chain Monte Carlo

• In high-dimensional spaces, rejection sampling and importance sampling are very inefficient

• An alternative is Markov Chain Monte Carlo (MCMC)

• It keeps a record of the current state and the proposal depends on that state

• Most common algorithms are the Metropolis-Hastings algorithm and Gibbs Sampling

36

Page 37: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Markov Chains Revisited

A Markov Chain is a distribution over discrete-state random variables so that

The graphical model of a Markov chain is this:

We will denote as a row vector

A Markov chain can also be visualized as a state transition diagram.

37

x1, . . . ,xM

p(x1, . . . ,xT ) = p(x1)p(x2 | x1) · · · = p(x1)TY

t=2

p(xt | xt�1)

p(xt | xt�1) ⇡t

T

Page 38: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

The State Transition Diagram

A33 A33

A11 A11k=1

k=2

k=3

time

t-2 t-1 t

38

states

Page 39: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Some Notions

• The Markov chain is said to be homogeneous if the transitions probabilities are all the same at every time step t (here we only consider homogeneous Markov chains)

• The transition matrix is row-stochastic, i.e. all entries are between 0 and 1 and all rows sum up to 1

• Observation: the probabilities of reaching the states can be computed using a vector-matrix multiplication

39

Page 40: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

The Stationary Distribution

The probability to reach state k is

Or, in matrix notation:

We say that is stationary if

Questions:

•How can we know that a stationary distributions exists?

•And if it exists, how do we know that it is unique?

40

⇡t = ⇡t�1A

⇡k,t =KX

i=1

⇡i,t�1Aik

⇡t = ⇡t�1⇡t

Page 41: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

The Stationary Distribution (Existence)

To find a stationary distribution we need to solve the eigenvector problem

The stationary distribution is then where is the eigenvector for which the eigenvalue is 1.

This eigenvector needs to be normalized so that it is a valid distribution.

Theorem (Perron-Frobenius): Every row-stochastic matrix has such an eigen vector, but this vector may not be unique.

41

ATv = v

⇡ = vT v

Page 42: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Stationary Distribution (Uniqueness)

• A Markov chain can have many stationary distributions

• Sufficient for a unique stationary distribution: we can reach every state from any other state in finite steps at non-zero probability (i.e. the chain is ergodic)

• This is equivalent to the property that the transition matrix is irreducible:

42

1 2 3 4

0.9

0.90.5 0.5

1.00.10.1

8i, j 9m (Am)ij > 0

Page 43: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Main Idea of MCMC

• So far, we specified the transition probabilities and analysed the resulting distribution

• This was used, e.g. in HMMs

Now:

• We want to sample from an arbitrary distribution

• To do that, we design the transition probabilities so that the resulting stationary distribution is our desired (target) distribution!

43

Page 44: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Detailed Balance

44

Definition: A transition distribution satisfies the property of detailed balance if

The chain is then said to be reversible.

⇡t

⇡iAij = ⇡jAji

⇡1

⇡3 ⇡1A13 + · · ·

⇡3A31 + · · ·

A31

t-1 t

A13

Page 45: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Making a Distribution Stationary

Theorem: If a Markov chain with transition matrix

A is irreducible and satisfies detailed balance wrt. the distribution , then is a stationary distribution of the chain.

Proof:

it follows .

This is a sufficient, but not necessary condition.

45

⇡ ⇡

KX

i=1

⇡iAij =KX

i=1

⇡jAji = ⇡j

KX

i=1

Aji = ⇡j 8j

⇡ = ⇡A

Page 46: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Sampling with a Markov Chain

The idea of MCMC is to sample state transitions based on a proposal distribution q.

The most widely used algorithm is the Metropolis-Hastings (MH) algorithm.

In MH, the decision whether to stay in a given state is based on a given probability.

If the proposal distribution is , then we stay in state with probability

46

q(x0 | x)x

0

min

✓1,

p̃(x0)q(x | x0)

p̃(x)q(x0 | x)

Unnormalized target distribution

Page 47: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

The Metropolis-Hastings Algorithm

• Initialize

• for

•define

•sample

•compute acceptance probability

•compute

•sample

•set new sample to

47

x

0

s = 0, 1, 2, . . .

x = x

s

x

0 ⇠ q(x0 | x)

↵ =p̃(x0)q(x | x0)

p̃(x)q(x0 | x)r = min(1,↵)

u ⇠ U(0, 1)

x

s+1 =

(x

0 if u < r

x

s if u � r

Page 48: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Why Does This Work?

We have to prove that the transition probability of the MH algorithm satisfies detailed balance wrt the target distribution.

Theorem: If is the transition probability of the MH algorithm, then

Proof:

48

pMH(x0 | x)

p(x)pMH(x0 | x) = p(x0)pMH(x | x0)

Page 49: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Why Does This Work?

We have to prove that the transition probability of the MH algorithm satisfies detailed balance wrt the target distribution.

Theorem: If is the transition probability of the MH algorithm, then

Note: All formulations are valid for discrete and for continuous variables!

49

pMH(x0 | x)

p(x)pMH(x0 | x) = p(x0)pMH(x | x0)

Page 50: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Choosing the Proposal

• A proposal distribution is valid if it gives a non-zero probability of moving to the states that have a non-zero probability in the target.

• A good proposal is the Gaussian, because it has a non-zero probability for all states.

• However: the variance of the Gaussian is important!

•with low variance, the sampler does not explore sufficiently, e.g. it is fixed to a particular mode

•with too high variance, the proposal is rejected too often, the samples are a bad approximation

50

Page 51: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Example

Target is a mixture of 2 1D Gaussians.

Proposal is a Gaussian with different variances.

51

0

200

400

600

800

1000 −100

−50

0

50

100

0

0.1

0.2

Samples

MH with N(0,1.0002) proposal

Iterations

0

200

400

600

800

1000 −100

−50

0

50

100

0

0.01

0.02

0.03

Samples

MH with N(0,8.0002) proposal

Iterations

0

200

400

600

800

1000 −100

−50

0

50

100

0

0.02

0.04

0.06

Samples

MH with N(0,500.0002) proposal

Iterations

Page 52: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Gibbs Sampling

• Initialize

• For

•Sample

•Sample

•...

•Sample

Idea: sample from the full conditional

This can be obtained, e.g. from the Markov blanket in graphical models.

52

{zi : i = 1, . . . ,M}⌧ = 1, . . . , T

z(⌧+1)1 ⇠ p(z1 | z(⌧)2 , . . . , z(⌧)M )

z(⌧+1)2 ⇠ p(z2 | z(⌧+1)

1 , . . . , z(⌧)M )

z(⌧+1)M ⇠ p(zM | z(⌧+1)

1 , . . . , z(⌧+1)M�1 )

Page 53: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Gibbs Sampling: Example

• Use an MRF on a binary image with edge potentials (“Ising model”) and node potentials

53

(xt) = N (yt | xt,�2)

(xs, xt) = exp(Jxsxt)

xt

yt

xs

xt 2 {�1, 1}

Page 54: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Gibbs Sampling: Example

• Use an MRF on a binary image with edge potentials (“Ising model”) and node potentials

• Sample each pixel in turn

54

(xt) = N (yt | xt,�2)

sample 1, Gibbs

−1

−0.5

0

0.5

1sample 5, Gibbs

−1

−0.5

0

0.5

1mean after 15 sweeps of Gibbs

−1

−0.5

0

0.5

1

(xs, xt) = exp(Jxsxt)

After 1 sample After 5 samples Average after 15 samples

Page 55: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Gibbs Sampling is a Special Case of MH

• The proposal distribution in Gibbs sampling is

• This leads to an acceptance rate of:

• Although the acceptance is 100%, Gibbs sampling does not converge faster, as it only updates one variable at a time.

55

q(x0 | x) = p(x0i | x�i)I(x0

�i = x�i)

↵ =p(x0)q(x | x0)

p(x)q(x0 | x) =p(x0

i | x0�i)p(x

0�i)p(xi | x0

�i)

p(xi | x�i)p(x�i)p(x0i | x�i)

= 1

Page 56: 14. Sampling Methods - Computer Vision Group - Home · PD Dr. Rudolph Triebel Computer Vision Group Machine Learning for Computer Vision Summary • There are mainly 4 different types

PD Dr. Rudolph TriebelComputer Vision Group

Machine Learning for Computer Vision

Summary

• Markov Chain Monte Carlo is a family of sampling algorithms that can sample from arbitrary distributions by moving in state space

• Most used methods are the Metropolis-Hastings (MH) and the Gibbs sampling method

• MH uses a proposal distribution and accepts a proposed state randomly

• Gibbs sampling does not use a proposal distribution, but samples from the full conditionals

56