8/10/2019 PF on Riemannian Manifolds
1/23
Chapter 17
Particle Filtering on Riemannian Manifolds.
Application to Covariance Matrices Tracking
Hichem Snoussi
17.1 Introduction
Given a dynamical system characterized by a state-space model, the objective of
the online Bayesian filtering is the estimation of the posterior marginal probability
of the hidden state given all the observations collected until the current time. The
nonlinear and/or the non Gaussian aspect of the prior transition distributions and
the observation model leads to intractable integrals when evaluating the marginals.
Therefore, one has to resort to approximate Monte Carlo schemes. Particle filtering
[1] is such an approximate Monte Carlo method estimating, recursively in time, the
marginal posterior distribution of the continuous hidden state of the system. The
particle filter provides a point mass approximation of these distributions by drawing
particles according to a proposal distribution and then weighting the particles in order
to fit the target distribution.
The particle filter method is usually applied to track a hidden state belonging to an
Euclidean space. The most popular scheme is to sample the particles according to a
random walk around the previous particles. However, in some tracking applications,
the state may be constrained to belong to a Riemannian manifold. Recently, some
works have been dedicated to design algorithms adapted to the Riemannian mani-fold constraints, based on differential geometry tools: Gradient-descent algorithm on
Grassmann manifold for object recognition [2], statistical analysis of diffusion tensor
MRI[3], geodesic-based deconvolution algorithms [4], tracking principal subspaces
[5], algorithms in Stiefel and Grassman manifolds[6,7], statistical analysis on man-
ifolds [810], optimization on matrix manifolds[11, 12], and a general scheme for
tracking fast-varying states on Riemannian manifolds in [13]. This chapter is devoted
to the application of this differential-geometric framework to design efficient target
H. Snoussi(B)Charles Delaunay Institute,
UMR STMR 6279 CNRS, University of Technology of Troyes,
12, rue Marie Curie, 10010, Troyes, France
e-mail: [email protected]
F. Nielsen and R. Bhatia (eds.),Matrix Information Geometry, 427
DOI: 10.1007/978-3-642-30232-9_17, Springer-Verlag Berlin Heidelberg 2013
8/10/2019 PF on Riemannian Manifolds
2/23
428 H. Snoussi
tracking algorithms. We particularly consider the case where the observation noise
covariance is unknown and time-varying. The Bayesian filtering objective is thus to
jointly estimate the hidden target state and the time-varying noise covariance. As
the noise covariance is a positive definite matrix, the Euclidean space is not suitable
when tracking this covariance. Instead, one should exploit the differential geometricproperties of the space of positive definite matrices, by constraining the estimated
matrix to move along the geodesics of this Riemannian manifold. The proposed
sequential Bayesian updating consists thus in drawing state samples while moving
on the manifold geodesics.
The chapter is organized as follows: Sect. 17.2 is a brief introduction to the particle
filtering method on the Euclidean spaces. In Sect. 17.3,we describe some concepts
of differential geometry. In Sect. 17.4,we present a general scheme for the particle
filtering method on a Riemannian manifold. Section 17.5is dedicated to the design
of a particle filter jointly tracking a target state belonging to an Euclidean space anda time-varying noise covariance modeling the evolution over time of the sensing
system imperfections.
17.2 Bayesian Filtering on Euclidean Spaces
In this section, we briefly recall the particle filter method for filtering in nonlinear
dynamical systems characterized in Euclidean spaces. It is an approximate MonteCarlo method estimating, recursively in time, the marginal posterior distribution of
the continuous hidden state of the system, given the observations. The particle filter
provides a point mass approximation of these distributions. For more details and a
comprehensive review of the particle filter see[1,14,15].
The observed system evolves in time according to the following nonlinear dynam-
ics: xt px(xt | xt1, ut)
yt py (yt | xt, ut), (17.1)
where yt Rny denotes the observation at time t, xt Rnx denotes the unknown
continuous state, and ut Udenotes a known control signal. The probability dis-
tribution px(xt | xt1, ut)models the stochastic transition dynamics of the hidden
state. Given the continuous state, the observations yt follow a stochastic model
py (yt | xt, ut), where the stochastic aspect reflects the observation noise.
The Bayesian filtering is based on the estimation of the posterior marginal prob-
ability p(xt | y1:t). The nonlinear and the non Gaussian aspect of the transition
distributions leads to intractable integrals when evaluating the marginals. Therefore,
one has to resort to Monte Carlo approximation where the joint posterior distribution
p(x0:t | y1:t) is approximated by the point-mass distribution of a set of weightedsamples (called particles){x
(i )0:t, w
(i )t }
Ni =1:
8/10/2019 PF on Riemannian Manifolds
3/23
17 Particle Filtering on Riemannian Manifolds 429
pN(x0:t | y1:t)=
Ni =1
w(i )t x(i )0:t
(dx 0:t),
wherex(i )0:t(dx 0:t)denotes the Dirac function, and dx 0:tis the Lebesgue measure.Based on the same set of particles, the marginal posterior probability (of interest)
p(xt | y1:t)can also be approximated as follows:
pN(xt | y1:t)=
Ni =1
w(i )t x(i )t
(dxt),
Backward estimation of the marginal state probability is also possible given the
particles{x(i )
0:t+t , w
(i )
t+t }N
i =1
:
p(xt | y1:t+t )
Ni =1
w(i )t+tx(i )t
(dx t),
In the Bayesian importance sampling (IS) method, the particles {x(i )0:t}
Ni =1 are
sampled according to a proposal distribution(x0:t | y1:t)and {w(i )t }are the corre-
sponding normalized importance weights:
w(i )t
p(y1:t | x(i )0:t)p(x
(i )0:t)
(x(i )0:t | y1:t)
.
17.2.1 Sequential Monte Carlo
Sequential Monte Carlo (SMC) consists of propagating the trajectories {x(i )0:t}
Ni =1 in
time without modifying the past simulated particles. This is possible for the class ofproposal distributions having the following form:
(x0:t | y1:t)= (x0:t1 | y1:t1)(xt | x0:t1, y1:t).
The normalized importance weights are then recursively computed in time as:
w(i )t w
(i )t1
p(yt | x(i )t )p(x
(i )t | x
(i )0:t1)
(x(i )t | x
(i )0:t1, y1:t)
. (17.2)
For sake of clarity, one can adopt the transition prior as the proposal distribution:
(x(i )t | x
(i )0:t1, y1:t)= px(xt | xt1, ut)
8/10/2019 PF on Riemannian Manifolds
4/23
430 H. Snoussi
in which case the weights are updated according to the likelihood function:
w(i )t w
(i )t1p(yt | x
(i )t ).
The particle filter algorithm (depicted in pseudo-code Algorithm 17.1) consists of2 steps: the sequential importance sampling step and the selection step. The selection
(resampling) step replaces the weighted particles by unweighted particles in order to
avoid the collapse of the Monte Carlo approximation caused by the variance increase
of the weights. It consists of selecting the trajectories {x(i )0:t}with probabilitiesw
(i )t .
The trajectories with weak weights are eliminated and the trajectories with strong
weights are multiplied. After the selection step, all the weights are equal to 1 /N.
Algorithm 17.1Particle filter algorithm on an Euclidean space1: function PF(PP)
2: Initializationx(i )0 p0(x)
3: for t=1 toT do(Sequential importance sampling)
4: fori =1, ...,N do(sample from the transition prior)
5: x(i )t px(xt | x
(i )t1, ut)
6: set( x(i )0:t)= ( x
(i )t ,x
(i )0:t1)
7: end for
8: Update the importance weights
9: for i =1, ...,N do(evaluate and normalize the weights)
10: w
(i )
t p(yt | x(i)
t )11: end for
12: Resampling:
13: Select with replacement from { x(i)0:t}
Ni =1 with probability {w
(i)t } to obtain N particles
{x(i )0:t}
Ni =1
14: end for
15: end function
17.3 Differential Geometry Tools
In order to have a self-content framework, we devote this section to the introduction of
some differentiable geometry tools related to the concept of Riemannian manifolds.
These tools are necessary to the design of the particle filter on Riemannian manifolds.
For further details on Riemannian geometry, refer to[16].
First, we need to define a topological manifold as follows:
Definition 17.1 A manifoldM of dimension n, or n-manifold, is a topological space
with the following properties:(i) M is Haussdorff,
(ii) M is locally Euclidean of dimensionn , and
(iii) M has a countable basis of open sets.
8/10/2019 PF on Riemannian Manifolds
5/23
17 Particle Filtering on Riemannian Manifolds 431
Fig. 17.1 Topological mani-
fold
U
2
M
1
Intuitively, a topological manifold is a set of points which can be considered locally
as a flat Euclidean space. In other words, each point p M has a neighborhood
Uhomeomorphic to an n-ball in Rn . Let be such an homeomorphism. The pair
(U,)is called a coordinate neighborhood: to p Uwe assign the n coordinates
1(p), 2(p), ..., n(p) of its image (p) in Rn (see Fig. 17.1). If p lies also in a
second neighborhood V, let(p)= [1(p), 2(p), ...,n (p)]be its correspondent
coordinate system. The transformation 1 on Rn given by:
1 : [1, ...,n] [1, ...,n ],
defines a local coordinate transformation on Rn from = [i ]to = [i ].In differential geometry, one is interested in intrinsic geometric properties which
are invariant with respect to the choice of the coordinate system. This can be
achieved by imposing smooth transformations between local coordinate systems
(see Fig. 17.2). The following definition of differentiable manifold formalizes this
concept in a global setting:
Definition 17.2 A differentiable (or smooth) manifoldM is a topological manifold
with a family U = {U,}of coordinate neighborhoods such that:
(1) theU cover M,(2) for any ,, if the neighborhoods intersection U U is non empty, then
1 and
1 are diffeomorphisms of the open sets(U U)and
(U U)ofRn ,
(3) any coordinate neighborhood (V,) meeting the property (2) with every
U, Uis itself in U
Tangent spaces. On a differentiable manifold, an important notion (in the sequel)
is the tangent space. The tangent space Tp(M) at a point p of the manifold M
is the vector space of the tangent vectors to the curves passing by the point p. It
is intuitively the vector space obtained by a local linearization around the point p.
More formally, let f : M R be a differentiable function on the manifold M
and : I M a curve on M, thedirectional derivativeof f along the curve
is written:
8/10/2019 PF on Riemannian Manifolds
6/23
432 H. Snoussi
1
1
U
2
2
M
U
1
Fig. 17.2 Differentiable manifold
Fig. 17.3 Tangent space on
the manifold
e1
M
2
1p
e2
d
dtf(t)=
fi
di
dt=
didt
i
f
where the derivative operator ei = i
can be considered as a vector belonging to
the tangent space at the point p. The tangent space is then the vector space spanned
by the differential operators i p:Tp(M)=
ci
i
p
| [c1, ..., cn ] Rn
,
where the differential operator ( i
)pcan be seen geometrically as the tangent vector
to the i th coordinate curve (fixing all j coordinates j =i and varying only the value
ofi ), see Fig. 17.3.
Vector fields and tensor fields. A vector field Xis an application M p Tp,
which assigns a tangent vector to each point of the manifold:
X : p S Xp Tp
8/10/2019 PF on Riemannian Manifolds
7/23
17 Particle Filtering on Riemannian Manifolds 433
Fig. 17.4 Vector field on the
manifold
M
p
Xp = Xpii
The vector fieldXcan bedefined by its n component functions {Xi }n
i =1 (see Fig. 17.4).X isC (smooth) if and only if all its scalar components (Xi )are C
.
A tensor field A of type [q, r] is an application which maps a point p Mto
some multilinear mapping Ap fromTr
p toTq
p:
A : p ApAp : Tp Tp
r direct products
Tp Tp
q direct products
The types [0, r] and [1, r] are respectively called tensor fields of covariant degreerand tensor fields of contravariant degree 1 and covariant degree r. For example, a
scalar product is a tensor field of type[0, 2]:
Tp Tp R
(Xp, Yp)< Xp, Yp > .
Riemannian metric. For each point p in M, assume that an inner product p is defined on the tangent space Tp(M). Thus, a mapping from the points of
the differentiable manifold to their inner product (bilinear form) is defined. If thismapping is smooth, then the pair (M, < , >p)is calledRiemannian manifold(see
Fig. 17.5). The Riemannian metric is thus a tensor field g which is, according to a
coordinate system{}, defined by the positive definite matrices G p:
Gi j (p)=p
On a manifoldM, an infinite number of Riemannian metrics may be defined. The
metric does not thus represent an intrinsic geometric property of the manifold.
Consider now a curve: [a, b] (S, g), its length||||is defined as:
8/10/2019 PF on Riemannian Manifolds
8/23
434 H. Snoussi
M
p q
Fig. 17.5 Riemannian metric
|||| =
ba
||d
dt||dt=
ba
gi ji j dt
Affine connections. An affine connection is an infinitesimal linear relation p,p
between the tangent spaces of two neighboring points p and p (see Fig. 17.6). It
can be defined by its n 3 connection coefficients ki j (with respect to the coordinate
system[
i
]) as follows:
p,p ((j )p)= (j )p di (ki j )p(k)p
Let p andq be two points on M anda curve linking p and q. If the tangent
vectorsX(t)meet the following relation along the curve:
X(t+d t)= (t),(t+dt)(X(t)),
then, X isparallelalongand is aparallel translationon (see Fig. 17.7).
The covariant derivative of a vector field X along a curve is defined as the
infinitesimal variation betweenX(t) and the parallel translation ofX(t+h) T(t+h)to the space T(t) along . The parallel translation is in fact necessary in order to
consider the limit of the difference of two vectors belonging to the same vector
space. The vectors X(t) and X(t+d t) belong to different tangent spaces and the
quantity d X(t)= X(t+dt)X(t) may not be defined (see Fig. 17.8). The covariant
derivative D Xdt
forms then a vector field along the curveand can be expressed as a
function of the connection coefficients as follows:
D Xdt = ((t+dt),(t)(X(t+ dt)) X(t))/dt= {Xk(t) + i (t)Xj (t)(ki j )(t)}(k)(t)
(17.3)
8/10/2019 PF on Riemannian Manifolds
9/23
17 Particle Filtering on Riemannian Manifolds 435
M
p
j
Tp
p
j
j
j
p,pdikijk
Tp
Fig. 17.6 Affine connections
(b)
(a)
(t)(t+dt)
X(a)
X(t) X(t+dt)
X(b)
(t) ,(t+dt)
Fig. 17.7 Parallel translation
The expression (17.3) of the covariant derivative along a curvecan be extended
to define the directional derivative along a tangent vectorDby considering the curve
whose tangent vector is D. The directional derivative, denoted by DX has the
following expression:
DX = Di {(iX
k)p + Xjp(
ki j )p}(k)p.
The covariant derivative along the curvecan then be written as:
8/10/2019 PF on Riemannian Manifolds
10/23
436 H. Snoussi
,p
X(t+dt)
(t+dt)
(t)
X(t)
X(t)
Tp
p
Fig. 17.8 Covariant derivative along a curve
D X(t)
dt= (t)X. (17.4)
Consider now two vector fields X and Y on the manifold M. The covariant
derivativeXY Tp(M) ofYwith respect to Xcan be defined by the following
expression:
XY = Xi {i Y
k +Y j ki j }k. (17.5)
The expression(17.5) of the covariant derivative can be used as a characterization
of the connection coefficients ki j . In fact, takingX =i and Y =j , the connection
coefficients are characterized as follows:
ij =ki jk.
A differentiable manifoldM is said to be flat if and only if there exists a coordinate
system[i ]such that the connection coefficients {ki j }are identically 0. This means
that all the coordinate vector fieldsi are parallel along any curveon M.
Riemannian connection. A Riemannian connection is an affine connection defined on a Riemannian manifold(M, g =)such thatX, Y,Z T(S), the
following property holds:
8/10/2019 PF on Riemannian Manifolds
11/23
17 Particle Filtering on Riemannian Manifolds 437
Z < X, Y >=+ < X, ZY > (17.6)
where the left hand side of the equation means the differential operator Zapplied to
the scalar function< X, Y >on the manifold.
Letbe a curve on the manifold M and D Xdt et DYdt the covariant derivatives ofX and Y along respectively. According to the expression (17.4) of the covariant
derivative and the fact that the differential operator (t) consists of deriving with
respect tot, one has the following interesting identity concerning the variation of the
scalar product on the manifold with a Riemannian connection:
d
dt< X(t), Y(t) >=+ < X(t),
DY(t)
dt>
The above equation means that the scalar product is conserved under a paralleltranslation (D X(t)dt
= DY(t)dt
=0):
< (X), (Y) >=< X, Y > .
A particular example is the Euclidean space which is a flat manifold characterized
by a Riemannian connection.
Geodesics. A geodesic between two endpoints (a) and(b) on a Riemannian
manifold (M, g, ) is a curve : [a, b] M which is locally defined as the
shortest curve on the manifold connecting these endpoints. More formally, the defi-
nition of a geodesic is given by:
Definition 17.3 The parametrized curve(t)is said to be a geodesic if its velocity
(tangent vector) d/dtis constant (parallel) along, that is if it satisfies the condition
(D/dt)(d/dt)= 0, fora
8/10/2019 PF on Riemannian Manifolds
12/23
438 H. Snoussi
Ep(X)
M
Tp(M)
p
X
Fig. 17.9 Exponential mapping on the manifold
D(p, q)= ||g|| =
ba
gi ji j dt. (17.8)
The geodesic distance can also be defined as the shortest distance (over smooth
curves) between two points on the manifold endowed by a Riemannian connection.Exponential mapping. The exponential mapping is a central concept when
designing filtering methods on Riemannian manifolds. In fact, it represents an inter-
esting tool to build a bridge between an Euclidean space and the Riemannian man-
ifold. For a point p and a tangent vector X Tp(M), let : t = (t) be the
geodesic such that (0) = p and ddt
(0) = X. The exponential mapping of X is
defined as Ep(X) = (1). In other words, the exponential mapping assigns to the
tangent vector Xthe endpoint of the geodesic whose velocity at time t = 0 is the
vector X(see Fig. 17.9). It can be shown that there exist an neighborhood Uof 0 in
Tp(M)and a neighborhoodV ofpin M such that Ep |Uis a diffeomorphism fromU toV. Also, note that since the velocityd/dtis constant along the geodesic(t),
its length L from pto Ep(X)is:
L =
10
d
dtdt=
10
Xdt = X.
The exponential mappingEp(X) corresponds thus to the unique point on the geodesic
whose distance from pis the length of the vectorX.
8/10/2019 PF on Riemannian Manifolds
13/23
17 Particle Filtering on Riemannian Manifolds 439
T(M) vt
xt=Ep(vt)
xt+1
xt1
M
vt+1
vt+2
Fig. 17.10 Markov chain on a Riemannian manifold
17.4 Particle Filtering on Riemannian Manifolds
17.4.1 General Scheme
The aim of this section is to propose a general scheme for the extension of the particle
filtering method on a Riemannian manifold. The hidden state x is constrained to lie
in a Riemannian manifold (M,g, ) endowed with a Riemannian metric g and
an affine connection . The system evolves according to the following nonlinear
dynamics: xt px(xt | xt1, ut), x M
yt py (yt | xt, ut), (17.9)
where the Markov chain (random walk) px(xt | xt1, ut) on the manifold M is
defined according to the following generating mechanism:
1. Draw a sample vton the tangent space Txt1M according to a pdf pv(.).
2. x is obtained by the exponential mapping of vt according to the affine
connection.
In other words, a random vector vt is drawn on the tangent space Txt1M
by the usual Euclidean random technics. Then, the exponential mapping allows
the transformation of this vector to a point xt on the Riemannian manifold. The
point xt is the endpoint of the geodesic starting from xt1 with a random initial
velocity vector vt. Figure 17.10illustrates the transition dynamics on a Riemannian
manifold M.As a generating stochastic mechanism is defined on the tangent space, the particle
filtering is naturally extended by means of the exponential mapping. It simply consists
in propagating the trajectories on the manifold by the random walk process, weighting
8/10/2019 PF on Riemannian Manifolds
14/23
440 H. Snoussi
the particles by the likelihood function and sampling with replacement. The proposed
general scheme is depicted in the pseudo-code Algorithm 17.2.
Algorithm 17.2Particle filter algorithm on a Riemannian manifold M1: function PF(PP)
2: Initializationx(i )0 p0(x)
3: for t=1 toT do(Sequential importance sampling)4: fori =1, ...,N do(sample from the random walk on M)
5: v(i )t pv (v) on Txt1M
6: x(i )t = Ex(i)t1
(v(i )t )
7: set( x(i )0:t)= ( x
(i )t ,x
(i )0:t1)
8: end for
9: Update the importance weights
10: for i =1, ...,N do(evaluate and normalize the weights)11: w
(i )t p(yt | x
(i)t )
12: end for
13: Resampling:
14: Select with replacement from { x(i)0:t}
Ni =1 with probability {w
(i)t } to obtain N particles
x(i )0:t}
Ni=1
15: end for
16: end function
17.4.2 Point Estimates
Based on particle trajectories{ x(i )0:T}, classical particle filtering algorithm provides a
simple way to approximate point estimates. In fact, any quantity of interesth(x)can
be estimated by its a posteriori expectation, minimizing the expected mean square
error. The empirical mean of the transformed particles h(x(i )t ) represents an unbiased
Mont-Carlo estimation of the a posteriori expectation. Averaging in the manifoldcontext is no more a valid operation: The empirical mean could be located outside
the manifold or the averaging itself does not have a meaning in the absence of a
summation operator on the manifold. In order to obtain a valid point estimate, one
should rather minimize the mean square error, where the error is evaluated by the
geodesic distance D on the manifold (17.8). Following the work of Frchet [17],
the point estimate can be defined by the intrinsic mean (also called Riemannian
barycenter). The intrinsic mean has the following expression:
xt =arg minxtME(D(xt,st))2 (17.10)=arg minxtM
(D(xt,st))
2p(st | y1..T)dst
8/10/2019 PF on Riemannian Manifolds
15/23
17 Particle Filtering on Riemannian Manifolds 441
where the expectation operator is computed with respect to the a posteriori probability
density p(st | y1..T)and a Riemannian measuredinduced by the Riemannian met-
ric.
Remark 17.2 Note that the above estimator (17.10) does not yield the a poste-riori expectation in general. This is only the case when the manifold is flat with
respect to a Riemannian connection (Euclidean space). In fact, assume that[i ]
is a coordinate system corresponding to null Christoffel coefficients (ki j = 0).
A geodesic between 2 points p and q is then a straight line with coordinates
(p) + t((q) (p)). The geodesic distance D(p, q) (17.8) simplifies to a
quadratic distance
(i (p) i (q))2 leading to the usual Euclidean estimator
xt = E[xt | y1..T].
Computation of the point estimate (17.10) involves an integration operation (with
respect to st M and according to the posterior distribution) and a constrainedoptimization operation on the manifold M. The integral can be approximated (as in
the Euclidean case) by an empirical weighted sum of the geodesic distances applied
on the particles which are yielded within their weights by the particle filter Algorithm
17.2. The point estimate is then computed by the minimization of an approximated
expectation expression:
xt =arg min
xtM
N
i =1 w(i )t (D(xt,s
(i )t ))
2 (17.11)
where s(i )t andw
(i )t are the particles and their weights computed recursively by the
particle filter algorithm.
Concerning the constrained optimization in (17.11), more specific differential
geometric considerations should be taken into account. First, existence and unique-
ness of the intrinsic mean is shown by Karcher [18] when the manifold M has a
non positive sectional curvature. Second, a gradient descent like algorithm can be
designed on the manifold based on the exponential mapping which plays again a key
role in transferring Euclidean technics to a Riemannian manifold context. Denotingby J(xt)the objective function to be minimized with respect to xt,
J(xt)=
Ni =1
w(i )t (D(xt,s
(i )t ))
2,
a gradient flowx(l)t , starting from an initial guessx
(l)t and converging to the solution
xt, can be defined by moving in the direction of the opposite of the objective function
derivativeJ(x
(l)
t ). As the function derivative J(x
(l)
t ) lies in the tangent spaceTx
(l)t
(M), the exponential mapping can be used to map the opposite derivative vector
tothenextpointx(l+1)t . The gradient-like descent algorithm is then derived as follows:
8/10/2019 PF on Riemannian Manifolds
16/23
442 H. Snoussi
Tx(l)t
(M)
M
x(l+1)t =E
x(l)t
(J)J x(l)t
Fig. 17.11 A gradient descent step on a Riemannian manifold
x(l+1)t =Ex(l)t
(J(x(l)t )) (17.12)
Figure 17.11illustrates an iteration of the gradient descent algorithm on a Rie-
mannian manifold.
Remark 17.3 The implementation of the proposed particle filter algorithm on the
manifoldM and the computation of the point estimates require an explicit expressionof the exponential mapping. In other words, solving geodesic differential Eq. (17.7)
conditions the feasibility of the manifold version of the particle filter algorithm.
Explicit analytic expressions are only available in some cases. Among these cases,
one can find:
Flat manifolds (with constant null connections ) where geodesics are straight
lines.
The set of Gaussian probability densities with fixed mean: This example will play
a central role in the next Sect. 17.5devoted to the target tracking application.
17.5 Application to Tracking with Unknown Time-Varying State
Covariance
This section is devoted to the application of the above developed framework to design
an efficient target tracking algorithm. The target state is assumed to be observed
according to the general system (17.9), where the observation equation is assumed
to have a general form depending on the sensing model. Concerning the transition
dynamics px(xt | xt1), we adopt a mean-scale mixture model. According to this
model, the hidden statext Rnx (belonging to an Euclidean space) has a Gaussian
distribution with a random mean tand a random covariance matrix t. The mean
8/10/2019 PF on Riemannian Manifolds
17/23
17 Particle Filtering on Riemannian Manifolds 443
follows a Gaussian random walk reflecting the time correlation of the system tra-
jectory and the covariance matrix follows a Generalized Gaussian random walk
on the Riemannian manifold of positive definite matrices S+. The transition prior is
thus defined by an augmented Markov state(t,t,x t)as follows:t N(t |t1, )
t GN(t | t1,)
xt N(xt |t,t)
(17.13)
Contrary to the usual assumption of a constant known covariance, the case of
a stochastic varying state covariance represents an elegant way to deal with abrupt
changes in state trajectories. In fact, this parametric family is able to describe the fat
tails and the skewness of the regression modelpx(xt | xt1, ut). In fact, the resulting
regression model can be written in an integral form:
px(xt | xt1, ut)=
p(t,t | xt1)N(xt; t,t)dtdt
showing that the regression model is an infinite continuous mixture of Gaussian mod-
els. This model can be considered as the extension of the scalar Generalized Hyper-
bolic distributions introduced by Barndorff-Nilesen [19] to a multivariate regression
context.
Before defining the Generalized Gaussian random walkGNof covariance matri-
ces and the particle filter algorithm, we introduce hereafter the expressions of the
Riemannian metric and the Riemannian connections in the special case of the Rie-
mannian manifold S+. Closed forms for the geodesic curves and geodesic distances
are also obtained, providing an efficient implementation of the general particle filter
scheme and the point estimate computation proposed in the previous Sect. 17.4.
17.5.1 Space of Positive Definite Matrices
A positive definite matrix can represent the covariance matrix of a zero-mean mul-
tivariate normal distributionN(0,). Therefore, the set of positive definite matrices
S+can be identified with the statistical model of multivariate Gaussian distributions.
Analyzing the geometric structure of parametric statistical models based on differ-
ential geometric tools was first suggested by Rao in 1945[20]. Since then, many
works have been dedicated to the analysis of statistical manifolds in the light of dif-
ferential geometry. In particular, the seminal work of Amari[21] led to a significant
advancement in this new field.
In the work of Rao [20], it was suggested that the Fisher Information matrixprovides a natural Riemannian metric on statistical manifolds. Denoting by p(.| )a point of a statistical parametric manifold, the Fisher metric is expressed as:
8/10/2019 PF on Riemannian Manifolds
18/23
444 H. Snoussi
gi j ()= E
log p(x |)
i
log p(x |)
j
=
X
log p(x |)
i
log p(x |)
jp(x |)d(x)
Assuming that the connection is symmetric and torsion free, the Riemannian con-
nection can be uniquely defined given the Riemannian metric. It is determined by
the following expression of its Christoffel symbols of the second kind:
ki j = gkm i j m =
1
2gkm (
gj m
i+gi m
jgi j
k),
whereg km are the elements ofg 1 (inverse Fisher matrix).
Applying the above expressions to the case of the manifoldS+ = {N(0;), >
0}of multivariate normal distributions, a closed form for the Riemannian metric is
derived as follows:
g(Bi,j ,Bk,l )=< Bi,j ,Bk,l >= 1
2trace{1Bi,j
1Bk,l }, i, j, k, l =1..nx
(17.14)
where{Bi,j }is the canonical basis of the tangent space TS+which coincides with
the space of real symmetric matrices S. The matrix Bi,j is expressed as follows:
Bi,j = 1ii , i = j
1i j
+1j i
, i = j
with 1i j standing for the matrix with 1 at rowi and column j and 0 elsewhere.
A closed form for the geodesic curve is derived in [22] for the general case of
multivariate normal distributions. Here, we only consider the case of multivariate
normal distributions with a fixed zero-mean. Denoting by (r s ), r s, s =1..nx the
components of the positive definite matrix and using the explicit expressions of the
Riemannian metric and connections, the Euler-Lagrange equations (17.7) reduces to
the following matrix-form differential equation:
d2(t)
dt2
d(t)
dt
1(t)d(t)
dt=0,
leading to the following closed form of a geodesic (t)starting from (0)with an
initial velocity (0):
(t)= (0)1/2 exp [tx](0)1/2, (17.15)
where the matrix x T(0)S+ is defined as,
(0)= (0)1/2x(0)1/2,
and exp [.] stands for the usual matrix exponentiation.
8/10/2019 PF on Riemannian Manifolds
19/23
17 Particle Filtering on Riemannian Manifolds 445
The geodesic distance between two matrices1and 2 D (length of the geodesic
curve between 1 and 2) has an explicit analytic form as well:
D(1,2)= Ln(1/21 2
1/21 )
=1
2
nxi =1log
2(i )
where Ln(.)is the matrix logarithm defined by Ln(exp[A])= Afor ASandiare the eigenvalues of the matrix
1/21 2
1/21 .
17.5.2 Particle Filter Implementation
Given the explicit analytic expression of the geodesic curve(17.15)for the space S+of positive definite matrices, the generating mechanism of a Gaussian random walk
GN(t | t1,)(as defined in Sect. 17.4), has the following steps:
1. Sample a Gaussian symmetric velocity matrix B S with a precision
( nx(nx+1)2
nx(nx+1)2
matrix): B N(0;)
2. The next matrix tis then obtained by:
t =Et1 (B)= 1/2t1exp
1/2t1 B
1/2t1
1/2t1
The proposed particle filter jointly estimate the hidden target positionx t(belong-
ing to an Euclidean space) and its covariance t (belonging to the Riemannian
manifold S+) as follows (see Algorithm 17.3):
1. Propagate the trajectories ((i )0:t1,
(i )0:t1,x
(i )0:t1) by generating the samples
((i )t ,
(i )t ,x
(i )t )according to the prior model (17.13).
2. Update the importance weights which are proportional to the likelihood function.
The explicit solution of the geodesic distance allows also the implementation of
the intrinsic mean for the tracking of the covariance matrix. In fact, approximatingthe expected error by the empirical weighted sum of geodesic distances, the point
estimate is defined as follows:
t = arg minS+
Ni =1
w(i )t (D(,
(i )t ))
2
= arg minS+
Ni =1
w(i )t
1
2trL n2(1/2
(i )t
1/2)
The gradient of the objective function, belonging to the tangent space TS+, has
the following expression:
8/10/2019 PF on Riemannian Manifolds
20/23
446 H. Snoussi
J()=
N
Ni =1
Ln(((i )t )
1) (17.16)
Given the explicit expression of both the gradient (17.16) and the exponentialmapping (17.15) on the manifoldS+, the gradient-descent algorithm scheme defined
in Sect. 17.4for the computation oft is efficiently implemented. Given an initial
guess (0)
, a gradient flow (l)
evolving towards the solution is defined as follows:
(l+1)
=E
(l) (J((l)
))
= ((l)
)1/2 exp
((l)
)1/2J((l)
)1/2
((l)
)1/2
Algorithm 17.3Particle filter algorithm for tracking
1: function PF(PP)
2: Initialization((i )0 ,
(i)0 ,x
(i )0 ) p0
3: for t=1 toT do(Sequential importance sampling)4: fori =1, ...,N do(sample from the random walk prior)
5: (i )t px(t |
(i )t1)
6: B(i )t N(0;) on T(i )t1
S+
7: (i )t = E(i )t1
(B(i )t )
8: x(i )t N((i)t ; (i)t )
9: set (i)t = (
(i )t ,
(i )t ,x
(i )t ) (Augmented state)
10: set ((i )0:t) = (
(i )t ,
(i )0:t1)
11: end for
12: Update the importance weights
13: for i =1, ...,N do(evaluate and normalize the weights)
14: w(i )t p(yt | x
(i)t )
15: end for
16: Resampling:
17: Select with replacement from{(i )0:t}
Ni =1 with probability{w
(i )t }
18: end for
19: end function
17.5.3 Simulation Results
We consider the tracking of a target moving over a 2-D field. The statex t = [xpt ,x
vt]
is formed by the position and the velocity of the target. For simplicity, we assume a
kinematic parametric model for the transition dynamics of the hidden state:
8/10/2019 PF on Riemannian Manifolds
21/23
17 Particle Filtering on Riemannian Manifolds 447
40 45 50 55 60 650
10
20
30
40
50
60
70
80
90True Trajectoy
Estimated trajectory
Rangebearing sensor
Fig. 17.12 Target tracking with unknown noise covariance. Note that only the selected nodes are
plotted in the figure
x
pt
xvt
=
1 0 Ts 0
0 1 0 Ts0 0 1 0
0 0 0 1
xpt1
xvt1
+
T2s /2 0
0 T2s /2
Ts 0
0 Ts
ut
where the sampling interval is Ts = 0.1s and ut is a zero-mean white Gaussian
noise.
The observations are obtained through a network of 400 range-bearing sensors
deployed randomly in the field under surveillance. At each time t, a selected node
(according to the proximity to the target) obtains an observation of the target position
through a range-bearing model:
yrtyt= psm xt+0.5arctan s2x2
s1x1
+ tvtwheresm = (s1, s2)and xt = (x1,x2)are the node and the target positions at time
t, p(set to 10) is the energy emitted by the target (measured at a reference distance
of 1 meter) and vtis a white Gaussian random vector. The corrupting noise has a
covariance tevolving in time as depicted in Fig. 17.13: constant for the first T/4s,
increasing with a linear slope for T/2s and constant for the lastT/4s.
The particle filter is applied to jointly estimate the target position and the noise
covariance matrix. Figure 17.12 illustrates the target tracking performances. Thetrajectory of the target is recovered with a mean square error of 0.39 m. Figure17.13
illustrates the performance of the algorithm to online track the covariance variation
over time. Note that, despite their fluctuation, the estimated covariance elements
8/10/2019 PF on Riemannian Manifolds
22/23
448 H. Snoussi
50 100 150 200 250 300 350 400 450 5000.2
0
0.2
0.4
0.6
0.8
1
1.2
1.4 True valueEstimated value
11
50 100 150 200 250 300 350 400 450 5000.2
0
0.2
0.4
0.6
0.8
1
1.2
1.4 True valueEstimated value
12
50 100 150 200 250 300 350 400 450 5000.2
0
0.2
0.4
0.6
0.8
1
1.2
1.4
Time (s)
True valueEstimated value
21
50 100 150 200 250 300 350 400 450 5000.2
0
0.2
0.4
0.6
0.8
1
1.2
1.4
Time (s)
Estimated valueTrue value
22
Fig. 17.13 Online estimation of the noise covariance elements
follow the tendency of the true covariance elements. The fluctuation of the estimated
noise covariance is mainly due to the fact that the data are less informative with
respect to the covariance matrix. In fact, unlike the target position estimation, the
online estimation of the covariance t is an ill-posed problem based on only one
observation yt. The success of the algorithm to approximately recover the tendency
of the covariance matrix is due to the Markov prior regularization defined by the
Generalized Gaussian random walkGN
(
t |
t1,
) defined in the previoussubsection.
17.6 Conclusion
A differential-geometric framework is proposed to implement the particle filtering
algorithm on Riemannian manifold. The exponential mapping plays a central role
in connecting the manifold-valued particles to the samples generated on the tangent
space by the usual random generating techniques on Euclidean spaces. The proposed
algorithm has been applied to jointly track the target position with the time-varying
noise covariance.
8/10/2019 PF on Riemannian Manifolds
23/23
17 Particle Filtering on Riemannian Manifolds 449
References
1. Doucet, A., Godsill, S., Andrieu, C.: On sequential Monte Carlo sampling methods for Bayesian
filtering. Stat. Comput.10(3), 197208 (2000)
2. Liu, X., Srivastava, A., Gallivan, K.: Optimal linear representations of images for object recog-nition. IEEE Pattern Anal. Mach. Intell. 25(5), 662666 (2004)
3. Lenglet, C., Rousson, M., Deriche, R., Faugeras, O.: Statistics on the manifold of multivari-
ate normal distributions: theory and application to diffusion tensor MRI processing. J. Math.
Imaging Vis.25, 423444 (2006)
4. Fiori, S.: Geodesic-based and projection-based neural blind deconvolution algorithms. Signal
process.88, 521538 (2008)
5. Srivastava, A., Klassen, E.: Bayesian and geometric subspace tracking. Adv. Appl. Probab.
36(1), 4356 (2004)
6. Edelman, A., Arias, T.A., Smith, S.T.: The geometry of algorithms with orthogonality con-
straints. SIAM J. Matrix Anal. Appl.20(2), 303353 (1998)
7. Manton, J.H.: Optimization algorithms exploiting unitary constraints. IEEE Trans. SignalProcess.50(3), 63565 (2002)
8. Bhattacharya, R., Patrangenaru, V.: Large sample theory of intrinsic and extrinsic sample means
on manifolds, I. Ann. Stat. 31(1), 129 (2003)
9. Bhattacharya, R., Patrangenaru, V.: Large sample theory of intrinsic and extrinsic sample means
on manifolds, II. Ann. Stat. 33(3), 12251259 (2005)
10. Pennec, X.: Intrinsic statistics on Riemannian manifolds: basic tools for geometric measure-
ments. J. Math. Imaging Vis.25(1), 127154 (2006)
11. Smith, S.T.: Optimization techniques on Riemannian manifolds, Hamiltonian and gradient
flows, algorithms and control. In: Bloch, A. (ed.) Fields Institute Communication, vol. 3.
American Mathematical Society: Providence, RI, pp. 113136 (1994)
12. Absil, P.-A., Mahony, R., Sepulchre, R.: Optimization Algorithms on Matrix Manifolds. Prince-ton University Press, New Jersey (2008)
13. Snoussi, H., Mohammad-Djafari, A.: Particle Filtering on Riemannian Manifolds. In:
Mohammad-Djafari, A. (ed.) Bayesian Inference and Maximum Entropy Methods, MaxEnt
Workshops, July 2006, pp. 219226, American Institute Physics
14. Doucet, A., de Freitas, N., Gordon, N.J.: Sequential Monte Carlo Methods in Practice. Springer,
Berlin (2001)
15. Andrieu, C., Doucet, A., Singh, S., Tadic, V.: Particle methods for change detection, system
identification, and control. Proc. IEEE92(3), 423438 (March 2004)
16. Boothby, W.M.: An Introduction to Differential Manifolds and Riemannian Geometry. Acad-
emic Press, Orlando (1986)
17. Frchet, M.: Leslments alatoires de nature quelconque dans un espace distanci. Ann. Inst.H, Poincar10, 215310 (1948)
18. Karcher, H.: Riemannian centre of mass and mollifier smoothing. Comm. Pure Appl. Math.
30, 509541 (1977)
19. Barndorff-Nielsen, O.: Exponentially decreasing distributions for the logarithm of particle size.
Proc. Roy. Soc. Lond.353, 401419 (1977)
20. Rao, C.R.: Information and accuracy attainable in the estimation of statistical parameters. Bull.
Calcutta Math. Soc.37, 8191 (1945)
21. Amari, S.: Differential-Geometrical Methods in Statistics, Volume 28 of Springer Lecture
Notes in Statistics. Springer, New York (1985)
22. Calvo, M., Oller, J.: An explicit solution of information geodesic equations for the multivariate
normal model. Stat. Decis.9, 119138 (1991)