Beruflich Dokumente
Kultur Dokumente
net/publication/317129116
CITATIONS READS
4 133
3 authors, including:
Fred Daum
Raytheon Company
113 PUBLICATIONS 2,076 CITATIONS
SEE PROFILE
All content following this page was uploaded by Fred Daum on 25 May 2017.
Abstract—We describe a new algorithm for stochastic particle flow filters using Gromov’s method. We derive a simple
exact formula for Q in certain special cases. The purpose of using stochastic particle flow is two fold: improve estimation
accuracy of the state vector and improve the accuracy of uncertainty quantification. Q is the covariance matrix of the
diffusion for particle flow corresponding to Bayes’ rule.
Keywords- particle filter, nonlinear filter, particle flow, transport problem, extended Kalman filter, Monge-Kantorovich
optimal transport
I. INTRODUCTION
Particle filters generally suffer from the curse of dimensionality owing to particle degeneracy, which is caused by
Bayes’ rule, as explained in [7]. The root cause of particle degeneracy is that the particles are in the wrong place to
compute Bayes’ rule efficiently. Particle flow is an algorithm to compute Bayes’ rule using the flow of particles
rather than as the multiplication of two functions. The purpose of the particle flow algorithm is to mitigate the
problem of particle degeneracy for particle filters, by moving the particles to the correct region in state space to
represent the conditional probability density efficiently. We derive a law of motion for the random set of particles,
analogous to physics. If the initial random set of particles corresponds to the prior density, then the final set of
particles resulting from the flow corresponds to the posteriori density. We can prove this under mild regularity
conditions on the densities (e.g., smooth nowhere vanishing densities on a convex subset of a d-dimensional smooth
manifold); see [29] for details. A refreshingly honest survey of standard particle filters is in [10], the best
introduction to which is [15], with more details on nonlinear filters in [4] and [11]. In this paper we derive a simple
exact formula for the covariance matrix (Q) of the diffusion for our flow, and we explain that this simple concrete
formula is an example of a very general theorem of Gromov [78]. Numerical experiments which show the utility of
this theory are given in our companion paper [68].
Signal Processing, Sensor/Information Fusion, and Target Recognition XXVI, edited by Ivan Kadar, Proc.
of SPIE Vol. 10200, 102000I · © 2017 SPIE · CCC code: 0277-786X/17/$18 · doi: 10.1117/12.2248723
∂p 1 ∂p
= − div( fp ) + div Q ( x) (3)
∂λ 2 ∂x
in which Q is the covariance matrix of the diffusion w = w(λ), and div(.) is the divergence of (.). The diffusion w(λ)
should not be confused with so-called “process noise” of the dynamical model of x(t). Process noise is a random
process that evolves in time, whereas w(λ) evolves with λ. Moreover, Q is not the covariance matrix of the process
noise, because w is not process noise. If we assume that the flow of particles is deterministic, then Q = 0 and the PDE
simplifies significantly. We have derived many deterministic flows as shown in figure 1 of [38] with details given in
the references listed here. Our deterministic flows are similar to optimal transport algorithms, but our particle flows
are much simpler and faster to compute, because we avoid solving a variational problem, as explained in [21].
As explained in [27], the best particle flows avoid explicitly normalizing the conditional probability density; we have
learned this from many numerical experiments as well as theory. It is easy to derive a PDE for particle flow that
avoids normalization of the conditional density. In particular, plug (1) into (3) and take the gradient, and we get the
following PDE for the flow f:
∂p
2 ∂ div (Q ) / 2 p
∂ log h ∂ log p ∂div( f ) ∂ log p ∂f
+
∂x
=−fT 2
− − (4)
∂x ∂x ∂x ∂x ∂x ∂x
This PDE does not depend on the normalization K(λ), because we killed it by taking the gradient with respect to x.
That is, multiply p(x, λ) or h(x) by any non-zero constant factor, and the PDE does not change at all. One can
simplify this equation by picking Q = 0 or else picking Q to have values that make nice things happen, in which case
we can actually solve (4) exactly as a formula. But in this paper we will compute a simple formula for Q to improve
estimation accuracy and also improve uncertainty quantification (UQ). There are many different ways to pick the
matrix Q, as explained in [39].
We would like to avoid computing the conditional probability density p itself, owing to severe numerical problems for
particles in the tails of the distribution for high dimensional state vectors, and it is easy to write the last term in (4)
using the log of p rather than p itself with the help of the following well known identity:
∂p
div T
∂x = div ∂ log p + ∂ log p ∂ log p (5)
p ∂x ∂x ∂x
Recall that our geodesic particle flow derived in [20] and [25] has the form of equation (2) with:
−1 T
∂ 2 log p ∂ log h
f = − 2 (6)
∂x ∂x
and the covariance matrix of the random process w(λ) is given by Q. That is,
[
Q = P −1 + λH T R −1H ] −1
[
H T R −1 H P −1 + λH T R −1 H ]
−1
(9)
This simple exact formula was recently derived jointly by Rob Fuentes and Eric Blake [44]. In equation (9) R
denotes the measurement noise covariance matrix, P is the error covariance matrix of x prior to a given measurement,
and H is the measurement sensitivity matrix for a linear measurement model: z = Hx + v, in which v is Gaussian zero
mean measurement noise. If the measurements are nonlinear in x, then we can compute H using the standard
linearization approach as in extended Kalman filters. Numerical experiments show that this is an excellent
approximation for the examples that we have studied so far [68]. In many practical applications, one can pick a
coordinate system in which the measurements are exactly linear in x, in which case our theory works without any
approximations. We have used this simple trick in many real world applications of EKFs with radars over the last
five decades; for example, see [4] and [67]. One can also derive a simple exact formula for Q, which is essentially the
same as (9), assuming Gaussian prior and Gaussian likelihood, without the additional assumption of geodesic flow.
However, the geodesic flow is generally better than the Gaussian flow for nonlinear non-Gaussian applications. The
Gaussian flow was derived in [5] assuming Q = 0, but it is easy to generalize this exact particle flow for non-zero Q
given by equation (9).
We can avoid computing large matrix inverses by using Woodbury’s matrix inversion formula as follows:
−1
1
(P −1
+ λH T R −1 H )
−1
= P − PH T R + HPH T HP (10)
λ
and hence (9) becomes:
T 1 T
−1
T −1 T 1 T
−1
(11)
Q = P − PH R + HPH HP H R H P − PH R + HPH HP
λ λ
To avoid dividing by zero when λ = 0 (which always happens at the start of the flow), we can use the equivalent form:
(12)
[
Q = P − λPH T R + λHPH T ( )
−1
] [
HP H T R −1 H P − λPH T R + λHPH T ( )
−1
HP ]
We need to compute the square root of the matrix Q in order to generate the random variables dw for use in the actual
stochastic flow (2). Such a real valued matrix square root always exists because Q is symmetric and positive semi-
definite (in theory). We have used the MATLAB matrix square root primitives to do this, but very small numerical
errors in computing Q can easily result in a non-symmetric matrix for Q (or maybe a matrix that is not positive semi-
definite). In such cases MATLAB will complain that the square root of Q does not exist, or it will give us a complex
valued square root. To avoid such numerical troubles, it is very easy to symmetrize Q immediately before computing
the matrix square root:
Q + QT
Q→ (13)
2
There are two important properties of our simple exact formula for Q in equation (9). First, Q is zero for
uninformative measurements. For example, let R be very large corresponding to useless data; inspection of (9) shows
that Q is essentially zero in this case. This is a crucial feature of any good Q for particle flow. Obviously the
particles should not move if the measurement is not informative! Hence, it is clearly a bad idea to ask for Q that is
always positive definite. More specifically, it is a bad idea to ask for a Q that is a positive multiple of the identity
matrix, say Q = αI, and compute a fixed α using Monte Carlo simulations analogous to tuning the process noise
covariance matrix for a Kalman filter. Nevertheless, several popular MCMC algorithms violate this obvious law of
good particle behavior. Second, inspection of equation (9) shows that Q is generally not full rank, because H is rarely
full rank in any practical applications. At first this seemed like an undesirable feature, because we generally prefer
positive definite covariance matrices for diffusion tensors. For example, we like positive definite Q for Kalman
filters, because this tends to stabilize the filter, because this makes the plant model obviously and robustly completely
controllable. Stability of the Kalman filter is guaranteed by a completely controllable plant model and a completely
observable measurement model, according to Kalman’s beautiful theorem [43]. More generally, for nonlinear filters,
a positive definite Q avoids a degenerate Fokker-Planck equation, although one can live with hypo-elliptic PDEs
(which correspond to controllable plant models). Upon further reflection, we see that Q cannot always be full rank,
because certain directions in state space are not informed by a given measurement, and hence we do not want the
particles to move in those directions. In summary, rather than being undesirable, it is crucial that Q is not always full
rank. The Tychonov regularization of Q mentioned in the last paragraph results in a very small motion of the
particles for non-informative measurements, but this has no practical effect on filter state vector estimation accuracy
or filter uncertainty quantification accuracy.
The simple exact formula for Q shown in equation (9) is a concrete example of the application of a very general
theorem of Gromov. In particular, on pages 148 to 149 of Gromov’s book [40], we find a rather surprising theorem
about the existence of solutions to linear underdetermined PDEs. In particular, Gromov shows that there exists a very
nice solution to this system of PDEs if and only if it is sufficiently underdetermined. A “very nice solution” in this
context means one that does not require integration, but rather only uses differentiation and algebra. A “sufficiently
underdetermined” system of PDEs means that the number of unknowns is at least the number of equations plus the
dimension of x. This is a very strange theorem for several reasons. First, solving PDEs without any integration
seems too good to be true. Second, when was the last time that you heard a theorem that requires the number of
unknowns to be sufficiently large? Also, there are several caveats: (1) Gromov’s solution to the PDE system is not
the most general solution, and it could be very far from the most general solution; and (2) the solution is formal and
local.
We were aware of Gromov’s theorem for over seven years, as discussed in [8], and it seemed that Gromov’s method
could be useful for us, because we were trying to solve a highly underdetermined linear first order PDE. But the
trouble was that our PDE was not sufficiently underdetermined to satisfy the hypotheses of Gromov’s theorem. In
particular, with zero diffusion, we had a single scalar-valued PDE with a d-dimensional unknown function. This is
highly underdetermined, but not underdetermined enough. We needed more unknowns! But our recent attempts to
compute Q solved this problem. Q is a whole matrix of unknowns, and posing the problem as jointly finding f and Q
has plenty of unknowns compared with the number of equations. In summary, two open problems solved each other:
(1) how to compute f and Q jointly, and (2) how to apply Gromov’s theorem to particle flow. In summary, Gromov’s
theorem does not apply to deterministic particle flow, but it always applies to stochastic particle flow. This is yet
another benefit of stochastic vs. deterministic methods.
∫η ( x, λ )dx = 0
Ω
(15)
This existence theorem is valid even for the weakest notion of solution, including formal and local solutions. But
Gromov’s theorem does not mention any such condition for the existence of a solution to the underdetermined system
of PDEs. Hence, Gromov’s theorem does not apply to (14), because it is not sufficiently underdetermined. This was
bothering us for many years. Moreover, the addition of more unknowns avoids the stringent requirement of equation
(15) by invoking Gromov’s theorem. That is, the extra degrees of freedom using non-zero Q allows us to balance the
RHS and LHS of our PDE and thereby avoid the requirement in (15). Furthermore, we conjecture that transport
problems using non-zero diffusion will be less sensitive to errors in computing the normalization of the conditional
probability density, because normalization of the density is equivalent to condition (15). This connection is explained
in [27].
The mathematical literature on optimal transport almost always assumes zero diffusion (Q = 0). In particular, Villani
mentions the possibility of random transport in passing, but there are no theorems or algorithms for such stochastic
transport in [45] or [46], and there is very little mention of stochastic transport in [47]. Moreover, there are very few
papers on stochastic optimal transport, with the conspicuous exception of [48], as far as we know. But it seems
intuitively that such stochastic transport would be useful for the reasons given above in connection with Gromov’s
theorem avoiding the stringent and unforgiving requirement of (15).
Gromov is one of the very few mathematicians who study underdetermined PDEs seriously (see the foreword to [40]
for elaboration on this point). Underdetermined linear PDEs are extremely simple, and hence they are considered
beneath contempt by most research mathematicians; for example, Trèves says that: “underdetermined systems [of
PDEs] are much easier to study than overdetermined ones (and in most cases, than determined ones also); they are
also much less interesting” on page 9 in [49]. Naturally, Gromov and Lanczos and Hilbert and Schmidt and we do
not agree with the last part of Trèves’ assertion. (The generalized inverse of linear differential operators was invented
by Hilbert & Schmidt several decades before Moore & Penrose reinvented the generalized inverse for matrices; see
Nashed [52] for a good historical survey). Moreover, an underdetermined PDE is “ill-posed” according to the
standard textbook definitions. The solution to this ill-posed problem in optimal transport theory is to ask for the
unique solution of the transport PDE that minimizes some convex functional of the flow of particles (e.g., Wasserstein
distance). But this solution comes at the extremely high computational cost of solving a variational problem. We
prefer to avoid solving such a variational problem.
To dispel the mystery about Gromov’s method, let us consider a nice simple example and show the exact solution to
the PDE. The simplest non-trivial example is the following underdetermined PDE:
∂q1 ∂q2
+ + q3 = η ( x) (16)
∂x1 ∂x2
in which η(x) is a given smooth function of x, and we have 3 unknowns and one scalar-valued equation. We can
write this in the more concise form: Lq = η, in which L is a linear differential operator and q is the vector of
unknowns. Let us try to find a solution of the form q = Mη, in which M is a linear differential operator, the existence
of which is promised by Gromov’s theorem. Plugging our solution into the PDE, we get LMη = η, for arbitrary
smooth functions η(x), which is the same as LM = 1. Taking the formal adjoint we get: M*L* = 1. This is the
equation that we need to solve for M. By inspection, we get the following exact simple solution:
T
∂η ∂η
q = Mη = − , ,η (17)
∂x2 ∂x1
∂ 2η ∂ 2η
− + +η = η (18)
∂x1∂x2 ∂x2∂x1
For functions η(x) that are at least twice continuously differentiable with respect to x, the mixed partials are equal,
and hence (18) is an identity for all smooth η(x). We can easily generalize this simple example to arbitrarily high
dimensions of x, as well as much more general linear operators L. But thinking about this simple example gives the
basic intuitive understanding of what is actually going on with Gromov’s method. Notice that this method would not
work unless the PDE was sufficiently underdetermined, as per Gromov’s theorem. Two unknowns and one equation
is not enough, but rather we need at least three unknowns and one equation for x with dimension d = 2.
We can generalize this example as follows. Consider the single scalar-valued linear underdetermined PDE:
div(r ) + bT s = η (19)
in which q = (r, s), where r is d dimensional, and s has arbitrarily high dimension. One possible family of exact
solutions to (19) is:
T T
∂η ∂θ
r = A + B + β (x) (20)
∂x ∂x
bη bbT
s= 2
+ I − 2 y (21)
b b
in which A and B are arbitrary skew-symmetric matrices that are independent of x, and both θ and η are arbitrary
functions of x that are twice continuously differentiable, and β is an arbitrary smooth function of x such that the ith
component of β is not a function of the ith element of x, and y is a completely arbitrary vector to match the dimension
of s. It is easy to check that this is an exact solution to (19) using the fact that the Hessians of η and θ are symmetric
matrices, because η and θ are smooth functions of x. Moreover, we can generalize this example to arbitrarily high
number of equations. This simple example makes it clear why Gromov’s theorem assumes that the number of
unknowns is at least equal to the number of equations plus the dimension of x. Apparently our solution in (20) and
(21) is more general than Gromov’s solution q = Mη.
Suppose that we have an exact solution (or an excellent approximation) of the following PDE:
r ) = η~
div(~ (22)
We can use this information to generalize our solution of (15) even further. In particular, add (22) to (19):
r ) + bT s = η + η~
div(r + ~ (23)
Now we can compute a family of solutions for (23) analogous to our solution given in (20) and (21) as follows:
T T
∂η ∂θ
r = A + B + β (x) - ~r (24)
∂x ∂x
b(η + η~ ) bbT
s= 2
+ I − 2 y (25)
b b
in which A, B, y, β, and θ have the same meanings as given above. This family of solutions is more general than that
given in (20) and (21). We can generalize this still further by taking linear combinations of solutions of PDEs like
(22), which results in:
b(η + ∑ wkη~k )
bbT
s= Λ
2
+ I − 2
y (27)
b b
in which Λ denotes the set of solutions of the PDE (22) that we wish to exploit, and wk is an arbitrary real number
that weights the kth solution of (22).
It is easy to manufacture a plethora of solutions of (22) by starting with ~r and computing the divergence to find out
what the corresponding η~ is. This is the reverse of the usual process of “solving” a PDE, and it is obviously much
easier. Solutions of (22) can be computed off-line and stored in memory to be used in real time. For d ≥ 2 the PDE in
(22) is underdetermined, and hence we have many solutions of (22) to choose from. The vector y can be chosen to
stabilize the particle flow (e.g., as shown in [16]), or we could choose y to minimize the sensitivity of the flow to
errors in the normalization of the conditional density (e.g., as shown in [26]). One can also imagine other methods to
choose y.
These solutions of our PDE generalize Gromov’s solution in several ways. First, we have not assumed as much
smoothness as Gromov assumed; in particular, we assume that the functions are twice continuously differentiable;
∞
that is, we work in C², whereas Gromov works in C . Second, Gromov assumes that the coefficients of the PDE are
constant (wrt x and λ), whereas we do not make such an assumption. Third, our solutions are apparently more general
than Gromov’s solutions, as explained above (e.g. equations (26) and (27)). It is often said that Gromov works in a
very general and abstract setting, which is certainly true, but nevertheless we have generalized Gromov’s solution of
linear constant coefficient underdetermined PDEs in the three specific ways explained above.
How is it possible that such a simple exact solution of PDEs eluded the grasp of Goursat [8] and Courant & Hilbert
[8] and Forsyth [8]? It is because these mathematicians did not consider underdetermined PDEs worth serious study,
because they are “ill-posed” and “non-physical.” Moreover, even if such mathematicians had studied
underdetermined PDEs, who would imagine that they needed the PDE to be “sufficiently underdetermined” in
Gromov’s sense? Furthermore, who would have imagined that underdetermined PDEs would be useful for anything?
In contrast with the simple examples given above in (16) to (27), Gromov’s book is famously difficult to read. In fact
the book by Spring [50] was intended to help mathematicians understand [40] by filling in all the details in the proofs,
and also by stating the theorems more completely and more precisely. But Spring’s book is of little help to normal
engineers, because it does not discuss the relevant theorem on page 149 in Gromov’s book at all, and it is inscrutable
for normal engineers who do not have jet bundles and fiber bundles for breakfast. One might instead consult Lanczos
[51] for a much more readable treatment of underdetermined PDEs, with simple explicit examples, although it
obviously does not cover Gromov’s work. One must be careful about the meaning of certain mathematical terms in
[40]; for example, when Gromov says “inverse” of a linear operator, he really means “right inverse” or sometimes
“left inverse.” Also, the theorem in Gromov’s book is shrouded in rather general and abstract structures, which
normal engineers do not have for breakfast. The more recent exposition on page 40 of Gromov’s paper [41] is
somewhat easier to understand than Gromov’s book [40]. In addition, Yang’s note [42] gives several examples that
make Gromov’s theory easier to grasp by normal engineers. Another important reason to study Yang’s paper is to
understand the generalization of Gromov’s method using exterior differential forms [42]. The generalization of
Gromov’s method suggested by Yang [42] is roughly analogous to our simple examples in equations (19) to (27),
except that we used linear algebra and elementary calculus rather than Cartan’s exterior calculus; also, Yang’s
examples give solutions to ODEs, whereas we solve PDEs. The solutions to the PDE (19) given here in (20) to (27)
were invented using Polya’s method, as explained on page 357 of Seiler’s superb book [53].
Both Hilbert and Cartan attempted to find solutions to underdetermined differential equations without using any
integration but only algebra and differentiation, similar to Gromov’s theorem, but without success. These
unsuccessful attempts are surveyed on page 339 of Seiler’s book [53]. In a certain sense Gromov’s theorem is
intuitively obvious once it is pointed out. In particular, as explained before, the plethora of unknowns play the role of
extra degrees of freedom, which we can use to balance the RHS and LHS of our PDE and avoid integration wrt x or λ;
the intuitive idea of Gromov’s theorem is that simple. Of course, explaining why we need exactly so many unknowns
FUTURE RESEARCH
For future research, we plan to ruthlessly exploit Gromov’s method (and generalizations thereof) to derive new
particle flow filter algorithms. First, we could compute Q and f jointly by solving the PDE (4) without assuming that
the last three terms in (4) sum to zero. There are many variations on this theme. Second, we could compute f and Q
jointly by solving (4) assuming that the conditional probability densities are from the multivariate exponential family
[4], rather than from the Gaussian family of probability densities. We can compute the exact solution of the Fokker-
Planck equation for the multivariate exponential family using separation of variables (see [56] and [77]). Using this
theory we can explicitly compute f and Q by solving a system of underdetermined linear algebraic equations, using
the Moore-Penrose inverse and the projection into the null space of the relevant linear operator, very similar to (27).
Hence, we consider this a very promising direction of research. Third, we could compute f and Q jointly by
assuming that the measurement is an arbitrary smooth nonlinear function of x with additive Gaussian noise and a
multivariate Gaussian prior probability density of x. This is obviously a special case of the second idea, but allows us
to obtain more explicit formulas.
Fourth, we could compute f and Q jointly by assuming that the last three terms in our PDE (4) sum to zero, and solve
for Q using the value of f given in (6) along with the constraint that Q is a symmetric positive semi-definite matrix.
For example, we could factorize Q using the Cholesky square root, which would enforce the obvious constraints for a
covariance matrix. This would result in a quadratic algebraic system of equations for SQRT(Q), which can be solved
explicitly as a formula using ideas similar to [19], at least in certain special cases. Recall that in [19] we solved a
vector Riccati equation explicitly as a formula for the particle flow. If Q is assumed to be invariant with respect to x,
then the PDE (4) is a linear algebraic system in Q, and one could also consider solving this linear algebraic system
directly without the explicit constraint on Q as a covariance matrix; this approach in fact works in certain cases,
including the Gaussian example given in this paper. In particular, in order to derive (9) we did not explicitly enforce
any constraint on Q as a covariance matrix. One can also use a homotopy method to compute SQRT(Q) starting with
the initial value of SQRT(Q) given by (9); this homotopy would morph both the PDE and the solution jointly. Other
initial values of SQRT(Q) are also possible (e.g., the identity matrix), but such methods do not exploit our knowledge
of exact solutions of our PDE for related problems. One can imagine many other methods to compute Q or SQRT(Q)
for general nonlinear filter problems with smooth nowhere vanishing conditional probability densities p, including
Newton’s method or quasi-Newton methods or hybrids of homotopy and Newton’s method. The control theory
literature on solving algebraic matrix Riccati equations is vast, and it can be used fruitfully here. For example, we
know how to solve the algebraic matrix Riccati equation for a unique positive semi-definite solution by integrating
the corresponding Riccati matrix differential equation to steady state, assuming a positive definite initial condition
and positive semi-definite matrix coefficients, which correspond to natural controllability conditions on the relevant
linear dynamical system; that is, no explicit constraints to obtain positive definite solutions are needed; this occurs
naturally under the right conditions. This control theory method is analogous to a homotopy algorithm described
earlier, in that we are integrating a matrix differential equation with respect to a scalar. We emphasize that these
methods do not make any assumptions about the dynamics or measurements or conditional density or likelihood other
than smoothness and nowhere vanishing conditional densities. In particular, we do not assume Gaussian priors or
densities from the multivariate exponential family or linear measurements or linear dynamics for the purpose of
computing Q or f. That is, this approach is much more general than several of the earlier methods suggested.
Fifth, it would be nice to have a geometric (i.e., coordinate free) version of our solutions to (4) and (15), perhaps
using exterior differential systems, as suggested by Yang [42]; however, whether this mathematical flourish is
actually useful to obtain new and better solutions in practice remains to be seen.
Sixth, we would like to compute the most general family of solutions to our PDE, rather than a series of special
families of solutions. Maybe completion of the PDE system to involution (in the style of Kähler) would be useful to
c(d )
e ≤ (29)
N
The so-called “constants” in these two bounds hide a multitude of details about the problem, including: smoothness of
the functions, coupling between components of the state vector, errors in the proposal densities, dimensional
dependence of sampling errors, etc.; for details of such fraud see [73] and [38]. Moreover, it is well known that QMC
works much better for low dimensional problems in practice, which has motivated the search for methods of
dimensional reduction or decoupling; for example, it is standard to apply a PCA to a data set before attempting QMC
sampling. Also, more recent research has improved the Koksma-Hlawka inequality by explicitly modeling the
smoothness of the functions as well as the coupling between components of the state vector. Finally, it is well known
that the Koksma-Hlawka inequality is far from tight, whereas the modern QMC bounds are much tighter, because
they actually include the key ingredients which the K-H bounds sweeps under the rug. For example, a more recent
bound on L² approximation error has the form:
ACKNOWLEDGEMENTS
The simple exact beautiful formula for Q (equation 9) was derived jointly by Dr. Rob Fuentes and Dr. Eric Blake,
and their work is gratefully acknowledged. We thank Professor Simon Maskell for telling us that even for examples
in which the state vector estimate of our particle flows (with Q = 0) is excellent, the particle distribution generally
underestimates the variance of the correct conditional density; this was a strong motivation for our research reported
here. Theoretical and numerical confirmation of Simon’s observation, along with a resolution of this issue, are
given in [68]. We thank Flávio de Melo for his nice clear proof of Gromov’s theorem, as well as excellent questions
and suggestions about how to actually apply Gromov’s method [76].
REFERENCES
[1] Uwe Hanebeck, Kai Briechle & Andreas Rauh, “Progressive Bayes: a new framework for nonlinear state estimation,” Proceedings of SPIE
Conference, Orlando Florida, April 2003.
[2] Simon Godsill & Tim Clapp, “Improvement strategies for Monte Carlo particle filters,” pages 139-158 in “Sequential Monte Carlo methods
in practice,” edited by Arnaud Doucet, Nando de Freitas & Neil Gordon, Springer-Verlag, 2001.
[3] Christian Musso, Nadia Oudjane & Francois Le Gland, “Improving regularised particle filters,” pages 247-271 in “Sequential Monte Carlo
methods in practice,” edited by Arnaud Doucet, Nando de Freitas & Neil Gordon, Springer-Verlag, 2001.
[4] Fred Daum, “Nonlinear filters: beyond the Kalman filter,” IEEE AES Systems magazine, special tutorial issue, pages 57 to 69, August
2005.
[5] Fred Daum, Jim Huang and Arjang Noushin, “Exact particle flow for nonlinear filters,” proceedings of SPIE Conference, Orlando Florida,
April 2010.
[6] Fred Daum & Jim Huang, “Numerical experiments for nonlinear filters with exact particle flow induced by log-homotopy,” proceedings of
SPIE conference, Orlando Florida, April 2010.
[7] Fred Daum & Jim Huang, “Particle degeneracy: root cause and solution,” Proceedings of SPIE Conference, Orlando Florida, April 2011.
[8] Fred Daum & Jim Huang, “Exact particle flow for nonlinear filters: seventeen dubious solutions to a first order linear underdetermined
PDE,” Proceedings of IEEE Asilomar conference, November 2010.
[9] Fred Daum, Jim Huang & Arjang Noushin, “Coulomb’s law particle flow for nonlinear filters,” Proceedings of SPIE Conference, San
Diego CA, August 2011.
[10] Arnaud Doucet & A. M. Johansen, “A tutorial on particle filtering and smoothing: fifteen years later,” Chapter 24 in “The Oxford handbook
of nonlinear filtering,” edited by Dan Crisan & Boris Rozovskiǐ, Oxford University Press, 2011.
[11] “The Oxford handbook of nonlinear filtering,” edited by Dan Crisan & Boris Rozovskiĭ, Oxford University Press, 2011.
[12] Fred Daum, Jim Huang & Arjang Noushin, “Numerical experiments for Coulomb’s law particle flow for nonlinear filters,” Proceedings of
SPIE Conference, San Diego CA, August 2011.
[13] Fred Daum & Jim Huang, “Hollywood log-homotopy: movies of particle flow for nonlinear filters,” Proceedings of SPIE conference,
Orlando Florida, April 2011.
[14] Fred Daum & Jim Huang, “Nonlinear filters with log-homotopy,” Proceedings of SPIE Conference on signal & data processing, San Diego
California, September 2007.
[15] Branko Ristic, Sanjeev Arulampalam & Neil Gordon, “Beyond the Kalman filter,” Artech House publishing, 2004.
[16] Fred Daum & Jim Huang, “Generalized particle flow for nonlinear filters,” Proceedings of SPIE Conference, Orlando Florida, April 2010.
[17] Fred Daum, “A new nonlinear filtering formula for discrete time measurements,” IEEE CDC Proceedings, Fort Lauderdale Florida,
December 1985.
[18] Fred Daum, “A new nonlinear filtering formula for non-Gaussian discrete time measurements,” IEEE CDC Proceedings, Athens Greece,
December 1986.
[19] Fred Daum and Jim Huang, “Zero curvature particle flow for nonlinear filters,” Proceedings of SPIE Conference, Baltimore, April 2013.
DownloadedViewFrom:
publicationhttp://spiedigitallibrary.org/pdfaccess.ashx?url=/data/conferences/spiep/92568/
stats on 05/25/2017 Terms of Use: http://spiedigitallibrary.org/ss/termsofuse.aspx