tags:
- algorithm
- integration
share: "true"
Wasserstein-Fisher-Rao Gradient Flow
Suppose we have two probability measures,
this is well-studied and has well-known restrictions. An alternative here would be the Fisher-Rao distance,
Note that this is distance is constrained by a reaction PDE instead of a continuity equation. We can see that
where
but this is irrelevant to the topic at hand.
The PDE constraining the Wasserstein-Fisher-Rao metric will intuitively be
i.e., we use velocity field
Then, if I have a measure
See, e.g., YWR23
At first, we may be interested in the KL-divergence function often used in gradient flows, which (for a given distribution
but if we keep
I don't think I need to center the weight derivative? Anyway, if
Suppose we now have
where
where
What's nice here is that this is "forgetful" of our initial statue
Take our Hilbert space
so our kernel becomes the Christoffel-Darboux kernel of
with
which means the embedding of an arbitrary function
Assume we want to match points and weights
^3be7e2
with
and somewhat less obviously, suppose we abuse notation to consider a diagonal matrix
which indicates the trace is operating on a matrix with only one nonzero column
Here we maintain control of three things:
We now assume data from
One can compare and contrast this with the minimization problem given in ^021c59. While they are very similar, note that the geometry of the static problem by
Suppose we have node
In this squared-exponential kernel case, we know that
We know that, if we set the RHS to zero, we might get a steady-state solution
The expressions on the right are necessitated by the form of the squared-exponential kernel.
If
Suppose
I'm not sure of either of these things, though.
We return to the original ODE when