All smooth -divergences are locally the same
22 Jul 2017
For discrete distributions and ,
the -divergence is defined as
where is a convex function satisfying
the condition .
If is a variation of , then
Provided is twice differentiable, we can develop it into Taylor series
and thus approximate the -divergence by a quadratic function
Comparing it with the Fisher metric, we see that it is the same quadratic
form scaled by a constant factor .
Note that not all -divergences are locally the same,
only the smooth ones.
For example, the total variation distance corresponds to
which is not quadratic around .
Special case: -divergence
The -divergence with having the form
is known as the -divergence. Noting that
has the properties
and ,
we obtain the approximation of the -divergence
which directly generalizes the result of Kullback
for the KL divergence and its reverse,
corresponding to and respectively.
Local approximation is exact for Pearson divergence
Pearson divergence is the -divergence with ,
corresponding to the generating function
We established the following quadratic approximation of the -divergence
that is valid for small . However, if we allow big deviations
, then we obtain Pearson divergence
(scaled by ),
Thus, Pearson divergence is the linear extension of the Fisher metric
from a local neighborhood to the whole space.
Consequently, local quadratic approximation is exact
for Pearson divergence, since .