On a Variational Definition for the Jensen-Shannon Symmetrization of Distances Based on the Information Radius |
| |
Authors: | Frank Nielsen |
| |
Affiliation: | Sony Computer Science Laboratories, Tokyo 141-0022, Japan; |
| |
Abstract: | We generalize the Jensen-Shannon divergence and the Jensen-Shannon diversity index by considering a variational definition with respect to a generic mean, thereby extending the notion of Sibson’s information radius. The variational definition applies to any arbitrary distance and yields a new way to define a Jensen-Shannon symmetrization of distances. When the variational optimization is further constrained to belong to prescribed families of probability measures, we get relative Jensen-Shannon divergences and their equivalent Jensen-Shannon symmetrizations of distances that generalize the concept of information projections. Finally, we touch upon applications of these variational Jensen-Shannon divergences and diversity indices to clustering and quantization tasks of probability measures, including statistical mixtures. |
| |
Keywords: | Jensen-Shannon divergence, diversity index, Ré nyi entropy, information radius, information projection, exponential family, Bregman divergence, Fenchel– Young divergence, Bregman information, q-exponential family, q-divergence, Bhattacharyya distance, centroid, clustering |
|
|