Joint Entropy Maximization in Kernel-Based Topographic Maps

Abstract

A new learning algorithm for kernel-based topographic map formation is introduced. The kernel parameters are adjusted individually so as to maximize the joint entropy of the kernel outputs. This is done by maximizing the differential entropies of the individual kernel outputs, given that the map's output redundancy, due to the kernel overlap, needs to be minimized. The latter is achieved by minimizing the mutual information between the kernel outputs. As a kernel, the (radial) incomplete gamma distribution is taken since, for a gaussian input density, the differential entropy of the kernel output will be maximal. Since the theoretically optimal joint entropy performance can be derived for the case of nonoverlapping gaussian mixture densities, a new clustering algorithm is suggested that uses this optimum as its “null” distribution. Finally, it is shown that the learning algorithm is similar to one that performs stochastic gradient descent on the Kullback-Leibler divergence for a heteroskedastic gaussian mixture density model.

Cite

Text

Van Hulle. "Joint Entropy Maximization in Kernel-Based Topographic Maps." Neural Computation, 2002. doi:10.1162/089976602760128054

Markdown

[Van Hulle. "Joint Entropy Maximization in Kernel-Based Topographic Maps." Neural Computation, 2002.](https://mlanthology.org/neco/2002/hulle2002neco-joint/) doi:10.1162/089976602760128054

BibTeX

@article{hulle2002neco-joint,
  title     = {{Joint Entropy Maximization in Kernel-Based Topographic Maps}},
  author    = {Van Hulle, Marc M.},
  journal   = {Neural Computation},
  year      = {2002},
  pages     = {1887-1906},
  doi       = {10.1162/089976602760128054},
  volume    = {14},
  url       = {https://mlanthology.org/neco/2002/hulle2002neco-joint/}
}