Infinite surprise – the iridescent personality of Kullback-Leibler divergence
This article is originally published at https://blogs.rstudio.com/tensorflow/Kullback-Leibler divergence is not just used to train variational autoencoders or Bayesian networks (and not just a hard-to-pronounce thing). It is a fundamental concept in information theory, put to use in a vast range of applications. Most interestingly, it's not always about constraint, regularization or compression. Quite on the contrary, sometimes it is about novelty, discovery and surprise.
Thanks for visiting r-craft.org
This article is originally published at https://blogs.rstudio.com/tensorflow/
Please visit source website for post related comments.