arXiv Analytics

Sign in

arXiv:2002.11187 [cs.LG]AbstractReferencesReviewsResources

Reliable Estimation of Kullback-Leibler Divergence by Controlling Discriminator Complexity in the Reproducing Kernel Hilbert Space

Sandesh Ghimire, Prashnna K Gyawali, Linwei Wang

Published 2020-02-25Version 1

Several scalable methods to compute the Kullback Leibler (KL) divergence between two distributions using their samples have been proposed and applied in large-scale machine learning models. While they have been found to be unstable, the theoretical root cause of the problem is not clear. In this paper, we study in detail a generative adversarial network based approach that uses a neural network discriminator to estimate KL divergence. We argue that, in such case, high fluctuations in the estimates are a consequence of not controlling the complexity of the discriminator function space. We provide a theoretical underpinning and remedy for this problem through the following contributions. First, we construct a discriminator in the Reproducing Kernel Hilbert Space (RKHS). This enables us to leverage sample complexity and mean embedding to theoretically relate the error probability bound of the KL estimates to the complexity of the neural-net discriminator. Based on this theory, we then present a scalable way to control the complexity of the discriminator for a consistent estimation of KL divergence. We support both our proposed theory and method to control the complexity of the RKHS discriminator in controlled experiments.

Related articles: Most relevant | Search more
arXiv:2011.01170 [cs.LG] (Published 2020-11-02, updated 2021-04-12)
Homeomorphic-Invariance of EM: Non-Asymptotic Convergence in KL Divergence for Exponential Families via Mirror Descent
arXiv:2002.04753 [cs.LG] (Published 2020-02-12)
A Random-Feature Based Newton Method for Empirical Risk Minimization in Reproducing Kernel Hilbert Space
arXiv:2007.06168 [cs.LG] (Published 2020-07-13)
Model Fusion with Kullback--Leibler Divergence