Skip to contents

The goal of kldest is to estimate Kullback-Leibler (KL) divergence DKL(P||Q)D_{KL}(P||Q) between two probability distributions PP and QQ based on:

  • a sample x1,...,xnx_1,...,x_n from PP and the probability density qq of QQ, or
  • samples x1,...,xnx_1,...,x_n from PP and y1,...,ymy_1,...,y_m from QQ.

The distributions PP and QQ may be uni- or multivariate, and they may be discrete, continuous or mixed discrete/continuous.

Different estimation algorithms are provided for continuous distributions, either based on nearest neighbour density estimation or kernel density estimation. Confidence intervals for KL divergence can also be computed, either via subsampling (preferred) or bootstrapping.

Installation

You can install kldest from CRAN:

Alternatively, can install the development version of kldest from GitHub with:

# install.packages("devtools")
devtools::install_github("niklhart/kldest")

A minimal example for KL divergence estimation

KL divergence estimation based on nearest neighbour density estimates is the most flexible approach.

Set a seed for reproducibility

KL divergence between 1-D Gaussians

Analytical KL divergence:

kld_gaussian(mu1 = 0, sigma1 = 1, mu2 = 1, sigma2 = 2^2)
#> [1] 0.4431472

Estimate based on two samples from these Gaussians:

X <- rnorm(100)
Y <- rnorm(100, mean = 1, sd = 2)
kld_est_nn(X, Y)
#> [1] 0.2169136

Estimate based on a sample from the first Gaussian and the density of the second:

q <- function(x) dnorm(x, mean = 1, sd =2)
kld_est_nn(X, q = q)
#> [1] 0.6374628

Uncertainty quantification via subsampling:

kld_ci_subsampling(X, q = q)
#> $est
#> [1] 0.6374628
#> 
#> $ci
#>      2.5%     97.5% 
#> 0.2601375 0.9008446

KL divergence between 2-D Gaussians

Analytical KL divergence between an uncorrelated and a correlated Gaussian:

kld_gaussian(mu1 = rep(0,2), sigma1 = diag(2),
             mu2 = rep(0,2), sigma2 = matrix(c(1,1,1,2),nrow=2))
#> [1] 0.5

Estimate based on two samples from these Gaussians:

X1 <- rnorm(100)
X2 <- rnorm(100)
Y1 <- rnorm(100)
Y2 <- Y1 + rnorm(100)
X <- cbind(X1,X2)
Y <- cbind(Y1,Y2)

kld_est_nn(X, Y)
#> [1] 0.3358918