p]

KL[q || p]

If log_p(z) = Log[p(z)] for distribution p, this Op approximates the negative Kullback-Leibler divergence.

elbo_ratio(log_p, q, n=100) = -1 * KL[q || p],
KL[q || p] = E[ Log[q(Z)] - Log[p(Z)] ]

Note that if p is a Distribution, then distributions.kl(q, p) may be defined and available as an exact result.

doc_TensorFlow
2016-10-14 12:42:37
Comments
Leave a Comment

Please login to continue.