KLqp
KLqp is a variational inference algorithm used to approximate posterior distributions in probabilistic models. It denotes the minimization of the Kullback-Leibler divergence D_KL(q(z; phi) || p(z|x)) between a chosen variational distribution q over latent variables z and the true posterior p(z|x), with x representing observed data. The variational distribution q is parameterized by phi, and the goal is to find phi that minimizes this divergence.
This objective is equivalent to maximizing the Evidence Lower Bound (ELBO): L(phi) = E_{q(z;phi)}[log p(x,z) - log q(z;phi)],
Optimization is typically performed with stochastic gradient methods. For continuous latent variables, the reparameterization trick is
KLqp is implemented in several probabilistic programming libraries, such as Edward and TensorFlow Probability, as part
Notes: KLqp stands for minimizing the KL divergence from q to p (the posterior). It is conceptually