Skip to content

Variational Inference with Rényi Divergence (VR)

This method generalizes the standard ELBO using \(\alpha\)-Rényi divergence.

Unlike VI implementation with LRT, here explicit weights \(w \sim \mathcal{N}(\mu, \text{softplus}(\rho))\) are sampled using weight perturbation during the forward pass. The objective is defined as:

\[ \mathcal{L}_{\text{VR}}(\theta, \alpha) = -\frac{1}{1-\alpha} \log \frac{1}{K} \sum_{k=1}^K \left( \frac{p(\mathcal{D}, w_k)}{q_\theta(w_k)} \right)^{1-\alpha} \]

The parameter \(\alpha\) (default 1.0) controls the bias-variance trade-off, allowing for more robust posterior approximations compared to standard Kullback-Leibler divergence.


Yingzhen Li, Richard E. Turner "Rényi Divergence Variational Inference" (2016)