1 research outputs found

    Refined α\alpha-Divergence Variational Inference via Rejection Sampling

    Full text link
    We present an approximate inference method, based on a synergistic combination of R\'enyi α\alpha-divergence variational inference (RDVI) and rejection sampling (RS). RDVI is based on minimization of R\'enyi α\alpha-divergence Dα(pq)D_\alpha(p||q) between the true distribution p(x)p(x) and a variational approximation q(x)q(x); RS draws samples from a distribution p(x)=p~(x)/Zpp(x) = \tilde{p}(x)/Z_{p} using a proposal q(x)q(x), s.t. Mq(x)p~(x),xMq(x) \geq \tilde{p}(x), \forall x. Our inference method is based on a crucial observation that D(pq)D_\infty(p||q) equals logM(θ)\log M(\theta) where M(θ)M(\theta) is the optimal value of the RS constant for a given proposal qθ(x)q_\theta(x). This enables us to develop a \emph{two-stage} hybrid inference algorithm. Stage-1 performs RDVI to learn qθq_\theta by minimizing an estimator of Dα(pq)D_\alpha(p||q), and uses the learned qθq_\theta to find an (approximately) optimal M~(θ)\tilde{M}(\theta). Stage-2 performs RS using the constant M~(θ)\tilde{M}(\theta) to improve the approximate distribution qθq_\theta and obtain a sample-based approximation. We prove that this two-stage method allows us to learn considerably more accurate approximations of the target distribution as compared to RDVI. We demonstrate our method's efficacy via several experiments on synthetic and real datasets.Comment: 6 pages, 1 figur
    corecore