Learning with Maximum A-Posteriori Perturbation Models

[edit]

Andreea Gane, Tamir Hazan, Tommi Jaakkola ;
Proceedings of the Seventeenth International Conference on Artificial Intelligence and Statistics, PMLR 33:247-256, 2014.

Abstract

Perturbation models are families of distributions induced from perturbations. They combine randomization of the parameters with maximization to draw unbiased samples. Unlike Gibbs’ distributions, a perturbation model defined on the basis of low order statistics still gives rise to high order dependencies. In this paper, we analyze, extend and seek to estimate such dependencies from data. In particular, we shift the modelling focus from the parameters of the Gibbs’ distribution used as a base model to the space of perturbations. We estimate dependent perturbations over the parameters using a hard-EM approach, cast in the form of inverse convex programs. Each inverse program confines the randomization to the parameter polytope responsible for generating the observed answer. We illustrate the method on several computer vision problems.

Related Material