• Daniel Jiwoong Im, Sungjin Ahn, Roland Memisevic, and Yoshua Bengio. Denoising Criterion for Variational Auto-Encoding Framework. arXiv preprint arXiv:1511.06406, 2015.

Summary

Auto-encoders are used to learn representations of observed data . They do so by minimizing the reconstruction error of a decoder , given a characterization of hidden variables from an encoder . Along with a penalty for regularization, this can be written as maximizing the objective

over the encoder and decoder’s parameters. However, it’s unclear what a principled choice of the penalty is, as well as the choice of the encoder for a given decoder . Variational auto-encoders (Kingma and Welling, 2014; Rezende et al., 2014) reinterpret the evidence lower bound prominent in variational inference for this purpose:

where is a model prior placed over the hidden variables of the encoder. The penalty term is automatically given as the KL divergence between the encoder and the prior, and the encoder can more generally be specified following variational inference literature. This is now a probabilistic encoder-decoder, with a principled objective given by a lower bound to the model evidence .

In this paper, Im et al. take ideas from denoising auto-encoders (Vincent et al., 2008) and apply it to variational auto-encoders. In particular, they consider a corrupted distribution , such that

With a distribution to model corrupted noise, the augmented is more expressive than the original without corruption: it is able to model additional uncertainty around the data inputs.

For inference, plugging this variational distribution into the ELBO leads to an analytically intractable objective, as the log-density cannot be computed. As a proxy, they propose maximizing the objective function

which they show is in fact an upper bound to the original ELBO. They also claim it is still a lower bound to the model evidence , and thus still a valid variational objective (I’ll comment on this later). Gradients proceed as typically done in the literature, either using the reparameterized gradient for differentiable latent variables, or the score function estimator in the more general case. They have an interesting Q&A format for discussing their experiments that I found kind of cool.

Discussion

This paper is eerily similar to a paper Rajesh Ranganath, Dave Blei, and I uploaded just two weeks earlier than theirs, on Hierarchical Variational Models. And that was uploaded only three weeks ago from today! Man, the field moves fast.

  • The idea is the same: place a prior on inputs to the variational distribution—treating them as variational latent variables—and proceed to marginalize them out. The new inputs are now the inputs to the prior distribution. In their case, the prior is on the data input for inference networks, which output local variational parameters. In our case, the prior is on variational parameters more generally. A prior on data input is the special case when the variational model uses an inference network: following our notation, we replace with the more powerful , and use an inference network to output prior hyperparameters (see end of Section 5).
  • The source of inspiration is different: they follow the standard procedure in denoising auto-encoders. We follow the Bayesian framework, viewing the variational distribution as a model of the posterior latent variables. These different sources of inspiration motivate difference choices of complexity for the prior: they only consider Bernoulli or Gaussian, whereas from the Bayesian interpretation, it is natural to study more complex priors such as the normalizing flow (Rezende and Mohamed, 2015) as a prior; we show the latter leads to more expressive approximations and also more scalable computation. (And in our more recent paper on the Variational Gaussian Process, we use an even cooler prior :^).
  • The variational objectives are different, and I have concerns that their objective is not in fact a lower bound on .¹ This means the values of the objective function are no longer a proxy for the model evidence. Among a few things, this invalidates the experimental results, as they compare to methods based on reported variational bounds. Further, maximizing their objective can lead to poor approximations, as there is no stopping the function to blow up to arbitrary values as the global maximum. Actually, it surprises me that the maximized variational objective values in the experiments are still close to those in . Would be very interested to hear from the authors on these issues.

Regardless of whether the theorems and experiments are correct, denoising variational auto-encoders is a much needed idea. It provides great intuition for why one should use hierarchical instantiations of variational models. Moreover it’s always interesting to see the same ideas inspired from a different field. These intuitions especially go nicely with our VGP paper, in which we make more explicit the auto-encoder connection for learning variational models with latent variables. Maybe we’ll add these details in a future revision.

Minor point: They claim that injecting noise both at the data input level (“denoising auto-encoder”) and the stochastic hidden layer level (“variational auto-encoder”) can be advantageous. However, they’re really one and the same, which I’m sure they’re aware of as they wrote Lemma 1 in this light. That is, the corrupted hidden variable is just another stochastic hidden layer, which is marginalized out. This makes the abstract and introduction a little confusing.

¹ In the proof of Lemma 1, when they try to show that upper bounds , there are three expectations expanded out in the first line. If the two expectations regarding the distribution are more explicitly written, it is clear why the ’s do not necessarily cancel:

When discussing this with Rajesh Ranganath, he points out a simple counterexample where : Consider a one-dimensional latent variable model, with . Let

Then is Bernoulli with probability , and

But , implying . In fact, taking 0.9 in the variational parameter to be arbitrarily close to 1 makes explode.