Conference item icon

Conference item

Monte Carlo variational auto-encoders

Abstract:
Variational auto-encoders (VAE) are popular deep latent variable models which are trained by maximizing an Evidence Lower Bound (ELBO). To obtain tighter ELBO and hence better variational approximations, it has been proposed to use importance sampling to get a lower variance estimate of the evidence. However, importance sampling is known to perform poorly in high dimensions. While it has been suggested many times in the literature to use more sophisticated algorithms such as Annealed Importance Sampling (AIS) and its Sequential Importance Sampling (SIS) extensions, the potential benefits brought by these advanced techniques have never been realized for VAE: the AIS estimate cannot be easily differentiated, while SIS requires the specification of carefully chosen backward Markov kernels. In this paper, we address both issues and demonstrate the performance of the resulting Monte Carlo VAEs on a variety of applications.
Publication status:
Published
Peer review status:
Peer reviewed

Actions


Access Document


Files:
Publication website:
http://proceedings.mlr.press/v139/thin21a.html

Authors



Publisher:
Journal of Machine Learning Research
Pages:
7258-7267
Series:
Proceedings of Machine Learning Research
Series number:
139
Publication date:
2021-07-01
Acceptance date:
2021-05-08
Event title:
38th International Conference on Machine Learning (ICML 2021)
Event location:
Virtual event
Event website:
https://icml.cc/Conferences/2021
Event start date:
2021-07-18
Event end date:
2021-07-24
ISSN:
2640-3498


Language:
English
Keywords:
Pubs id:
1258138
Local pid:
pubs:1258138
Deposit date:
2022-12-08

Terms of use



Views and Downloads






If you are the owner of this record, you can report an update to it here: Report update to this record

TO TOP