About

Log in?

DTU users get better search results including licensed content and discounts on order fees.

Anyone can log in and get personalized features such as favorites, tags and feeds.

Log in as DTU user Log in as non-DTU user No thanks

DTU Findit

Conference paper

How to Train Deep Variational Autoencoders and Probabilistic Ladder Networks

From

University of Copenhagen1

Aalto University2

Department of Applied Mathematics and Computer Science, Technical University of Denmark3

Cognitive Systems, Department of Applied Mathematics and Computer Science, Technical University of Denmark4

Variational autoencoders are a powerful framework for unsupervised learning. However, previous work has been restricted to shallow models with one or two layers of fully factorized stochastic latent variables, limiting the flexibility of the latent representation. We propose three advances in training algorithms of variational autoencoders, for the first time allowing to train deep models of up to five stochastic layers, (1) using a structure similar to the Ladder network as the inference model, (2) warm-up period to support stochastic units staying active in early training, and (3) use of batch normalization.

Using these improvements we show state-of-the-art log-likelihood results for generative modeling on several benchmark datasets.

Language: English
Year: 2016
Proceedings: 33rd International Conference on Machine Learning (ICML 2016)International Conference on Machine Learning
Series: Jmlr: Workshop and Conference Proceedings
ISSN: 19387228
Types: Conference paper
ORCIDs: Winther, Ole

DTU users get better search results including licensed content and discounts on order fees.

Log in as DTU user

Access

Analysis