Deep Learning Seminar 2018

From Chrome
Revision as of 19:05, 16 April 2018 by Matteo (Talk | contribs)

Jump to: navigation, search

Seminars Aims & Organization

Deep Learning is nowadays becoming the dominant approach in learning cognitive systems which are nowdays able to recognize patterns in data (e.g., images, text, and sounds) or to perform end to end learning of complex behaviors.

In this first edition of the Deep Learning Seminars, held at Politecnico di Milano on the 23rd of February 2018,


Speaker: Luigi Malagò, Title: Variational AutoEncoder: An Introduction and Recent Perspectives Abstract: Variational AutoEncoders are generative models which consist of two neural networks. The first one is an encoder, whose purpose is to map the inputs to the parameters of a probability density function over the latent space, the second one in cascade is a decoder, which maps latent variables to probability density functions over the observation space. Variational AutoEncoders are usually trained using variational inference approaches, in particular by maximizing a lower-bound for the log-likelihood of the model, since training the model directly by optimizing the log-likelihood, is not computationally efficient. The current research in this field covers different directions, among which we have the use of richer models over the latent variables and the definition of sharper bounds for the loss function. In this presentation we focus our attention on the characterization of the geometrical properties of the statistical model for the latent variables learned during training. In our work we are interested is exploiting the intrinsic properties of the latent model, to evaluate the impact of the hyper-parameters of a Variational AutoEncoder over the learned geometry, and at the same time get insights for the design of more robust and efficient training procedures.