Difference between revisions of "Deep Learning Seminar 2018"
(→Seminars Aims & Organization) |
(→Seminars Aims & Organization) |
||
Line 17: | Line 17: | ||
'''Speaker''': Jonathan Masci, NNAISENSE | '''Speaker''': Jonathan Masci, NNAISENSE | ||
+ | |||
'''Title''': Deep Learning on Graphs and Manifolds | '''Title''': Deep Learning on Graphs and Manifolds | ||
+ | |||
'''Abstract''': | '''Abstract''': | ||
Line 23: | Line 25: | ||
'''Speaker''': Francesco Visin, DeepMind | '''Speaker''': Francesco Visin, DeepMind | ||
+ | |||
'''Title''': Graph Networks | '''Title''': Graph Networks | ||
+ | |||
'''Abstract''': | '''Abstract''': |
Revision as of 19:12, 16 April 2018
Seminars Aims & Organization
Deep Learning is nowadays becoming the dominant approach in learning cognitive systems which are nowdays able to recognize patterns in data (e.g., images, text, and sounds) or to perform end to end learning of complex behaviors.
In this first edition of the Deep Learning Seminars, held at Politecnico di Milano on the 23rd of February 2018, three invited speakers will present their research on new trends on learning in non Euclidean spaces.
Speaker: Luigi Malagò, Machine Learning and Optimization Group at the Romanian Institute of Science and Technology (RIST)
Title: Variational AutoEncoder: An Introduction and Recent Perspectives
Abstract: Variational AutoEncoders are generative models which consist of two neural networks. The first one is an encoder, whose purpose is to map the inputs to the parameters of a probability density function over the latent space, the second one in cascade is a decoder, which maps latent variables to probability density functions over the observation space. Variational AutoEncoders are usually trained using variational inference approaches, in particular by maximizing a lower-bound for the log-likelihood of the model, since training the model directly by optimizing the log-likelihood, is not computationally efficient. The current research in this field covers different directions, among which we have the use of richer models over the latent variables and the definition of sharper bounds for the loss function. In this presentation we focus our attention on the characterization of the geometrical properties of the statistical model for the latent variables learned during training. In our work we are interested is exploiting the intrinsic properties of the latent model, to evaluate the impact of the hyper-parameters of a Variational AutoEncoder over the learned geometry, and at the same time get insights for the design of more robust and efficient training procedures.
Speaker: Jonathan Masci, NNAISENSE
Title: Deep Learning on Graphs and Manifolds
Abstract:
Speaker: Francesco Visin, DeepMind
Title: Graph Networks
Abstract: