Advances Deep Learning
06/02/2019: Rooms for lectures have been changed due to increased, with respect to planned, audience!!!! 06/02/2019: This website is online, please check for material, updates, and announcements here!
Contents
Course Aim and Objectives
Nowadays deep learning spans multiple fields in science and engineering, from autonomous driving to human machine interaction. Deep networks have outperformed traditional hand-crafted algorithms, and achieved human performance in solving many complex tasks, such as natural language processing and image recognition. A plethora of papers presenting the success of deep learning in different scenarios is steadily being published, and most often papers frame on a few, very successful, architectures. These architectures are nowadays becoming de facto standards in deep learning such as: convolutional neural networks (CNN), long-short time memories (LSTM), generative adversarial networks (GAN), graph networks, to name a few examples.
Our goal is to provide the students with the skills to understand, become familiar, and use for their research the most successful architectural patterns in deep neural networks. This is intended as an advanced course, thus proficiency in neural networks and basic notions of non linear optimization, i.e., backpropagation, and image/signal processing are assumed as pre-requirement to the participants. For instance, teachers will provide the basics to understand recurrent neural networks and convolutional networks, but feed forward architectures and backpropagation are assumed to be well known notions.
The course presents a general and unified view over most successful architectural patterns in deep neural networks, and the machine learning problems they are naturally suited for. Each lecture will be organized around one of the most relevant deep learning architectures or learning problems (described below), and will be structured as follows:
- a formulation of the addressed problem and brief recall of the required notions and mathematical background when needed
- a unified presentation of the considered model and possible theoretical results
- an overview of the most relevant/successful applications of the considered model
- a brief hands-on demo session to show how to use these models in Python
- an overview of research directions and papers presenting the most important achievements in the literature
These lectures will be planned in a coordinated manner to:
- better exploit the different background of the instructors (Matteucci: Machine Learning and Graphical Models, Boracchi: Image Analysis and Processing, Giusti: Image Classification and Applications)
- better present how these models interact to address the most challenging applications
Detailed Schedule of Lectures
The course is organized in 4 lectures (5 hours each) where the teachers illustrate the models and the theory behind them, and 2 seminars (planned as half day so far) where PhD students, as part of the evaluation, present some papers to the class.
Note: Seminars are part of the lectures an thus they count for PhD attendance.
- Recurrent deep architectures and NLP (07/02/2019 from 09:00 to 14:00 in N.1.2): models for sequence to sequence learning and hierarchical sequence to sequence modeling, long-short term memories, gated recurrent units, memory networks, attention mechanism in natural language and video sequences processing.
- Convolutional Neural Networks for advanced visual recognition tasks (14/02/2019 from 09:00 to 14:00 in N.1.2): Convolutional neural networks, fully convolutional networks for segmentation, object detection, region-proposal networks, semantic annotation, adversarial examples, image captioning, understanding what CNNs learn.
- Unsupervised deep models (21/02/2019 from 09:00 to 14:00 in N.1.2): Generative Models (Pixel RNN/CNN), Variational Autoencoders, Generative Adversarial Networks. Image Restoration through Deep-Learning models, adversarial examples attack.
- Graph Models (28/02/2019 from 09:00 to 14:00 in Sala Seminari - Ed. 20): Graph convolutional networks, graph to vector, vector to graph, graph to graph modeling, applications in shape analysis and non Euclidean spaces.
- Student Seminars 1 (14/03/2019 from 09:00 to 18:00 in Sala Conferenze - Ed. 20): PhD students seminars on advanced visual recognition tasks and unsupervised deep models
- Student Seminars 2 (28/03/2019 from 09:00 to 14:00 in Aula Seminari Alessandra Alario Ed. 21): PhD students seminars on Graphical Models and recurrent deep architectures
Course Evaluation
Final assessment differs for PhD students, who will be evaluated on a Pass or Fail scale, and MS students, who will be evaluated on a 30 grades scale for their 5 CFU exams. In particular:
- PhD students will be evaluated on class participation (75% of lectures attendance is mandatory) and on a seminar on the papers proposed by the teachers after the lectures.
- MS students will be evaluated on a small group project to be agreed with the teachers according to this schedule:
- At the end of each the seminar day, according to the topic, a group of students presents the project to the audience (topic + dataset + model expected to be used + evaluation criteria) and receives comments.
- By July each group presents a paper (up to 6 papges) reporting the outcome of the project (topic + dataset + modl actually used + comparison with respect to state of art + discussion of results)
- End of July, or end of August, each group discuss the paper with the teachers
The project option is available also for PhD students not presenting a paper on the defined dates upon request.
Course Material
The course material includes the teachers slides, references to the literature, and the papers to base seminars upon.
Teachers Slides
- Course Intro: Course aims and objectives, course organization, schedule, evaluation procedures and the lik.
- Recurrent Neural Networks: Introduction to sequence modeling, recurrent neural networks, vaniscing gradient, long short-term memories
- Sequence 2 Sequence: Models for Sequence to Sequence modeling, attention mechanisms, Neural Turing Machines
Reference to Literature
... coming soon ...