Difference between revisions of "Artificial Neural Networks and Deep Learning"

From Chrome
Jump to: navigation, search
Line 96: Line 96:
  
 
* [https://www.deib.polimi.it/ita/personale/dettagli/267262 Matteo Matteucci]: the course teacher <!--and this is his [https://politecnicomilano.webex.com/join/matteo.matteucci webex room]-->
 
* [https://www.deib.polimi.it/ita/personale/dettagli/267262 Matteo Matteucci]: the course teacher <!--and this is his [https://politecnicomilano.webex.com/join/matteo.matteucci webex room]-->
* [https://www.deib.polimi.it/ita/personale/dettagli/549640 Giacomo Boracchi]: the course co-teacher and this is his [https://politecnicomilano.webex.com/join/giacomo.boracchi webex room]  
+
* [https://www.deib.polimi.it/ita/personale/dettagli/549640 Giacomo Boracchi]: the course co-teacher <!-- and this is his [https://politecnicomilano.webex.com/join/giacomo.boracchi webex room]-->
* [https://www.deib.polimi.it/ita/personale/dettagli/846174 Francesco Lattari]: the course teaching assistant and this is his [https://politecnicomilano.webex.com/join/francesco.lattari webex room]
+
* [https://www.deib.polimi.it/ita/personale/dettagli/846174 Francesco Lattari]: the course teaching assistant <!-- and this is his [https://politecnicomilano.webex.com/join/francesco.lattari webex room]-->
 
* [https://www.deib.polimi.it/ita/personale/dettagli/920490 Eugenio Lomurno]: the course teaching assistant and this is his [https://politecnicomilano.webex.com/join/eugenio.lomurno webex room]
 
* [https://www.deib.polimi.it/ita/personale/dettagli/920490 Eugenio Lomurno]: the course teaching assistant and this is his [https://politecnicomilano.webex.com/join/eugenio.lomurno webex room]
  
Line 109: Line 109:
 
* Providing an overview of the most successful applications with particular emphasis on models for solving visual recognition tasks.
 
* Providing an overview of the most successful applications with particular emphasis on models for solving visual recognition tasks.
  
We have compiled a detailed syllabus of the course student can use to double check their preparation against before the exam.  
+
We have compiled a detailed syllabus of the course that students can use to double-check their preparation before the exam.  
 
* [[Media:AN2DL_Syllabus_2021.pdf|[2020/2021] Course Syllabus]]: a detailed list of topics covered by the course and which students are expected to know when approaching the exam
 
* [[Media:AN2DL_Syllabus_2021.pdf|[2020/2021] Course Syllabus]]: a detailed list of topics covered by the course and which students are expected to know when approaching the exam
 
<!--* [[Media:AN2DL_Syllabus.pdf|[2019/2020] Course Syllabus]]: a detailed list of topics covered by the course and which students are expected to know when approaching the exam-->
 
<!--* [[Media:AN2DL_Syllabus.pdf|[2019/2020] Course Syllabus]]: a detailed list of topics covered by the course and which students are expected to know when approaching the exam-->
Line 115: Line 115:
 
===Detailed course schedule===
 
===Detailed course schedule===
  
A detailed schedule of the course can be found here; topics are just indicative while days and teachers are correct up to some last minute change (I will notify you by email). Please note that not all days we have lectures!!
+
A detailed schedule of the course can be found here; topics are just indicative while days and teachers are correct up to some last-minute change (I will notify you by email). Please note that not all days we have lectures!!
  
 
  Note: Lecture timetable interpretation
 
  Note: Lecture timetable interpretation
  * On Wednesday, in T.2.1, Team 1, starts at 15:15, ends at 17:00
+
  * On Wednesday, in 9.0.1, starts at 16:30, ends at 18:00  
  * On Wednesday, in T.2.1, Team 2, starts at 17:30, ends at 19:15
+
  * On Thursday, in 20.S.1, starts at 14:30, ends at 16:00
* On Thursday, in teacher webex room, starts at 16:30, ends at 19:15
+
  
Note: Teams division is based on your Codice Persona (and should minimize overlap)
+
Lectures will be recorded and shared afterward, no streaming of lectures is foreseen.
* Team 1: odd Codice Persona
+
 
* Team 2: even Codice Persona
+
For a [https://boracchi.faculty.polimi.it/teaching/AN2DLCalendar_CS.htm course google calendar] you might look here!
  
 
<!--
 
<!--
Line 370: Line 369:
 
* 2 home projects in the form of a "Kaggle style" challenge practicing the topics of the course graded up to 5/30 each
 
* 2 home projects in the form of a "Kaggle style" challenge practicing the topics of the course graded up to 5/30 each
  
The final score will sum the grade of the written exam and the grade of the home projects. Home projects are not compulsory and they are issued only once a year.
+
The final score will sum up the grade of the written exam and the grade of the home projects. Home projects are not compulsory and they are issued only once a year.
 
<!--
 
<!--
 
You can find here one [[Media:AN2DL_ExamExample.pdf|example of the exam text]] to get a flavor of what to expect in the written examination.
 
You can find here one [[Media:AN2DL_ExamExample.pdf|example of the exam text]] to get a flavor of what to expect in the written examination.
Line 377: Line 376:
 
==Teaching Material (the textbook)==  
 
==Teaching Material (the textbook)==  
  
Lectures will be based on material from different sources, teachers will provide their slides to students as soon they are available. As a general reference you can check the following text, but keep in mind that teachers will not follow it strictly
+
Lectures will be based on material from different sources, teachers will provide their slides to students as soon they are available. As a general reference, you can check the following textbook, but keep in mind that teachers will not follow it strictly
  
 
* [http://www.deeplearningbook.org/ Deep Learning]. Ian Goodfellow, Yoshua Bengio, and Aaron Courville, MIT Press, 2016.
 
* [http://www.deeplearningbook.org/ Deep Learning]. Ian Goodfellow, Yoshua Bengio, and Aaron Courville, MIT Press, 2016.
 
<!-- The remaining material about the course is available through [http://webeep.polimi.it WeBeep].-->
 
<!-- The remaining material about the course is available through [http://webeep.polimi.it WeBeep].-->
  
Regarding the python programming language, we will provide you the basics about numpy and python scripting in case you want some introductory material you can check here
+
Regarding the Python programming language, we will provide you with the basics about NumPy and python scripting in case you want some introductory material you can check here
  
 
* [https://docs.python.org/3/tutorial/ Python tutorials]: these are the official python tutorials, we suggest 3.An Informal Introduction to Python (Numbers, Strings, Lists), 4.More Control Flow Tools (if, for, range, functions), 5.Data Structures (More on lists, Dictionaries), 9.Classes.
 
* [https://docs.python.org/3/tutorial/ Python tutorials]: these are the official python tutorials, we suggest 3.An Informal Introduction to Python (Numbers, Strings, Lists), 4.More Control Flow Tools (if, for, range, functions), 5.Data Structures (More on lists, Dictionaries), 9.Classes.
Line 397: Line 396:
 
*[[Media:AN2DL_06_2021_BeyondSeq2Seq.pdf|[2021/2022] Beyond Sequence 2 Sequence Learning]]: Neural Turing Machines, attention mechanisms, the Transformer.
 
*[[Media:AN2DL_06_2021_BeyondSeq2Seq.pdf|[2021/2022] Beyond Sequence 2 Sequence Learning]]: Neural Turing Machines, attention mechanisms, the Transformer.
  
Slides from the lectures by Giacomo Boracchi are available in [https://boracchi.faculty.polimi.it/teaching/AN2DL.htm his webpage]; for your convenience I am giving pointers to the slide here for you (in case you note discrepancies please notify me)
+
Slides from the lectures by Giacomo Boracchi are available in [https://boracchi.faculty.polimi.it/teaching/AN2DL.htm his webpage]; for your convenience, I am giving pointers to the slide here for you (in case you note discrepancies please notify me)
 
* [2021/2022] [https://boracchi.faculty.polimi.it/teaching/AN2DL/2021_AN2DL_Lez1_ImageClassification.pdf The Image Classification Problem]
 
* [2021/2022] [https://boracchi.faculty.polimi.it/teaching/AN2DL/2021_AN2DL_Lez1_ImageClassification.pdf The Image Classification Problem]
 
* [2021/2022] [https://boracchi.faculty.polimi.it/teaching/AN2DL/2021_AN2DL_Lez2_CNN.pdf Convolutional Neural Networks]
 
* [2021/2022] [https://boracchi.faculty.polimi.it/teaching/AN2DL/2021_AN2DL_Lez2_CNN.pdf Convolutional Neural Networks]
Line 406: Line 405:
 
* [2021/2022] [https://boracchi.faculty.polimi.it/teaching/AN2DL/2021_AN2DL_Lez6_CNN_for_object_detection_Metric_Learning.pdf Object Detction Networks and Metric Learning]
 
* [2021/2022] [https://boracchi.faculty.polimi.it/teaching/AN2DL/2021_AN2DL_Lez6_CNN_for_object_detection_Metric_Learning.pdf Object Detction Networks and Metric Learning]
  
Slides from the practicals by Francesco Lattari and Eugenio Lomurno
+
Slides from the practicals by Francesco Lattari and Eugenio Lomurno will be published here after each lab session:
* [https://drive.google.com/drive/folders/1zO_E89H3mjc7nysBkGUt_MmCn1Mb2DzP?usp=sharing Python and convolutions]: a Colab notebook with a crash course on Python 3 to prepare for the practicals and extra material about convolutions!
+
<!--* [https://drive.google.com/drive/folders/1zO_E89H3mjc7nysBkGUt_MmCn1Mb2DzP?usp=sharing Python and convolutions]: a Colab notebook with a crash course on Python 3 to prepare for the practicals and extra material about convolutions!
 
* [https://drive.google.com/drive/folders/1-WFu2S-MBOrNsFRy1IWEptF_hTMLmic9?usp=sharing Feed Forward Neural Networks]: NumPy Arrays, TensorFlow2 Tensors, and Feed Forward Neural Network
 
* [https://drive.google.com/drive/folders/1-WFu2S-MBOrNsFRy1IWEptF_hTMLmic9?usp=sharing Feed Forward Neural Networks]: NumPy Arrays, TensorFlow2 Tensors, and Feed Forward Neural Network
 
* [https://drive.google.com/drive/folders/1-gaItYI8i01STVPrUcqZ8vxK27yGFENn?usp=sharing Dealing with Overfitting]: Hold Out, Early Stopping and Cross-Validation
 
* [https://drive.google.com/drive/folders/1-gaItYI8i01STVPrUcqZ8vxK27yGFENn?usp=sharing Dealing with Overfitting]: Hold Out, Early Stopping and Cross-Validation
Line 414: Line 413:
 
* [https://drive.google.com/drive/folders/1dxfaXo68dWnrS6xHY0yLvUDjTwDPFN6h?usp=sharing Autoencoders and Segmentation]: Exercises on Autoencoders and Multiclass Segmentation  
 
* [https://drive.google.com/drive/folders/1dxfaXo68dWnrS6xHY0yLvUDjTwDPFN6h?usp=sharing Autoencoders and Segmentation]: Exercises on Autoencoders and Multiclass Segmentation  
 
* [https://drive.google.com/drive/folders/1rZQUrs6Rb1dGg5-KLg-y9tT1yw09gr7G?usp=sharing Dealing with Time Series]: Recurrent Neural Networks, Long Short-Term Memories, and other models.
 
* [https://drive.google.com/drive/folders/1rZQUrs6Rb1dGg5-KLg-y9tT1yw09gr7G?usp=sharing Dealing with Time Series]: Recurrent Neural Networks, Long Short-Term Memories, and other models.
 +
-->
  
 
=== External Sources ===
 
=== External Sources ===
 
* [https://towardsdatascience.com/back-to-basics-deriving-back-propagation-on-simple-rnn-lstm-feat-aidan-gomez-c7f286ba973d Deriving Back Propagation on simple RNN/LSTM]: a tutorial step by step to derive the backpropagation formulas
 
* [https://towardsdatascience.com/back-to-basics-deriving-back-propagation-on-simple-rnn-lstm-feat-aidan-gomez-c7f286ba973d Deriving Back Propagation on simple RNN/LSTM]: a tutorial step by step to derive the backpropagation formulas
* [https://karpathy.github.io/2019/04/25/recipe/ A Recipe for Training Neural Networks]: a neat post on Andrej Karpathy blog about do's and don'ts in training neural networks
+
* [https://karpathy.github.io/2019/04/25/recipe/ A Recipe for Training Neural Networks]: a neat post on Andrej Karpathy's blog about do's and don'ts in training neural networks
 
+
  
 
<!--*[[Media:AN2DL_Lab1_2020_KerasIntroduction.zip|[2020/2021] Introduction to Keras]]: Introduction to Keras and Tensorflow2 (slides + notebook)
 
<!--*[[Media:AN2DL_Lab1_2020_KerasIntroduction.zip|[2020/2021] Introduction to Keras]]: Introduction to Keras and Tensorflow2 (slides + notebook)
Line 431: Line 430:
 
*[[Media:AN2DL_Lab3_2019_KerasRNN.zip|[2019/2020] Recurrent architectures in Keras]]: How to build, train, and evaluate recurrent neural architectures in Keras and Tensorflow2 (slides + notebook)
 
*[[Media:AN2DL_Lab3_2019_KerasRNN.zip|[2019/2020] Recurrent architectures in Keras]]: How to build, train, and evaluate recurrent neural architectures in Keras and Tensorflow2 (slides + notebook)
 
-->
 
-->
<!--
 
===Past Online Exams===
 
  
As for the last semester, Politecnico di Milano exams will be held online. We have not yet decided the digital environment we are going to use for the exams, but you can have a flavor of what to expect by looking at past semester exam calls.  
+
===Exams===
 +
 
 +
Politecnico di Milano exams will be held digitally in presence via the [https://remoteexam.polimi.it/ Politecnico di Milano remote exam platform] using the Safe Exams Browser, it works only on Windows and iOS so please be prepared and test it in advance. You can have a flavor of what to expect by looking at some past exam calls.  
 
   
 
   
 
*[[Media:AN2DL_20200619.pdf| 19/06/2020 Online Exam]]
 
*[[Media:AN2DL_20200619.pdf| 19/06/2020 Online Exam]]
Line 440: Line 439:
 
*[[Media:AN2DL_20200903.pdf| 03/09/2020 Online Exam]]
 
*[[Media:AN2DL_20200903.pdf| 03/09/2020 Online Exam]]
  
'''Note''': this year written exams will be graded 26 points instead of 27. This is because the past year challenges were 5 points while this year they are 6 points worth. This means that the maximum mark is still 26+6=32. You can get an extra point if you participate in the second phase of the Image Segmentation Challenge thus the full mark could reach in principle 33. To get a laude it is required to pass 30.
+
'''Note''': written exams will be graded 20 points plus 10 points are given by 2 software challenges issues only during the semester. Laude will be given to students who, beside getting the highest grade, will show participation in class, will perform particularly well in the challenges (this includes the quality of the report), will submit ahead of time the written exams.  
  
 +
<!--
 
===Lab software setup===
 
===Lab software setup===
  

Revision as of 16:04, 12 September 2022


The following are last-minute news you should be aware of ;-)

07/08/2022: Grades for the 06/07/2022 call are available here!
06/03/2021: Grades for the 2021/2021 homeworks are available here!
24/11/2021: Lecture on 09/12/2021 moved to 16/12/2021
24/11/2021: Change of classroom on the 15/12 from T2.1 to B4.3
12/11/2021: Here is the link to the First Homework!!! (read the registration rules, they have been updated)
12/11/2021: First homework is coming out! Register to submit your solutions!
07/11/2021: We have restructured the notebooks from the labs, please check the new organization and material
13/10/2021: Final grade for 2020/2021 year are here
30/09/2021: Video and slides updated + notebooks published
24/09/2021: Removed last year practicals, this year they will be different!!! Python crash course moved at the end of the material
22/09/2021: Added colab crash course on Python 3 in the material session
18/09/2021: We skip exercising on 22/09 and added exercising on 15/12 I updated the schedule!
16/09/2021: Updated material on the first lectures + added reference to python tutorials in the material section
15/09/2021: Lectures start today!
14/09/2021: Website under maintenance ... come back later


Course Aim & Organization

Neural networks are mature, flexible, and powerful non-linear data-driven models that have successfully been applied to solve complex tasks in science and engineering. The advent of the deep learning paradigm, i.e., the use of (neural) network to simultaneously learn an optimal data representation and the corresponding model, has further boosted neural networks and the data-driven paradigm.

Nowadays, deep neural networks can outperform traditional hand-crafted algorithms, achieving human performance in solving many complex tasks, such as natural language processing, text modeling, gene expression modeling, and image recognition. The course provides a broad introduction to neural networks (NN), starting from the traditional feedforward (FFNN) and recurrent (RNN) neural networks, till the most successful deep-learning models such as convolutional neural networks (CNN) and long short-term memories (LSTM).

The course's major goal is to provide students with the theoretical background and the practical skills to understand and use NN, and at the same time become familiar and with Deep Learning for solving complex engineering problems.

Teachers

The course is composed of a blending of lectures and exercises by the course teachers and a teaching assistant.

Course Program and Syllabus

This goal is pursued in the course by:

  • Presenting major theoretical results underpinning NN (e.g., universal approx, vanishing/exploding gradient, etc.)
  • Describing the most important algorithms for NN training (e.g., backpropagation, adaptive gradient algorithms, etc.)
  • Illustrating the best practices on how to successfully train and use these models (e.g., dropout, data augmentation, etc.)
  • Providing an overview of the most successful Deep Learning architectures (e.g., CNNs, sparse and dense autoencoder, LSTMs for sequence to sequence learning, etc.)
  • Providing an overview of the most successful applications with particular emphasis on models for solving visual recognition tasks.

We have compiled a detailed syllabus of the course that students can use to double-check their preparation before the exam.

  • [2020/2021] Course Syllabus: a detailed list of topics covered by the course and which students are expected to know when approaching the exam

Detailed course schedule

A detailed schedule of the course can be found here; topics are just indicative while days and teachers are correct up to some last-minute change (I will notify you by email). Please note that not all days we have lectures!!

Note: Lecture timetable interpretation
* On Wednesday, in 9.0.1, starts at 16:30, ends at 18:00 
* On Thursday, in 20.S.1, starts at 14:30, ends at 16:00

Lectures will be recorded and shared afterward, no streaming of lectures is foreseen.

For a course google calendar you might look here!


Course Evaluation

Course evaluation is composed of two parts:

  • A written examination covering the whole program graded up to 20/30
  • 2 home projects in the form of a "Kaggle style" challenge practicing the topics of the course graded up to 5/30 each

The final score will sum up the grade of the written exam and the grade of the home projects. Home projects are not compulsory and they are issued only once a year.

Teaching Material (the textbook)

Lectures will be based on material from different sources, teachers will provide their slides to students as soon they are available. As a general reference, you can check the following textbook, but keep in mind that teachers will not follow it strictly

  • Deep Learning. Ian Goodfellow, Yoshua Bengio, and Aaron Courville, MIT Press, 2016.

Regarding the Python programming language, we will provide you with the basics about NumPy and python scripting in case you want some introductory material you can check here

  • Python tutorials: these are the official python tutorials, we suggest 3.An Informal Introduction to Python (Numbers, Strings, Lists), 4.More Control Flow Tools (if, for, range, functions), 5.Data Structures (More on lists, Dictionaries), 9.Classes.

Course Slides

Slides from the lectures by Matteo Matteucci

Slides from the lectures by Giacomo Boracchi are available in his webpage; for your convenience, I am giving pointers to the slide here for you (in case you note discrepancies please notify me)

Slides from the practicals by Francesco Lattari and Eugenio Lomurno will be published here after each lab session:

External Sources


Exams

Politecnico di Milano exams will be held digitally in presence via the Politecnico di Milano remote exam platform using the Safe Exams Browser, it works only on Windows and iOS so please be prepared and test it in advance. You can have a flavor of what to expect by looking at some past exam calls.

Note: written exams will be graded 20 points plus 10 points are given by 2 software challenges issues only during the semester. Laude will be given to students who, beside getting the highest grade, will show participation in class, will perform particularly well in the challenges (this includes the quality of the report), will submit ahead of time the written exams.