Skip to content

Introductory Course to the field of Deep Learning, commonly known as Deep Neural Networks. Throughout this course, participants will delve into the theoretical foundations of Deep Learning models, understanding their mechanisms and potential applications.

Notifications You must be signed in to change notification settings

eealanoca/deep_learning_course

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 

Repository files navigation

Intro to Deep Learning

Introductory Course to the field of Deep Learning, commonly known as Deep Neural Networks. Throughout this course, participants will delve into the theoretical foundations of Deep Learning models, understanding their mechanisms and potential applications. By the end of the program, attendees will have acquired the skills to construct and train models to address real-world problems.

Requirements

Installing PyTorch (on Google Server)

Go to https://colab.research.google.com and sign in with your Google account. If you do not have a Google account you can create one. From there you can create a new notebook.

Install using Conda (for local installation)

pip3 install torch torchvision torchaudio

Course Organization

1. Fundamentals

Introduction, AI vs ML vs DL. Why DL now?

1.1. Modern Neural Networks

  • Perceptron, activation functions, and matrix representation.
  • UAT, Feed-Forward Networks, and output function (softmax).
  • Gradient Descent to find network parameters.
  • Computational graphs and the BackPropagation algorithm.
  • Tensors, Einstein Notation, and Tensor Chain Rule.
  • Cross-Entropy and Backpropagation with Tensors.
  • Practical aspects of training and FF Network in pytorch.

Readings: Chapter 2. Linear Algebra, Chapter 3. Probability and Information Theory, Chapter 6. Deep Feedforward Networks

1.2. Initialization, Regularization, and Optimization

  • Generalization, Test-Dev-Train set, and Regularization.
  • Ensemble, Dropout, and Gradient Vanishing.
  • Parameter Initialization and Normalization.
  • Optimization Algorithms, SGD with Momentum, RMSProp, Adam.

Readings: Chapter 7. Regularization for Deep Learning, Chapter 8. Optimization for Training Deep Models, Chapter 11. Practical Methodology

2. Convolutional Neural Networks (CNN)

  • Introduction to Convolutional Networks.
  • Well-known architectures: AlexNet, VGG, GoogLeNet, ResNet, DenseNet.

Readings: Chapter 9. Convolutional Networks, Chapter 12. Applications

3. Recurrent Neural Networks (RNN)

  • Introduction to Recurrent Networks.
  • Recurrent Network Architecture.
  • Auto-regression, Language Modeling, and Seq-to-Seq Architectures.
  • Gated and Memory Cell RNNs: GRU and LSTM.

Readings: Chapter 10. Sequence Modeling: Recurrent and Recursive Nets, Chapter 12. Applications

4. Advanced Topics

  • Neural Attention.
  • Transformers.
  • Variational Autoencoders.
  • Generative Adversarial Networks.
  • Vision Transformers (ViT).

Readings: Chapter 14. Autoencoders, Chapter 20. Deep Generative Models, An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale

5. Reinforcement Learning

  • Markov decision processes, returns, and policies.
  • Markov process, Markov reward process, Markov decision process, Partially observable Markov decision process, Policy.
  • Expected return.
  • State and action values, Optimal policy, Bellman equations.
  • Tabular reinforcement learning.
  • Dynamic programming, Monte Carlo methods, Temporal difference methods.
  • Fitted Q-learning.
  • Double Q-learning and double deep Q-network.
  • Policy gradient methods.
  • Derivation of gradient update REINFORCE algorithm, Baselines, State-dependent baselines.
  • Actor-critic methods.
  • Offline reinforcement learning.

Readings: Chapters 2 - 13. Reinforcement Learning: An Introduction

Books

There is no mandatory textbook for the course. Some lectures will include suggested readings from "Deep Learning" by Ian Goodfellow, Yoshua Bengio, and Aaron Courville; however, it is not necessary to purchase a copy as it is available online for free.

  1. Deep Learning by Ian Goodfellow, Yoshua Bengio, and Aaron Courville (fundamental bibliography for the course)
  2. Deep Learning for Vision Systems by Mohamed Elgendy
  3. Probabilistic and Statistical Models for Outlier Detection by Charu Aggarwal
  4. Pattern Recognition and Machine Learning by Chistopher M. Bishop.
  5. Speech and Language Processing by Daniel Jurafsky and James Martin
  6. Notes on Deep Learning for NLP by Antoine J.-P. Tixier
  7. AutoML: Methods, Systems, Challenges edited by Frank Hutter, Lars Kotthoff, and Joaquin Vanschoren
  8. Reinforcement Learning: An Introduction by Richard S. Sutton and Andrew G. Barto.

About

Introductory Course to the field of Deep Learning, commonly known as Deep Neural Networks. Throughout this course, participants will delve into the theoretical foundations of Deep Learning models, understanding their mechanisms and potential applications.

Topics

Resources

Stars

Watchers

Forks