Skip to content

Latest commit

 

History

History
186 lines (136 loc) · 6.45 KB

seminar-topics.md

File metadata and controls

186 lines (136 loc) · 6.45 KB

Legend for book sources

  • PDL: Probabilstic Deep Learning models
  • SR: Statistical Rethinking: A Bayesian Course with Examples in R and Stan (Second Edition)
  • TF Machine Learning with Tensor Flow

Classical Deep Learning

Deep Learning Frameworks (Sven Laur)

Concepts

  • Hardware model behind Deep learning
  • Tensorflow
  • Pytorch

Additional materials

  • ??

Feed-Forward neural network architectures (PDL 25-61, TF 276-294)

Concepts

  • Fully connected neural networks
  • RELU and sigmoid neurons
  • Max pooling
  • Convolutional neural networks

Additional materials

  • ??

Recurrent neural network architectures (TF 335-364)

Concepts

  • LSTM

Additional materials:

Optimisation methods (PDL 63-90)

Concepts

  • Learning rate
  • Gradient decent
  • Stochastic gradient decent
  • Automatic differentation
  • Static and dynamic backpropagation
  • Hyperparameter optimisation

Additional materials

  • ??

Maximum likelihood and loss functions (PDL 91-127)

Concepts

  • Maximum likelihood
  • Maximum aposteriori probability
  • Loss function derivation
  • Crossentropy and negative log-likelihood
  • Kullback-Leibler divergence
  • Mean square error and negative log-likelihood

Additional materials about the interpretation of Kullback-Leibler divergence

Attaching probability to outcomes

Main method: Use data to fix a single model and use the model to assign probabilities to observations. Does not work well if there are many near optimal models with widely different predictions.

Probabilistic models for continuous data (PDL129-145)

Concepts

  • Negative log-likelihood as loss
  • Prediction intervals from predicted distribution
  • Models with varying error terms aka Heteroskedasticity
  • Predicting mean and variance

Additional materials

  • ??

Probabilistic models for count data (PDL 145 - 156, SR 323-380)

Concepts

  • Binomial distribution
  • Poisson distribution
  • Zero-Inflated Poisson distribution
  • Negative binomial distribution
  • Logistic regression
  • Poisson regression
  • Diagnostic methods

Additional materials

Mixture models (PDL 157-166, SR 359-366, 369-397)

Concepts

  • Multinomial distribution
  • Discretised logistic mixture distribution
  • Regression with discretised logistic mixture distribution
  • WaveNet and PixelCNN

Additional materials

Normalising flows (PDL 166 - 193)

  • Transformation functions (bijectors)
  • Probability density function and Jacobian
  • Maximum likelihood to estimate parameters of Transformations
  • Neural networks as Transformation functions
  • Glow model for faces

Additional materials

Working with posteriors

Main method Use data to fix the weight of individual models and use these to average over predictions. Allow to measure the uncertainty due to variability of training data.

Basics of Bayesian inference (PDL 197-228)

Concepts

  • Model averaging
  • Bayes formula and corresponding inference rules
  • Coin-tossing example
  • Bayesian linear regression model

Additional materials

Variational inference (PDL 229-245)

Concepts

  • Kullback-Leibler difference
  • Parametric posterior approximation
  • Variational inference for a single neuron
  • Bayes Backprop algorithm in practice
  • Stochastic Variational Inference

Additional materials

Monte-Carlo dropout (PDL 245-263)

Concepts

  • Dropout layer as a regulariser
  • Monte-Carlo dropout architecture
  • Uncertainty measures for Bayesian classification

Additional materials

Markov-Chain-Monte-Carlo methods (SR 263-296)

Concepts

  • Metropolis algorithm
  • Gibbs sampling
  • Hamiltonian Monte Carlo
  • Adaptation, burn-in and convergence diagnostics

Additional materials