🐦 Repository for Kording lab teaching session by Titipat and friends
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Type Name Latest commit message Commit time
Failed to load latest commit information.


Kording lab teaching 2015

Repository for lab teaching at Kording's Lab.

Session 1

Presentation covers how to get up running Spark on Quest (Northwestern Computing Clusters). Notebook goes through Map Reduce algorithm where we provide an example on a simple text file. We then implement stochatic gradient descent to train Logistic Regression in order to classify MNIST dataset. Here is a link to a presentation file.

Session 2

This tutorial is one of the Deep learning tutorial by Kording lab. In this presentation, we will go through simple neural network architecture and math behind it including forward propagation and back propagation (ref. Andrew Ng lecture). We will then code the forward propagation and back propagation algorithm and test it on MNIST examples. If we have time, we will go through existing python packages that implement Neural Network (e.g. Lasagne, scikit-neuralnetwork, pybrain, nolearn.

Session 3

In this session, we will go over Recurrent Neural Network (RNN) and its application in Natural Language Processing (NLP). We'll go through architecture of the RNN also a bit of Convolutional Neural Network (CNN) presented by Pavan. We use blog posts and CS224d: Deep Learning for Natural Language Processing as main materials. After the lecture, We will go over implementing RNN for application in NLP.

Session 4 (Luca Lonini and Daniel Acuna)

In this tutorial we show how to use TensorFlow to build and train 2 models: a Softmax Regression Model and a CNN. The example shows how to train the networks and how save the variables and the computation graph for display in the TensorFlow board. Further info can be found here: TensorFlow.org

Session 5 (Pavan Ramkumar and Luca Lonini)

In this session, we go through various kind of ConvNet visualization techniques. We focus on techniques to visualize the entire neural code, layerwise neural codes, individual units, as well as individual layers and units conditioned on specific images.

Here is a list of papers we discuss.

We also provide an iTorch demo of how to specify and train a conventional CNN (LeNet) using Torch.