Skip to content

Deep Learning model training visualized using mutual information between the input and the output

License

Notifications You must be signed in to change notification settings

liadmagen/DLIT_WC

Repository files navigation

Investigating Deep Learning models training using Information Theory visualization (Web Components version)

This research project aim is to investigate and learn ways to improve the general accuracy of deep learning models on an unseen data.

Motivation

while using percision/F1-Statistic score as an objective metric, mutual information of the hidden layer is being examined during the training processs. Preplexity should b

  • To find out how much data is 'enough' for a Machine Learning Model (For example, Deep Learning) in order to achieve good generalized resulsts -
  • Which characteristics are relevant for the input data (i.e. Variance) in order to achieve better scores?
  • Does running a model through multiple epochs really compress the model and overcome the bias/variance problem (prevents overfitting)?

Related work

This work is heavily based on the work and theories of Prof. Geoff I Webb (Monash University) and Prof. Naftali Tishby (Hebrew University).

We focus on Deep Learning models, using different network architectures, in order to achieve the following:

  • to provide an interactive visualization of Deep learning network's hidden layers training phase using mutual information
  • Storing data of different training executions results in order to be analyzed
  • To measure and analyze the impact of variance/Std Deviation differences of the training data, on the hold-out sets accuarcy scores

Background

Since 'Deep Learning' is practically a learnable mapping function between input and ouput, this project visualizes the training process, where the model slowly finds the correct mapping parameters, using stochastic gradient descent based algorithm, until reaching the maximal bound. Unlike other visualizations, where the layers are visualized directly or using dimension reduction techniques (http://colah.github.io/posts/2014-10-Visualizing-MNIST/), here we're interested in the general information that the hidden layers' nodes contain, regarding both the inputs and the outputs.

According to the information bottleneck theory (Tishbey, Schwartz-Ziv.), during the training process the mutual information about the input and the output is divided between the layer in a way that resemles a form of an encoder and a decoder, until the data compress at a specific point.

Lastly, it is important to let the model run long time, on many epochs, in order to test the hypothesis that the model can and will reach the upper information bottleneck theory bound.

Related research:

methods

Setup

Prerequisites

First, install Polymer CLI using npm (we assume you have pre-installed node.js).

npm install -g polymer-cli

Second, install Bower using npm

npm install -g bower
Initialize project from template
mkdir my-app
cd my-app
polymer init polymer-2-starter-kit

Start the development server

This command serves the app at http://127.0.0.1:8081 and provides basic URL routing for the app:

polymer serve

Build

The polymer build command builds your Polymer application for production, using build configuration options provided by the command line or in your project's polymer.json file.

You can configure your polymer.json file to create multiple builds. This is necessary if you will be serving different builds optimized for different browsers. You can define your own named builds, or use presets. See the documentation on building your project for production for more information.

The Polymer Starter Kit is configured to create three builds using the three supported presets:

"builds": [
  {
    "preset": "es5-bundled"
  },
  {
    "preset": "es6-bundled"
  },
  {
    "preset": "es6-unbundled"
  }
]

Builds will be output to a subdirectory under the build/ directory as follows:

build/
  es5-bundled/
  es6-bundled/
  es6-unbundled/
  • es5-bundled is a bundled, minified build with a service worker. ES6 code is compiled to ES5 for compatibility with older browsers.
  • es6-bundled is a bundled, minified build with a service worker. ES6 code is served as-is. This build is for browsers that can handle ES6 code - see building your project for production for a list.
  • es6-unbundled is an unbundled, minified build with a service worker. ES6 code is served as-is. This build is for browsers that support HTTP/2 push.

Run polymer help build for the full list of available options and optimizations. Also, see the documentation on the polymer.json specification and building your Polymer application for production.

Preview the build

This command serves your app. Replace build-folder-name with the folder name of the build you want to serve.

polymer serve build/build-folder-name/

Run tests

This command will run Web Component Tester against the browsers currently installed on your machine:

polymer test

If running Windows you will need to set the following environment variables:

  • LAUNCHPAD_BROWSERS
  • LAUNCHPAD_CHROME

Read More here daffl/launchpad

Adding a new view

You can extend the app by adding more views that will be demand-loaded e.g. based on the route, or to progressively render non-critical sections of the application. Each new demand-loaded fragment should be added to the list of fragments in the included polymer.json file. This will ensure those components and their dependencies are added to the list of pre-cached components and will be included in the build.

About

Deep Learning model training visualized using mutual information between the input and the output

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published