Skip to content

PyTorch implementation of the paper Decoupled Neural Interfaces using Synthetic Gradients

License

Notifications You must be signed in to change notification settings

PanPapag/Decoupled-Neural-Interfaces-using-Synthetic-Gradients

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

13 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Decoupled-Neural-Interfaces-using-Synthetic-Gradients

A PyTorch implementation of the paper Decoupled Neural Interfaces using Synthetic Gradients. This repo is a modification of dni.pytorch. In this implementation both classification and synethic gradient models are Feed-Forward NN which can be modified rather simple just by changing the hidden_layer_sizes parameters.

Introduction

Training directed neural networks typically requires forward-propagating data through a computation graph, followed by backpropagating error signal, to produce weight updates. All layers, or more generally, modules, of the network are therefore locked, in the sense that they must wait for the remainder of the network to execute forwards and propagate error backwards before they can be updated. In this work we break this constraint by decoupling modules by introducing a model of the future computation of the network graph. These models predict what the result of the modelled subgraph will produce using only local information. In particular we focus on modelling error gradients: by using the modelled synthetic gradient in place of true backpropa- gated error gradients we decouple subgraphs, and can update them independently and asynchronously i.e. we realise decoupled neural interfaces.

Datasets and Models

Currently this repo only test MNIST dataset with Feed-Forward NN. Feel free to contribute and build experiments with other datasets (Cifar10, Shakespeare, etc) and models (CNN, RNN, etc.).

Training

train.py can make all the work for you. To experiment with hyperaparameters check the command-line arguments.

About

PyTorch implementation of the paper Decoupled Neural Interfaces using Synthetic Gradients

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages