Skip to content

sinzlab/orbit_transfer

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Can Functional Transfer Methods Capture Simple Inductive Biases? -- Code

This repository contains the implementation of all methods that we discuss in our publication Can Functional Transfer Methods Capture Simple Inductive Biases?

⚙️ Usage:

The experiments from that paper that use the methods from this repository can be found in orbit_transfer_recipes.

The experiments further require installation of:

⭐ Features:

Models:

models/cnn.py: Modular architecture of a plain CNN model

models/group_cnn.py: Implementation of a group-equivariant CNN model

models/group_equivariant_layers.py: Layers used in the G-CNN model

models/learned_equiv.py: Implementation of the Orbit model

models/mlp.py: Modular architecture of a plain MLP model

models/vit.py: Simplified implementation of a small VIT model

Transfer methods:

We implemented all transfer methods that we discussed in the paper as main-loop-modules (see nntransfer). This includes: attention transfer, knowledge distillation, representational distance learning and orbit transfer.

Simple MNIST-1D experiments:

A less modular and flexible implementation of all transfer methods and corresponding training can be found in trainer/simple_train.py and trainer/forward_methods.py. Fitting models are given in models/mnist_1d.py.

🐛 Report bugs

In case you find a bug, please create an issue or contact any of the contributors.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages