Skip to content

vark1/gradiatorjs

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

70 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

GradiatorJS

GradiatorJS is a lightweight, from-scratch neural network library built in typescript which you can use to create and train deep neural nets directly in your browser. It is an autodiff engine with backpropagation which is implemented over a dynamically build DAG i.e. the library will automatically compute the gradients of all parameters with respect to a final output (like a loss fn). This allows you to define dynamic neural nets (including Fully Connected and Convolutional Neural Networks), train these networks on custom datasets using mini-batch gradient descent etc etc.

This project is done with a pedagogical intention in mind (to help concepts feel intuitive and easy to understand). You can use this library in your own projects using the npm package.

Note: While this library can be used on its own, it is being used as the core engine for Visualizer, an interactive platform for building, training, and seeing inside your models in real-time.

Features:

  • Dynamic Computation Graph: At its core is a Val object that tracks all operations to build a computation graph on the fly, enabling automatic backpropagation. It can store anything from a scalar to an ND array.
  • Modern Network Layers: Comes with all the essential layers needed for modern image classification tasks:
    • Dense: FC layer
    • Conv2D: Convolutional Layer
    • MaxPooling2D: Max Pool Layer
    • Flatten: Flattening Layer
  • Standard Activation Functions: ReLU, Sigmoid, Tanh, Softmax.
  • Loss Functions: Standard loss functions like crossEntropyLoss_binary, crossEntropyLoss_categorical and a numerically stable crossEntropyLoss_softmax for multi-class problems.
  • Accuracy Functions: calcBinaryAccuracy and calcMultiClassAccuracy.
  • Operations: add, sub, mul, dot, pow, div, divElementWise, negate, abs, exp, log, sum, mean, conv2d, maxPool2d. (all with broadcasting!)
  • Async training: Ability to pause, resume, stop training using state management.
  • Mini-Batch GD: getMiniBatch which creates shuffled batches.
  • Activation data for a sample batch: While training, it passes activation data of a sample in callback, which can be used to visualize the network's current state.
  • Pure TypeScript/JavaScript: No external dependencies for the core engine, making it lightweight and easy to understand.

Installation

Using npm:

npm install gradiatorjs

Getting Started

Creating a binary classification network

import {layer, act, loss, model} from 'gradiatorjs';

const model = new model.Sequential(
    // Assuming an input shape of [batch, 64, 64, 3] for cat images
    new layer.Flatten(),
    new layer.Dense(12288, 20, act.relu),
    new layer.Dense(20, 7, act.relu),
    new layer.Dense(7, 5, act.relu),
    new layer.Dense(5, 1, act.sigmoid)
);

const loss = loss.crossEntropyLoss_binary(predictions, true_labels);

Creating a Convolutional Network

import { layer, act, loss, model } from 'gradiatorjs';

const model = new model.Sequential(
    // Assuming an input shape of [batch, 28, 28, 1] for a dataset like MNIST
    new layer.Conv(1, 6, 5, 1, 0, act.relu), // in_channels, out_channels, kernel, stride, padding, activation
    new layer.MaxPool2D(2, 2),
    new layer.Conv(6, 16, 5, 1, 0, act.relu),
    new layer.MaxPool2D(2, 2),
    new layer.Flatten(),
    new layer.Dense(256, 120, act.relu),
    new layer.Dense(120, 84, act.relu),
    new layer.Dense(84, 10)
);

const loss = loss.crossEntropyLoss_softmax(logits, one_hot_labels);

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published