Compyute
is a toolbox for building and training and analyzing neural networks only using NumPy
/CuPy
under the hood to perform computations.
All you need is to pip install .
git clone https://github.com/dakofler/Compyute
pip install .
As of CuPy
v13, the package does not require a GPU toolkit to be installed, so Compyute
can be used on CPU-only machines. If you want to make use of GPUs, make sure to install the CUDA Toolkit following the installation guide of CuPy
(https://docs.cupy.dev/en/stable/install.html).
There are example-notebooks included that show how to use the toolbox and its Tensor
object.
Similar to PyTorch
, in Compyute
a Tensor
-object represents the central block that keeps track of data and gradients. However, unlike PyTorch
, Compyute
does not support autograd to compute gradients (yet?). Instead the computation of gradients is defined within a model's layers and functions.
import compyute as cp
# create a tensor from a list of lists, the data type is inferred automatically
x = cp.tensor([[4, 5, 6], [7, 8, 9]])
# alternatively, define data types
x = cp.tensor([1, 2, 3], dtype=cp.int32)
x = cp.tensor([1, 2, 3], dtype="int64")
# change datatypes
y = x.as_type(cp.float32)
y = x.float()
y = x.int()
# define the device the tensor is stored on
c = cp.tensor([1, 2, 3], device=cp.cuda)
c = cp.tensor([1, 2, 3], device="cuda")
# change devices
c.to_device(cp.cpu)
c.to_device("cpu")
c = c.cpu()
The Tensor
object supports most operations also known from PyTorch
tensors or NumPy
arrays. Here are some examples:
# basic arithmetic
z = x + y
z = x - y
z = x * y
z = x / y
x += 10
# operations from linear algebra
z = x @ y
z = cp.inner(x, y)
z = cp.outer(x, y)
# aggregate elements of a tensor
z = cp.sum(x, axis=0)
z = cp.mean(x, axis=0)
# functions for initializing tensors
z = cp.ones(shape=(10, 10))
z = cp.zeros(shape=(10, 10))
z = cp.full(shape=(10, 10), value=99)
z = cp.random.normal(shape=(10, 10))
The framework offers some utility functions to preprocess ...
from compyute.preprocessing import split_train_val_test
train, val, test = split_train_val_test(x, ratio_val=0.25, ratio_test=0.25)
... and encode data before using it for training.
from compyute.preprocessing.text import BPETokenizer
tokenizer = BPETokenizer()
tokenizer.fit(data, vocab_size=400)
Models can be built using predefined modules (such as Linear
or ReLU
) and containers (such as Sequential
).
import compyute.nn as nn
model = nn.Sequential(
nn.Linear(4, 16),
nn.ReLU(),
nn.Linear(16, 3),
)
Alternatively, models can also be built entirely from scratch by defining custom classes that inherit from the Container
class.
With custom models, the user defines what modules to use and how data and gradients flow through the network. Models are generally composed of one or more Modules
. Compyute
provides a variety of modules such as activation, normalization, linear, convolutional and recurrent layers with more to come.
When inheriting from predefined containers, such as Sequential
, the forward-function is already defined (in the case of Sequential
, Modules are processed in the order specified in the arguments). This way, you can create reusable blocks.
import compyute.nn as nn
# create a block with custom arguments by inheriting from the 'Sequential' container
class MyConvBlock(nn.Sequential):
def __init__(self, in_channels, out_channels):
conv = nn.Convolution2D(in_channels, out_channels, kernel_size=3)
relu = nn.ReLU()
bn = Batchnorm1d(out_channels)
super().__init__(conv, relu, bn)
When you want to define a custom forward-behaviour, the Container
base class can be used.
import compyute.nn as nn
# create a model from scratch by inheriting from the 'Container' base class
class MyModel(nn.Container):
def __init__(self):
super().__init__()
# define your modules
self.lin1 = nn.Linear(4, 16)
self.relu = nn.ReLU()
self.bn = nn.Batchnorm1d(16)
self.lin2 = nn.Linear(16, 3)
def forward(self, x):
# define the forward pass
x = self.lin1(x)
x = self.relu(x)
x = self.bn(x)
x = self.lin2(x)
# define the backward pass
def backward(dy):
dy = self.lin2.backward(dy)
dy = self.bn.backward(dy)
dy = self.relu.backward(dy)
dy = self.lin1.backward(dy)
return dy
# register the models backward function
self._backward = backward
return x
All modules can be trained and updated using common optimizer algorithms, such as SGD or Adam. Callbacks offer extended functionality such as tracking the loss-values during training. An easy and approchable way is to use a Trainer
object.
from compyute.nn.trainer import Trainer
from compyute.nn.callbacks import EarlyStopping, History, Progressbar
# define trainer
trainer = Trainer(
model=my_model,
optimizer="sgd",
loss_function="crossentropy",
metric_function="accuracy",
callbacks=[EarlyStopping(), History(), Progressbar()]
)
# train model
trainer.train(X_train, y_train, epochs=10)
Alternatively, you can write your own training loop.
epochs = 100
batch_size = 32
train_dl = nn.DataLoader(X_train, y_train, batch_size)
val_dl = nn.DataLoader(X_val, y_val, batch_size)
loss_func = nn.CrossEntropy()
optim = nn.optimizers.SGD(model.parameters)
for epoch in range(epochs):
# training
with model.training():
for x, y in train_dl():
# forward pass
y_pred = model(x)
_ = loss_func(y_pred, y)
# backward pass
optim.reset_grads() # reset all gradients
model.backward(loss_func.backward()) # compute new gradients
optim.step() # update parameters
# validiation
val_loss = 0
for x, y in val_dl():
y_pred = model(x)
val_loss += loss_func(y_pred, y).item()
val_loss /= len(val_dl)
print(f"epoch {epoch}: {val_loss=:.4f}")
Model checkpoints can also be saved and loaded later on.
nn.save_module(model, "my_model.cp")
Daniel Kofler - AI Research Associate (dkofler@outlook.com)
I use this project to gain a deeper understanding of the inner workings of neural networks. Therefore, project is a work in progress and possibly will forever be, as I am planning to constantly add new features and optimizations. The code is by far not perfect, as I am still learning with every new feature that is added. If you have any suggestions or find any bugs, please don't hesitate to contact me.
Cheers,
Daniel