The PyTorch implementation of Generative Pre-trained Transformers (GPTs) using Kolmogorov-Arnold Networks (KANs) for language modeling
-
Updated
Jun 11, 2024 - Python
The PyTorch implementation of Generative Pre-trained Transformers (GPTs) using Kolmogorov-Arnold Networks (KANs) for language modeling
This project is dedicated to the implementation and research of Kolmogorov-Arnold convolutional networks. The repository includes implementations of 1D, 2D, and 3D convolutions with different kernels, ResNet-like and DenseNet-like models, training code based on accelerate/PyTorch, as well as scripts for experiments with CIFAR-10 and Tiny ImageNet.
TKAN: Temporal Kolmogorov-Arnold Networks
Improved LBFGS and LBFGS-B optimizers in PyTorch.
KANs for text classification on GLUE tasks
A multi-agent deep reinforcement learning model to de-traffic our lives
Combine B-Spline (BS) and Radial Basic Function (RBF) in Kolmogorov-Arnold Networks (KANs)
This is the repo for the MixKABRN Neural Network (Mixture of Kolmogorov-Arnold Bit Retentive Networks), and an attempt at first adapting it for training on text, and later adjust it for other modalities.
An implementation of the KAN architecture using learnable activation functions for knowledge distillation on the MNIST handwritten digits dataset. The project demonstrates distilling a three-layer teacher KAN model into a more compact two-layer student model, comparing the performance impacts of distillation versus non-distilled models.
DL model deployment using docker, API deployment with FastAPI, and MLOps using WandB for overhead-mnist dataset
Add a description, image, and links to the kolmogorov-arnold-networks topic page so that developers can more easily learn about it.
To associate your repository with the kolmogorov-arnold-networks topic, visit your repo's landing page and select "manage topics."