Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
-
Updated
Jul 25, 2024 - Python
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
all kinds of text classification models and more with deep learning
Implementation / replication of DALL-E, OpenAI's Text to Image Transformer, in Pytorch
A collection of important graph embedding, classification and representation learning papers with implementations.
A simple but complete full-attention transformer with a set of promising experimental features from various papers
A TensorFlow Implementation of the Transformer: Attention Is All You Need
Graph Attention Networks (https://arxiv.org/abs/1710.10903)
Pytorch implementation of the Graph Attention Network model by Veličković et. al (2017, https://arxiv.org/abs/1710.10903)
Keras Attention Layer (Luong and Bahdanau scores).
Show, Attend, and Tell | a PyTorch Tutorial to Image Captioning
Reformer, the efficient Transformer, in Pytorch
Multilingual Automatic Speech Recognition with word-level timestamps and confidence
Implementation of LambdaNetworks, a new approach to image recognition that reaches SOTA with less compute
To eventually become an unofficial Pytorch implementation / replication of Alphafold2, as details of the architecture get released
Sequence-to-sequence framework with a focus on Neural Machine Translation based on PyTorch
Implementation of 🦩 Flamingo, state-of-the-art few-shot visual question answering attention net out of Deepmind, in Pytorch
Implementation of various self-attention mechanisms focused on computer vision. Ongoing repository.
Implementation of SoundStorm, Efficient Parallel Audio Generation from Google Deepmind, in Pytorch
An implementation of Performer, a linear attention-based transformer, in Pytorch
Add a description, image, and links to the attention-mechanism topic page so that developers can more easily learn about it.
To associate your repository with the attention-mechanism topic, visit your repo's landing page and select "manage topics."