Implementation of Alphafold 3 in Pytorch
-
Updated
Jul 22, 2024 - Python
Implementation of Alphafold 3 in Pytorch
Implementation of MagViT2 Tokenizer in Pytorch
Implementation of Toolformer, Language Models That Can Use Tools, by MetaAI
Implementation of the LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens Paper
Implementation of MambaFormer in Pytorch ++ Zeta from the paper: "Can Mamba Learn How to Learn? A Comparative Study on In-Context Learning Tasks"
Implementation of Agent Attention in Pytorch
Implementation of MeshGPT, SOTA Mesh generation using Attention, in Pytorch
An implementation of local windowed attention for language modeling
Implementation of Diffusion Policy, Toyota Research's supposed breakthrough in leveraging DDPMs for learning policies for real-world Robotics
Implementation of a single layer of the MMDiT, proposed in Stable Diffusion 3, in Pytorch
Zeta implemantion of "Rethinking Attention: Exploring Shallow Feed-Forward Neural Networks as an Alternative to Attention Layers in Transformers"
Pytorch Implementation of the sparse attention from the paper: "Generating Long Sequences with Sparse Transformers"
Integrating Mamba/SSMs with Transformer for Enhanced Long Context and High-Quality Sequence Modeling
PyTorch Implementation of Jamba: "Jamba: A Hybrid Transformer-Mamba Language Model"
My implementation of the model KosmosG from "KOSMOS-G: Generating Images in Context with Multimodal Large Language Models"
Implementation of the conditionally routed attention in the CoLT5 architecture, in Pytorch
Implementation of Band Split Roformer, SOTA Attention network for music source separation out of ByteDance AI Labs
Implementation of "PaLM2-VAdapter:" from the multi-modal model paper: "PaLM2-VAdapter: Progressively Aligned Language Model Makes a Strong Vision-language Adapter"
Unofficial implementation of iTransformer - SOTA Time Series Forecasting using Attention networks, out of Tsinghua / Ant group
Implementation of Q-Transformer, Scalable Offline Reinforcement Learning via Autoregressive Q-Functions, out of Google Deepmind
Add a description, image, and links to the attention-mechanisms topic page so that developers can more easily learn about it.
To associate your repository with the attention-mechanisms topic, visit your repo's landing page and select "manage topics."