moe
Here are 30 public repositories matching this topic...
PyTorch implementation of moe, which stands for mixture of experts
-
Updated
Feb 11, 2021 - Python
[ICLR 2023] "Sparse MoE as the New Dropout: Scaling Dense and Self-Slimmable Transformers" by Tianlong Chen*, Zhenyu Zhang*, Ajay Jaiswal, Shiwei Liu, Zhangyang Wang
-
Updated
Feb 28, 2023 - Python
Meet Moe, a discord bot, written in modern python!
-
Updated
Apr 24, 2023 - Python
pytorch open-source library for the paper "AdaTT Adaptive Task-to-Task Fusion Network for Multitask Learning in Recommendations"
-
Updated
Aug 15, 2023 - Python
Batch download high quality videos from https://twist.moe
-
Updated
Sep 30, 2023 - Python
Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*
-
Updated
Dec 14, 2023 - Python
⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training
-
Updated
Feb 26, 2024 - Python
Restoring Spatially-Heterogeneous Distortions using Mixture of Experts Network (ACCV 2020)
-
Updated
Mar 30, 2024 - Python
ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward experts. We released a collection of ModuleFormer-based Language Models (MoLM) ranging in scale from 4 billion to 8 billion parameters.
-
Updated
Apr 10, 2024 - Python
Official repository for paper "MATERobot: Material Recognition in Wearable Robotics for People with Visual Impairments" at ICRA 2024, Best Paper Finalist on Human-Robot Interaction
-
Updated
Apr 11, 2024 - Python
A simple project that help visualize expert router choices for text generation
-
Updated
Apr 17, 2024 - Python
PyTorch Re-Implementation of "The Sparsely-Gated Mixture-of-Experts Layer" by Noam Shazeer et al. https://arxiv.org/abs/1701.06538
-
Updated
Apr 19, 2024 - Python
Community Implementation of the paper: "Multi-Head Mixture-of-Experts" In PyTorch
-
Updated
Apr 27, 2024 - Python
中文Mixtral混合专家大模型(Chinese Mixtral MoE LLMs)
-
Updated
Apr 30, 2024 - Python
This is the repo for the MixKABRN Neural Network (Mixture of Kolmogorov-Arnold Bit Retentive Networks), and an attempt at first adapting it for training on text, and later adjust it for other modalities.
-
Updated
May 14, 2024 - Python
Improve this page
Add a description, image, and links to the moe topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the moe topic, visit your repo's landing page and select "manage topics."