Skip to content
/ moegpt Public

Mixture of Experts (MoE) implementation for minGPT

License

Notifications You must be signed in to change notification settings

zxaall/moegpt

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

moegpt

Mixture of Experts (MoE) implementation for minGPT by Val Krigan

My contribution here is only MoE implementations plus some necessary changes to hook it up. It supports training in soft MoE and sparce modes. First one is used for routers pretraining.

The rest of the code was borrowed from here:

https://github.com/karpathy/minGPT

About

Mixture of Experts (MoE) implementation for minGPT

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages