- Attention is all you need: https://arxiv.org/abs/1706.03762
- Annotated transformer (Pytorch): https://nlp.seas.harvard.edu/2018/04/03/attention.html
- einsum:
- videos:
- Youtube: https://youtu.be/ULY6pncbRY8
- Bilibili: https://www.bilibili.com/video/BV1ee411g7Sv
- code snippets: https://numpy.org/doc/stable/reference/generated/numpy.einsum.html
- einops: https://github.com/arogozhnikov/einops/
- videos:
- Jax:
- Jax: https://jax.readthedocs.io/en/latest/index.html
- Haiku: https://dm-haiku.readthedocs.io/en/latest/index.html
- Haiku101: Haiku库的基本使用逻辑 - 谷雨的文章 - 知乎 https://zhuanlan.zhihu.com/p/471892075
- reference implementations I used yesterday:
- haiku: this is cleaner but more functional & the transformer is not complete
- flax: this is more complex and easier to get lost
- elegy: perhaps use it for training loop
-
Notifications
You must be signed in to change notification settings - Fork 0
Oaklight/jax-transformer
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
About
No description or website provided.
Topics
Resources
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published