Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Port of DeepSpeed's overview and feature discussion. #27

Closed
wants to merge 3 commits into from

Conversation

ShadenSmith
Copy link
Contributor

I left a few TODOs as placeholders for the other documentation assignments.

The other major TODO is to ensure that our documentation (especially this README.md landing page) is nicely aligned with the blog post. We should be able to read the blog post and easily connect claims to our documentation here.

@ShadenSmith ShadenSmith added the documentation Improvements or additions to documentation label Feb 6, 2020
@ShadenSmith
Copy link
Contributor Author

Closing as this was included in the Megatron PR.

@ShadenSmith ShadenSmith closed this Feb 7, 2020
@ShadenSmith ShadenSmith deleted the readme_pr branch February 7, 2020 21:35
jeffra pushed a commit to jeffra/DeepSpeed that referenced this pull request May 15, 2020
…ster_weights

Load checkpoints with different DP degree
rraminen pushed a commit to rraminen/DeepSpeed that referenced this pull request Apr 28, 2021
* Support nvidia bert dataset

* Format fixes

* E2E run of Nvidia Data with SQUAD 90.6 F1

* Minor fixes

* Update README

* Update README
rraminen pushed a commit to rraminen/DeepSpeed that referenced this pull request Jul 26, 2021
liamcli pushed a commit to determined-ai/DeepSpeed that referenced this pull request Sep 27, 2021
* Pull changes from DeepSpeed

* Update op builder compatibility

* Update sparse_attn.py

Co-authored-by: sid <sidney.black@aleph-alpha.de>
delock referenced this pull request in delock/DeepSpeedSYCLSupport Nov 8, 2022
commit 747f4202c55b50431fb1d3434cafd7332322a037 (HEAD, origin/xpu-main, origin/HEAD)
Author: Guo Yejun <yejun.guo@intel.com>
Date:   Thu Oct 20 19:43:48 2022 +0800

    transformee.py: use torch.gelu instead of small op combination (#27)
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
documentation Improvements or additions to documentation
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Table of contents in README.md Port DeepSpeed overview documentation
2 participants