Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Megatron tutorial #19

Closed
ShadenSmith opened this issue Feb 5, 2020 · 0 comments · Fixed by #30
Closed

Megatron tutorial #19

ShadenSmith opened this issue Feb 5, 2020 · 0 comments · Fixed by #30
Assignees
Labels
documentation Improvements or additions to documentation

Comments

@ShadenSmith
Copy link
Contributor

We need to port the Megatron tutorial into docs/ and then update links to it in README.md, etc.

@ShadenSmith ShadenSmith added the documentation Improvements or additions to documentation label Feb 5, 2020
@ShadenSmith ShadenSmith self-assigned this Feb 5, 2020
@ShadenSmith ShadenSmith linked a pull request Feb 6, 2020 that will close this issue
@samyam samyam closed this as completed in #30 Feb 7, 2020
rraminen pushed a commit to rraminen/DeepSpeed that referenced this issue Apr 28, 2021
* change squad baseline to use new apex
rraminen pushed a commit to rraminen/DeepSpeed that referenced this issue Apr 28, 2021
* update bing_bert example to use sparse attention (microsoft#19)

* update bing_bert example to use sparse transformer

* Updated teh BertSparseSelfAttention example based on the ST updates

* updated bing_bert example based on final updates for Sparse Attention; also added un/pad of Bert layer input

* updated based on Tunji's comment: added a separate script for SA

* fixed a typo

* added an exception when both transformer kernel and SA are set together.

* fixed an issue with last PR: removed keyword self for function call as it was moved out of class

Co-authored-by: Arash Ashari <arashari@microsoft.com>
Co-authored-by: arashashari <arash.ashari@gmail.com>
rraminen pushed a commit to rraminen/DeepSpeed that referenced this issue Jul 2, 2021
pengwa pushed a commit to pengwa/DeepSpeed that referenced this issue Oct 14, 2022
Liangliang-Ma pushed a commit to Liangliang-Ma/DeepSpeed that referenced this issue Nov 6, 2023
* Reuse hf_model list among tests to avoid slow loading

* try to debug test skip

* another attempt to print test failure

* another attempt

* more attempt to print skip reason

* revert changes that are temporary

* remove extra flag for pytest

* add a dummy test to test pytest

* test skip message

* put old test and temp test together to compare

* try to find out the reason skip message are not printed

* comment all skips

* check skip in common.py

* revert last commits

* shorten name to show skip message

* change test name

* expand number of columns to 120 when running pytest

* detect deepspeed installation

* add test code for environment

* change pytorch version 2.1.0==>2.0.1

* add py-cpuinfo as requiiremetns to dev

* install py-cpuinfo manually

* Change COLUMNS to 140 to allow display of pytest skip message

* ping pytorch to 2.0.1

* add pip list before install deepspeed

* install cpuinfo before install deepspeed

* change workflow to work with pytorch 2.1

* add torch install to CI workflow

* install py-cpuinfo

* enforce autotp test on single socket instance

* enforce 2 ranks in cpu autotp tests

* enable tests that can only run on torch 2.1 or above

* make build faster

* remove -j make option

* add back skip for codegen

* check UT result

* update tutorial
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
documentation Improvements or additions to documentation
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant