Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

DeepSpeed to DeepSpeed converter for changing tp/pp #27

Closed
wants to merge 5 commits into from

Conversation

tjruwase
Copy link

Addressing #23
@stas00, FYI

@stas00
Copy link

stas00 commented Jan 23, 2022

@tjruwase, it appears that EAI have also been doing work in that direction, albeit in one direction only: many to 1, but it appears not to work exactly from a quick read of the issue.

EleutherAI/gpt-neox#466

So perhaps some collaboration can be had - hoping that they have already solved some of the problems you have encountered and vice versa.

@tjruwase
Copy link
Author

@stas00, thanks for this pointer. Sure, it would be nice to chat with them after you and I chat.

@tjruwase
Copy link
Author

@stas00, can you confirm that it is okay to close this because of bigscience-workshop/Megatron-DeepSpeed#238

@stas00
Copy link

stas00 commented Jan 24, 2022

yes, it'd be easier to test directly in the BigScience fork and then backport to the MSFT fork. So it's definitely ok to close this one.

@tjruwase tjruwase closed this Jan 24, 2022
@stas00
Copy link

stas00 commented Jan 24, 2022

this PR has moved to bigscience-workshop/Megatron-DeepSpeed#239

hyoo pushed a commit to hyoo/Megatron-DeepSpeed that referenced this pull request Apr 21, 2023
saforem2 added a commit to saforem2/Megatron-DeepSpeed that referenced this pull request Oct 11, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants