Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Allow deepspeed without distributed launcher #2204

Conversation

pacman100
Copy link
Contributor

@pacman100 pacman100 commented Dec 1, 2023

Copy link
Collaborator

@muellerzr muellerzr left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LG2M as long as the user reports back it works too.

@muellerzr muellerzr marked this pull request as ready for review December 1, 2023 14:08
@muellerzr muellerzr merged commit 0e48b23 into huggingface:main Dec 1, 2023
23 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

ver 4.35.2 transformers.Trainer, deepspeed backward(loss) is not used if distributed_state is NO
2 participants