Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[wav2vec2] fix --gradient_checkpointing #13964

Merged
merged 1 commit into from
Nov 11, 2021
Merged

Conversation

stas00
Copy link
Contributor

@stas00 stas00 commented Oct 11, 2021

This PR fixes --gradient_checkpointing in wav2vec2 examples.

@sgugger

Copy link
Collaborator

@sgugger sgugger left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The research projects are not actively maintained and pinned under a Transformers version, which is why I did not update them, cc @patrickvonplaten

@stas00
Copy link
Contributor Author

stas00 commented Oct 11, 2021

it was failing deepspeed tests for wav2vec2 - which I still run manually to check that Deepspeed didn't break something. wav2vec2 is a complex model with several manual fixes to make DS work.

but I hear you about the pinned version! This makes sense. I was running those with master, so should probably add a skip rule to match the pinned version. Which would make it much more difficult for me to run the tests.

@stas00
Copy link
Contributor Author

stas00 commented Nov 11, 2021

@patrickvonplaten, do we merge this or close the PR? Thank you!

@patrickvonplaten patrickvonplaten merged commit 77262ef into master Nov 11, 2021
@patrickvonplaten patrickvonplaten deleted the wav2vec2-grad-ckpt branch November 11, 2021 16:50
Albertobegue pushed a commit to Albertobegue/transformers that referenced this pull request Jan 27, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

3 participants