Skip to content

Conversation

patil-suraj
Copy link
Contributor

This PR lowers the tolerance for gradient checkpointing test.

@HuggingFaceDocBuilderDev
Copy link

HuggingFaceDocBuilderDev commented Sep 27, 2022

The documentation is not available anymore as the PR was closed or merged.

@patil-suraj patil-suraj merged commit 84b9df5 into main Sep 29, 2022
@patil-suraj patil-suraj deleted the fix-grad-ckpt-test branch September 29, 2022 09:57
@@ -199,7 +199,7 @@ def prepare_init_args_and_inputs_for_common(self):

def test_gradient_checkpointing(self):
init_dict, inputs_dict = self.prepare_init_args_and_inputs_for_common()
model = self.model_class(**init_dict)
model = self.model_class(**init_dict).eval()
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

nice! Let's be extra careful in the future to not just lower the test tolerance. Here I think it really was dropout that was messing up the grads

prathikr pushed a commit to prathikr/diffusers that referenced this pull request Oct 26, 2022
prathikr pushed a commit to prathikr/diffusers that referenced this pull request Oct 26, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants