Test step with DDP #3639
-
Having refactored my code to avoid iterable datasets I've now got DDP training working (I also had to set ulimit to prevent another crash). However now it crashes at the test step. The message implies DDP is not needed for testing - but I don't see any mention in the documentation of how to disable DDP once training has complete (plus I would assume that trainer.test() would do this if it were required). Is there something I should be doing different - this is my train / test code - the test dateloader has batchsize=1
|
Beta Was this translation helpful? Give feedback.
Replies: 3 comments
-
So looks like |
Beta Was this translation helpful? Give feedback.
-
Does not look like a but to me. The error seems to come from the underlying PyTorch DistributedDataParallel, not Lightning. |
Beta Was this translation helpful? Give feedback.
-
Yes that makes sense. |
Beta Was this translation helpful? Give feedback.
Does not look like a but to me.
The error seems to come from the underlying PyTorch DistributedDataParallel, not Lightning.
It seems we can't really do anything. But freezing the model before test should not be needed, test does not alter the weights and puts model into eval mode anyway. Makes sense?