Is your feature request related to a problem? Please describe.
Issure reported by internal QA:
Currently, when image number is less than GPU number, we only check the case that partition dataset for ranks before training, then some GPU rank will have "epoch_length = 0":
https://github.com/Project-MONAI/MONAI/blob/dev/monai/engines/workflow.py#L274
But if we use DistributedSampler, it splits data at runtime, would be nice to add some check in it.
Is your feature request related to a problem? Please describe.
Issure reported by internal QA:
Currently, when image number is less than GPU number, we only check the case that partition dataset for ranks before training, then some GPU rank will have "epoch_length = 0":
https://github.com/Project-MONAI/MONAI/blob/dev/monai/engines/workflow.py#L274
But if we use
DistributedSampler, it splits data at runtime, would be nice to add some check in it.