You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Aug 28, 2025. It is now read-only.
# Calculate total stepstb_size=self.hparams.train_batch_size*max(1, self.trainer.gpus)
ab_size=self.trainer.accumulate_grad_batches*float(self.trainer.max_epochs)
self.total_steps= (len(train_loader.dataset) //tb_size) //ab_size
If I'm not mistaken, it should be something along the lines of:
# Calculate total stepstb_size=self.hparams.train_batch_size*max(1, self.trainer.gpus)
ab_size=tb_size*self.trainer.accumulate_grad_batchesself.total_steps=int((len(train_loader.dataset) /ab_size) *float(self.trainer.max_epochs))
In the first version, on MRPC (3668 instances), with 30 epochs, 32 batch size, 1 gpu and 1 batch accumulation total_steps amounts to 3; in the second version, it amounts to 3438.