Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fixing issue "Training beyond specified 't_total' steps with schedule 'warmup_linear'" reported in #556 #604

Merged
merged 3 commits into from
Jun 14, 2019

Conversation

samuelbroscheit
Copy link
Contributor

Fixing the issues reported in #556

Reason for issue was that num_optimzation_steps was computed from example size, which is different from actual size of dataloader when an example is chunked into multiple instances.

Solution in this pull request is to compute num_optimization_steps directly from len(data_loader).

Reason for issue was that optimzation steps where computed from example size, which is different from actual size of dataloader when an example is chunked into multiple instances.

Solution in this pull request is to compute num_optimization_steps directly from len(data_loader).
@samuelbroscheit samuelbroscheit changed the title Fixing the issues reported in https://github.com/huggingface/pytorch-pretrained-BERT/issues/556 Fixing issue "Training beyond specified 't_total' steps with schedule 'warmup_linear'" reported in #556 May 11, 2019
examples/run_classifier.py Outdated Show resolved Hide resolved
@lukovnikov
Copy link
Contributor

Looks good.

@thomwolf
Copy link
Member

Ok merging thanks, sorry for the delay!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

4 participants