Skip to content

Commit

Permalink
Add --curriculum (fixes #533)
Browse files Browse the repository at this point in the history
Summary: Pull Request resolved: #554

Differential Revision: D14300596

Pulled By: myleott

fbshipit-source-id: f38c8e58daef99d5e4b97dd423e4142e4294a4f0
  • Loading branch information
myleott authored and facebook-github-bot committed Mar 4, 2019
1 parent 1fd0a6f commit 2ad1178
Show file tree
Hide file tree
Showing 3 changed files with 11 additions and 7 deletions.
2 changes: 2 additions & 0 deletions fairseq/options.py
Expand Up @@ -233,6 +233,8 @@ def add_dataset_args(parser, train=False, gen=False):
group.add_argument('--max-sentences-valid', type=int, metavar='N',
help='maximum number of sentences in a validation batch'
' (defaults to --max-sentences)')
group.add_argument('--curriculum', default=0, type=int, metavar='N',
help='don\'t shuffle batches for first N epochs')
if gen:
group.add_argument('--gen-subset', default='test', metavar='SPLIT',
help='data subset to generate (train, valid, test)')
Expand Down
6 changes: 3 additions & 3 deletions fairseq/tasks/fairseq_task.py
Expand Up @@ -90,9 +90,9 @@ def dataset(self, split):
return self.datasets[split]

def get_batch_iterator(
self, dataset, max_tokens=None, max_sentences=None, max_positions=None,
ignore_invalid_inputs=False, required_batch_size_multiple=1,
seed=1, num_shards=1, shard_id=0, num_workers=0,
self, dataset, max_tokens=None, max_sentences=None, max_positions=None,
ignore_invalid_inputs=False, required_batch_size_multiple=1,
seed=1, num_shards=1, shard_id=0, num_workers=0,
):
"""
Get an iterator that yields batches of data from the given dataset.
Expand Down
10 changes: 6 additions & 4 deletions train.py
Expand Up @@ -120,13 +120,15 @@ def main(args, init_distributed=False):

def train(args, trainer, task, epoch_itr):
"""Train the model for one epoch."""

# Update parameters every N batches

# Initialize data iterator
itr = epoch_itr.next_epoch_itr(fix_batches_to_gpus=args.fix_batches_to_gpus)
update_freq = args.update_freq[epoch_itr.epoch - 1] \
if epoch_itr.epoch <= len(args.update_freq) else args.update_freq[-1]

# Initialize data iterator
itr = epoch_itr.next_epoch_itr(
fix_batches_to_gpus=args.fix_batches_to_gpus,
shuffle=(epoch_itr.epoch >= args.curriculum),
)
itr = iterators.GroupedIterator(itr, update_freq)
progress = progress_bar.build_progress_bar(
args, itr, epoch_itr.epoch, no_progress_bar='simple',
Expand Down

0 comments on commit 2ad1178

Please sign in to comment.