Skip to content

Commit

Permalink
Revert "PL: --adafactor option (huggingface#6776)"
Browse files Browse the repository at this point in the history
This reverts commit 1179f87.
  • Loading branch information
fabiocapsouza committed Nov 15, 2020
1 parent 2c12d5a commit 8bf3ca8
Show file tree
Hide file tree
Showing 2 changed files with 1 addition and 12 deletions.
12 changes: 1 addition & 11 deletions examples/lightning_base.py
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,6 @@
PreTrainedTokenizer,
)
from transformers.optimization import (
Adafactor,
get_cosine_schedule_with_warmup,
get_cosine_with_hard_restarts_schedule_with_warmup,
get_linear_schedule_with_warmup,
Expand Down Expand Up @@ -138,15 +137,7 @@ def configure_optimizers(self):
"weight_decay": 0.0,
},
]
if self.hparams.adafactor:
optimizer = Adafactor(
optimizer_grouped_parameters, lr=self.hparams.learning_rate, scale_parameter=False, relative_step=False
)

else:
optimizer = AdamW(
optimizer_grouped_parameters, lr=self.hparams.learning_rate, eps=self.hparams.adam_epsilon
)
optimizer = AdamW(optimizer_grouped_parameters, lr=self.hparams.learning_rate, eps=self.hparams.adam_epsilon)
self.opt = optimizer

scheduler = self.get_lr_scheduler()
Expand Down Expand Up @@ -260,7 +251,6 @@ def add_model_specific_args(parser, root_dir):
parser.add_argument("--num_train_epochs", dest="max_epochs", default=3, type=int)
parser.add_argument("--train_batch_size", default=32, type=int)
parser.add_argument("--eval_batch_size", default=32, type=int)
parser.add_argument("--adafactor", action="store_true")


class LoggingCallback(pl.Callback):
Expand Down
1 change: 0 additions & 1 deletion examples/seq2seq/test_seq2seq_examples.py
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,6 @@
CUDA_AVAILABLE = torch.cuda.is_available()
CHEAP_ARGS = {
"label_smoothing": 0.2,
"adafactor": True,
"early_stopping_patience": 2,
"logger_name": "default",
"length_penalty": 0.5,
Expand Down

0 comments on commit 8bf3ca8

Please sign in to comment.