From 5210c59bfb506b24e5aa4922886e5f0c15535a36 Mon Sep 17 00:00:00 2001 From: lewtun Date: Sat, 1 May 2021 22:11:53 +0200 Subject: [PATCH] Fix examples in M2M100 docstrings Replaces `tok` with `tokenizer` so examples can run with copy-paste --- src/transformers/models/m2m_100/modeling_m2m_100.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/src/transformers/models/m2m_100/modeling_m2m_100.py b/src/transformers/models/m2m_100/modeling_m2m_100.py index 5d01e09129826..20c4aea990ecd 100755 --- a/src/transformers/models/m2m_100/modeling_m2m_100.py +++ b/src/transformers/models/m2m_100/modeling_m2m_100.py @@ -566,7 +566,7 @@ def _init_weights(self, module): >>> model_inputs = tokenizer(text_to_translate, return_tensors='pt') >>> # translate to French - >>> gen_tokens = model.generate( **model_inputs, forced_bos_token_id=tok.get_lang_id("fr")) + >>> gen_tokens = model.generate( **model_inputs, forced_bos_token_id=tokenizer.get_lang_id("fr")) >>> print(tokenizer.batch_decode(gen_tokens, skip_special_tokens=True)) """ @@ -1272,7 +1272,7 @@ def forward( >>> model_inputs = tokenizer(text_to_translate, return_tensors='pt') >>> # translate to French - >>> gen_tokens = model.generate( **model_inputs, forced_bos_token_id=tok.get_lang_id("fr")) + >>> gen_tokens = model.generate( **model_inputs, forced_bos_token_id=tokenizer.get_lang_id("fr")) >>> print(tokenizer.batch_decode(gen_tokens, skip_special_tokens=True)) """ return_dict = return_dict if return_dict is not None else self.config.use_return_dict