Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
modify double considering special tokens in
language_modeling.py
(#…
…11275) * Update language_modeling.py in "class TextDatasetForNextSentencePrediction(Dataset)", double considering "self.tokenizer.num_special_tokens_to_add(pair=True)" so, i remove self.block_size, and add parameter for "def create_examples_from_document". like "class LineByLineWithSOPTextDataset" do * Update language_modeling.py
- Loading branch information