Skip to content

Commit

Permalink
fix mpt not support left padding issue. (#172)
Browse files Browse the repository at this point in the history
Signed-off-by: Lv, Kaokao <kaokao.lv@intel.com>
  • Loading branch information
lkk12014402 committed Aug 25, 2023
1 parent 925051c commit 24ae58e
Showing 1 changed file with 1 addition and 2 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,6 @@
import logging
import os
import sys
sys.path.append("/data2/lkk/llama/test_pr/intel-extension-for-transformers")
import transformers
from transformers.modeling_utils import unwrap_model
from dataclasses import dataclass, field
Expand Down Expand Up @@ -505,6 +504,7 @@ def main():
torch_dtype=model_dtype,
low_cpu_mem_usage=True,
)
tokenizer.padding_side = "left" # allow batched inference, while mpt series don't support
else:
raise ValueError(
"Must provide model_name_or_path to load a pretrained CausalLM model."
Expand Down Expand Up @@ -546,7 +546,6 @@ def main():

if tokenizer.pad_token_id is None:
tokenizer.pad_token_id = tokenizer.eos_token_id
tokenizer.padding_side = "left" # Allow batched inference

raw_datasets, preprocess_function = preprocess_dataset(raw_datasets, tokenizer, data_args, finetune_args)
column_names = list(raw_datasets["train"].features)
Expand Down

0 comments on commit 24ae58e

Please sign in to comment.