-
Notifications
You must be signed in to change notification settings - Fork 2.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Can not use custom sentence for QG #14
Comments
I think the problem is caused by the tokenization. The training examples should be tokenized as follows: tokenizer = BertTokenizer.from_pretrained(
args.bert_model, do_lower_case=args.do_lower_case)
r_list = []
for idx, line in enumerate(chunk):
tk_list = tokenizer.tokenize(line)
r_list.append((idx, tk_list)) For example, the sentence |
Hey @donglixp, |
Hi @aretius , If you would like to directly run the provided fine-tuned checkpoint, the same preprocessing pipeline is recommended. For custom fine-tuning, other toolkits should also work as long as both fine-tuning and inference use the similar input formats. |
Problem
Hi. I want to try the QG using
decode_seq2seq.py
. It works when I try use the sample data. But when I use another data, it encounter Key Error: 'H.E.Note
Question
Terminal Output
The text was updated successfully, but these errors were encountered: