--decode_to_file does not create output file #48
Comments
It should be sufficient to just use |
Actually, I did.
But I couldn't find any output file anywhere. |
Now, I see. It's in the tmp folder. Thanks! |
While translating a file, I got the following error. What could this be connected with:
|
O, that's only a character issue while translating a word containing the character "È"! |
Hi all,
I think this is caused by the OOVs like @mehmedes said. I'm not sure whether the vocab.bpe.32000 in this data can cover the pre-processed training data in English to German translation task. I used my data, where the training data contains some unknown words out of vocabulary. I instead rewrote the code to something like:
Where I can ensure the Thank you |
Does this only happen with BPE, or with the standard "tokens_32k" too? We don't have a built-in tokenizer for BPE, it was used only for papers to have perplexities comparable with other papers. It cannot be detokenized, so I believe it's better to use our own tokenizer. Or is the problem the same? |
During inference, I'm not able to create a file containing the inference output.
I've tried
--decode_to_file
, but no output file is being created...The text was updated successfully, but these errors were encountered: