You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm working with large sentences or documents, but without commas or special characters, because I cleaned the text in previous steps, So, I am sending sentences like the example below, but very large:
"amor edificio casa perro" --> expected output ---> "love building house dog"
As you can notice, there is no presence of commas or dots, will your model need this characters for translate in chunks?
The text was updated successfully, but these errors were encountered:
I removed stopwords and removed special characters as well, I am working in a sentiment analysis report, I think some techniques for detect the sentiment score does not need contextual polarity, that is the reason I am calculating the score based in the average of scores of all words present by document.
I think this could work:
for translation in model.translate_stream(text, show_progress_bar=False, chunk_size=16, source_lang='es', target_lang='en'):
translated_text.append(translation)
Hi @nreimers
Will the method Translate split or tokenize the input text in sentences?
model.translate(text, source_lang='es', target_lang='en')
I'm working with large sentences or documents, but without commas or special characters, because I cleaned the text in previous steps, So, I am sending sentences like the example below, but very large:
"amor edificio casa perro" --> expected output ---> "love building house dog"
As you can notice, there is no presence of commas or dots, will your model need this characters for translate in chunks?
The text was updated successfully, but these errors were encountered: