Join GitHub today
GitHub is home to over 20 million developers working together to host and review code, manage projects, and build software together.
Have you thought about the situation of a very long words? #4
Since i am trying to make a training data for Chinese fonts, I happened to prepare a text file including a very long "words" consist of about 3000 Chinese characters.
This comment has been minimized.
This comment has been minimized.Show comment Hide comment
Indeed, I did not have this case in mind. As you must have understood, the algorithm will try to fit each word in the current line. If the word does not fit, a new line will be used to fit it. This is a very simple way to proceed, and I agree that very long words could break it, as they would be too long to fit it an entire line.
One possible way around it would be to decrease the font size specifically for this word to fit in, but I think that would just mess things up: different font sizes would mean different inter-character spacings metrics, and
I do not clearly understand why you would require a 3000 character long word for training. Couldn't you use 3000 independent characters, separated by a whitespace?
Thanks for your feedback!