You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
First of all, thank you very much for the great work. Would really appreciate this feature. However, as I played around with it, I could not get started on re-training a word embedding model. Here are the steps to reproduce the error:
# run the sample script first to generate a model
$ bash word-vector-example.sh
# after the model is generated from the sample script
# we run
./fasttext skipgram -input data/fil9 -inputModel result/fil9.bin -output retrained -incr
The output displayed was:
Update args
Load dict from trained model
Load dict from training data
Read 124M words
Number of words: 218316
Number of labels: 0
Merge dict
Read 124M words
Number of words: 0
Number of labels: 0
terminate called after throwing an instance of 'std::invalid_argument'
what(): Empty vocabulary. Try a smaller -minCount value.
Aborted
I have tried to adjust -minCount option, but it did not work.
After looking at the code, I feel that the error has something to do with
dict_->addDict(dictInData, false);
from line 744 of fasttext.cc, but I am not sure about the exact cause of this problem. In fact, I am a bit confused by why we can load dictionary from the raw text we are supposed to train on.
The text was updated successfully, but these errors were encountered:
First of all, thank you very much for the great work. Would really appreciate this feature. However, as I played around with it, I could not get started on re-training a word embedding model. Here are the steps to reproduce the error:
The output displayed was:
I have tried to adjust
-minCount
option, but it did not work.After looking at the code, I feel that the error has something to do with
from line 744 of
fasttext.cc
, but I am not sure about the exact cause of this problem. In fact, I am a bit confused by why we can load dictionary from the raw text we are supposed to train on.The text was updated successfully, but these errors were encountered: