You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Actually does this even make sense? Keras' weighting mechanism only applies to training samples, not at test time. Wouldn't scaling a word vector just turn it into another word? Have to see if there's precedent in the literature for this.
It is troubling that in the experimental results in #20 a bag-of-words SVM does as well as a neural LSTM, because that indicates that I'm not able to effectively make use of sequential data. It's possible that the issue is that I'm treating all words as equally informative in the neural network models. TF-IDF weighting is one way to address this. Would max pooling in a convent model (#17) be another?
Optionally weight input vectors by TF-IDF. Should be easy to get from scikit-learn.
The text was updated successfully, but these errors were encountered: