You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello the h20 team. Thanks for this wonderful package!
I was simply wondering if there is a tutorial somewhere that shows how we can use h20 to perform a full naive bayes text classification model.
I see a small example here, but in my dataset I have many documents (think headlines of articles), so the matrix representation of the bag-of-word processing would be (in a regular R session) a sparse matrix or a document term matrix.
Can h20 manage/create that? or somehow h2o can only work with a dataframe with a small number of dummies for each selected word already created by the user?
Thanks!
The text was updated successfully, but these errors were encountered:
Thank you for the kind words about H2O. But unfortunately, we do not currently have an example that you’re looking for.
One optimization you can try is to save the representation as an svmlight (aka libsvm) file, which is a sparse format, and load it in. Another option is to use PCA to reduce the dimensionality, and then use NB or a better algorithm like RF or GBM.
Hello the h20 team. Thanks for this wonderful package!
I was simply wondering if there is a tutorial somewhere that shows how we can use
h20
to perform a full naive bayes text classification model.I see a small example here, but in my dataset I have many documents (think headlines of articles), so the matrix representation of the bag-of-word processing would be (in a regular R session) a sparse matrix or a document term matrix.
Can
h20
manage/create that? or somehowh2o
can only work with a dataframe with a small number of dummies for each selected word already created by the user?Thanks!
The text was updated successfully, but these errors were encountered: