New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Process killed by the system #12
Comments
Yeah, the problem is most likely that you are feeding the whole dev set as one batch, which is theoretically possible, but takes a lot of memory. You can break up the dev set into multiple batches of fixed size and then combine the results. |
I'm able to make it work now. Thank you very much. |
@lenhhoxung86 After feeding multiple batches of dev data, how did you deal with the summary step? Since the summary is for each batch. |
@junchaozheng Hi, you can do like this:
But in general, I have to say that word2vec and CNN are not really a good choice for text classification. From my experience, tf-idf is always the best feature to deal with text. |
Hello @lenhhoxung86 @junchaozheng we are seeing the same issue wondering what does your |
Hello, I've tested your code on my own data of 20,000 examples, and the result is quite good. I have another data of 300,000 examples. Each example is a short sentence of approximate 20 words. When tested on this new dataset, the first 100 steps are fine. However, it stops on the evaluation step. The dev set has more than 60,000 examples, and the message when it stops is "Killed".
I guess the reason is that the number of examples of the dev set is very large, so it consumes a lot of memory. Is that true? And how can I fix that?
Thank you very much.
The text was updated successfully, but these errors were encountered: