-
Notifications
You must be signed in to change notification settings - Fork 836
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Out of memory error when reading from large files. #160
Comments
A more formal stack trace from a newer node version:
|
The problem seems isolated to the file reader stream not being paused properly... |
will be solved via #161 |
SFO1502579255M:~ 502579255$ head -n 92006 /Users/502579255/ELK/kibana-4.4.1-darwin-x64/Kibana_Essentials-master/tweet.json | /usr/local/bin/elasticdump --bulk=true --input $ --output=http://localhost:9200/ <--- Last few GCs ---> 15098 ms: Scavenge 1405.7 (1458.1) -> 1405.7 (1458.1) MB, 19.0 / 0 ms (+ 54.7 ms in 1 steps since last GC) [allocation failure] [incremental marking delaying mark-sweep]. <--- JS stacktrace ---> ==== JS stack trace ========================================= Security context: 0x314dfe6e3ac1 FATAL ERROR: CALL_AND_RETRY_LAST Allocation failed - process out of memory |
@craighawki what version of node and elasticdump are you using? |
@craighawki it might also be the case that you have a very large document (over 1GB?) |
i get the same issue, may i know how do you solved it? 1145104 ms: Mark-sweep 1389.2 (1434.0) -> 1389.1 (1434.0) MB, 975.4 / 0.0 ms [allocation failure] [scavenge might not succeed]. <--- JS stacktrace ---> |
Your node process has run our of RAM... perhaps you are trying to import many large documents (all of which have to be parsed). Reduce how many documents you import per batch via |
I got the same Issue while testing, Run the protractor test and got following error: 91873 ms: Mark-sweep 1391.5 (1437.1) -> 1391.5 (1437.1) MB, 1263.2 / 0.0 ms [allocation failure] [scavenge might not succeed]. |
It appears that the latest version of elasticdump (1.0.0) has a problem reading from large input files.
I tested with smaller files, and it appears that elastcdump reads the first 100 lines from the input file, writes them to the output, then on the next read from the input, it tries to read the entire remainder of the input.
For example:
and
and
as I keep increasing the size of the input file, the pattern persists: the second read/write iteration attempts to read the entire remainder of the input. Once the input file gets large enough, elasticdump will run out of memory trying to read it all in.
The text was updated successfully, but these errors were encountered: