New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
fscrawler throws error when using flag --loop 1 #547
Comments
Thanks for reporting with all details! |
I reproduced it hardly as it seems to be a race condition. Not sure yet how to fix it properly. Will try. |
fscrawler throws error when using flag --loop 1
It always throws an error for me. Some more info: I'm running it in a ubuntu docker container with openjdk-8-jdk
|
fscrawler throws error when using flag --loop 1
fscrawler throws error when using flag --loop 1
I just ran it 3 times: Config: {
"name" : "documents",
"fs" : {
"url" : "/Users/dpilato/Documents/Elasticsearch/work/fscrawler/547/documents"
}
} Command to run it: $ ./fscrawler-2.5-SNAPSHOT/bin/fscrawler --config_dir ~/Documents/Elasticsearch/work/fscrawler/547/config documents --debug --loop 1 --restart First run
That helped me BTW to discover another issue (might be related) which is that the REST Client does not close. I had to type CTRL+C to stop it:
Second run
This time I don't have to CTRL+C but it generated the error you are also getting. Third run
It worked again but I had to CTRL+C:
I'm going to try to force a call to |
So I found a way to fix that. Instead of calling Going to push the fix soon. |
We were closing the bulk processors with `close()` which was trying to immediately exit the bulk processor. Calling `awaitClose(30, TimeUnit.SECONDS)` is giving 30 more seconds to the bulk processor to flush first all existing request before actually closing. I thought that `close()` was doing the same behind the scene but apparently not. Closes #547.
@dadoonet Many thanks! |
Hi, I hope someone could help me with this error.
Every time I run fscrawler with
--loop 1
i got the errorGot a hard failure when executing the bulk request
and the data doesn't get send to elasticsearch. If I run fscrawler without that option all works fine and I can see the data.I'm using the following fscrawler 2.5 snapshot
fscrawler-2.5-20180215.233518-30.zip
The text was updated successfully, but these errors were encountered: