New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Request Timeout during Import #85
Comments
Hey @curiousc0w, |
@orangejulius Also, I was told that the production branches would be stable and not be updated although I'm having some issues querying certain place names. Here's the link to what I'm referring to. |
Ah yes, clearing the |
@orangejulius |
It's quite large, well over 200 million records. Our full planet build comes to 150GB, so with just OA, you're probably looking at around 100GB. |
@orangejulius So I just ran into the timeout error/bug again. It looped through the error I posted above for 5-10 minutes and the import re-started again at |
It will take quite a while: our full build takes almost two days. We run all of our importers in parallel, and until recently OA was the one that took the longest. Since then we've actually split the OA import into two parts which we run in parallel. Now that I think about it, I've seen this timeout issue when the machine doing the importing is overloaded. This can be because the combination of Elasticsearch and the importer is too much for one machine, or because Elasticsearch is running out of memory, so definitely make sure you're watching the utilization of your hardware. Our dev cluster uses 4 r3.xlarge instances, for a total of 120GB of memory. Near the end of the import process, when there's lots of data in play, they are using most of their 4 CPUs' capacity. |
@orangejulius |
@orangejulius I've tried to query for an address in Toronto, Ontario. And I received the following error. I've searched around and can't find any issues that other users have had with this in Pelias.
|
that might be Elasticsearch telling you it's completely overloaded. You can probably find more info in the API logs (they go to the console by default), or the elasticsearch logs (I can never remember where they go, and it depends on your system) |
@orangejulius
It may very well be an overload. I have another VM doing an import so RAM is running very close to full capacity. |
I'm in the process of installing Pelias without vagrant on my Ubuntu machine as well as a Centos machine.
On the Ubuntu machine I tried to
npm install
OpenAddresses andnode import.js
usingnode
version v0.10.38. Which resulted in the following error.This was resolved by switching to a newer version of
node
(v0.12.0). I could import a large dataset like Australia (countrywide) without an issue. I've also tried using both master and production branches.I tried to do this on a Centos machine and got the above error with both versions of node. Would this be an issue with the OS I am using or a RAM problem?
The text was updated successfully, but these errors were encountered: