New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ES hadoop problem finding the correct cluster nodes #636
Comments
Is PIG host can access to all your ES hosts? You can see your ES cluster topology by querying "/_cat/nodes", be sure these IPs are accessible from host where PIG is located. |
Hello and thanks for the responser.
It doesnt seem to find the nodes in the cluster using localhost. Howerer when i trying to GET localhost:9200 from ssh i get no exception:
|
I am not sure but I believe elasticsearch for Hadoop is working like Zookeeper: 1- Hadoop client (Pig, Spark...) use elasticsearch JAR lib and query node(s) you specify in the configuration 2- A ES node will answer who is the master node in the cluster 3- Hadoop client will use this IP So, use _cat API to get the real master IP of your ES nodes. |
With the _cat/nodes i can see the 4 nodes of es cluster. I have also tried older versions. |
So from the Pig host, you can curl without problem the IP given by _cat/nodes ? Also, enable log tracing: https://www.elastic.co/guide/en/elasticsearch/hadoop/current/logging.html, maybe this will give some clues. |
I run this command to enable logging:
log4j.properties:
test.pig:
The only debug messages i get:
And the error:
|
Hello,
I cannot read or write data in elasticsearch using the following pig script:
After running the above script i get the following error:
For some reason the es hadoop lib makes invalid request (host/ip:9200 instead of host:9200). I am using latest stable version of ES hadoop (2.2.0) and i have a 3 node cluster of elasticsearch (2.1.0). Hadoop version: 2.4.0.2.1.5.0-695
Pig version: 0.12.1.2.1.5.0-695
Thanks in Advance
The text was updated successfully, but these errors were encountered: