You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Elastic Search v 2.x stopped supporting the DOT ('.') character in the field names. There has been various discussions going on to support and handle the side effects of this change.
^ Here the field name "adv.id" contains '.' dot character. While trying to insert using Hive insert with ES-Hadoop Connector, results in the following error:
Caused by: org.elasticsearch.hadoop.rest.EsHadoopInvalidRequest: Found unrecoverable error [127.0.0.1:9200] returned Bad Request(400) - Field name [adv.id] cannot contain '.'; Bailing out..
^ And the entire Hive Job will be failed due to the error.
Meanwhile Logstash supports this via de_dot as well as we can even use Ruby block to replace the DOT character with something else before writing to Elastic Search. I couldn't find a similar feature that can be used to achieve the same with ES-Hadoop Connector.
It would be great if there a feature/configuration that can be used to replace the DOT character in the field names with some other character or just ignore the document which contains the DOT character before writing to Elastic Search.
The text was updated successfully, but these errors were encountered:
What kind an issue is this?
Feature description
Elastic Search v 2.x stopped supporting the DOT ('.') character in the field names. There has been various discussions going on to support and handle the side effects of this change.
elastic/elasticsearch#17759
elastic/elasticsearch#15714
elastic/elasticsearch#15951
Reproduce the issue
Example Document:
app { "adv.id": "efT3Fg5JnvJVs57IOnc" }
^ Here the field name "adv.id" contains '.' dot character. While trying to insert using Hive insert with ES-Hadoop Connector, results in the following error:
Caused by: org.elasticsearch.hadoop.rest.EsHadoopInvalidRequest: Found unrecoverable error [127.0.0.1:9200] returned Bad Request(400) - Field name [adv.id] cannot contain '.'; Bailing out..
^ And the entire Hive Job will be failed due to the error.
Meanwhile Logstash supports this via de_dot as well as we can even use Ruby block to replace the DOT character with something else before writing to Elastic Search. I couldn't find a similar feature that can be used to achieve the same with ES-Hadoop Connector.
It would be great if there a feature/configuration that can be used to replace the DOT character in the field names with some other character or just ignore the document which contains the DOT character before writing to Elastic Search.
The text was updated successfully, but these errors were encountered: