New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Array types not supported in automatic mapping #847
Comments
Additionally, simply running basic operations on this fails:
|
Even basic operations on it fail:
|
To quote the page you referred to:
In other words see, https://www.elastic.co/guide/en/elasticsearch/hadoop/current/configuration.html#cfg-field-info in particular |
Array types in Elasticsearch aren't necessarily special.
for example, because of the analyzer, you could have:
And they are essentially functionally equivalent.
This is defined here:
https://www.elastic.co/guide/en/elasticsearch/reference/1.4/mapping-array-type.html
The mapping would look like this:
But Spark and elasticsearch-hadoop breaks on the mapping.
It gets the mapping as: "string"
But it's not a string. it's list[string]
The type conversion here is wrong:
https://www.elastic.co/guide/en/elasticsearch/hadoop/current/spark.html#spark-type-conversion
Spark does not detect it as a list[string].
This means that functions like explode() in Hive won't work on this document properly.
The text was updated successfully, but these errors were encountered: