You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm running into issues with spark_apply and nested columns. For example the snippet below produces the following error.
Error: org.apache.spark.sql.AnalysisException: cannot resolve 'from_json(vals)' due to data type mismatch: Input schema bigint must be a struct, an array or a map.;
'Project [a#152, b#153, from_json(LongType, vals#154, Some(America/New_York)) AS vals#201, d#155]
I'm curious where the Some(America/New_York) piece comes from given that this is an array of integers.
The error appears to be an issue with serialization of nested columns, vals in this case, even though spark_apply is just passing that column through and not trying to operate on it. The NA value in the field b that is used in the calculation seems to trigger the issue.
I'm running into issues with
spark_apply
and nested columns. For example the snippet below produces the following error.I'm curious where the
Some(America/New_York)
piece comes from given that this is an array of integers.The error appears to be an issue with serialization of nested columns,
vals
in this case, even thoughspark_apply
is just passing that column through and not trying to operate on it. The NA value in the fieldb
that is used in the calculation seems to trigger the issue.On the other hand, the same calculation without the NA value runs through okay.
The text was updated successfully, but these errors were encountered: