New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Getting Avro error while trying to load data from EMR spark dataframe to redshift table. #222
Comments
This looks like a duplicate of #180 and is likely a problem with dependency conflicts in your environment. |
I am using EMR 4.6.0 and spark 1.6.1. #180 is for the older version of the spark. From: Josh Rosen notifications@github.com This looks like a duplicate of #180#180 and is likely a problem with dependency conflicts in your environment. — |
Which versions of |
Hello, I am using following dependency.
Thanks Which versions of spark-redshift, spark-avro, and avro-mapred are you using? — |
Please give this a try using the 1.0.1 release of this library and re-open if this still an issue. |
16/06/27 17:00:14 ERROR InsertIntoHadoopFsRelation: Aborting task.
java.lang.IncompatibleClassChangeError: Found interface org.apache.hadoop.mapreduce.TaskAttemptContext, but class was expected
at org.apache.avro.mapreduce.AvroKeyOutputFormat.getRecordWriter(AvroKeyOutputFormat.java:85)
at com.databricks.spark.avro.AvroOutputWriter.(AvroOutputWriter.scala:82)
at com.databricks.spark.avro.AvroOutputWriterFactory.newInstance(AvroOutputWriterFactory.scala:31)
at org.apache.spark.sql.sources.DefaultWriterContainer.initWriters(commands.scala:470)
at org.apache.spark.sql.sources.BaseWriterContainer.executorSideSetup(commands.scala:360)
at org.apache.spark.sql.sources.InsertIntoHadoopFsRelation.org$apache$spark$sql$sources$InsertIntoHadoopFsRelation$$writeRows$1(commands.scala:172)
at org.apache.spark.sql.sources.InsertIntoHadoopFsRelation$$anonfun$insert$1.apply(commands.scala:160)
at org.apache.spark.sql.sources.InsertIntoHadoopFsRelation$$anonfun$insert$1.apply(commands.scala:160)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:63)
at org.apache.spark.scheduler.Task.run(Task.scala:70)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
The text was updated successfully, but these errors were encountered: