Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

java.lang.ClassCastException: scala.coll ection.Iterator$$anon$11 cannot be cast to scala.Tuple2 #518

Closed
mavencode01 opened this issue Jul 22, 2015 · 1 comment

Comments

@mavencode01
Copy link

Am getting this exception each time I run my job ?

Any idea why ?

Exception in thread "main" org.apache.spark.SparkException: Job aborted due to s
tage failure: Task 2 in stage 0.0 failed 4 times, most recent failure: Lost task
 2.3 in stage 0.0 (TID 13, 10.0.0.209): java.lang.ClassCastException: scala.coll
ection.Iterator$$anon$11 cannot be cast to scala.Tuple2
        at org.apache.spark.sql.DataFrame$$anonfun$33.apply(DataFrame.scala:1189
)
        at org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$17.a
pply(RDD.scala:686)
        at org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$17.a
pply(RDD.scala:686)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:
35)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:
35)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:
35)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:
35)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:
35)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)

Sample code


public static void main(String[] args) throws UnsupportedEncodingException, FileNotFoundException, IOException {
        int cpus = 30;

        SparkConf sparkConf = new SparkConf(true).setAppName("SparkQueryApp")
                 .setMaster("spark://192.168.1.201:7077")
                 .set("es.nodes", "192.168.1.209")
                 .set("es.nodes.discovery", "false")
                 .set("es.cluster", "es-dev")       
                 .set("es.scroll.size", "5000")
                .setJars(JavaSparkContext.jarOfClass(Demo.class))
                .set("spark.serializer", "org.apache.spark.serializer.KryoSerializer")
                .set("spark.default.parallelism", String.valueOf(cpus * 2))
                .set("spark.logConf", "true");


        SparkContext sparkCtx = new SparkContext(sparkConf);        
        SQLContext sqlContext = new SQLContext(sparkCtx);

        DataFrame df = JavaEsSparkSQL.esDF(sqlContext, "rt-logs/test");

        df.registerTempTable("Log");


        DataFrame sqlDFTest = sqlContext.sql("SELECT created, count(instances) FROM Log where id  = 100000 group by instances");

        List<String> results = sqlDFTest.javaRDD().map(new Function<Row, String>(){

            static final long serialVersionUID = 1L;

            @Override
            public String call(Row row) throws Exception {                  
                return "Created" + row.getString(0) + " Count " + row.getInt(1);
            }               
        }).collect();

        try (Writer writer = new BufferedWriter(new OutputStreamWriter(new FileOutputStream("result3.txt"), "utf-8"))) {

            for (String row : results) {                
                writer.write(row);
            }

            writer.write(sqlDFTest.first().toString());
        }
    }

@costin
Copy link
Member

costin commented Jul 27, 2015

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants