Skip to content

Commit

Permalink
Add addSparkListener to JavaSparkContext
Browse files Browse the repository at this point in the history
  • Loading branch information
JoshRosen committed Jan 20, 2015
1 parent 163ba19 commit 25988f3
Showing 1 changed file with 10 additions and 1 deletion.
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@ import org.apache.hadoop.mapreduce.{InputFormat => NewInputFormat}

import org.apache.spark._
import org.apache.spark.AccumulatorParam._
import org.apache.spark.annotation.Experimental
import org.apache.spark.annotation.{DeveloperApi, Experimental}
import org.apache.spark.api.java.JavaSparkContext.fakeClassTag
import org.apache.spark.broadcast.Broadcast
import org.apache.spark.rdd.{EmptyRDD, HadoopRDD, NewHadoopRDD, RDD}
Expand Down Expand Up @@ -688,6 +688,15 @@ class JavaSparkContext(val sc: SparkContext)
sc.clearFiles()
}

/**
* :: DeveloperApi ::
* Register a listener to receive up-calls from events that happen during execution.
*/
@DeveloperApi
def addSparkListener(listener: SparkListener): Unit = {
sc.addSparkListener(listener)
}

/**
* Returns the Hadoop configuration used for the Hadoop code (e.g. file systems) we reuse.
*/
Expand Down

0 comments on commit 25988f3

Please sign in to comment.