diff --git a/src/main/scala/pyspark_cassandra/Pickling.scala b/src/main/scala/pyspark_cassandra/Pickling.scala index 3ca6a6f..413418d 100644 --- a/src/main/scala/pyspark_cassandra/Pickling.scala +++ b/src/main/scala/pyspark_cassandra/Pickling.scala @@ -50,13 +50,6 @@ class Pickling extends PicklingUtils { Pickler.registerCustomPickler(classOf[UDTValue], UDTValuePickler) Pickler.registerCustomPickler(classOf[DriverUDTValue], DriverUDTValuePickler) Pickler.registerCustomPickler(classOf[DataFrame], DataFramePickler) - - try{ - // Check if type converter for UUIDHolder already registered - TypeConverter.forType(typeTag[UUIDHolder]) - }catch { - case e:IllegalArgumentException => TypeConverter.registerConverter(UnpickledUUIDConverter) - } } } diff --git a/src/main/scala/pyspark_cassandra/PythonHelper.scala b/src/main/scala/pyspark_cassandra/PythonHelper.scala index 9cdb6cc..4e68967 100644 --- a/src/main/scala/pyspark_cassandra/PythonHelper.scala +++ b/src/main/scala/pyspark_cassandra/PythonHelper.scala @@ -17,26 +17,23 @@ package pyspark_cassandra import pyspark_cassandra.Utils._ import pyspark_util.Pickling._ import pyspark_util.Conversions._ - import java.lang.Boolean import java.util.{ List => JList, Map => JMap } - import scala.collection.JavaConversions._ - import org.apache.spark.SparkContext import org.apache.spark.api.java.{ JavaRDD, JavaSparkContext } import org.apache.spark.rdd.RDD import org.apache.spark.streaming.api.java.JavaDStream import org.apache.spark.streaming.dstream.DStream - import com.datastax.driver.core.ConsistencyLevel import com.datastax.spark.connector._ import com.datastax.spark.connector.rdd._ import com.datastax.spark.connector.streaming.toDStreamFunctions import com.datastax.spark.connector.writer._ +import com.datastax.spark.connector.types.TypeConverter class PythonHelper() { - + TypeConverter.registerConverter(UnpickledUUIDConverter) implicit val pickling = new Pickling() /* ----------------------------------------------------------------------- */