From c1d030253c8abbb249d6211cedf9053a8534a2ff Mon Sep 17 00:00:00 2001 From: Yang Lei Date: Tue, 13 Oct 2015 15:22:03 -0400 Subject: [PATCH] upgrade to Spark 1.5.1 --- README.md | 1 + cloudant-spark-sql/build.sbt | 2 +- .../src/main/scala/com/cloudant/spark/CloudantDatasource.scala | 2 +- .../spark/CloudantPartitionedPrunedFilteredDatasource.scala | 2 +- .../com/cloudant/spark/CloudantPrunedFilteredDatasource.scala | 2 +- .../src/main/scala/com/cloudant/spark/DefaultSource.scala | 2 +- .../spark/riak/RiakPartitionedPrunedFilteredDatasource.scala | 2 +- spark-test/build.sbt | 2 +- 8 files changed, 8 insertions(+), 7 deletions(-) diff --git a/README.md b/README.md index fdb138e..575e356 100644 --- a/README.md +++ b/README.md @@ -36,6 +36,7 @@ Spark Version | Release # | Binary Location 1.3.1 | v1.3.1.2 | [Location] (https://github.com/cloudant/spark-cloudant/releases/download/v1.3.1.2/cloudant-spark.jar) 1.4.0 | v1.4.0.0 | [Location] (https://github.com/cloudant/spark-cloudant/releases/download/1.4.0.0/cloudant-spark.jar) 1.4.1 | v1.4.1.3 | [Location] (https://github.com/cloudant/spark-cloudant/releases/download/v1.4.1.3/cloudant-spark.jar) +1.5.1 | v1.5.1.0 | [Location] (https://github.com/yanglei99/spark-cloudant/releases/download/v1.5.1.0/cloudant-spark.jar) ### Build from source: diff --git a/cloudant-spark-sql/build.sbt b/cloudant-spark-sql/build.sbt index 06b9ca6..86b7c01 100644 --- a/cloudant-spark-sql/build.sbt +++ b/cloudant-spark-sql/build.sbt @@ -14,7 +14,7 @@ resolvers ++= Seq( ) libraryDependencies ++= { - val sparkV = "1.4.1" + val sparkV = "1.5.1" val sprayV = "1.3.2" val playJsonV = "2.2.3" Seq( diff --git a/cloudant-spark-sql/src/main/scala/com/cloudant/spark/CloudantDatasource.scala b/cloudant-spark-sql/src/main/scala/com/cloudant/spark/CloudantDatasource.scala index 5403a25..ad9582f 100644 --- a/cloudant-spark-sql/src/main/scala/com/cloudant/spark/CloudantDatasource.scala +++ b/cloudant-spark-sql/src/main/scala/com/cloudant/spark/CloudantDatasource.scala @@ -17,8 +17,8 @@ package com.cloudant.spark import org.apache.spark.rdd.RDD import org.apache.spark.sql.SQLContext -import org.apache.spark.sql.catalyst.expressions.Row import org.apache.spark.sql.types._ +import org.apache.spark.sql.Row import org.apache.spark.sql.sources.{TableScan, RelationProvider, BaseRelation} import com.cloudant.spark.common._ diff --git a/cloudant-spark-sql/src/main/scala/com/cloudant/spark/CloudantPartitionedPrunedFilteredDatasource.scala b/cloudant-spark-sql/src/main/scala/com/cloudant/spark/CloudantPartitionedPrunedFilteredDatasource.scala index d781c16..8107caf 100644 --- a/cloudant-spark-sql/src/main/scala/com/cloudant/spark/CloudantPartitionedPrunedFilteredDatasource.scala +++ b/cloudant-spark-sql/src/main/scala/com/cloudant/spark/CloudantPartitionedPrunedFilteredDatasource.scala @@ -18,7 +18,7 @@ package com.cloudant.spark import org.apache.spark.{SparkContext, SparkConf} import org.apache.spark.rdd.RDD import org.apache.spark.sql.SQLContext -import org.apache.spark.sql.catalyst.expressions.Row +import org.apache.spark.sql.Row import org.apache.spark.sql.types._ import org.apache.spark.sql.sources._ import scala.collection.mutable.ArrayBuffer diff --git a/cloudant-spark-sql/src/main/scala/com/cloudant/spark/CloudantPrunedFilteredDatasource.scala b/cloudant-spark-sql/src/main/scala/com/cloudant/spark/CloudantPrunedFilteredDatasource.scala index 17e223d..2e3af6c 100644 --- a/cloudant-spark-sql/src/main/scala/com/cloudant/spark/CloudantPrunedFilteredDatasource.scala +++ b/cloudant-spark-sql/src/main/scala/com/cloudant/spark/CloudantPrunedFilteredDatasource.scala @@ -18,7 +18,7 @@ package com.cloudant.spark import org.apache.spark.{SparkContext, SparkConf} import org.apache.spark.rdd.RDD import org.apache.spark.sql.SQLContext -import org.apache.spark.sql.catalyst.expressions.Row +import org.apache.spark.sql.Row import org.apache.spark.sql.types._ import org.apache.spark.sql.sources._ import com.cloudant.spark.common._ diff --git a/cloudant-spark-sql/src/main/scala/com/cloudant/spark/DefaultSource.scala b/cloudant-spark-sql/src/main/scala/com/cloudant/spark/DefaultSource.scala index 7d1daed..ff04822 100644 --- a/cloudant-spark-sql/src/main/scala/com/cloudant/spark/DefaultSource.scala +++ b/cloudant-spark-sql/src/main/scala/com/cloudant/spark/DefaultSource.scala @@ -18,7 +18,7 @@ package com.cloudant.spark import org.apache.spark.{SparkContext, SparkConf} import org.apache.spark.rdd.RDD import org.apache.spark.sql.SQLContext -import org.apache.spark.sql.catalyst.expressions.Row +import org.apache.spark.sql.Row import org.apache.spark.sql.types._ import org.apache.spark.sql.sources._ import scala.collection.mutable.ArrayBuffer diff --git a/cloudant-spark-sql/src/main/scala/com/cloudant/spark/riak/RiakPartitionedPrunedFilteredDatasource.scala b/cloudant-spark-sql/src/main/scala/com/cloudant/spark/riak/RiakPartitionedPrunedFilteredDatasource.scala index d84b9c0..5dc006c 100644 --- a/cloudant-spark-sql/src/main/scala/com/cloudant/spark/riak/RiakPartitionedPrunedFilteredDatasource.scala +++ b/cloudant-spark-sql/src/main/scala/com/cloudant/spark/riak/RiakPartitionedPrunedFilteredDatasource.scala @@ -18,7 +18,7 @@ package com.cloudant.spark.riak import org.apache.spark.{SparkContext, SparkConf} import org.apache.spark.rdd.RDD import org.apache.spark.sql.SQLContext -import org.apache.spark.sql.catalyst.expressions.Row +import org.apache.spark.sql.Row import org.apache.spark.sql.types._ import org.apache.spark.sql.sources._ import scala.collection.mutable.ArrayBuffer diff --git a/spark-test/build.sbt b/spark-test/build.sbt index 356dc87..9716e1d 100644 --- a/spark-test/build.sbt +++ b/spark-test/build.sbt @@ -14,7 +14,7 @@ resolvers ++= Seq( ) libraryDependencies ++= { - val sparkV = "1.4.1" + val sparkV = "1.5.1" Seq( "org.apache.spark" %% "spark-core" % sparkV % "provided", "org.apache.spark" %% "spark-sql" % sparkV % "provided"