From 06f062dc701ef132298b09699411c83f5498912a Mon Sep 17 00:00:00 2001 From: Patrick Wendell Date: Wed, 11 Jun 2014 14:11:26 -0700 Subject: [PATCH 1/2] HOTFIX: PySpark tests should be order insensitive --- python/pyspark/sql.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/python/pyspark/sql.py b/python/pyspark/sql.py index b4e9618cc25b5..960d0a82448aa 100644 --- a/python/pyspark/sql.py +++ b/python/pyspark/sql.py @@ -117,7 +117,7 @@ def parquetFile(self, path): >>> srdd = sqlCtx.inferSchema(rdd) >>> srdd.saveAsParquetFile(parquetFile) >>> srdd2 = sqlCtx.parquetFile(parquetFile) - >>> srdd.collect() == srdd2.collect() + >>> sorted(srdd.collect()) == sorted(srdd2.collect()) True """ jschema_rdd = self._ssql_ctx.parquetFile(path) @@ -141,7 +141,7 @@ def table(self, tableName): >>> srdd = sqlCtx.inferSchema(rdd) >>> sqlCtx.registerRDDAsTable(srdd, "table1") >>> srdd2 = sqlCtx.table("table1") - >>> srdd.collect() == srdd2.collect() + >>> sorted(srdd.collect()) == sorted(srdd2.collect()) True """ return SchemaRDD(self._ssql_ctx.table(tableName), self) @@ -293,7 +293,7 @@ def saveAsParquetFile(self, path): >>> srdd = sqlCtx.inferSchema(rdd) >>> srdd.saveAsParquetFile(parquetFile) >>> srdd2 = sqlCtx.parquetFile(parquetFile) - >>> srdd2.collect() == srdd.collect() + >>> sorted(srdd2.collect()) == sorted(srdd.collect()) True """ self._jschema_rdd.saveAsParquetFile(path) @@ -307,7 +307,7 @@ def registerAsTable(self, name): >>> srdd = sqlCtx.inferSchema(rdd) >>> srdd.registerAsTable("test") >>> srdd2 = sqlCtx.sql("select * from test") - >>> srdd.collect() == srdd2.collect() + >>> sorted(srdd.collect()) == sorted(srdd2.collect()) True """ self._jschema_rdd.registerAsTable(name) From 1eb54871a2475091ebbd183dc4a57d156bf36626 Mon Sep 17 00:00:00 2001 From: Patrick Wendell Date: Wed, 11 Jun 2014 14:12:15 -0700 Subject: [PATCH 2/2] False change --- sql/core/src/main/scala/org/apache/spark/sql/SchemaRDD.scala | 1 + 1 file changed, 1 insertion(+) diff --git a/sql/core/src/main/scala/org/apache/spark/sql/SchemaRDD.scala b/sql/core/src/main/scala/org/apache/spark/sql/SchemaRDD.scala index 7ad8edf5a5a6e..44b19bca460b0 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/SchemaRDD.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/SchemaRDD.scala @@ -33,6 +33,7 @@ import org.apache.spark.api.java.JavaRDD import java.util.{Map => JMap} /** + * ***FALSE CHANGE*** * :: AlphaComponent :: * An RDD of [[Row]] objects that has an associated schema. In addition to standard RDD functions, * SchemaRDDs can be used in relational queries, as shown in the examples below.