Skip to content

Commit

Permalink
[SPARK-2627] [PySpark] minor indentation changes
Browse files Browse the repository at this point in the history
  • Loading branch information
nchammas committed Aug 6, 2014
1 parent 983d963 commit 274b238
Showing 1 changed file with 9 additions and 7 deletions.
16 changes: 9 additions & 7 deletions python/pyspark/tests.py
Original file line number Diff line number Diff line change
Expand Up @@ -594,7 +594,8 @@ def test_oldhadoop(self):
"mapred.output.format.class": "org.apache.hadoop.mapred.SequenceFileOutputFormat",
"mapred.output.key.class": "org.apache.hadoop.io.IntWritable",
"mapred.output.value.class": "org.apache.hadoop.io.MapWritable",
"mapred.output.dir": basepath + "/olddataset/"}
"mapred.output.dir": basepath + "/olddataset/"
}
self.sc.parallelize(dict_data).saveAsHadoopDataset(conf)
input_conf = {"mapred.input.dir": basepath + "/olddataset/"}
old_dataset = sorted(self.sc.hadoopRDD(
Expand Down Expand Up @@ -624,11 +625,13 @@ def test_newhadoop(self):
valueConverter="org.apache.spark.api.python.WritableToDoubleArrayConverter").collect())
self.assertEqual(result, array_data)

conf = {"mapreduce.outputformat.class":
conf = {
"mapreduce.outputformat.class":
"org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat",
"mapred.output.key.class": "org.apache.hadoop.io.IntWritable",
"mapred.output.value.class": "org.apache.spark.api.python.DoubleArrayWritable",
"mapred.output.dir": basepath + "/newdataset/"}
"mapred.output.key.class": "org.apache.hadoop.io.IntWritable",
"mapred.output.value.class": "org.apache.spark.api.python.DoubleArrayWritable",
"mapred.output.dir": basepath + "/newdataset/"
}
self.sc.parallelize(array_data).saveAsNewAPIHadoopDataset(
conf,
valueConverter="org.apache.spark.api.python.DoubleArrayToWritableConverter")
Expand Down Expand Up @@ -1012,8 +1015,7 @@ class NumPyTests(PySparkTestCase):
"""General PySpark tests that depend on numpy """

def test_statcounter_array(self):
x = self.sc.parallelize(
[np.array([1.0, 1.0]), np.array([2.0, 2.0]), np.array([3.0, 3.0])])
x = self.sc.parallelize([np.array([1.0, 1.0]), np.array([2.0, 2.0]), np.array([3.0, 3.0])])
s = x.stats()
self.assertSequenceEqual([2.0, 2.0], s.mean().tolist())
self.assertSequenceEqual([1.0, 1.0], s.min().tolist())
Expand Down

0 comments on commit 274b238

Please sign in to comment.