From ac6b82c36f4cc5a40b23ec1cad81249fde32c3cf Mon Sep 17 00:00:00 2001 From: Takeshi YAMAMURO Date: Wed, 2 Mar 2016 10:17:17 +0900 Subject: [PATCH] Use foreach not map --- .../apache/spark/sql/execution/datasources/text/TextSuite.scala | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/text/TextSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/text/TextSuite.scala index 3277dca9d5c54..59e0e6a7cfa00 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/text/TextSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/text/TextSuite.scala @@ -60,7 +60,7 @@ class TextSuite extends QueryTest with SharedSQLContext { test("SPARK-13503 Support to specify the option for compression codec for TEXT") { val testDf = sqlContext.read.text(testFile) - Seq("bzip2", "deflate", "gzip").map { codecName => + Seq("bzip2", "deflate", "gzip").foreach { codecName => val tempDir = Utils.createTempDir() val tempDirPath = tempDir.getAbsolutePath() testDf.write.option("compression", codecName).mode(SaveMode.Overwrite).text(tempDirPath)