From 4b35c8d54a17f7fb867ae0ecabfaabe887bd5cc5 Mon Sep 17 00:00:00 2001 From: Jongyoul Lee Date: Thu, 25 Jun 2015 12:04:37 +0900 Subject: [PATCH] [ZEPPELIN-18] Running pyspark without deploying python libraries to every yarn node - Dummy for trigger --- .../main/java/org/apache/zeppelin/spark/SparkInterpreter.java | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/spark/src/main/java/org/apache/zeppelin/spark/SparkInterpreter.java b/spark/src/main/java/org/apache/zeppelin/spark/SparkInterpreter.java index bf229cadecc..a3670533c4b 100644 --- a/spark/src/main/java/org/apache/zeppelin/spark/SparkInterpreter.java +++ b/spark/src/main/java/org/apache/zeppelin/spark/SparkInterpreter.java @@ -278,7 +278,7 @@ public SparkContext createSparkContext() { for (String lib : pythonLibs) { File libFile = new File(zeppelinPythonLibPath, lib); if (libFile.exists()) { - pythonLibUris.add(libFile.toURI().toString()); + pythonLibUris.add(libFile.toURI().toString()); } } pythonLibUris.trimToSize();