-
Notifications
You must be signed in to change notification settings - Fork 28k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[SPARK-2454] Do not ship spark home to Workers #1734
Changes from 3 commits
09272b7
188fc5d
1c2532c
f71f391
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -84,7 +84,7 @@ def __init__(self, master=None, appName=None, sparkHome=None, pyFiles=None, | |
@param serializer: The serializer for RDDs. | ||
@param conf: A L{SparkConf} object setting Spark properties. | ||
@param gateway: Use an existing gateway and JVM, otherwise a new JVM | ||
will be instatiated. | ||
will be instantiated. | ||
|
||
|
||
>>> from pyspark.context import SparkContext | ||
|
@@ -126,8 +126,6 @@ def _do_init(self, master, appName, sparkHome, pyFiles, environment, batchSize, | |
self._conf.setMaster(master) | ||
if appName: | ||
self._conf.setAppName(appName) | ||
if sparkHome: | ||
self._conf.setSparkHome(sparkHome) | ||
if environment: | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I don't see why we should disable this. There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. This is part of an effort to shift away from calling setSparkHome within Spark completely. After all, all that does is set |
||
for key, value in environment.iteritems(): | ||
self._conf.setExecutorEnv(key, value) | ||
|
@@ -144,7 +142,7 @@ def _do_init(self, master, appName, sparkHome, pyFiles, environment, batchSize, | |
# the classpath or an external config file | ||
self.master = self._conf.get("spark.master") | ||
self.appName = self._conf.get("spark.app.name") | ||
self.sparkHome = self._conf.get("spark.home", None) | ||
self.sparkHome = os.environ.get("SPARK_HOME") | ||
for (k, v) in self._conf.getAll(): | ||
if k.startswith("spark.executorEnv."): | ||
varName = k[len("spark.executorEnv."):] | ||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
is this needed to make the tests work?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yeah, otherwise all tests with
local-cluster
fail