diff --git a/bin/spark-class b/bin/spark-class index 5f047b3a777f2..629843333f2cc 100755 --- a/bin/spark-class +++ b/bin/spark-class @@ -170,6 +170,7 @@ if [ -n "$SPARK_SUBMIT_BOOTSTRAP_DRIVER" ]; then shift exec "$RUNNER" org.apache.spark.deploy.SparkSubmitDriverBootstrapper "$@" else + # Note: The format of this command is closely echoed in SparkSubmitDriverBootstrapper.scala if [ -n "$SPARK_PRINT_LAUNCH_COMMAND" ]; then echo -n "Spark Command: " 1>&2 echo "$RUNNER" -cp "$CLASSPATH" $JAVA_OPTS "$@" 1>&2 diff --git a/bin/spark-submit b/bin/spark-submit index fd8f4b8e80de9..32c911cd0438b 100755 --- a/bin/spark-submit +++ b/bin/spark-submit @@ -46,7 +46,7 @@ export SPARK_SUBMIT_PROPERTIES_FILE=${SPARK_SUBMIT_PROPERTIES_FILE:-"$DEFAULT_PR # For client mode, the driver will be launched in the same JVM that launches # SparkSubmit, so we may need to read the properties file for any extra class # paths, library paths, java options and memory early on. Otherwise, it will -# be too late by the time the JVM has started. +# be too late by the time the driver JVM has started. if [[ "$SPARK_SUBMIT_DEPLOY_MODE" == "client" && -f "$SPARK_SUBMIT_PROPERTIES_FILE" ]]; then # Parse the properties file only if the special configs exist