From 830e042c17fcc6e743d0af41d6c2c4bf704ef67c Mon Sep 17 00:00:00 2001 From: "Zhang, Liye" Date: Tue, 4 Nov 2014 17:48:41 +0800 Subject: [PATCH] [CORE][minor] minor code cleanup and small modifications --- .../org/apache/spark/deploy/worker/ExecutorRunner.scala | 5 +---- .../scala/org/apache/spark/ui/jobs/JobProgressListener.scala | 2 +- 2 files changed, 2 insertions(+), 5 deletions(-) diff --git a/core/src/main/scala/org/apache/spark/deploy/worker/ExecutorRunner.scala b/core/src/main/scala/org/apache/spark/deploy/worker/ExecutorRunner.scala index 8ba6a01bbcb97..b788572f1b7ec 100644 --- a/core/src/main/scala/org/apache/spark/deploy/worker/ExecutorRunner.scala +++ b/core/src/main/scala/org/apache/spark/deploy/worker/ExecutorRunner.scala @@ -47,7 +47,7 @@ private[spark] class ExecutorRunner( val executorDir: File, val workerUrl: String, val conf: SparkConf, - var state: ExecutorState.Value) + var state: ExecutorState.ExecutorState) extends Logging { val fullId = appId + "/" + execId @@ -133,15 +133,12 @@ private[spark] class ExecutorRunner( // parent process for the executor command builder.environment.put("SPARK_LAUNCH_WITH_SCALA", "0") process = builder.start() - val header = "Spark Executor Command: %s\n%s\n\n".format( - command.mkString("\"", "\" \"", "\""), "=" * 40) // Redirect its stdout and stderr to files val stdout = new File(executorDir, "stdout") stdoutAppender = FileAppender(process.getInputStream, stdout, conf) val stderr = new File(executorDir, "stderr") - Files.write(header, stderr, UTF_8) stderrAppender = FileAppender(process.getErrorStream, stderr, conf) state = ExecutorState.RUNNING diff --git a/core/src/main/scala/org/apache/spark/ui/jobs/JobProgressListener.scala b/core/src/main/scala/org/apache/spark/ui/jobs/JobProgressListener.scala index e3223403c17f4..cdca69efc5a56 100644 --- a/core/src/main/scala/org/apache/spark/ui/jobs/JobProgressListener.scala +++ b/core/src/main/scala/org/apache/spark/ui/jobs/JobProgressListener.scala @@ -61,7 +61,7 @@ class JobProgressListener(conf: SparkConf) extends SparkListener with Logging { val stageIdToInfo = new HashMap[StageId, StageInfo] // Number of completed and failed stages, may not actually equal to completedStages.size and - // failedStages.size respectively due to completedStage and failedStages only maintain the latest + // failedStages.size respectively due to completedStages and failedStages only maintain the latest // part of the stages, the earlier ones will be removed when there are too many stages for // memory sake. var numCompletedStages = 0