From f2b597022b4fc4023c238e5b5a9824946f84f84e Mon Sep 17 00:00:00 2001 From: lianhuiwang Date: Fri, 23 May 2014 22:02:57 +0800 Subject: [PATCH 1/4] bugfix worker DriverStateChanged state should match DriverState.FAILED --- core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala | 2 ++ 1 file changed, 2 insertions(+) diff --git a/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala b/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala index fb9cc116cd08b..e7ff855010488 100755 --- a/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala +++ b/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala @@ -311,6 +311,8 @@ private[spark] class Worker( state match { case DriverState.ERROR => logWarning(s"Driver $driverId failed with unrecoverable exception: ${exception.get}") + case DriverState.FAILED => + logWarning(s"Driver $driverId failed with state $state") case DriverState.FINISHED => logInfo(s"Driver $driverId exited successfully") case DriverState.KILLED => From 480ce949a83c0d854078b38f5665f3369cf759eb Mon Sep 17 00:00:00 2001 From: lianhuiwang Date: Sat, 24 May 2014 23:24:37 +0800 Subject: [PATCH 2/4] address aarondav comments --- .../main/scala/org/apache/spark/deploy/worker/Worker.scala | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala b/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala index e7ff855010488..e3533bdf8e03e 100755 --- a/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala +++ b/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala @@ -312,11 +312,13 @@ private[spark] class Worker( case DriverState.ERROR => logWarning(s"Driver $driverId failed with unrecoverable exception: ${exception.get}") case DriverState.FAILED => - logWarning(s"Driver $driverId failed with state $state") + logWarning(s"Driver $driverId exited with failure") case DriverState.FINISHED => logInfo(s"Driver $driverId exited successfully") case DriverState.KILLED => logInfo(s"Driver $driverId was killed by user") + case _ => + logDebug(s"Driver $driverId changed state to $state") } masterLock.synchronized { master ! DriverStateChanged(driverId, state, exception) From 5f6bbb7119ecd188af4967ac15f3ff1986ad400d Mon Sep 17 00:00:00 2001 From: Wang Lianhui Date: Wed, 16 Jul 2014 22:40:03 +0800 Subject: [PATCH 3/4] missing document about spark.deploy.retainedDrivers The configuration on spark.deploy.retainedDrivers is undocumented but actually used https://github.com/apache/spark/blob/master/core/src/main/scala/org/apache/spark/deploy/master/Master.scala#L60 --- docs/spark-standalone.md | 14 ++++++++++++++ 1 file changed, 14 insertions(+) diff --git a/docs/spark-standalone.md b/docs/spark-standalone.md index f5c0f7cef83d2..6ed128b9fe66a 100644 --- a/docs/spark-standalone.md +++ b/docs/spark-standalone.md @@ -156,6 +156,20 @@ SPARK_MASTER_OPTS supports the following system properties: + + + + + + + + + + From 64660fd0a7844e4fc6603257e6db14f1b084843c Mon Sep 17 00:00:00 2001 From: Wang Lianhui Date: Thu, 17 Jul 2014 09:36:08 +0800 Subject: [PATCH 4/4] address pwendell's comments --- docs/spark-standalone.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/docs/spark-standalone.md b/docs/spark-standalone.md index 6ed128b9fe66a..ad8b6c0e51a78 100644 --- a/docs/spark-standalone.md +++ b/docs/spark-standalone.md @@ -160,14 +160,14 @@ SPARK_MASTER_OPTS supports the following system properties:
Property NameDefaultMeaning
spark.deploy.retainedApplications200 + The number of completedApps to retain. If this cap is exceeded, then the oldest completedApps will be removed.
+
spark.deploy.retainedDrivers200 + The number of completedDrivers to retain. If this cap is exceeded, then the oldest completedDrivers will be removed.
+
spark.deploy.spreadOut truespark.deploy.retainedApplications 200 - The number of completedApps to retain. If this cap is exceeded, then the oldest completedApps will be removed.
+ The maximum number of completed applications to display. Older applications will be dropped from the UI to maintain this limit.
spark.deploy.retainedDrivers 200 - The number of completedDrivers to retain. If this cap is exceeded, then the oldest completedDrivers will be removed.
+ The maximum number of completed drivers to display. Older drivers will be dropped from the UI to maintain this limit.