Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[SPARK-17119][Core]allow the history server to delete .inprogress files(configurable) #16293

Closed
wants to merge 3 commits into from

Conversation

cnZach
Copy link
Contributor

@cnZach cnZach commented Dec 15, 2016

What changes were proposed in this pull request?

The History Server (HS) currently only considers completed applications when deleting event logs from spark.history.fs.logDirectory (since SPARK-6879). This means that over time, .inprogress files (from failed jobs, jobs where the SparkContext is not closed, spark-shell exits etc...) can accumulate and impact the HS.

Instead of having to manually delete these files, this change add a configurable feature to let user decide if the .inprogress files should also be deleted after a period of time:
spark.history.fs.cleaner.deleteInProgress.enabled
spark.history.fs.cleaner.noProgressMaxAge

How was this patch tested?

verified with manual tests
unit tests added in FsHistoryProviderSuite.scala but I am not able to run ./dev/run-tests for the whole project on my laptop, failed on SparkSinkSuite and network related tests uner org.apache.spark.network.* (all due to java.io.IOException: Failed to connect to /<my_laptop_ip>:62343).

[info] SparkSinkSuite:
[info] - Success with ack *** FAILED *** (1 minute)
[info] java.io.IOException: Error connecting to /0.0.0.0:62298
[info] at org.apache.avro.ipc.NettyTransceiver.getChannel(NettyTransceiver.java:261)

doc

monitoring.md is also updated

@AmplabJenkins
Copy link

Can one of the admins verify this patch?

@vanzin
Copy link
Contributor

vanzin commented Dec 15, 2016

@cnZach could you close this? This was already implemented in SPARK-8617. I don't think we need a config option.

@splinepalash
Copy link

Hi Respected contributors,

Thanks for all of your hard work to make this platform really stable.

I'm almost new to Spark and building an application using it. I was facing problems to initiate my new spark jobs if any previous job is failed due to some problem. Could anyone Please help me how I can use this functionality? Is it packaged in any Spark official realest?

Best regards,
Palash Gupta

@cnZach
Copy link
Contributor Author

cnZach commented Dec 19, 2016

as there's already some changes implemented in SPARK-8617, close this PR.

@cnZach cnZach closed this Dec 19, 2016
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants