-
Notifications
You must be signed in to change notification settings - Fork 28.9k
SPARK-1860: Do not cleanup application work/ directories by default #800
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
This causes an unrecoverable error for applications that are running for longer than 7 days that have jars added to the SparkContext, as the jars are cleaned up even though the application is still running.
|
Merged build triggered. |
|
Merged build started. |
|
Sure - might be good to have it off by default. /cc @velvia. |
|
@aarondav what about just not cleaning up the data if the app is still running? In the future we should probably assess the TTL based on the finish time of the app, not the start time. |
|
Merged build finished. All automated tests passed. |
|
All automated tests passed. |
|
This patch is intended as a hotfix in the hopes that it can make it into the 1.0 release. Avoiding cleaning up running applications seems like the better solution in general, but is out of scope of this PR. |
|
I agree, okay let's just pull in this fix and we can hopefully patch the bigger issue later. |
|
LGTM |
|
btw - regarding the branch name - I don't think this was too shitty of a default. I'd actually like to have this on by default if we can get it into working order, because otherwise users will only find out once it's too late that they are out of disk space :P |
This causes an unrecoverable error for applications that are running for longer than 7 days that have jars added to the SparkContext, as the jars are cleaned up even though the application is still running. Author: Aaron Davidson <[email protected]> Closes #800 from aarondav/shitty-defaults and squashes the following commits: a573fbb [Aaron Davidson] SPARK-1860: Do not cleanup application work/ directories by default (cherry picked from commit bb98eca) Signed-off-by: Patrick Wendell <[email protected]>
|
Just a little miffed because it took some time to figure out why our executors suddenly started failing with jar-not-found errors :) I'd prefer a full solution; last modified time runs into an issue if the executor lies dormant for a week. You might say, "that's unlikely", but I'd say, "it'll happen to someone, and they'll be a little miffed." The worker should have enough state to figure out which executors are currently active, though I'm not sure if the problem is made more difficult by multi-worker scenarios. |
This causes an unrecoverable error for applications that are running for longer than 7 days that have jars added to the SparkContext, as the jars are cleaned up even though the application is still running. Author: Aaron Davidson <[email protected]> Closes apache#800 from aarondav/shitty-defaults and squashes the following commits: a573fbb [Aaron Davidson] SPARK-1860: Do not cleanup application work/ directories by default
…se message (apache#800) [HADP-58798] Add utility functions to facilitate logging spark diagnose message
This causes an unrecoverable error for applications that are running for longer
than 7 days that have jars added to the SparkContext, as the jars are cleaned up
even though the application is still running.