Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[SPARK-24821][Core] Fail fast when submitted job compute on a subset of all the partitions for a barrier stage #21918

Closed
wants to merge 1 commit into from

Conversation

jiangxb1987
Copy link
Contributor

What changes were proposed in this pull request?

Check on DAGScheduler.submitJob() to make sure we are not launching a barrier stage on only a subset of all the partitions(one example is the first() operation), otherwise shall fail fast.

How was this patch tested?

Add new test case in BarrierStageOnSubmittedSuite.

@holdensmagicalunicorn
Copy link

@jiangxb1987, thanks! I am a bot who has found some folks who might be able to help with the review:@mateiz, @rxin and @kayousterhout

@SparkQA
Copy link

SparkQA commented Jul 30, 2018

Test build #93792 has finished for PR 21918 at commit b93d212.

  • This patch fails Spark unit tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@jiangxb1987
Copy link
Contributor Author

retest this please

@SparkQA
Copy link

SparkQA commented Jul 31, 2018

Test build #93816 has finished for PR 21918 at commit b93d212.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@mengxr
Copy link
Contributor

mengxr commented Jul 31, 2018

@jiangxb1987 See my comment #21927 (comment).

@mengxr
Copy link
Contributor

mengxr commented Aug 2, 2018

@jiangxb1987 This should be addressed by #21927 . Could you close this PR?

@jiangxb1987 jiangxb1987 closed this Aug 2, 2018
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
4 participants