Skip to content

Conversation

@jiangxb1987
Copy link
Contributor

@jiangxb1987 jiangxb1987 commented Oct 16, 2017

What changes were proposed in this pull request?

Update the config spark.files.ignoreEmptySplits, rename it and make it internal.

This is followup of #19464

How was this patch tested?

Exsiting tests.

@jiangxb1987
Copy link
Contributor Author

Copy link
Member

@HyukjinKwon HyukjinKwon left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM too

@liutang123
Copy link
Contributor

It looks better.

@SparkQA
Copy link

SparkQA commented Oct 16, 2017

Test build #82787 has finished for PR 19504 at commit bcb3dbd.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

.set(HADOOP_RDD_IGNORE_EMPTY_SPLITS, true)
sc = new SparkContext(conf)

def testIgnoreEmptySplits(
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

testIgnoreEmptySplits(
       data = Array.empty[Tuple2[String, String]],
       actualPartitionNum = 1,
       expectedPartitionNum = 0)

=>

testIgnoreEmptySplits(
       data = Array.empty[(String, String)],
       actualPartitionNum = 1,
       expectedPartitionNum = 0)

@cloud-fan
Copy link
Contributor

LGTM, merging to master!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

6 participants