-
Notifications
You must be signed in to change notification settings - Fork 28.9k
[SPARK-12077] [SQL] change the default plan for single distinct #10075
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
Test build #46994 has finished for PR 10075 at commit
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I actually prefer the old name better. Two years from now, I'm not sure how much context we will have about "1.5".
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think the big difference is that Oracle releases every x years, while we release every 3 month...
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
But we are doing the same thing as maintaining compatibility and no (much) regressions, otherwise we should just remove this flag (it's not public right now).
This reverts commit 192a04d.
|
LGTM pending jenkins. |
|
Test build #2137 has finished for PR 10075 at commit
|
|
Test build #47023 has finished for PR 10075 at commit
|
|
Test build #2142 has finished for PR 10075 at commit
|
|
LGTM. Merging to master and branch 1.6. |
Use try to match the behavior for single distinct aggregation with Spark 1.5, but that's not scalable, we should be robust by default, have a flag to address performance regression for low cardinality aggregation. cc yhuai nongli Author: Davies Liu <[email protected]> Closes #10075 from davies/agg_15. (cherry picked from commit 96691fe) Signed-off-by: Yin Huai <[email protected]>
Use try to match the behavior for single distinct aggregation with Spark 1.5, but that's not scalable, we should be robust by default, have a flag to address performance regression for low cardinality aggregation.
cc @yhuai @nongli