Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Enable kudo serializer by default #12222

Open
wants to merge 3 commits into
base: branch-25.04
Choose a base branch
from

Conversation

liurenjie1024
Copy link
Collaborator

Closes #12202 .

Enable kudo serializer by default, and contains several fix due to shuffle size change.

Signed-off-by: liurenjie1024 <liurenjie2008@gmail.com>
@liurenjie1024
Copy link
Collaborator Author

build

Copy link
Collaborator

@abellina abellina left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@liurenjie1024 can you provide performance numbers for NDS with and without MT shuffle?

@sameerz sameerz added the performance A performance related task/issue label Feb 25, 2025
@liurenjie1024
Copy link
Collaborator Author

build

@@ -500,7 +500,7 @@ class AdaptiveQueryExecSuite
val conf = new SparkConf()
.set(SQLConf.ADAPTIVE_EXECUTION_ENABLED.key, "true")
.set(SQLConf.LOCAL_SHUFFLE_READER_ENABLED.key, "true")
.set(SQLConf.AUTO_BROADCASTJOIN_THRESHOLD.key, "400")
.set(SQLConf.AUTO_BROADCASTJOIN_THRESHOLD.key, "50")
Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This change is required since shuffle size changed.

@@ -99,7 +99,7 @@ class GpuLoreSuite extends SparkQueryCompareTestSuite with FunSuiteWithTempDir w
}

test("AQE broadcast") {
doTestReplay("90[*]") { spark =>
doTestReplay("93[*]") { spark =>
Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Same as above, shuffle size change leads to plan change.

@@ -2052,7 +2052,7 @@ val SHUFFLE_COMPRESSION_LZ4_CHUNK_SIZE = conf("spark.rapids.shuffle.compression.
.internal()
.startupOnly()
.booleanConf
.createWithDefault(false)
.createWithDefault(true)
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

For spark.rapids.shuffle.kudo.serializer.measure.buffer.copy.enabled, should we enable that by default?

Copy link
Collaborator

@revans2 revans2 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The changes themselves look fine. I mostly want to see the performance numbers to show that it is at least as good as the old code. I know we have done some of that in the past and that there have been a lot of optimizations recently so it should be good. But this is a big change so I want to see it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
performance A performance related task/issue
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Enable kudo serializer by default.
5 participants