Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature] Reduce redundant shuffle for spark dynamic bucket writes #3222

Open
2 tasks done
wForget opened this issue Apr 17, 2024 · 2 comments
Open
2 tasks done

[Feature] Reduce redundant shuffle for spark dynamic bucket writes #3222

wForget opened this issue Apr 17, 2024 · 2 comments
Labels
enhancement New feature or request

Comments

@wForget
Copy link
Member

wForget commented Apr 17, 2024

Search before asking

  • I searched in the issues and found nothing similar.

Motivation

Dynamic bucket writing does two shuffles, the first repartitionByKeyPartitionHash seems unnecessary, It seems to be only used to determine assignId. However, assignId can be calculated through partitionHash/keyHash/numParallelism/numAssigners, we do not need to do extra shuffle. Can we remove it?

Solution

No response

Anything else?

No response

Are you willing to submit a PR?

  • I'm willing to submit a PR!
@wForget wForget added the enhancement New feature or request label Apr 17, 2024
@wForget
Copy link
Member Author

wForget commented Apr 17, 2024

@YannByron could you please take a look?

@wForget wForget changed the title [Feature] Reduce redundant shuffle for dynamic bucket writes [Feature] Reduce redundant shuffle for spark dynamic bucket writes Apr 17, 2024
@JingsongLi
Copy link
Contributor

it is hard, Perhaps different assigners will have the same bucket data

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants