Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Improment] The issue of indexing large volumes of data tables. #2253

Closed
1 of 2 tasks
wg1026688210 opened this issue Nov 3, 2023 · 1 comment
Closed
1 of 2 tasks
Labels
enhancement New feature or request

Comments

@wg1026688210
Copy link
Contributor

wg1026688210 commented Nov 3, 2023

Search before asking

  • I searched in the issues and found nothing similar.

Motivation

Now we have seen significant improvements in query performance through index creation. However, we have encountered several issues during the indexing process.

  1. Some tasks before sorting and writing to Paimon are getting stuck waiting for data for a long time.
  2. Due to data skewness, some indexing tasks take a long time to execute.
  3. The flink batch Job of building index will global failover after taskManager oom
  4. FilesTable can not query the paritions which has built index

Solution

No response

Anything else?

No response

Are you willing to submit a PR?

  • I'm willing to submit a PR!
@wg1026688210 wg1026688210 added the enhancement New feature or request label Nov 3, 2023
@wg1026688210
Copy link
Contributor Author

issue 1,2 has been fixed
related
#3081
#2749

issue 3 need a remote suffle service for task failover when using flink

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant