Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[TPC-H] Workers get restarted after runnning out of memory during multiple queries at scale 1000 #1367

Open
6 tasks
hendrikmakait opened this issue Feb 8, 2024 · 4 comments
Assignees
Labels

Comments

@hendrikmakait
Copy link
Member

hendrikmakait commented Feb 8, 2024

At scale 1000, all of these queries have workers getting restarted after running out of memory. We should investigate the cause and see if we're missing optimizations, have chosen a poor join order, or whether there are any other issues with these queries.

@phofl
Copy link
Contributor

phofl commented Feb 9, 2024

Query 18 most likely dies because our source dataset is weird. We have files that have 50mbs in memory and files that have 380mbs in memory. The latter is relatively big for our small machines (8GB of ram). This gets worse through our strategy of combining multiple partitions when we drop columns, we end up combining a few large ones which makes them even bigger.

I don't know how we want to proceed exactly, but the varying partitions are probably not very good for what we want to do here.

Edit: This is not compression related, the difference scales down to compressed file sizes

@fjetter
Copy link
Member

fjetter commented Feb 12, 2024

Varying partition sizes are very realistic and we shouldn't micro optimize our code to only run on extremely homogeneous datasets

@phofl
Copy link
Contributor

phofl commented Feb 12, 2024

I agree, but this is hard to change with the Current read_parquet

@phofl
Copy link
Contributor

phofl commented Feb 12, 2024

See #1376 for 17 and 18

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

3 participants