Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Start to use explicit memory limits in the parquet chunked reader #9991

Merged
merged 3 commits into from
Jan 30, 2024

Conversation

revans2
Copy link
Collaborator

@revans2 revans2 commented Dec 7, 2023

This depends on rapidsai/cudf#14360

Signed-off-by: Robert (Bobby) Evans <bobby@apache.org>
@revans2 revans2 added the cudf_dependency An issue or PR with this label depends on a new feature in cudf label Dec 7, 2023
@sameerz sameerz added the reliability Features to improve reliability or bugs that severly impact the reliability of the plugin label Dec 10, 2023
Comment on lines +2584 to +2585
val passReadLimit = if (useSubPageChunked) {
4 * chunkSizeByteLimit
Copy link
Collaborator

@ttnghia ttnghia Jan 26, 2024

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can we set some multiplier constant or configurable constant for it, instead of hard coding 4X like this? In the (near) future, chunked ORC reader may have benefit from it.

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm not sure where would be a good place to put this. It is a magic number based on our estimates that we will take 4x the target batch size as the amount of memory that we are allowed to use. GpuDataProducer is the only place that might be common between them, but that is not a proper place for it. I could create a static object to hold it somewhere. GpuConventionMagicNumbers or something.

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Okay then let's see what can we do the next time with the ORC chunked reader. I'm fine to leave this as-is for now.

@revans2 revans2 marked this pull request as ready for review January 29, 2024 22:03
@revans2
Copy link
Collaborator Author

revans2 commented Jan 29, 2024

build

@revans2
Copy link
Collaborator Author

revans2 commented Jan 30, 2024

@ttnghia could you please take another look?

.doc("Enable a chunked reader where possible for reading data that is smaller " +
"than the typical row group/page limit. Currently this only works for parquet.")
.booleanConf
.createWithDefault(false)
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

We should have a followup issue to turn this on in 24.04.

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
cudf_dependency An issue or PR with this label depends on a new feature in cudf reliability Features to improve reliability or bugs that severly impact the reliability of the plugin
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

5 participants