Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[SPARK-12992][SQL]: Update parquet reader to support more types when decoding to ColumnarBatch. #10908

Closed
wants to merge 2 commits into from

Conversation

nongli
Copy link
Contributor

@nongli nongli commented Jan 26, 2016

This patch implements support for more types when doing the vectorized decode. There are
a few more types remaining but they should be very straightforward after this. This code
has a few copy and paste pieces but they are difficult to eliminate due to performance
considerations.

Specifically, this patch adds support for:

  • String, Long, Byte types
  • Dictionary encoding for those types.

@SparkQA
Copy link

SparkQA commented Jan 26, 2016

Test build #50047 has finished for PR 10908 at commit ea1f406.

  • This patch fails Spark unit tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@SparkQA
Copy link

SparkQA commented Jan 27, 2016

Test build #50174 has finished for PR 10908 at commit 2c8aaad.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds no public classes.


@Override
public byte readByte() {
throw new UnsupportedOperationException("only readInts is valid.");
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

readInts -> readBytes

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This should be readInts. The only valid read* APIs that doesn't also decode definition levels is used to decode dictionary ids, which are always ints. I updated the comment for readIntegers() to try to capture this.

…decoding to ColumnarBatch.

This patch implements support for more types when doing the vectorized decode. There are
a few more types remaining but they should be very straightfoward after this. This code
has a few copy and paste pieces but they are difficult to eliminate due to performance
considerations.

Specifically, this patch adds support for:
  - String, Long, Byte types
  - Dictionary encoding for those types.
@SparkQA
Copy link

SparkQA commented Feb 1, 2016

Test build #50499 has finished for PR 10908 at commit 2ea2d54.

  • This patch fails Spark unit tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@SparkQA
Copy link

SparkQA commented Feb 3, 2016

Test build #2493 has finished for PR 10908 at commit 2ea2d54.

  • This patch passes all tests.
  • This patch does not merge cleanly.
  • This patch adds no public classes.

@davies
Copy link
Contributor

davies commented Feb 3, 2016

LGTM, merging this into master, thanks!

@asfgit asfgit closed this in 21112e8 Feb 3, 2016
@nongli nongli deleted the spark-12992 branch February 3, 2016 19:17
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
3 participants