Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature] Spark 3.1.1 Support #1617

Closed
1 of 2 tasks
johnc1231 opened this issue Mar 9, 2021 · 7 comments
Closed
1 of 2 tasks

[Feature] Spark 3.1.1 Support #1617

johnc1231 opened this issue Mar 9, 2021 · 7 comments

Comments

@johnc1231
Copy link

johnc1231 commented Mar 9, 2021

What kind an issue is this?

  • Bug report. If you’ve found a bug, please provide a code snippet or test to reproduce it below.
    The easier it is to track down the bug, the faster it is solved.
  • Feature Request. Start by telling us what problem you’re trying to solve.
    Often a solution already exists! Don’t send pull requests to implement new features without
    first getting our support. Sometimes we leave features out on purpose to keep the project small.

Feature description

I'm wondering what the plan is for Spark 3.1.1 support. I've been able to build elasticsearch-spark-30 for Spark 3.1.1 locally with only a minor tweak to the current code (seems like compactLogs method is gone in 3.1.1, so needed a change here. compactLogs was removed here )

@koertkuipers
Copy link

did you get test/check to pass?

spark switched from javax.servlet to jakarta.servlet for servlet-api and this seems to have broken tests for me.
i get:

java.lang.NoClassDefFoundError: javax/servlet/http/HttpSessionIdListener
at org.sparkproject.jetty.server.handler.ContextHandler.<clinit>(ContextHandler.java:121) 

@johnc1231
Copy link
Author

Good point, I haven't run all the tests. I just swapped it in for my use case and was able to use it successfully.

I think it wouldn't be too hard to handle the changes, but I'm wondering how this project handles backwards incompatible changes in Spark. Do you just accept a PR to get Spark 3.1.1 working, at the expense of Spark 3.0.0, or do we have to maintain both?

@jbaiera
Copy link
Member

jbaiera commented Mar 10, 2021

I ran into these two problems while getting the Spark 3.0 support PR finished recently. I aired on the side of not addressing the backwards compatibility problem at the time in order to prioritize getting 3.0 support out the door.

We might be able to get around the backwards compatibility requirements if 3.1.1 support goes out in v8.0.0 only, but there's no guarantee of when 8.0 will fully land as GA. It's likely to be quite a ways off and I'd like to avoid (any more) long delays to version support. Spark 3.1.1 support should probably land in a 7.x release, which means it needs to be backwards compatible with 3.0.0. Them's the brakes unfortunately, but I don't think it's impossible to do.

@pan3793
Copy link

pan3793 commented May 28, 2021

Have any plan to support DataSourceV2?

@jbaiera
Copy link
Member

jbaiera commented Jun 1, 2021

@pan3793 I think that would be awesome to support, but I've had a hard time finding documentation for the API's and haven't had much time to dig into the source for them yet. Definitely something we'd like to tackle going forward.

@pan3793
Copy link

pan3793 commented Jun 2, 2021

@masseyke
Copy link
Member

masseyke commented Feb 2, 2022

Closing this because support for spark 3.1 was added in #1807. We've got #1801 to track support for DataSourceV2.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

5 participants