-
Notifications
You must be signed in to change notification settings - Fork 297
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add hadoop aws/gcp jar to the spark default image #1908
Conversation
Signed-off-by: Kevin Su <pingsutw@apache.org>
Codecov ReportAll modified and coverable lines are covered by tests ✅
Additional details and impacted files@@ Coverage Diff @@
## master #1908 +/- ##
=======================================
Coverage 62.81% 62.81%
=======================================
Files 307 307
Lines 22984 22984
Branches 3490 3490
=======================================
Hits 14438 14438
Misses 8124 8124
Partials 422 422 ☔ View full report in Codecov by Sentry. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for fixing! Appreciate it!
This is a band-aid. How will this work on gcp? It should be at the user level imo |
Signed-off-by: Kevin Su <pingsutw@apache.org>
Signed-off-by: Kevin Su <pingsutw@apache.org>
* Add hadoop-aws jar to the spark default image Signed-off-by: Kevin Su <pingsutw@apache.org> * no-cache-dir Signed-off-by: Kevin Su <pingsutw@apache.org> --------- Signed-off-by: Kevin Su <pingsutw@apache.org>
* Add hadoop-aws jar to the spark default image Signed-off-by: Kevin Su <pingsutw@apache.org> * no-cache-dir Signed-off-by: Kevin Su <pingsutw@apache.org> --------- Signed-off-by: Kevin Su <pingsutw@apache.org> Signed-off-by: Rafael Raposo <rafaelraposo@spotify.com>
TL;DR
Add hadoop s3 and gcs dependencies to the default spark image. Spark needs these jar to read the data from s3 / gcs.
Type
Are all requirements met?
Complete description
Spark example:
Tracking Issue
NA
Follow-up issue
NA