Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[SPARK-13403] [SQL] Pass hadoopConfiguration to HiveConf constructors. #11273

Conversation

rdblue
Copy link
Contributor

@rdblue rdblue commented Feb 19, 2016

This commit updates the HiveContext so that sc.hadoopConfiguration is used to instantiate its internal instances of HiveConf.

I tested this by overriding the S3 FileSystem implementation from spark-defaults.conf as "spark.hadoop.fs.s3.impl" (to avoid HADOOP-12810).

@rxin
Copy link
Contributor

rxin commented Feb 20, 2016

cc @marmbrus

@marmbrus
Copy link
Contributor

Seems reasonable.

@marmbrus
Copy link
Contributor

ok to test

@SparkQA
Copy link

SparkQA commented Feb 22, 2016

Test build #51669 has finished for PR 11273 at commit 32a3dcf.

  • This patch fails Scala style tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@srowen
Copy link
Member

srowen commented Feb 23, 2016

FWIW also seems reasonable to me; making a new conf is rarely correct. @rdblue can you rebase?

@vanzin
Copy link
Contributor

vanzin commented Mar 14, 2016

Also, could you fix the title to follow the Spark convention?
https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark#ContributingtoSpark-PullRequest

@rdblue rdblue changed the title SPARK-13403: Pass hadoopConfiguration to HiveConf constructors. [SPARK-13403] [SQL] Pass hadoopConfiguration to HiveConf constructors. Mar 16, 2016
sc.hadoopConfiguration may contain Hadoop-specific configuration
properties that are not used by SparkSQL's HiveContext because it is not
passed when constructing instances of HiveConf.
@rdblue rdblue force-pushed the SPARK-13403-new-hive-conf-from-hadoop-conf branch from 32a3dcf to f0f6cee Compare March 16, 2016 23:42
@rdblue
Copy link
Contributor Author

rdblue commented Mar 17, 2016

Sorry for the delay, @vanzin and @srowen! I didn't get notified that you had commented. I've rebased this on master and fixed the PR title.

@rdblue rdblue closed this Mar 17, 2016
@rdblue rdblue reopened this Mar 17, 2016
@SparkQA
Copy link

SparkQA commented Mar 17, 2016

Test build #53377 has finished for PR 11273 at commit f0f6cee.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@rxin
Copy link
Contributor

rxin commented Mar 17, 2016

Thanks - I'm going to merge this.

@asfgit asfgit closed this in 5faba9f Mar 17, 2016
@rdblue
Copy link
Contributor Author

rdblue commented Mar 17, 2016

Thanks for the reviews!

roygao94 pushed a commit to roygao94/spark that referenced this pull request Mar 22, 2016
This commit updates the HiveContext so that sc.hadoopConfiguration is used to instantiate its internal instances of HiveConf.

I tested this by overriding the S3 FileSystem implementation from spark-defaults.conf as "spark.hadoop.fs.s3.impl" (to avoid [HADOOP-12810](https://issues.apache.org/jira/browse/HADOOP-12810)).

Author: Ryan Blue <blue@apache.org>

Closes apache#11273 from rdblue/SPARK-13403-new-hive-conf-from-hadoop-conf.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
6 participants