Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[SPARK-22369][PYTHON][DOCS] Exposes catalog API documentation in PySpark #19596

Closed
wants to merge 1 commit into from

Conversation

HyukjinKwon
Copy link
Member

What changes were proposed in this pull request?

This PR proposes to add a link from spark.catalog(..) to Catalog and expose Catalog APIs in PySpark as below:

2017-10-29 12 25 46

2017-10-29 12 26 33

Note that this is not shown in the list on the top - https://spark.apache.org/docs/latest/api/python/pyspark.sql.html#module-pyspark.sql

2017-10-29 12 30 58

This is basically similar with DataFrameReader and DataFrameWriter.

How was this patch tested?

Manually built the doc.

@SparkQA
Copy link

SparkQA commented Oct 28, 2017

Test build #83170 has finished for PR 19596 at commit 82fe78b.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@HyukjinKwon
Copy link
Member Author

cc @rxin, do you think it's okay to expose PySpark's Catalog class API doc?

@rxin
Copy link
Contributor

rxin commented Oct 29, 2017

Yea definitely.

@HyukjinKwon
Copy link
Member Author

cc @holdenk and @viirya, mind taking a look please? I remember I had few talks about Sphinx and __all__ and I believe you guys are right reviewers.

@viirya
Copy link
Member

viirya commented Oct 30, 2017

Don't we like to add it to the list of Important classes of Spark SQL and DataFrames?

@viirya
Copy link
Member

viirya commented Oct 30, 2017

I've generated the Python docs. Looks good.

@HyukjinKwon
Copy link
Member Author

HyukjinKwon commented Oct 30, 2017

Thanks, @viirya. I wasn't sure if I should add it to the list. My intention was .. this one is like DataFrameReader and DataFrameWriter (supposed to be used via spark.read) and I wanted to.. like hide the package path pyspark.sql.Catalog (instead of pyspark.sql.catalog.Catalog) in the doc, and I just decided the smallest change I could think for this issue.

I am fine with adding it too. It's easy to add it if anyone feels strongly about this. I guess it is also true that it might be one of the "Important classes" anyway. Please let me know.

@SparkQA
Copy link

SparkQA commented Nov 2, 2017

Test build #83334 has finished for PR 19596 at commit 82fe78b.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@rxin
Copy link
Contributor

rxin commented Nov 2, 2017

Merging in master.

@asfgit asfgit closed this in 41b6012 Nov 2, 2017
@HyukjinKwon
Copy link
Member Author

Thank you @rxin and @viirya.

@HyukjinKwon HyukjinKwon deleted the SPARK-22369 branch January 2, 2018 03:41
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
4 participants