-
Notifications
You must be signed in to change notification settings - Fork 28.6k
[SPARK-3430] [PySpark] [Doc] generate PySpark API docs using Sphinx #2292
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
When I tried this, I got a lot of warnings saying that Sphinx couldn't import the PySpark modules:
I guess this is because When I do
it picks up the imports and runs. |
thanks, I had update the PR description. |
Maybe we can just stick that command inside the Makefile... |
Yes, I tried to convert most of the markup (but not all of them), what does this mean in epydocs? |
The markup like |
FYI: Those ones are broken in epydoc, too, so in the case of |
I guess we have some choices for the markup language dialect that we use for our docstrings. I tend to prefer the Google style to ReStructuredText / Javadoc, since it's a little less cluttered to edit / read. We can revisit this later, though; I don't think it's important to rewrite all of the docstrings as part of this patch. |
@JoshRosen @nchammas I had addressed all the comments, please take a look again. |
Oh, sorry I didn't mean the badly formatted doc strings should be fixed in this PR. At least not the ones that are also bad in epydoc. That probably should be left for another PR. There are many strings that need fixing:
Etc... |
Hmm, the Makefile PYTHONPATH didn't seem to work. Maybe we should add those directories to sys.path in # If extensions (or modules to document with autodoc) are in another directory,
# add these directories to sys.path here. If the directory is relative to the
# documentation root, use os.path.abspath to make it absolute, like shown here.
sys.path.insert(0, os.path.abspath('.')) |
@JoshRosen The Makefile has been fixed. |
Jenkins, test this please. |
Jenkins, retest this please. |
Jenkins, test this please. |
QA tests have started for PR 2292 at commit
|
QA tests have finished for PR 2292 at commit
|
Weird, I've never seen Hive fail to initialize like this before. |
Jenkins, test this please. |
QA tests have started for PR 2292 at commit
|
QA tests have finished for PR 2292 at commit
|
QA tests have started for PR 2292 at commit
|
QA tests have finished for PR 2292 at commit
|
Conflicts: python/pyspark/serializers.py
QA tests have started for PR 2292 at commit
|
QA tests have finished for PR 2292 at commit
|
@JoshRosen I had moved the docs to python/docs/, rebased with master. I think it's ready to merge, please take another look, thanks. |
QA tests have started for PR 2292 at commit
|
Tests timed out after a configured wait of |
QA tests have started for PR 2292 at commit
|
Tests timed out after a configured wait of |
QA tests have started for PR 2292 at commit
|
QA tests have finished for PR 2292 at commit
|
QA tests have started for PR 2292 at commit
|
QA tests have finished for PR 2292 at commit
|
This looks good to me; I'm going to merge this into master but leave the JIRA open so that we remember to eventually remove the epydocs / etc. |
Using Sphinx to generate API docs for PySpark.
requirement: Sphinx
The generated API docs will be located at python/docs/_build/html/index.html
It can co-exists with those generated by Epydoc.
This is the first working version, after merging in, then we can continue to improve it and replace the epydoc finally.