[ 
https://issues.apache.org/jira/browse/SPARK-19551?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15945685#comment-15945685
 ] 

Arthur Tacca commented on SPARK-19551:
--------------------------------------

Thanks, I needed the reminder! In fact the person that generated their own 
build of the docs got back to me; I hope they don't mind me pasting what they 
said here:


I've compiled the documentation using Sphinx (ver 1.3.5.). I have a foggy 
memory on this as it's been a while, but I recall I had to rollback to older 
version of sphinx to have copybutton.js to work properly - this is what allows 
us to toggle the ``>>>`` mark in the python code 

- (example) 
http://takwatanabe.me/pyspark/generated/generated/mllib.classification.DenseVector.html#mllib.classification.DenseVector
- (js file) 
https://github.com/scipy/scipy-sphinx-theme/blob/master/_theme/scipy/static/js/copybutton.js

Otherwise, I simply just used the ``autosummary`` directive offered by Sphinx 
(http://www.sphinx-doc.org/en/stable/ext/autosummary.html). You can see how I 
used these in the *.rst files in 
https://github.com/wtak23/pyspark/tree/master/source.
For instance, in order to create the **entire** html subtrees from the link 
http://takwatanabe.me/pyspark/pyspark.ml.html , all I had to have was this rst 
file: 
https://raw.githubusercontent.com/wtak23/pyspark/master/source/pyspark.ml.rst

Once you have PySpark directory included in your $PYTHONPATH envvar, you should 
be able to simply run ``make html`` using the Makefile from the github branch 
below.

https://github.com/wtak23/pyspark/tree/master 



> Theme for PySpark documenation could do with improving
> ------------------------------------------------------
>
>                 Key: SPARK-19551
>                 URL: https://issues.apache.org/jira/browse/SPARK-19551
>             Project: Spark
>          Issue Type: Documentation
>          Components: Documentation, PySpark
>    Affects Versions: 2.1.0
>            Reporter: Arthur Tacca
>            Priority: Minor
>
> I have found the Python Spark documentation hard to navigate for two reasons:
> * Each page in the documentation is huge, because the whole of the 
> documentation is split up into only a few chunks.
> * The methods for each class is not listed in a short form, so the only way 
> to look through them is to browse past the full documentation for all methods 
> (including parameter lists, examples, etc.).
> This has irritated someone enough that they have done [their own build of the 
> pyspark documentation|http://takwatanabe.me/pyspark/index.html]. In 
> comparison to the official docs they are a delight to use. But of course it 
> is not clear whether they'll be kept up to date, which is why I'm asking here 
> that the official docs are improved. Perhaps that site could be used as 
> inspiration? I don't know much about these things, but it appears that the 
> main change they have made is to switch to the "read the docs" theme.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to