This is an automated email from the ASF dual-hosted git repository. janardhan pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/systemml-website.git
commit 5459acc87e622c2fa52079c4e4aac7768a2b5bbf Author: Niketan Pansare <[email protected]> AuthorDate: Thu Mar 21 10:24:02 2019 -0700 [DOC] Updated the setup instructions for macos and windows Path for Spark and Hadoop to navigate some common problems around .exe execution privileges. Closes #65. --- _src/install-systemml.html | 60 +++++++++++++++++++++++++++++++++++++++++++++- 1 file changed, 59 insertions(+), 1 deletion(-) diff --git a/_src/install-systemml.html b/_src/install-systemml.html index 6665363..03f23c6 100644 --- a/_src/install-systemml.html +++ b/_src/install-systemml.html @@ -54,11 +54,69 @@ limitations under the License. <p class="indent">Apache Spark 2.x</p> <p class="indent">Set SPARK_HOME to a location where Spark 2.x is installed.</p> + <div id="prerequisite-tabs"> + <ul> + <li><a href="#prerequisite-tabs-1">MacOS/Linux</a></li> + <li><a href="#prerequisite-tabs-2">Windows</a></li> + </ul> + + <div id="prerequisite-tabs-1"> + 1) Java <br /> + Make sure Java version is >= 1.8 and JAVA_HOME environment variable is set: + {% highlight bash %} +java -version +export JAVA_HOME="$(/usr/libexec/java_home)"{% endhighlight %} + + 2) Spark <br /> + Download Spark from <a href="https://spark.apache.org/downloads.html">https://spark.apache.org/downloads.html</a> and move to home directory, and extract. Also, set environment variables to point to the extracted directory + {% highlight bash %} +export SPARK_HOME="$HOME/spark-2.1.0-bin-hadoop2.7" +export HADOOP_HOME=$SPARK_HOME +export SPARK_LOCAL_IP=127.0.0.1{% endhighlight %} + + 3) Python and Jupyter <br /> + Download and install Anaconda Python 3+ from <a href="https://www.anaconda.com/distribution/#download-section">https://www.anaconda.com/distribution/#download-section</a> (includes jupyter, and pip) + {% highlight bash %} +export PYSPARK_DRIVER_PYTHON=jupyter +export PYSPARK_DRIVER_PYTHON_OPTS='notebook' $SPARK_HOME/bin/pyspark --master local[*] --driver-memory 8G{% endhighlight %} + </div> + + <div id="prerequisite-tabs-2"> + 1) Java <br /> + Make sure Java version is >= 1.8. Also, set JAVA_HOME environment variable and include %JAVA_HOME%\bin in the environment variable PATH: + {% highlight bash %} +java -version +ls "%JAVA_HOME%"{% endhighlight %} + + 2) Spark <br /> + Download Spark from <a href="https://spark.apache.org/downloads.html">https://spark.apache.org/downloads.html</a> and extract. Set the environment variable SPARK_HOME to point to the extracted directory. <br /> + + 3) Install winutils <br /> +- Download winutils.exe from <a href="http://github.com/steveloughran/winutils/raw/master/hadoop-2.6.0/bin/winutils.exe">http://github.com/steveloughran/winutils/raw/master/hadoop-2.6.0/bin/winutils.exe</a> <br /> +- Place it in c:\winutils\bin <br /> +- Set environment variable HADOOP_HOME to point to c:\winutils <br /> +- Add c:\winutils\bin to the environment variable PATH. <br /> +- Finally, modify permission of hive directory that will be used by Spark and check if Spark is correctly installed: + + {% highlight bash %} +winutils.exe chmod 777 /tmp/hive +%SPARK_HOME%\bin\spark-shell +%SPARK_HOME%\bin\pyspark --master local[*] --driver-memory 8G{% endhighlight %} + + 3) Python and Jupyter <br /> + Download and install Anaconda Python 3+ from <a href="https://www.anaconda.com/distribution/#download-section">https://www.anaconda.com/distribution/#download-section</a> (includes jupyter, and pip) + {% highlight bash %} +set PYSPARK_DRIVER_PYTHON=jupyter +set PYSPARK_DRIVER_PYTHON_OPTS=notebook +%SPARK_HOME%\bin\pyspark --master local[*] --driver-memory 8G{% endhighlight %} + </div> + + </div> </div> <!-- Step 2 --> <div class="col col-12"> - <h3><span class="circle">2</span>Setup</h3> + <h3><span class="circle">2</span>Setup SystemML</h3> </div> <div id="setup-tabs">
