Author: moon
Date: Sat Dec 17 15:48:31 2016
New Revision: 1774779
URL: http://svn.apache.org/viewvc?rev=1774779&view=rev
Log:
ZEPPELIN-1798
Modified:
zeppelin/site/docs/0.7.0-SNAPSHOT/atom.xml
zeppelin/site/docs/0.7.0-SNAPSHOT/interpreter/flink.html
zeppelin/site/docs/0.7.0-SNAPSHOT/quickstart/install_with_flink_and_spark_cluster.html
zeppelin/site/docs/0.7.0-SNAPSHOT/rss.xml
zeppelin/site/docs/0.7.0-SNAPSHOT/search_data.json
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/atom.xml
URL:
http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/atom.xml?rev=1774779&r1=1774778&r2=1774779&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/atom.xml (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/atom.xml Sat Dec 17 15:48:31 2016
@@ -4,7 +4,7 @@
<title>Apache Zeppelin</title>
<link href="http://zeppelin.apache.org/" rel="self"/>
<link href="http://zeppelin.apache.org"/>
- <updated>2016-12-17T21:52:14+09:00</updated>
+ <updated>2016-12-17T07:48:11-08:00</updated>
<id>http://zeppelin.apache.org</id>
<author>
<name>The Apache Software Foundation</name>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/interpreter/flink.html
URL:
http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/interpreter/flink.html?rev=1774779&r1=1774778&r2=1774779&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/interpreter/flink.html (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/interpreter/flink.html Sat Dec 17
15:48:31 2016
@@ -250,7 +250,7 @@ wget http://www.gutenberg.org/ebooks/10.
</code></pre></div>
<div class="highlight"><pre><code class="scala"><span class="o">%</span><span
class="n">flink</span>
<span class="k">case</span> <span class="k">class</span> <span
class="nc">WordCount</span><span class="o">(</span><span
class="n">word</span><span class="k">:</span> <span
class="kt">String</span><span class="o">,</span> <span
class="n">frequency</span><span class="k">:</span> <span
class="kt">Int</span><span class="o">)</span>
-<span class="k">val</span> <span class="n">bible</span><span
class="k">:</span><span class="kt">DataSet</span><span class="o">[</span><span
class="kt">String</span><span class="o">]</span> <span class="k">=</span> <span
class="n">env</span><span class="o">.</span><span
class="n">readTextFile</span><span class="o">(</span><span
class="s">"10.txt.utf-8"</span><span class="o">)</span>
+<span class="k">val</span> <span class="n">bible</span><span
class="k">:</span><span class="kt">DataSet</span><span class="o">[</span><span
class="kt">String</span><span class="o">]</span> <span class="k">=</span> <span
class="n">benv</span><span class="o">.</span><span
class="n">readTextFile</span><span class="o">(</span><span
class="s">"10.txt.utf-8"</span><span class="o">)</span>
<span class="k">val</span> <span class="n">partialCounts</span><span
class="k">:</span> <span class="kt">DataSet</span><span class="o">[</span><span
class="kt">WordCount</span><span class="o">]</span> <span class="k">=</span>
<span class="n">bible</span><span class="o">.</span><span
class="n">flatMap</span><span class="o">{</span>
<span class="n">line</span> <span class="k">=></span>
<span
class="s">"""\b\w+\b"""</span><span
class="o">.</span><span class="n">r</span><span class="o">.</span><span
class="n">findAllIn</span><span class="o">(</span><span
class="n">line</span><span class="o">).</span><span class="n">map</span><span
class="o">(</span><span class="n">word</span> <span class="k">=></span>
<span class="nc">WordCount</span><span class="o">(</span><span
class="n">word</span><span class="o">,</span> <span class="mi">1</span><span
class="o">))</span>
Modified:
zeppelin/site/docs/0.7.0-SNAPSHOT/quickstart/install_with_flink_and_spark_cluster.html
URL:
http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/quickstart/install_with_flink_and_spark_cluster.html?rev=1774779&r1=1774778&r2=1774779&view=diff
==============================================================================
---
zeppelin/site/docs/0.7.0-SNAPSHOT/quickstart/install_with_flink_and_spark_cluster.html
(original)
+++
zeppelin/site/docs/0.7.0-SNAPSHOT/quickstart/install_with_flink_and_spark_cluster.html
Sat Dec 17 15:48:31 2016
@@ -277,13 +277,15 @@ Clone Zeppelin.</p>
<div class="highlight"><pre><code class="text language-text"
data-lang="text">cd zeppelin
</code></pre></div>
<p>Package Zeppelin.</p>
-<div class="highlight"><pre><code class="text language-text"
data-lang="text">mvn clean package -DskipTests -Pspark-1.6 -Dflink.version=1.1.2
+<div class="highlight"><pre><code class="text language-text"
data-lang="text">mvn clean package -DskipTests -Pspark-1.6
-Dflink.version=1.1.3 -Pscala-2.10
</code></pre></div>
<p><code>-DskipTests</code> skips build tests- you're not developing
(yet), so you don't need to do tests, the clone version <em>should</em>
build.</p>
<p><code>-Pspark-1.6</code> tells maven to build a Zeppelin with Spark 1.6.
This is important because Zeppelin has its own Spark interpreter and the
versions must be the same.</p>
-<p><code>-Dflink.version=1.1.2</code> tells maven specifically to build
Zeppelin with Flink version 1.1.2.</p>
+<p><code>-Dflink.version=1.1.3</code> tells maven specifically to build
Zeppelin with Flink version 1.1.3.</p>
+
+<p>-<code>-Pscala-2.10</code> tells maven to build with Scala v2.10.</p>
<p><strong>Note:</strong> You may wish to include additional build flags such
as <code>-Ppyspark</code> or <code>-Psparkr</code>. See <a
href="https://github.com/apache/zeppelin#build">the build section of github for
more details</a>.</p>
@@ -312,7 +314,7 @@ As long as you didn't edit any code,
<p>Create a new notebook named "Flink Test" and copy and paste the
following code.</p>
<div class="highlight"><pre><code class="scala language-scala"
data-lang="scala"><span class="o">%</span><span class="n">flink</span> <span
class="c1">// let Zeppelin know what interpreter to use.</span>
-<span class="k">val</span> <span class="n">text</span> <span
class="k">=</span> <span class="n">env</span><span class="o">.</span><span
class="n">fromElements</span><span class="o">(</span><span class="s">"In
the time of chimpanzees, I was a monkey"</span><span class="o">,</span>
<span class="c1">// some lines of text to analyze</span>
+<span class="k">val</span> <span class="n">text</span> <span
class="k">=</span> <span class="n">benv</span><span class="o">.</span><span
class="n">fromElements</span><span class="o">(</span><span class="s">"In
the time of chimpanzees, I was a monkey"</span><span class="o">,</span>
<span class="c1">// some lines of text to analyze</span>
<span class="s">"Butane in my veins and I'm out to cut the
junkie"</span><span class="o">,</span>
<span class="s">"With the plastic eyeballs, spray paint the
vegetables"</span><span class="o">,</span>
<span class="s">"Dog food stalls with the beefcake
pantyhose"</span><span class="o">,</span>
@@ -393,13 +395,13 @@ As long as you didn't edit any code,
<p>Building from source is recommended where possible, for simplicity in this
tutorial we will download Flink and Spark Binaries.</p>
<p>To download the Flink Binary use <code>wget</code></p>
-<div class="highlight"><pre><code class="bash language-bash"
data-lang="bash">wget <span
class="s2">"http://mirror.cogentco.com/pub/apache/flink/flink-1.0.3/flink-1.0.3-bin-hadoop24-scala_2.10.tgz"</span>
-tar -xzvf flink-1.0.3-bin-hadoop24-scala_2.10.tgz
+<div class="highlight"><pre><code class="bash language-bash"
data-lang="bash">wget <span
class="s2">"http://mirror.cogentco.com/pub/apache/flink/flink-1.1.3/flink-1.1.3-bin-hadoop24-scala_2.10.tgz"</span>
+tar -xzvf flink-1.1.3-bin-hadoop24-scala_2.10.tgz
</code></pre></div>
-<p>This will download Flink 1.0.3, compatible with Hadoop 2.4. You do not
have to install Hadoop for this binary to work, but if you are using Hadoop,
please change <code>24</code> to your appropriate version.</p>
+<p>This will download Flink 1.1.3, compatible with Hadoop 2.4. You do not
have to install Hadoop for this binary to work, but if you are using Hadoop,
please change <code>24</code> to your appropriate version.</p>
<p>Start the Flink Cluster.</p>
-<div class="highlight"><pre><code class="bash language-bash"
data-lang="bash">flink-1.0.3/bin/start-cluster.sh
+<div class="highlight"><pre><code class="bash language-bash"
data-lang="bash">flink-1.1.3/bin/start-cluster.sh
</code></pre></div>
<h6>Building From source</h6>
@@ -407,11 +409,11 @@ tar -xzvf flink-1.0.3-bin-hadoop24-scala
<p>See the <a
href="https://github.com/apache/flink/blob/master/README.md">Flink Installation
guide</a> for more detailed instructions.</p>
-<p>Return to the directory where you have been downloading, this tutorial
assumes that is <code>$HOME</code>. Clone Flink, check out release-1.0, and
build.</p>
+<p>Return to the directory where you have been downloading, this tutorial
assumes that is <code>$HOME</code>. Clone Flink, check out release-1.1.3-rc2,
and build.</p>
<div class="highlight"><pre><code class="text language-text"
data-lang="text">cd $HOME
git clone https://github.com/apache/flink.git
cd flink
-git checkout release-1.0
+git checkout release-1.1.3-rc2
mvn clean install -DskipTests
</code></pre></div>
<p>Start the Flink Cluster in stand-alone mode</p>
@@ -427,8 +429,8 @@ mvn clean install -DskipTests
<p>(if binaries)
<code>
-flink-1.0.3/bin/stop-cluster.sh
-flink-1.0.3/bin/start-cluster.sh
+flink-1.1.3/bin/stop-cluster.sh
+flink-1.1.3/bin/start-cluster.sh
</code></p>
<p>(if built from source)
@@ -446,11 +448,11 @@ build-target/bin/start-cluster.sh
<p>Using binaries is also</p>
<p>To download the Spark Binary use <code>wget</code></p>
-<div class="highlight"><pre><code class="bash language-bash"
data-lang="bash">wget <span
class="s2">"http://mirrors.koehn.com/apache/spark/spark-1.6.1/spark-1.6.1-bin-hadoop2.4.tgz"</span>
-tar -xzvf spark-1.6.1-bin-hadoop2.4.tgz
-mv spark-1.6.1-bin-hadoop4.4 spark
+<div class="highlight"><pre><code class="bash language-bash"
data-lang="bash">wget <span
class="s2">"http://d3kbcqa49mib13.cloudfront.net/spark-1.6.3-bin-hadoop2.6.tgz"</span>
+tar -xzvf spark-1.6.3-bin-hadoop2.6.tgz
+mv spark-1.6.3-bin-hadoop2.6 spark
</code></pre></div>
-<p>This will download Spark 1.6.1, compatible with Hadoop 2.4. You do not
have to install Hadoop for this binary to work, but if you are using Hadoop,
please change <code>2.4</code> to your appropriate version.</p>
+<p>This will download Spark 1.6.3, compatible with Hadoop 2.6. You do not
have to install Hadoop for this binary to work, but if you are using Hadoop,
please change <code>2.6</code> to your appropriate version.</p>
<h6>Building From source</h6>
@@ -460,7 +462,7 @@ mv spark-1.6.1-bin-hadoop4.4 spark
<p>Return to the directory where you have been downloading, this tutorial
assumes that is $HOME. Clone Spark, check out branch-1.6, and build.
<strong>Note:</strong> Recall, we're only checking out 1.6 because it is
the most recent Spark for which a Zeppelin profile exists at
- the time of writing. You are free to check out other version, just make sure
you build Zeppelin against the correct version of Spark.</p>
+ the time of writing. You are free to check out other version, just make sure
you build Zeppelin against the correct version of Spark. However if you use
Spark 2.0, the word count example will need to be changed as Spark 2.0 is not
compatible with the following examples.</p>
<div class="highlight"><pre><code class="text language-text"
data-lang="text">cd $HOME
</code></pre></div>
<p>Clone, check out, and build Spark version 1.6.x.</p>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/rss.xml
URL:
http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/rss.xml?rev=1774779&r1=1774778&r2=1774779&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/rss.xml (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/rss.xml Sat Dec 17 15:48:31 2016
@@ -5,8 +5,8 @@
<description>Apache Zeppelin - The Apache Software
Foundation</description>
<link>http://zeppelin.apache.org</link>
<link>http://zeppelin.apache.org</link>
- <lastBuildDate>2016-12-17T21:52:14+09:00</lastBuildDate>
- <pubDate>2016-12-17T21:52:14+09:00</pubDate>
+ <lastBuildDate>2016-12-17T07:48:11-08:00</lastBuildDate>
+ <pubDate>2016-12-17T07:48:11-08:00</pubDate>
<ttl>1800</ttl>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/search_data.json
URL:
http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/search_data.json?rev=1774779&r1=1774778&r2=1774779&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/search_data.json (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/search_data.json Sat Dec 17 15:48:31 2016
@@ -226,7 +226,7 @@
"/interpreter/flink.html": {
"title": "Flink Interpreter for Apache Zeppelin",
- "content" : "<!--Licensed under the Apache License, Version 2.0 (the
"License");you may not use this file except in compliance with the
License.You may obtain a copy of the License
athttp://www.apache.org/licenses/LICENSE-2.0Unless required by applicable law
or agreed to in writing, softwaredistributed under the License is distributed
on an "AS IS" BASIS,WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND,
either express or implied.See the License for the specific language governing
permissions andlimitations under the License.-->Flink interpreter for Apache
ZeppelinOverviewApache Flink is an open source platform for distributed stream
and batch data processing. Flinkâs core is a streaming dataflow engine that
provides data distribution, communication, and fault tolerance for distributed
computations over data streams. Flink also builds batch processing on top of
the streaming engine, overlaying native iteration support, managed memory, and
program opt
imization.How to start local Flink cluster, to test the interpreterZeppelin
comes with pre-configured flink-local interpreter, which starts Flink in a
local mode on your machine, so you do not need to install anything.How to
configure interpreter to point to Flink clusterAt the
&quot;Interpreters&quot; menu, you have to create a new Flink
interpreter and provide next properties: property value Description
host local host name of running JobManager. 'local' runs
flink in local mode (default) port 6123 port of running JobManager
For more information about Flink configuration, you can find it here.How to
test it&#39;s workingIn example, by using the Zeppelin notebook is from
Till Rohrmann&#39;s presentation Interactive data analysis with Apache
Flink for Apache Flink Meetup.%shrm 10.txt.utf-8wget
http://www.gutenberg.org/ebooks/10.txt.utf-8%flinkcase class WordCount(word:
String, frequency: Int)val bible:DataSet[String] = en
v.readTextFile(&quot;10.txt.utf-8&quot;)val partialCounts:
DataSet[WordCount] = bible.flatMap{ line =&gt;
&quot;&quot;&quot;bw+b&quot;&quot;&quot;.r.findAllIn(line).map(word
=&gt; WordCount(word, 1))// line.split(&quot;
&quot;).map(word =&gt; WordCount(word, 1))}val wordCounts =
partialCounts.groupBy(&quot;word&quot;).reduce{ (left, right)
=&gt; WordCount(left.word, left.frequency + right.frequency)}val result10 =
wordCounts.first(10).collect()",
+ "content" : "<!--Licensed under the Apache License, Version 2.0 (the
"License");you may not use this file except in compliance with the
License.You may obtain a copy of the License
athttp://www.apache.org/licenses/LICENSE-2.0Unless required by applicable law
or agreed to in writing, softwaredistributed under the License is distributed
on an "AS IS" BASIS,WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND,
either express or implied.See the License for the specific language governing
permissions andlimitations under the License.-->Flink interpreter for Apache
ZeppelinOverviewApache Flink is an open source platform for distributed stream
and batch data processing. Flinkâs core is a streaming dataflow engine that
provides data distribution, communication, and fault tolerance for distributed
computations over data streams. Flink also builds batch processing on top of
the streaming engine, overlaying native iteration support, managed memory, and
program opt
imization.How to start local Flink cluster, to test the interpreterZeppelin
comes with pre-configured flink-local interpreter, which starts Flink in a
local mode on your machine, so you do not need to install anything.How to
configure interpreter to point to Flink clusterAt the
&quot;Interpreters&quot; menu, you have to create a new Flink
interpreter and provide next properties: property value Description
host local host name of running JobManager. 'local' runs
flink in local mode (default) port 6123 port of running JobManager
For more information about Flink configuration, you can find it here.How to
test it&#39;s workingIn example, by using the Zeppelin notebook is from
Till Rohrmann&#39;s presentation Interactive data analysis with Apache
Flink for Apache Flink Meetup.%shrm 10.txt.utf-8wget
http://www.gutenberg.org/ebooks/10.txt.utf-8%flinkcase class WordCount(word:
String, frequency: Int)val bible:DataSet[String] = be
nv.readTextFile(&quot;10.txt.utf-8&quot;)val partialCounts:
DataSet[WordCount] = bible.flatMap{ line =&gt;
&quot;&quot;&quot;bw+b&quot;&quot;&quot;.r.findAllIn(line).map(word
=&gt; WordCount(word, 1))// line.split(&quot;
&quot;).map(word =&gt; WordCount(word, 1))}val wordCounts =
partialCounts.groupBy(&quot;word&quot;).reduce{ (left, right)
=&gt; WordCount(left.word, left.frequency + right.frequency)}val result10 =
wordCounts.first(10).collect()",
"url": " /interpreter/flink.html",
"group": "interpreter",
"excerpt": "Apache Flink is an open source platform for distributed
stream and batch data processing."
@@ -557,7 +557,7 @@
"/quickstart/install_with_flink_and_spark_cluster.html": {
"title": "Install Zeppelin with Flink and Spark in cluster mode",
- "content" : "<!--Licensed under the Apache License, Version 2.0 (the
"License");you may not use this file except in compliance with the
License.You may obtain a copy of the License
athttp://www.apache.org/licenses/LICENSE-2.0Unless required by applicable law
or agreed to in writing, softwaredistributed under the License is distributed
on an "AS IS" BASIS,WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND,
either express or implied.See the License for the specific language governing
permissions andlimitations under the License.-->Install with flink and spark
clusterThis tutorial is extremely entry-level. It assumes no prior knowledge of
Linux, git, or other tools. If you carefully type what I tell you when I tell
you, you should be able to get Zeppelin running.Installing Zeppelin with Flink
and Spark in cluster modeThis tutorial assumes the user has a machine (real or
virtual with a fresh, minimal installation of Ubuntu 14.04.3 Server.Note: On
the size requ
irements of the Virtual Machine, some users reported trouble when using the
default virtual machine sizes, specifically that the hard drive needed to be at
least 16GB- other users did not have this issue.There are many good tutorials
on how to install Ubuntu Server on a virtual box, here is one of themRequired
ProgramsAssuming the minimal install, there are several programs that we will
need to install before Zeppelin, Flink, and Spark.gitopenssh-serverOpenJDK
7Maven 3.1+For git, openssh-server, and OpenJDK 7 we will be using the apt
package manager.gitFrom the command prompt:sudo apt-get install
gitopenssh-serversudo apt-get install openssh-serverOpenJDK 7sudo apt-get
install openjdk-7-jdk openjdk-7-jre-libA note for those using Ubuntu 16.04: To
install openjdk-7 on Ubuntu 16.04, one must add a repository. Sourcesudo
add-apt-repository ppa:openjdk-r/ppasudo apt-get updatesudo apt-get install
openjdk-7-jdk openjdk-7-jre-libMaven 3.1+Zeppelin requires maven version 3.x.
The version
available in the repositories at the time of writing is 2.x, so maven must be
installed manually.Purge any existing versions of maven.sudo apt-get purge
maven maven2Download the maven 3.3.9 binary.wget
&quot;http://www.us.apache.org/dist/maven/maven-3/3.3.9/binaries/apache-maven-3.3.9-bin.tar.gz&quot;Unarchive
the binary and move to the /usr/local directory.tar -zxvf
apache-maven-3.3.9-bin.tar.gzsudo mv ./apache-maven-3.3.9 /usr/localCreate
symbolic links in /usr/bin.sudo ln -s /usr/local/apache-maven-3.3.9/bin/mvn
/usr/bin/mvnInstalling ZeppelinThis provides a quick overview of Zeppelin
installation from source, however the reader is encouraged to review the
Zeppelin Installation GuideFrom the command prompt:Clone Zeppelin.git clone
https://github.com/apache/zeppelin.gitEnter the Zeppelin root directory.cd
zeppelinPackage Zeppelin.mvn clean package -DskipTests -Pspark-1.6
-Dflink.version=1.1.2-DskipTests skips build tests- you&#39;re not
developing (yet), so you don&am
p;#39;t need to do tests, the clone version should build.-Pspark-1.6 tells
maven to build a Zeppelin with Spark 1.6. This is important because Zeppelin
has its own Spark interpreter and the versions must be the
same.-Dflink.version=1.1.2 tells maven specifically to build Zeppelin with
Flink version 1.1.2.Note: You may wish to include additional build flags such
as -Ppyspark or -Psparkr. See the build section of github for more
details.Note: You can build against any version of Spark that has a Zeppelin
build profile available. The key is to make sure you check out the matching
version of Spark to build. At the time of this writing, Spark 1.6 was the most
recent Spark version available.Note: On build failures. Having installed
Zeppelin close to 30 times now, I will tell you that sometimes the build fails
for seemingly no reason.As long as you didn&#39;t edit any code, it is
unlikely the build is failing because of something you did. What does tend to
happen, is some dependency
that maven is trying to download is unreachable. If your build fails on this
step here are some tips:- Don&#39;t get discouraged.- Scroll up and read
through the logs. There will be clues there.- Retry (that is, run the mvn clean
package -DskipTests -Pspark-1.6 again)- If there were clues that a dependency
couldn&#39;t be downloaded wait a few hours or even days and retry again.
Open source software when compiling is trying to download all of the
dependencies it needs, if a server is off-line there is nothing you can do but
wait for it to come back.- Make sure you followed all of the steps carefully.-
Ask the community to help you. Go here and join the user mailing list. People
are there to help you. Make sure to copy and paste the build output (everything
that happened in the console) and include that in your message.Start the
Zeppelin daemon.bin/zeppelin-daemon.sh startUse ifconfig to determine the host
machine&#39;s IP address. If you are not familiar with how to do
this, a fairly comprehensive post can be found here.Open a web-browser on a
machine connected to the same network as the host (or in the host operating
system if using a virtual machine). Navigate to http://yourip:8080, where
yourip is the IP address you found in ifconfig.See the Zeppelin tutorial for
basic Zeppelin usage. It is also advised that you take a moment to check out
the tutorial notebook that is included with each Zeppelin install, and to
familiarize yourself with basic notebook functionality.Flink TestCreate a new
notebook named &quot;Flink Test&quot; and copy and paste the following
code.%flink // let Zeppelin know what interpreter to use.val text =
env.fromElements(&quot;In the time of chimpanzees, I was a
monkey&quot;, // some lines of text to analyze&quot;Butane in my
veins and I&#39;m out to cut the junkie&quot;,&quot;With the
plastic eyeballs, spray paint the vegetables&quot;,&quot;Dog food
stalls with the beefcake pantyh
ose&quot;,&quot;Kill the headlights and put it in
neutral&quot;,&quot;Stock car flamin&#39; with a loser in the
cruise control&quot;,&quot;Baby&#39;s in Reno with the Vitamin
D&quot;,&quot;Got a couple of couches, sleep on the love
seat&quot;,&quot;Someone came in sayin&#39; I&#39;m insane to
complain&quot;,&quot;About a shotgun wedding and a stain on my
shirt&quot;,&quot;Don&#39;t believe everything that you
breathe&quot;,&quot;You get a parking violation and a maggot on your
sleeve&quot;,&quot;So shave your face with some mace in the
dark&quot;,&quot;Savin&#39; all your food stamps and
burnin&#39; down the trailer park&quot;,&quot;Yo, cut
it&quot;)/* The meat and potatoes: this tells Flink to iterate
through the elements, in this case strings, transform the string to
lower case and split the string at white space into individual words
then f
inally aggregate the occurrence of each word. This creates the count
variable which is a list of tuples of the form (word,
occurances)counts.collect().foreach(println(_)) // execute the script and
print each element in the counts list*/val counts = text.flatMap{
_.toLowerCase.split(&quot;W+&quot;) }.map { (_,1)
}.groupBy(0).sum(1)counts.collect().foreach(println(_)) // execute the script
and print each element in the counts listRun the code to make sure the built-in
Zeppelin Flink interpreter is working properly.Spark TestCreate a new notebook
named &quot;Spark Test&quot; and copy and paste the following
code.%spark // let Zeppelin know what interpreter to use.val text =
sc.parallelize(List(&quot;In the time of chimpanzees, I was a
monkey&quot;, // some lines of text to analyze&quot;Butane in my veins
and I&#39;m out to cut the junkie&quot;,&quot;With the plastic
eyeballs, spray paint the vegetables&quot;,&quot;Dog food stall
s with the beefcake pantyhose&quot;,&quot;Kill the headlights and put
it in neutral&quot;,&quot;Stock car flamin&#39; with a loser in the
cruise control&quot;,&quot;Baby&#39;s in Reno with the Vitamin
D&quot;,&quot;Got a couple of couches, sleep on the love
seat&quot;,&quot;Someone came in sayin&#39; I&#39;m insane to
complain&quot;,&quot;About a shotgun wedding and a stain on my
shirt&quot;,&quot;Don&#39;t believe everything that you
breathe&quot;,&quot;You get a parking violation and a maggot on your
sleeve&quot;,&quot;So shave your face with some mace in the
dark&quot;,&quot;Savin&#39; all your food stamps and
burnin&#39; down the trailer park&quot;,&quot;Yo, cut
it&quot;))/* The meat and potatoes: this tells spark to iterate
through the elements, in this case strings, transform the string to
lower case and split the string at white space into ind
ividual words then finally aggregate the occurrence of each word.
This creates the count variable which is a list of tuples of the form (word,
occurances)*/val counts = text.flatMap {
_.toLowerCase.split(&quot;W+&quot;) } .map { (_,1) }
.reduceByKey(_ + _)counts.collect().foreach(println(_)) //
execute the script and print each element in the counts listRun the code to
make sure the built-in Zeppelin Flink interpreter is working properly.Finally,
stop the Zeppelin daemon. From the command prompt run:bin/zeppelin-daemon.sh
stopInstalling ClustersFlink ClusterDownload BinariesBuilding from source is
recommended where possible, for simplicity in this tutorial we will download
Flink and Spark Binaries.To download the Flink Binary use wgetwget
&quot;http://mirror.cogentco.com/pub/apache/flink/flink-1.0.3/flink-1.0.3-bin-hadoop24-scala_2.10.tgz&quot;tar
-xzvf flink-1.0.3-bin-hadoop24-scala_2.10.tgzThis will download Flink 1.
0.3, compatible with Hadoop 2.4. You do not have to install Hadoop for this
binary to work, but if you are using Hadoop, please change 24 to your
appropriate version.Start the Flink
Cluster.flink-1.0.3/bin/start-cluster.shBuilding From sourceIf you wish to
build Flink from source, the following will be instructive. Note that if you
have downloaded and used the binary version this should be skipped. The
changing nature of build tools and versions across platforms makes this section
somewhat precarious. For example, Java8 and Maven 3.0.3 are recommended for
building Flink, which are not recommended for Zeppelin at the time of writing.
If the user wishes to attempt to build from source, this section will provide
some reference. If errors are encountered, please contact the Apache Flink
community.See the Flink Installation guide for more detailed
instructions.Return to the directory where you have been downloading, this
tutorial assumes that is $HOME. Clone Flink, check out relea
se-1.0, and build.cd $HOMEgit clone https://github.com/apache/flink.gitcd
flinkgit checkout release-1.0mvn clean install -DskipTestsStart the Flink
Cluster in stand-alone modebuild-target/bin/start-cluster.shEnsure the cluster
is upIn a browser, navigate to http://yourip:8082 to see the Flink Web-UI.
Click on &#39;Task Managers&#39; in the left navigation bar. Ensure
there is at least one Task Manager present.If no task managers are present,
restart the Flink cluster with the following commands:(if
binaries)flink-1.0.3/bin/stop-cluster.shflink-1.0.3/bin/start-cluster.sh(if
built from
source)build-target/bin/stop-cluster.shbuild-target/bin/start-cluster.shSpark
1.6 ClusterDownload BinariesBuilding from source is recommended where
possible, for simplicity in this tutorial we will download Flink and Spark
Binaries.Using binaries is alsoTo download the Spark Binary use wgetwget
&quot;http://mirrors.koehn.com/apache/spark/spark-1.6.1/spark-1.6.1-bin-hadoop2.4.tgz&quot;t
ar -xzvf spark-1.6.1-bin-hadoop2.4.tgzmv spark-1.6.1-bin-hadoop4.4 sparkThis
will download Spark 1.6.1, compatible with Hadoop 2.4. You do not have to
install Hadoop for this binary to work, but if you are using Hadoop, please
change 2.4 to your appropriate version.Building From sourceSpark is an
extraordinarily large project, which takes considerable time to download and
build. It is also prone to build failures for similar reasons listed in the
Flink section. If the user wishes to attempt to build from source, this
section will provide some reference. If errors are encountered, please contact
the Apache Spark community.See the Spark Installation guide for more detailed
instructions.Return to the directory where you have been downloading, this
tutorial assumes that is $HOME. Clone Spark, check out branch-1.6, and
build.Note: Recall, we&#39;re only checking out 1.6 because it is the most
recent Spark for which a Zeppelin profile exists at the time of writing. You
are free to
check out other version, just make sure you build Zeppelin against the
correct version of Spark.cd $HOMEClone, check out, and build Spark version
1.6.x.git clone https://github.com/apache/spark.gitcd sparkgit checkout
branch-1.6mvn clean package -DskipTestsStart the Spark clusterReturn to the
$HOME directory.cd $HOMEStart the Spark cluster in stand alone mode, specifying
the webui-port as some port other than 8080 (the webui-port of
Zeppelin).spark/sbin/start-master.sh --webui-port 8082Note: Why --webui-port
8082? There is a digression toward the end of this document that explains
this.Open a browser and navigate to http://yourip:8082 to ensure the Spark
master is running.Toward the top of the page there will be a URL:
spark://yourhost:7077. Note this URL, the Spark Master URI, it will be needed
in subsequent steps.Start the slave using the URI from the Spark master
WebUI:spark/sbin/start-slave.sh spark://yourhostname:7077Return to the root
directory and start the Zeppelin daemon.
cd $HOMEzeppelin/bin/zeppelin-daemon.sh startConfigure InterpretersOpen a web
browser and go to the Zeppelin web-ui at http://yourip:8080.Now go back to the
Zeppelin web-ui at http://yourip:8080 and this time click on anonymous at the
top right, which will open a drop-down menu, select Interpreters to enter
interpreter configuration.In the Spark section, click the edit button in the
top right corner to make the property values editable (looks like a pencil).The
only field that needs to be edited in the Spark interpreter is the master
field. Change this value from local[*] to the URL you used to start the slave,
mine was spark://ubuntu:7077.Click Save to update the parameters, and click OK
when it asks you about restarting the interpreter.Now scroll down to the Flink
section. Click the edit button and change the value of host from local to
localhost. Click Save again.Reopen the examples and execute them again (I.e.
you need to click the play button at the top of the screen, or the bu
tton on the paragraph .You should be able check the Flink and Spark webuis (at
something like http://yourip:8081, http://yourip:8082, http://yourip:8083) and
see that jobs have been run against the clusters.Digression Sorry to be vague
and use terms such as &#39;something like&#39;, but exactly what web-ui
is at what port is going to depend on what order you started things. What is
really going on here is you are pointing your browser at specific ports, namely
8081, 8082, and 8083. Flink and Spark all want to put their web-ui on port
8080, but are well behaved and will take the next port available. Since
Zeppelin started first, it will get port 8080. When Flink starts (assuming you
started Flink first), it will try to bind to port 8080, see that it is already
taken, and go to the next one available, hopefully 8081. Spark has a webui for
the master and the slave, so when they start they will try to bind to 8080
already taken by Zeppelin), then 8081 (already taken by Flin
k&#39;s webui), then 8082. If everything goes smoothy and you followed the
directions precisely, the webuis should be 8081 and 8082. It is possible to
specify the port you want the webui to bind to (at the command line by passing
the --webui-port &lt;port&gt; flag when you start the Flink and Spark,
where &lt;port&gt; is the port you want to see that webui on. You
can also set the default webui port of Spark and Flink (and Zeppelin) in the
configuration files, but this is a tutorial for novices and slightly out of
scope.Next StepsCheck out the tutorial for more cool things you can do with
your new toy!Join the community, ask questions and contribute! Every little bit
helps.",
+ "content" : "<!--Licensed under the Apache License, Version 2.0 (the
"License");you may not use this file except in compliance with the
License.You may obtain a copy of the License
athttp://www.apache.org/licenses/LICENSE-2.0Unless required by applicable law
or agreed to in writing, softwaredistributed under the License is distributed
on an "AS IS" BASIS,WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND,
either express or implied.See the License for the specific language governing
permissions andlimitations under the License.-->Install with flink and spark
clusterThis tutorial is extremely entry-level. It assumes no prior knowledge of
Linux, git, or other tools. If you carefully type what I tell you when I tell
you, you should be able to get Zeppelin running.Installing Zeppelin with Flink
and Spark in cluster modeThis tutorial assumes the user has a machine (real or
virtual with a fresh, minimal installation of Ubuntu 14.04.3 Server.Note: On
the size requ
irements of the Virtual Machine, some users reported trouble when using the
default virtual machine sizes, specifically that the hard drive needed to be at
least 16GB- other users did not have this issue.There are many good tutorials
on how to install Ubuntu Server on a virtual box, here is one of themRequired
ProgramsAssuming the minimal install, there are several programs that we will
need to install before Zeppelin, Flink, and Spark.gitopenssh-serverOpenJDK
7Maven 3.1+For git, openssh-server, and OpenJDK 7 we will be using the apt
package manager.gitFrom the command prompt:sudo apt-get install
gitopenssh-serversudo apt-get install openssh-serverOpenJDK 7sudo apt-get
install openjdk-7-jdk openjdk-7-jre-libA note for those using Ubuntu 16.04: To
install openjdk-7 on Ubuntu 16.04, one must add a repository. Sourcesudo
add-apt-repository ppa:openjdk-r/ppasudo apt-get updatesudo apt-get install
openjdk-7-jdk openjdk-7-jre-libMaven 3.1+Zeppelin requires maven version 3.x.
The version
available in the repositories at the time of writing is 2.x, so maven must be
installed manually.Purge any existing versions of maven.sudo apt-get purge
maven maven2Download the maven 3.3.9 binary.wget
&quot;http://www.us.apache.org/dist/maven/maven-3/3.3.9/binaries/apache-maven-3.3.9-bin.tar.gz&quot;Unarchive
the binary and move to the /usr/local directory.tar -zxvf
apache-maven-3.3.9-bin.tar.gzsudo mv ./apache-maven-3.3.9 /usr/localCreate
symbolic links in /usr/bin.sudo ln -s /usr/local/apache-maven-3.3.9/bin/mvn
/usr/bin/mvnInstalling ZeppelinThis provides a quick overview of Zeppelin
installation from source, however the reader is encouraged to review the
Zeppelin Installation GuideFrom the command prompt:Clone Zeppelin.git clone
https://github.com/apache/zeppelin.gitEnter the Zeppelin root directory.cd
zeppelinPackage Zeppelin.mvn clean package -DskipTests -Pspark-1.6
-Dflink.version=1.1.3 -Pscala-2.10-DskipTests skips build tests- you&#39;re
not developing (yet),
so you don&#39;t need to do tests, the clone version should
build.-Pspark-1.6 tells maven to build a Zeppelin with Spark 1.6. This is
important because Zeppelin has its own Spark interpreter and the versions must
be the same.-Dflink.version=1.1.3 tells maven specifically to build Zeppelin
with Flink version 1.1.3.--Pscala-2.10 tells maven to build with Scala
v2.10.Note: You may wish to include additional build flags such as -Ppyspark or
-Psparkr. See the build section of github for more details.Note: You can build
against any version of Spark that has a Zeppelin build profile available. The
key is to make sure you check out the matching version of Spark to build. At
the time of this writing, Spark 1.6 was the most recent Spark version
available.Note: On build failures. Having installed Zeppelin close to 30 times
now, I will tell you that sometimes the build fails for seemingly no reason.As
long as you didn&#39;t edit any code, it is unlikely the build is failing
because of
something you did. What does tend to happen, is some dependency that maven is
trying to download is unreachable. If your build fails on this step here are
some tips:- Don&#39;t get discouraged.- Scroll up and read through the
logs. There will be clues there.- Retry (that is, run the mvn clean package
-DskipTests -Pspark-1.6 again)- If there were clues that a dependency
couldn&#39;t be downloaded wait a few hours or even days and retry again.
Open source software when compiling is trying to download all of the
dependencies it needs, if a server is off-line there is nothing you can do but
wait for it to come back.- Make sure you followed all of the steps carefully.-
Ask the community to help you. Go here and join the user mailing list. People
are there to help you. Make sure to copy and paste the build output (everything
that happened in the console) and include that in your message.Start the
Zeppelin daemon.bin/zeppelin-daemon.sh startUse ifconfig to determine the host
mach
ine&#39;s IP address. If you are not familiar with how to do this, a
fairly comprehensive post can be found here.Open a web-browser on a machine
connected to the same network as the host (or in the host operating system if
using a virtual machine). Navigate to http://yourip:8080, where yourip is the
IP address you found in ifconfig.See the Zeppelin tutorial for basic Zeppelin
usage. It is also advised that you take a moment to check out the tutorial
notebook that is included with each Zeppelin install, and to familiarize
yourself with basic notebook functionality.Flink TestCreate a new notebook
named &quot;Flink Test&quot; and copy and paste the following
code.%flink // let Zeppelin know what interpreter to use.val text =
benv.fromElements(&quot;In the time of chimpanzees, I was a
monkey&quot;, // some lines of text to analyze&quot;Butane in my
veins and I&#39;m out to cut the junkie&quot;,&quot;With the
plastic eyeballs, spray paint the veget
ables&quot;,&quot;Dog food stalls with the beefcake
pantyhose&quot;,&quot;Kill the headlights and put it in
neutral&quot;,&quot;Stock car flamin&#39; with a loser in the
cruise control&quot;,&quot;Baby&#39;s in Reno with the Vitamin
D&quot;,&quot;Got a couple of couches, sleep on the love
seat&quot;,&quot;Someone came in sayin&#39; I&#39;m insane to
complain&quot;,&quot;About a shotgun wedding and a stain on my
shirt&quot;,&quot;Don&#39;t believe everything that you
breathe&quot;,&quot;You get a parking violation and a maggot on your
sleeve&quot;,&quot;So shave your face with some mace in the
dark&quot;,&quot;Savin&#39; all your food stamps and
burnin&#39; down the trailer park&quot;,&quot;Yo, cut
it&quot;)/* The meat and potatoes: this tells Flink to iterate
through the elements, in this case strings, transform the string to
lower case and s
plit the string at white space into individual words then finally
aggregate the occurrence of each word. This creates the count variable
which is a list of tuples of the form (word,
occurances)counts.collect().foreach(println(_)) // execute the script and
print each element in the counts list*/val counts = text.flatMap{
_.toLowerCase.split(&quot;W+&quot;) }.map { (_,1)
}.groupBy(0).sum(1)counts.collect().foreach(println(_)) // execute the script
and print each element in the counts listRun the code to make sure the built-in
Zeppelin Flink interpreter is working properly.Spark TestCreate a new notebook
named &quot;Spark Test&quot; and copy and paste the following
code.%spark // let Zeppelin know what interpreter to use.val text =
sc.parallelize(List(&quot;In the time of chimpanzees, I was a
monkey&quot;, // some lines of text to analyze&quot;Butane in my veins
and I&#39;m out to cut the junkie&quot;,&quot;With the plastic eyeba
lls, spray paint the vegetables&quot;,&quot;Dog food stalls with the
beefcake pantyhose&quot;,&quot;Kill the headlights and put it in
neutral&quot;,&quot;Stock car flamin&#39; with a loser in the
cruise control&quot;,&quot;Baby&#39;s in Reno with the Vitamin
D&quot;,&quot;Got a couple of couches, sleep on the love
seat&quot;,&quot;Someone came in sayin&#39; I&#39;m insane to
complain&quot;,&quot;About a shotgun wedding and a stain on my
shirt&quot;,&quot;Don&#39;t believe everything that you
breathe&quot;,&quot;You get a parking violation and a maggot on your
sleeve&quot;,&quot;So shave your face with some mace in the
dark&quot;,&quot;Savin&#39; all your food stamps and
burnin&#39; down the trailer park&quot;,&quot;Yo, cut
it&quot;))/* The meat and potatoes: this tells spark to iterate
through the elements, in this case strings, transform the
string to lower case and split the string at white space into individual
words then finally aggregate the occurrence of each word. This
creates the count variable which is a list of tuples of the form (word,
occurances)*/val counts = text.flatMap {
_.toLowerCase.split(&quot;W+&quot;) } .map { (_,1) }
.reduceByKey(_ + _)counts.collect().foreach(println(_)) //
execute the script and print each element in the counts listRun the code to
make sure the built-in Zeppelin Flink interpreter is working properly.Finally,
stop the Zeppelin daemon. From the command prompt run:bin/zeppelin-daemon.sh
stopInstalling ClustersFlink ClusterDownload BinariesBuilding from source is
recommended where possible, for simplicity in this tutorial we will download
Flink and Spark Binaries.To download the Flink Binary use wgetwget
&quot;http://mirror.cogentco.com/pub/apache/flink/flink-1.1.3/flink-1.1.3-bin-hadoop24-scala_2.10.tgz&quot;tar
-xzvf
flink-1.1.3-bin-hadoop24-scala_2.10.tgzThis will download Flink 1.1.3,
compatible with Hadoop 2.4. You do not have to install Hadoop for this binary
to work, but if you are using Hadoop, please change 24 to your appropriate
version.Start the Flink Cluster.flink-1.1.3/bin/start-cluster.shBuilding From
sourceIf you wish to build Flink from source, the following will be
instructive. Note that if you have downloaded and used the binary version this
should be skipped. The changing nature of build tools and versions across
platforms makes this section somewhat precarious. For example, Java8 and Maven
3.0.3 are recommended for building Flink, which are not recommended for
Zeppelin at the time of writing. If the user wishes to attempt to build from
source, this section will provide some reference. If errors are encountered,
please contact the Apache Flink community.See the Flink Installation guide for
more detailed instructions.Return to the directory where you have been
downloading,
this tutorial assumes that is $HOME. Clone Flink, check out
release-1.1.3-rc2, and build.cd $HOMEgit clone
https://github.com/apache/flink.gitcd flinkgit checkout release-1.1.3-rc2mvn
clean install -DskipTestsStart the Flink Cluster in stand-alone
modebuild-target/bin/start-cluster.shEnsure the cluster is upIn a browser,
navigate to http://yourip:8082 to see the Flink Web-UI. Click on &#39;Task
Managers&#39; in the left navigation bar. Ensure there is at least one Task
Manager present.If no task managers are present, restart the Flink cluster with
the following commands:(if
binaries)flink-1.1.3/bin/stop-cluster.shflink-1.1.3/bin/start-cluster.sh(if
built from
source)build-target/bin/stop-cluster.shbuild-target/bin/start-cluster.shSpark
1.6 ClusterDownload BinariesBuilding from source is recommended where
possible, for simplicity in this tutorial we will download Flink and Spark
Binaries.Using binaries is alsoTo download the Spark Binary use wgetwget
&quot;http://d3kbc
qa49mib13.cloudfront.net/spark-1.6.3-bin-hadoop2.6.tgz&quot;tar -xzvf
spark-1.6.3-bin-hadoop2.6.tgzmv spark-1.6.3-bin-hadoop2.6 sparkThis will
download Spark 1.6.3, compatible with Hadoop 2.6. You do not have to install
Hadoop for this binary to work, but if you are using Hadoop, please change 2.6
to your appropriate version.Building From sourceSpark is an extraordinarily
large project, which takes considerable time to download and build. It is also
prone to build failures for similar reasons listed in the Flink section. If
the user wishes to attempt to build from source, this section will provide some
reference. If errors are encountered, please contact the Apache Spark
community.See the Spark Installation guide for more detailed
instructions.Return to the directory where you have been downloading, this
tutorial assumes that is $HOME. Clone Spark, check out branch-1.6, and
build.Note: Recall, we&#39;re only checking out 1.6 because it is the most
recent Spark for which a
Zeppelin profile exists at the time of writing. You are free to check out
other version, just make sure you build Zeppelin against the correct version of
Spark. However if you use Spark 2.0, the word count example will need to be
changed as Spark 2.0 is not compatible with the following examples.cd
$HOMEClone, check out, and build Spark version 1.6.x.git clone
https://github.com/apache/spark.gitcd sparkgit checkout branch-1.6mvn clean
package -DskipTestsStart the Spark clusterReturn to the $HOME directory.cd
$HOMEStart the Spark cluster in stand alone mode, specifying the webui-port as
some port other than 8080 (the webui-port of
Zeppelin).spark/sbin/start-master.sh --webui-port 8082Note: Why --webui-port
8082? There is a digression toward the end of this document that explains
this.Open a browser and navigate to http://yourip:8082 to ensure the Spark
master is running.Toward the top of the page there will be a URL:
spark://yourhost:7077. Note this URL, the Spark Master URI, it w
ill be needed in subsequent steps.Start the slave using the URI from the Spark
master WebUI:spark/sbin/start-slave.sh spark://yourhostname:7077Return to the
root directory and start the Zeppelin daemon.cd
$HOMEzeppelin/bin/zeppelin-daemon.sh startConfigure InterpretersOpen a web
browser and go to the Zeppelin web-ui at http://yourip:8080.Now go back to the
Zeppelin web-ui at http://yourip:8080 and this time click on anonymous at the
top right, which will open a drop-down menu, select Interpreters to enter
interpreter configuration.In the Spark section, click the edit button in the
top right corner to make the property values editable (looks like a pencil).The
only field that needs to be edited in the Spark interpreter is the master
field. Change this value from local[*] to the URL you used to start the slave,
mine was spark://ubuntu:7077.Click Save to update the parameters, and click OK
when it asks you about restarting the interpreter.Now scroll down to the Flink
section. Click the
edit button and change the value of host from local to localhost. Click Save
again.Reopen the examples and execute them again (I.e. you need to click the
play button at the top of the screen, or the button on the paragraph .You
should be able check the Flink and Spark webuis (at something like
http://yourip:8081, http://yourip:8082, http://yourip:8083) and see that jobs
have been run against the clusters.Digression Sorry to be vague and use terms
such as &#39;something like&#39;, but exactly what web-ui is at what
port is going to depend on what order you started things. What is really going
on here is you are pointing your browser at specific ports, namely 8081, 8082,
and 8083. Flink and Spark all want to put their web-ui on port 8080, but are
well behaved and will take the next port available. Since Zeppelin started
first, it will get port 8080. When Flink starts (assuming you started Flink
first), it will try to bind to port 8080, see that it is already taken, and go t
o the next one available, hopefully 8081. Spark has a webui for the master
and the slave, so when they start they will try to bind to 8080 already taken
by Zeppelin), then 8081 (already taken by Flink&#39;s webui), then 8082. If
everything goes smoothy and you followed the directions precisely, the webuis
should be 8081 and 8082. It is possible to specify the port you want the
webui to bind to (at the command line by passing the --webui-port
&lt;port&gt; flag when you start the Flink and Spark, where
&lt;port&gt; is the port you want to see that webui on. You can
also set the default webui port of Spark and Flink (and Zeppelin) in the
configuration files, but this is a tutorial for novices and slightly out of
scope.Next StepsCheck out the tutorial for more cool things you can do with
your new toy!Join the community, ask questions and contribute! Every little bit
helps.",
"url": " /quickstart/install_with_flink_and_spark_cluster.html",
"group": "tutorial",
"excerpt": "Tutorial is valid for Spark 1.6.x and Flink 1.1.2"