Great tutorial! Thanks a lot ^^

On 4 November 2015 at 17:12, Leonard Wolters <leon...@sagent.io> wrote:

> Indeed very nice! Thanks
> On Nov 4, 2015 5:04 PM, "Till Rohrmann" <trohrm...@apache.org> wrote:
>
>> Really cool tutorial Trevor :-)
>>
>> On Wed, Nov 4, 2015 at 3:26 PM, Robert Metzger <rmetz...@apache.org>
>> wrote:
>>
>>> For those interested, Trevor wrote a blog post describing how to setup
>>> Spark, Flink and Zeppelin, both locally and on clusters:
>>> http://trevorgrant.org/2015/11/03/apache-casserole-a-delicious-big-data-recipe-for-the-whole-family/
>>> Thanks Trevor for the great tutorial!
>>>
>>> On Thu, Oct 22, 2015 at 4:23 PM, Till Rohrmann <trohrm...@apache.org>
>>> wrote:
>>>
>>>> Hi Trevor,
>>>>
>>>> that’s actually my bad since I only tested my branch against a remote
>>>> cluster. I fixed the problem (not properly starting the
>>>> LocalFlinkMiniCluster) so that you can now use Zeppelin also in local
>>>> mode. Just check out my branch again.
>>>>
>>>> Cheers,
>>>> Till
>>>> ​
>>>>
>>>> On Wed, Oct 21, 2015 at 10:00 PM, Trevor Grant <
>>>> trevor.d.gr...@gmail.com> wrote:
>>>>
>>>>> Hey Till,
>>>>>
>>>>> I cloned your branch of Zeplin and while it will compile, it fails
>>>>> tests on timeout, which consequently was the same issue I was having when
>>>>> trying to use Zeppelin.
>>>>>
>>>>> Ideas?
>>>>>
>>>>>
>>>>> -------------------------------------------------------------------------------
>>>>> Test set: org.apache.zeppelin.flink.FlinkInterpreterTest
>>>>>
>>>>> -------------------------------------------------------------------------------
>>>>> Tests run: 2, Failures: 0, Errors: 2, Skipped: 0, Time elapsed:
>>>>> 100.347 sec <<< FAILURE! - in 
>>>>> org.apache.zeppelin.flink.FlinkInterpreterTest
>>>>> org.apache.zeppelin.flink.FlinkInterpreterTest  Time elapsed: 100.347
>>>>> sec  <<< ERROR!
>>>>> java.util.concurrent.TimeoutException: Futures timed out after [100000
>>>>> milliseconds]
>>>>> at
>>>>> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>>>>> at
>>>>> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>>>>> at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>>>>> at
>>>>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>>>>> at scala.concurrent.Await$.result(package.scala:107)
>>>>> at
>>>>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.getLeaderIndex(FlinkMiniCluster.scala:171)
>>>>> at
>>>>> org.apache.flink.runtime.minicluster.LocalFlinkMiniCluster.getLeaderRPCPort(LocalFlinkMiniCluster.scala:132)
>>>>> at
>>>>> org.apache.zeppelin.flink.FlinkInterpreter.getPort(FlinkInterpreter.java:136)
>>>>> at
>>>>> org.apache.zeppelin.flink.FlinkInterpreter.open(FlinkInterpreter.java:98)
>>>>> at
>>>>> org.apache.zeppelin.flink.FlinkInterpreterTest.setUp(FlinkInterpreterTest.java:42)
>>>>>
>>>>> org.apache.zeppelin.flink.FlinkInterpreterTest  Time elapsed: 100.347
>>>>> sec  <<< ERROR!
>>>>> java.lang.NullPointerException: null
>>>>> at
>>>>> org.apache.zeppelin.flink.FlinkInterpreter.close(FlinkInterpreter.java:221)
>>>>> at
>>>>> org.apache.zeppelin.flink.FlinkInterpreterTest.tearDown(FlinkInterpreterTest.java:48)
>>>>>
>>>>>
>>>>>
>>>>> Trevor Grant
>>>>> Data Scientist
>>>>> https://github.com/rawkintrevo
>>>>> http://stackexchange.com/users/3002022/rawkintrevo
>>>>>
>>>>> *"Fortunate is he, who is able to know the causes of things."  -Virgil*
>>>>>
>>>>>
>>>>> On Wed, Oct 21, 2015 at 11:57 AM, Till Rohrmann <trohrm...@apache.org>
>>>>> wrote:
>>>>>
>>>>>> Hi Trevor,
>>>>>>
>>>>>> in order to use Zeppelin with a different Flink version in local
>>>>>> mode, meaning that Zeppelin starts a LocalFlinkMiniCluster when
>>>>>> executing your jobs, you have to build Zeppelin and change the
>>>>>> flink.version property in the zeppelin/flink/pom.xml file to the
>>>>>> version you want to use.
>>>>>>
>>>>>> If you want to let Zeppelin submit jobs to a remote cluster, you
>>>>>> should build Zeppelin with the version of your cluster. That’s because
>>>>>> internally Zeppelin will use this version to construct a JobGraph
>>>>>> which is then submitted to the cluster. In order to configure the remote
>>>>>> cluster, you have to go the *Interpreter* page and scroll down to
>>>>>> the *flink* section. There you have to specify the address of your
>>>>>> cluster under *host* and the port under *port*. This should then be
>>>>>> used to submit jobs to the Flink cluster.
>>>>>>
>>>>>> I hope this answers your question.
>>>>>>
>>>>>> Btw: If you want to use Zeppelin with the latest Flink 0.10-SNAPSHOT
>>>>>> version, you should checkout my branch
>>>>>> https://github.com/tillrohrmann/incubator-zeppelin/tree/flink-0.10-SNAPSHOT
>>>>>> where I’ve made the necessary changes.
>>>>>>
>>>>>> Cheers,
>>>>>> Till
>>>>>> ​
>>>>>>
>>>>>> On Wed, Oct 21, 2015 at 5:00 PM, Trevor Grant <
>>>>>> trevor.d.gr...@gmail.com> wrote:
>>>>>>
>>>>>>> I'm setting up some Flink/Spark/Zeppelin at work.  Spark+Zeppelin
>>>>>>> seems to be relatively well supported and configurable but the Flink is 
>>>>>>> not
>>>>>>> so much.
>>>>>>>
>>>>>>> I want Zeppelin to run against my 0.10 build instead of the 0.6
>>>>>>> build that ships with Zeppelin.  My best guess at the moment on how to
>>>>>>> accomplish this is to create a symbolic link from the /opt/zepplin/flink
>>>>>>> folder to /opt/flink-0.10, but this feels dirty and wrong.
>>>>>>>
>>>>>>> Does anyone out there have any experience connecting Zeppelin to a
>>>>>>> non-prepackaged Flink build?
>>>>>>>
>>>>>>> I feel like there is a great opporutnity for a HOWTO write up if non
>>>>>>> currently exists.
>>>>>>>
>>>>>>> I'm asking on the Zeppelin user mailing list too as soon as I am
>>>>>>> added.
>>>>>>>
>>>>>>> Thanks for any help
>>>>>>>
>>>>>>> tg
>>>>>>>
>>>>>>>
>>>>>>> Trevor Grant
>>>>>>> Data Scientist
>>>>>>> https://github.com/rawkintrevo
>>>>>>> http://stackexchange.com/users/3002022/rawkintrevo
>>>>>>>
>>>>>>> *"Fortunate is he, who is able to know the causes of things."
>>>>>>>  -Virgil*
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>

Reply via email to