Indeed very nice! Thanks On Nov 4, 2015 5:04 PM, "Till Rohrmann" <trohrm...@apache.org> wrote:
> Really cool tutorial Trevor :-) > > On Wed, Nov 4, 2015 at 3:26 PM, Robert Metzger <rmetz...@apache.org> > wrote: > >> For those interested, Trevor wrote a blog post describing how to setup >> Spark, Flink and Zeppelin, both locally and on clusters: >> http://trevorgrant.org/2015/11/03/apache-casserole-a-delicious-big-data-recipe-for-the-whole-family/ >> Thanks Trevor for the great tutorial! >> >> On Thu, Oct 22, 2015 at 4:23 PM, Till Rohrmann <trohrm...@apache.org> >> wrote: >> >>> Hi Trevor, >>> >>> that’s actually my bad since I only tested my branch against a remote >>> cluster. I fixed the problem (not properly starting the >>> LocalFlinkMiniCluster) so that you can now use Zeppelin also in local >>> mode. Just check out my branch again. >>> >>> Cheers, >>> Till >>> >>> >>> On Wed, Oct 21, 2015 at 10:00 PM, Trevor Grant <trevor.d.gr...@gmail.com >>> > wrote: >>> >>>> Hey Till, >>>> >>>> I cloned your branch of Zeplin and while it will compile, it fails >>>> tests on timeout, which consequently was the same issue I was having when >>>> trying to use Zeppelin. >>>> >>>> Ideas? >>>> >>>> >>>> ------------------------------------------------------------------------------- >>>> Test set: org.apache.zeppelin.flink.FlinkInterpreterTest >>>> >>>> ------------------------------------------------------------------------------- >>>> Tests run: 2, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 100.347 >>>> sec <<< FAILURE! - in org.apache.zeppelin.flink.FlinkInterpreterTest >>>> org.apache.zeppelin.flink.FlinkInterpreterTest Time elapsed: 100.347 >>>> sec <<< ERROR! >>>> java.util.concurrent.TimeoutException: Futures timed out after [100000 >>>> milliseconds] >>>> at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219) >>>> at >>>> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223) >>>> at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107) >>>> at >>>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53) >>>> at scala.concurrent.Await$.result(package.scala:107) >>>> at >>>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.getLeaderIndex(FlinkMiniCluster.scala:171) >>>> at >>>> org.apache.flink.runtime.minicluster.LocalFlinkMiniCluster.getLeaderRPCPort(LocalFlinkMiniCluster.scala:132) >>>> at >>>> org.apache.zeppelin.flink.FlinkInterpreter.getPort(FlinkInterpreter.java:136) >>>> at >>>> org.apache.zeppelin.flink.FlinkInterpreter.open(FlinkInterpreter.java:98) >>>> at >>>> org.apache.zeppelin.flink.FlinkInterpreterTest.setUp(FlinkInterpreterTest.java:42) >>>> >>>> org.apache.zeppelin.flink.FlinkInterpreterTest Time elapsed: 100.347 >>>> sec <<< ERROR! >>>> java.lang.NullPointerException: null >>>> at >>>> org.apache.zeppelin.flink.FlinkInterpreter.close(FlinkInterpreter.java:221) >>>> at >>>> org.apache.zeppelin.flink.FlinkInterpreterTest.tearDown(FlinkInterpreterTest.java:48) >>>> >>>> >>>> >>>> Trevor Grant >>>> Data Scientist >>>> https://github.com/rawkintrevo >>>> http://stackexchange.com/users/3002022/rawkintrevo >>>> >>>> *"Fortunate is he, who is able to know the causes of things." -Virgil* >>>> >>>> >>>> On Wed, Oct 21, 2015 at 11:57 AM, Till Rohrmann <trohrm...@apache.org> >>>> wrote: >>>> >>>>> Hi Trevor, >>>>> >>>>> in order to use Zeppelin with a different Flink version in local mode, >>>>> meaning that Zeppelin starts a LocalFlinkMiniCluster when executing >>>>> your jobs, you have to build Zeppelin and change the flink.version >>>>> property in the zeppelin/flink/pom.xml file to the version you want >>>>> to use. >>>>> >>>>> If you want to let Zeppelin submit jobs to a remote cluster, you >>>>> should build Zeppelin with the version of your cluster. That’s because >>>>> internally Zeppelin will use this version to construct a JobGraph >>>>> which is then submitted to the cluster. In order to configure the remote >>>>> cluster, you have to go the *Interpreter* page and scroll down to the >>>>> *flink* section. There you have to specify the address of your >>>>> cluster under *host* and the port under *port*. This should then be >>>>> used to submit jobs to the Flink cluster. >>>>> >>>>> I hope this answers your question. >>>>> >>>>> Btw: If you want to use Zeppelin with the latest Flink 0.10-SNAPSHOT >>>>> version, you should checkout my branch >>>>> https://github.com/tillrohrmann/incubator-zeppelin/tree/flink-0.10-SNAPSHOT >>>>> where I’ve made the necessary changes. >>>>> >>>>> Cheers, >>>>> Till >>>>> >>>>> >>>>> On Wed, Oct 21, 2015 at 5:00 PM, Trevor Grant < >>>>> trevor.d.gr...@gmail.com> wrote: >>>>> >>>>>> I'm setting up some Flink/Spark/Zeppelin at work. Spark+Zeppelin >>>>>> seems to be relatively well supported and configurable but the Flink is >>>>>> not >>>>>> so much. >>>>>> >>>>>> I want Zeppelin to run against my 0.10 build instead of the 0.6 build >>>>>> that ships with Zeppelin. My best guess at the moment on how to >>>>>> accomplish >>>>>> this is to create a symbolic link from the /opt/zepplin/flink folder to >>>>>> /opt/flink-0.10, but this feels dirty and wrong. >>>>>> >>>>>> Does anyone out there have any experience connecting Zeppelin to a >>>>>> non-prepackaged Flink build? >>>>>> >>>>>> I feel like there is a great opporutnity for a HOWTO write up if non >>>>>> currently exists. >>>>>> >>>>>> I'm asking on the Zeppelin user mailing list too as soon as I am >>>>>> added. >>>>>> >>>>>> Thanks for any help >>>>>> >>>>>> tg >>>>>> >>>>>> >>>>>> Trevor Grant >>>>>> Data Scientist >>>>>> https://github.com/rawkintrevo >>>>>> http://stackexchange.com/users/3002022/rawkintrevo >>>>>> >>>>>> *"Fortunate is he, who is able to know the causes of things." >>>>>> -Virgil* >>>>>> >>>>>> >>>>> >>>> >>> >> >