Hi All,

I gave https://github.com/elbamos/Zeppelin-With-R another try and was able
to get that working as well, and that does indeed work with the official
precompiled spark with hadoop 2.6.

I compiled Zeppelin-With-R with:

mvn package install -DskipTests

Had to reset my R home in the spark interpreter to:

/usr/lib64/R/

All then works fine.

Cheers!







On Thu, Dec 17, 2015 at 1:10 AM, Corneau Damien <cornead...@gmail.com>
wrote:

> First of all, I would start by saying that if you have trouble with
> https://github.com/datalayer/zeppelin-R
> You better ask directly on that repository or ask echarles for some help
> since it isn't part of the https://github.com/apache/incubator-zeppelin
> repository.
> I think he will be able to give you better answers regarding your error.
>
> @Amos
> Zeppelin is an Open Source project, and we welcome any type of
> contributions, including helping on the mailing list since we can't answer
> every thread.
> Your remark about @FelixCheung has nothing to do here, and it doesn't help
> resolving @csuser issue.
> Furthermore, you shouldn't have to tell him what to do or not to do, your
> grudges whatever it is has nothing to do in this mailing list.
>
> On Thu, Dec 17, 2015 at 2:30 AM, Amos B. Elberg <amos.elb...@me.com>
> wrote:
>
>> CS:   What you’re doing is compiling two versions of Zeppelin from source
>> on top of a binary of a third version.  That’s going to give you trouble.
>>
>> The R Interpreter you’re using doesn’t interface with Zeppelin’s spark
>> installation at all.  All it shares is the name.  So, none of the things
>> you’ve been doing, with recompiling Zeppelin or Spark or whatever, is
>> actually having any impact on R working with hive.  R working or not
>> working, for you, with hive, is incidental.
>>
>> I suggest you start from a clean installation and install this
>> https://github.com/elbamos/Zeppelin-With-R from source.
>>
>> You should not need to specify -Pyarn, -Phive, etc. etc.   The R
>> interpreter in the package will use the same Spark as the rest of Zeppelin.
>>
>> Just mvn package install -DskipTests to install.
>>
>> At runtime, set the environment variable SPARK_HOME to point to your
>> existing, separately compiled, installation of Spark.  Zeppelin should try
>> to use Hive by default, and the R interpreter will use whatever the rest of
>> Zeppelin uses.
>>
>> Also — @FelixCheung, you have no business trying to provide support for
>> anyone on this project, and you certainly have no business giving anyone
>> advice about using R with it.
>>
>>
>> From: cs user <acldstk...@gmail.com> <acldstk...@gmail.com>
>> Reply: users@zeppelin.incubator.apache.org
>> <users@zeppelin.incubator.apache.org>
>> <users@zeppelin.incubator.apache.org>
>> Date: December 16, 2015 at 5:27:20 AM
>> To: users@zeppelin.incubator.apache.org
>> <users@zeppelin.incubator.apache.org>
>> <users@zeppelin.incubator.apache.org>
>> Subject:  Re: Zeppelin+spark+R+hive
>>
>> Hi All,
>>
>> Many thanks for getting back to me. I've managed to get this working by
>> downloading the tagged spark 1.5.2 release and compiling it with:
>>
>> ./make-distribution.sh --name custom-spark --tgz -Phadoop-2.6
>> -Dhadoop.version=2.6.0 -Pyarn -Phive -Phive-thriftserver -Psparkr
>>
>> I've then downloaded the source for this version of zeppelin:
>>
>> https://github.com/datalayer/zeppelin-R
>>
>> Then compiled it with (based on the readme from the above project):
>>
>> mvn clean install -Pyarn -Pspark-1.5 -Dspark.version=1.5.2
>> -Dhadoop.version=2.6.0 -Phadoop-2.6 -Ppyspark -Dmaven.findbugs.enable=false
>> -Drat.skip=true -Dcheckstyle.skip=true -DskipTests -pl
>> '!flink,!ignite,!phoenix,!postgresql,!tajo,!hive,!cassandra,!lens,!kylin'
>>
>> Within Zeppelin this allows spark to run with yarn, as well as the
>> ability to use the R interpreter with hive.
>>
>> Hope this helps someone else :-)
>>
>> Cheers!
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> On Tue, Dec 15, 2015 at 5:37 PM, Sourav Mazumder <
>> sourav.mazumde...@gmail.com> wrote:
>>
>>> I believe that is not going to solve the problem.
>>>
>>> If you need to run spark on Yarn (assuming that it is your requirement)
>>> ensure that you run it in Yarn Client mode. Yarn Clustre mode is not
>>> supported with Zeppelin yet.
>>>
>>> Regards,
>>> Sourav
>>>
>>>
>>> On Tue, Dec 15, 2015 at 9:32 AM, Felix Cheung <felixcheun...@hotmail.com
>>> > wrote:
>>>
>>>> If you are not using YARN, try building your Spark distribution without
>>>> this:
>>>>  -Pyarn
>>>> ?
>>>>
>>>>
>>>>
>>>> On Tue, Dec 15, 2015 at 12:31 AM -0800, "cs user" <acldstk...@gmail.com
>>>> > wrote:
>>>>
>>>> Hi Folks,
>>>>
>>>> We've been playing around with this project:
>>>>
>>>> https://github.com/datalayer/zeppelin-R
>>>>
>>>> However when we try and write a notebook using R which requires hive,
>>>> we run into the following:
>>>>
>>>> Error in value[[3L]](cond): Spark SQL is not built with Hive support
>>>>
>>>> This is when we are using the pre compiled spark with hadoop 2.6
>>>> support.
>>>>
>>>> To work around this, I've tried recompiling spark with hive support.
>>>> Accessing the hive context within an R notebook now works fine.
>>>>
>>>> However, it is then impossible to run existing notebooks which try to
>>>> submit jobs via yarn, the following error is encountered:
>>>>
>>>> java.lang.NoSuchMethodException:
>>>> org.apache.spark.repl.SparkILoop$SparkILoopInterpreter.classServerUri() at
>>>> java.lang.Class.getMethod(Class.java:1678) at
>>>> org.apache.zeppelin.spark.SparkInterpreter.createSparkContext(SparkInterpreter.java:271)
>>>> at
>>>> org.apache.zeppelin.spark.SparkInterpreter.getSparkContext(SparkInterpreter.java:145)
>>>> at
>>>> org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:464)
>>>> at
>>>> org.apache.zeppelin.interpreter.ClassloaderInterpreter.open(ClassloaderInterpreter.java:74)
>>>> at
>>>> org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:68)
>>>> at
>>>> org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:92)
>>>> at
>>>> org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:292)
>>>> at org.apache.zeppelin.scheduler.Job.run(Job.java:170) at
>>>> org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:118)
>>>> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) at
>>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178)
>>>> at
>>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292)
>>>> at
>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>> at
>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>> at java.lang.Thread.run(Thread.java:745)
>>>>
>>>> If I switch back to the old spark home, these jobs then work fine
>>>> again.
>>>>
>>>> I am compiling our custom version of spark with the following:
>>>>
>>>> ./make-distribution.sh --name custom-spark --tgz -Phadoop-2.6
>>>> -Dhadoop.version=2.6.0 -Pyarn -Phive -Phive-thriftserver
>>>>
>>>> Are there any other switches I need to add to overcome the above error?
>>>>
>>>> Thanks!
>>>>
>>>>
>>>>
>>>
>>
>

Reply via email to