Re: How do I configure R interpreter in Zeppelin?

2017-04-26 Thread Ruslan Dautkhanov
Thanks for feedback.

%spark.r
print("Hello World!")
 throws exception [2].

Understood - I'll try to remove -Pr and rebuild Zeppelin. Yep, I used a
fresh master snapshot.
( I have't seen anything in maven build logs that could indicate a problem
around R interpreter)
Will update this email thread with result after rebuilding Zeppelin without
-Pr


[2]

spark.r interpreter not found
org.apache.zeppelin.interpreter.InterpreterException: spark.r interpreter
not found at
org.apache.zeppelin.interpreter.InterpreterFactory.getInterpreter(InterpreterFactory.java:417)
at org.apache.zeppelin.notebook.Note.run(Note.java:620) at
org.apache.zeppelin.socket.NotebookServer.persistAndExecuteSingleParagraph(NotebookServer.java:1781)
at
org.apache.zeppelin.socket.NotebookServer.runParagraph(NotebookServer.java:1741)
at
org.apache.zeppelin.socket.NotebookServer.onMessage(NotebookServer.java:288)
at
org.apache.zeppelin.socket.NotebookSocket.onWebSocketText(NotebookSocket.java:59)
at
org.eclipse.jetty.websocket.common.events.JettyListenerEventDriver.onTextMessage(JettyListenerEventDriver.java:128)
at
org.eclipse.jetty.websocket.common.message.SimpleTextMessage.messageComplete(SimpleTextMessage.java:69)
at
org.eclipse.jetty.websocket.common.events.AbstractEventDriver.appendMessage(AbstractEventDriver.java:65)
at
org.eclipse.jetty.websocket.common.events.JettyListenerEventDriver.onTextFrame(JettyListenerEventDriver.java:122)




-- 
Ruslan Dautkhanov

On Wed, Apr 26, 2017 at 2:13 PM, moon soo Lee  wrote:

> Zeppelin includes two R interpreter implementations.
>
> One used to activated by -Psparkr the other -Pr.
> Since https://github.com/apache/zeppelin/pull/2215, -Psparkr is activated
> by default. And if you're trying to use sparkR, -Psparkr (activated by
> default in master branch) is implementation you might be more interested.
>
> So you can just try use with %spark.r prefix.
> Let me know if it works for you.
>
> Thanks,
> moon
>
> On Wed, Apr 26, 2017 at 12:11 AM Ruslan Dautkhanov 
> wrote:
>
>> Hi moon soo Lee,
>>
>> Cloudera's Spark doesn't have $SPARK_HOME/bin/sparkR
>> Would Zeppelin still enable its sparkR interpreter then?
>>
>> Built Zeppelin using
>>
>> $ mvn clean package -DskipTests -Pspark-2.1 -Ppyspark
>>> -Dhadoop.version=2.6.0-cdh5.10.1 -Phadoop-2.6 -Pyarn *-Pr*
>>> -Pvendor-repo -Pscala-2.10 -pl '!...,!...' -e
>>
>>
>> . . .
>>> [INFO] Zeppelin: *R Interpreter*  SUCCESS
>>> [01:01 min]
>>> [INFO] 
>>> 
>>> [INFO] BUILD SUCCESS
>>> [INFO] 
>>> 
>>> [INFO] Total time: 11:28 min
>>
>>
>> None of the R-related interpreters show up nevertheless.
>>
>> This is including latest Zeppelin snapshot and was the same on previous
>> releases of Zeppelin.
>> So something is missing on our side.
>>
>> R and R packages mentioned in http://zeppelin.apache.org/
>> docs/0.8.0-SNAPSHOT/interpreter/r.html
>> are installed on the servers that runs Zeppelin (and Spark driver as it
>> is yarn-client).
>>
>> I guess either above build options are wrong or there is another
>> dependency I missed.
>> conf/zeppelin-site.xml has R related interpreters mentioned - [1] but
>> none of them
>> show up once Zeppelin starts up.
>>
>> Any ideas?
>>
>>
>> Thank you,
>> Ruslan
>>
>>
>> [1]
>>
>> 
>>>   zeppelin.interpreters
>>>   org.apache.zeppelin.spark.PySparkInterpreter,org.
>>> apache.zeppelin.spark.SparkInterpreter,org.apache.zeppelin.
>>> *rinterpreter.RRepl*,org.apache.zeppelin.rinterpreter.*KnitR*
>>> ,org.apache.zeppelin.spark.*SparkRInterpreter*
>>> ,org.apache.zeppelin.spark.SparkSqlInterpreter,org.
>>> apache.zeppelin.spark.DepInterpreter,org.apache.
>>> zeppelin.markdown.Markdown,org.apache.zeppelin.angular.
>>> AngularInterpreter,org.apache.zeppelin.shell.
>>> ShellInterpreter,org.apache.zeppelin.file.HDFSFileInterpreter,org.
>>> apache.zeppelin.flink.FlinkInterpreter,,org.apache.zeppelin.python.
>>> PythonInterpreter,org.apache.zeppelin.lens.LensInterpreter,
>>> org.apache.zeppelin.ignite.IgniteInterpreter,org.apache.zeppelin.ignite.
>>> IgniteSqlInterpreter,org.apache.zeppelin.cassandra.
>>> CassandraInterpreter,org.apache.zeppelin.geode.GeodeOqlInterpreter,org.
>>> apache.zeppelin.postgresql.PostgreSqlInterpreter,org.
>>> apache.zeppelin.jdbc.JDBCInterpreter,org.apache.zeppelin.kylin.
>>> KylinInterpreter,org.apache.zeppelin.elasticsearch.
>>> ElasticsearchInterpreter,org.apache.zeppelin.scalding.
>>> ScaldingInterpreter,org.apache.zeppelin.alluxio.
>>> AlluxioInterpreter,org.apache.zeppelin.hbase.
>>> HbaseInterpreter,org.apache.zeppelin.livy.LivySparkInterpreter,org.
>>> apache.zeppelin.livy.LivyPySparkInterpreter,org.apache.zeppelin.livy.
>>> LivySparkRInterpreter,org.apache.zeppelin.livy.
>>> LivySparkSQLInterpreter,org.apache.zeppelin.bigquery.
>>> BigQueryInterpreter
>>>   Comma separated interpreter configurations. First
>>> interpreter become a def

Re: How do I configure R interpreter in Zeppelin?

2017-04-26 Thread moon soo Lee
Zeppelin includes two R interpreter implementations.

One used to activated by -Psparkr the other -Pr.
Since https://github.com/apache/zeppelin/pull/2215, -Psparkr is activated
by default. And if you're trying to use sparkR, -Psparkr (activated by
default in master branch) is implementation you might be more interested.

So you can just try use with %spark.r prefix.
Let me know if it works for you.

Thanks,
moon

On Wed, Apr 26, 2017 at 12:11 AM Ruslan Dautkhanov 
wrote:

> Hi moon soo Lee,
>
> Cloudera's Spark doesn't have $SPARK_HOME/bin/sparkR
> Would Zeppelin still enable its sparkR interpreter then?
>
> Built Zeppelin using
>
> $ mvn clean package -DskipTests -Pspark-2.1 -Ppyspark
>> -Dhadoop.version=2.6.0-cdh5.10.1 -Phadoop-2.6 -Pyarn *-Pr* -Pvendor-repo
>> -Pscala-2.10 -pl '!...,!...' -e
>
>
> . . .
>> [INFO] Zeppelin: *R Interpreter*  SUCCESS
>> [01:01 min]
>> [INFO]
>> 
>> [INFO] BUILD SUCCESS
>> [INFO]
>> 
>> [INFO] Total time: 11:28 min
>
>
> None of the R-related interpreters show up nevertheless.
>
> This is including latest Zeppelin snapshot and was the same on previous
> releases of Zeppelin.
> So something is missing on our side.
>
> R and R packages mentioned in
> http://zeppelin.apache.org/docs/0.8.0-SNAPSHOT/interpreter/r.html
> are installed on the servers that runs Zeppelin (and Spark driver as it is
> yarn-client).
>
> I guess either above build options are wrong or there is another
> dependency I missed.
> conf/zeppelin-site.xml has R related interpreters mentioned - [1] but none
> of them
> show up once Zeppelin starts up.
>
> Any ideas?
>
>
> Thank you,
> Ruslan
>
>
> [1]
>
> 
>>   zeppelin.interpreters
>>
>> org.apache.zeppelin.spark.PySparkInterpreter,org.apache.zeppelin.spark.SparkInterpreter,org.apache.zeppelin.
>> *rinterpreter.RRepl*,org.apache.zeppelin.rinterpreter.*KnitR*
>> ,org.apache.zeppelin.spark.*SparkRInterpreter*
>> ,org.apache.zeppelin.spark.SparkSqlInterpreter,org.apache.zeppelin.spark.DepInterpreter,org.apache.zeppelin.markdown.Markdown,org.apache.zeppelin.angular.AngularInterpreter,org.apache.zeppelin.shell.ShellInterpreter,org.apache.zeppelin.file.HDFSFileInterpreter,org.apache.zeppelin.flink.FlinkInterpreter,,org.apache.zeppelin.python.PythonInterpreter,org.apache.zeppelin.lens.LensInterpreter,org.apache.zeppelin.ignite.IgniteInterpreter,org.apache.zeppelin.ignite.IgniteSqlInterpreter,org.apache.zeppelin.cassandra.CassandraInterpreter,org.apache.zeppelin.geode.GeodeOqlInterpreter,org.apache.zeppelin.postgresql.PostgreSqlInterpreter,org.apache.zeppelin.jdbc.JDBCInterpreter,org.apache.zeppelin.kylin.KylinInterpreter,org.apache.zeppelin.elasticsearch.ElasticsearchInterpreter,org.apache.zeppelin.scalding.ScaldingInterpreter,org.apache.zeppelin.alluxio.AlluxioInterpreter,org.apache.zeppelin.hbase.HbaseInterpreter,org.apache.zeppelin.livy.LivySparkInterpreter,org.apache.zeppelin.livy.LivyPySparkInterpreter,org.apache.zeppelin.livy.LivySparkRInterpreter,org.apache.zeppelin.livy.LivySparkSQLInterpreter,org.apache.zeppelin.bigquery.BigQueryInterpreter
>>   Comma separated interpreter configurations. First
>> interpreter become a default
>> 
>
>
>
>
>
> --
> Ruslan Dautkhanov
>
> On Sun, Mar 19, 2017 at 1:07 PM, moon soo Lee  wrote:
>
>> Easiest way to figure out what your environment needs is,
>>
>> 1. run SPARK_HOME/bin/sparkR in your shell and make sure it works in the
>> same host where Zeppelin going to run.
>> 2. try use %spark.r in Zeppelin with SPARK_HOME configured. Normally it
>> should work when 1) works without problem, otherwise take a look error
>> message and error log to get more informations.
>>
>> Thanks,
>> moon
>>
>>
>> On Sat, Mar 18, 2017 at 8:47 PM Shanmukha Sreenivas Potti <
>> shanmu...@utexas.edu> wrote:
>>
>> I'm not 100% sure as I haven't set it up but it looks like I'm using
>>> Zeppelin preconfigured with Spark and I've also taken a snapshot of the
>>> Spark Interpreter configuration that I have access to/using in Zeppelin.
>>> This interpreter comes with SQL and Python integration and I'm figuring out
>>> how do I get to use R.
>>>
>>> On Sat, Mar 18, 2017 at 8:06 PM, moon soo Lee  wrote:
>>>
>>> AFAIK, Amazon EMR service has an option that launches Zeppelin
>>> (preconfigured) with Spark. Do you use Zeppelin provided by EMR or are you
>>> setting up Zeppelin separately?
>>>
>>> Thanks,
>>> moon
>>>
>>> On Sat, Mar 18, 2017 at 4:13 PM Shanmukha Sreenivas Potti <
>>> shanmu...@utexas.edu> wrote:
>>>
>>> ​​
>>> Hi Moon,
>>>
>>> Thanks for responding. Exporting Spark_home is exactly where I have a
>>> problem. I'm using Zeppelin notebook with Spark on EMR clusters from an AWS
>>> account on cloud. I'm not the master account holder for that AWS account
>>> but I'm guessing I'm a client account with limited access probably. Can I
>>> still do it?
>>>
>>> If yes, can you expla

Re: How do I configure R interpreter in Zeppelin?

2017-04-26 Thread Ruslan Dautkhanov
Hi moon soo Lee,

Cloudera's Spark doesn't have $SPARK_HOME/bin/sparkR
Would Zeppelin still enable its sparkR interpreter then?

Built Zeppelin using

$ mvn clean package -DskipTests -Pspark-2.1 -Ppyspark
> -Dhadoop.version=2.6.0-cdh5.10.1 -Phadoop-2.6 -Pyarn *-Pr* -Pvendor-repo
> -Pscala-2.10 -pl '!...,!...' -e


. . .
> [INFO] Zeppelin: *R Interpreter*  SUCCESS
> [01:01 min]
> [INFO]
> 
> [INFO] BUILD SUCCESS
> [INFO]
> 
> [INFO] Total time: 11:28 min


None of the R-related interpreters show up nevertheless.

This is including latest Zeppelin snapshot and was the same on previous
releases of Zeppelin.
So something is missing on our side.

R and R packages mentioned in
http://zeppelin.apache.org/docs/0.8.0-SNAPSHOT/interpreter/r.html
are installed on the servers that runs Zeppelin (and Spark driver as it is
yarn-client).

I guess either above build options are wrong or there is another dependency
I missed.
conf/zeppelin-site.xml has R related interpreters mentioned - [1] but none
of them
show up once Zeppelin starts up.

Any ideas?


Thank you,
Ruslan


[1]


>   zeppelin.interpreters
>
> org.apache.zeppelin.spark.PySparkInterpreter,org.apache.zeppelin.spark.SparkInterpreter,org.apache.zeppelin.
> *rinterpreter.RRepl*,org.apache.zeppelin.rinterpreter.*KnitR*
> ,org.apache.zeppelin.spark.*SparkRInterpreter*
> ,org.apache.zeppelin.spark.SparkSqlInterpreter,org.apache.zeppelin.spark.DepInterpreter,org.apache.zeppelin.markdown.Markdown,org.apache.zeppelin.angular.AngularInterpreter,org.apache.zeppelin.shell.ShellInterpreter,org.apache.zeppelin.file.HDFSFileInterpreter,org.apache.zeppelin.flink.FlinkInterpreter,,org.apache.zeppelin.python.PythonInterpreter,org.apache.zeppelin.lens.LensInterpreter,org.apache.zeppelin.ignite.IgniteInterpreter,org.apache.zeppelin.ignite.IgniteSqlInterpreter,org.apache.zeppelin.cassandra.CassandraInterpreter,org.apache.zeppelin.geode.GeodeOqlInterpreter,org.apache.zeppelin.postgresql.PostgreSqlInterpreter,org.apache.zeppelin.jdbc.JDBCInterpreter,org.apache.zeppelin.kylin.KylinInterpreter,org.apache.zeppelin.elasticsearch.ElasticsearchInterpreter,org.apache.zeppelin.scalding.ScaldingInterpreter,org.apache.zeppelin.alluxio.AlluxioInterpreter,org.apache.zeppelin.hbase.HbaseInterpreter,org.apache.zeppelin.livy.LivySparkInterpreter,org.apache.zeppelin.livy.LivyPySparkInterpreter,org.apache.zeppelin.livy.LivySparkRInterpreter,org.apache.zeppelin.livy.LivySparkSQLInterpreter,org.apache.zeppelin.bigquery.BigQueryInterpreter
>   Comma separated interpreter configurations. First
> interpreter become a default
> 





-- 
Ruslan Dautkhanov

On Sun, Mar 19, 2017 at 1:07 PM, moon soo Lee  wrote:

> Easiest way to figure out what your environment needs is,
>
> 1. run SPARK_HOME/bin/sparkR in your shell and make sure it works in the
> same host where Zeppelin going to run.
> 2. try use %spark.r in Zeppelin with SPARK_HOME configured. Normally it
> should work when 1) works without problem, otherwise take a look error
> message and error log to get more informations.
>
> Thanks,
> moon
>
>
> On Sat, Mar 18, 2017 at 8:47 PM Shanmukha Sreenivas Potti <
> shanmu...@utexas.edu> wrote:
>
> I'm not 100% sure as I haven't set it up but it looks like I'm using
>> Zeppelin preconfigured with Spark and I've also taken a snapshot of the
>> Spark Interpreter configuration that I have access to/using in Zeppelin.
>> This interpreter comes with SQL and Python integration and I'm figuring out
>> how do I get to use R.
>>
>> On Sat, Mar 18, 2017 at 8:06 PM, moon soo Lee  wrote:
>>
>> AFAIK, Amazon EMR service has an option that launches Zeppelin
>> (preconfigured) with Spark. Do you use Zeppelin provided by EMR or are you
>> setting up Zeppelin separately?
>>
>> Thanks,
>> moon
>>
>> On Sat, Mar 18, 2017 at 4:13 PM Shanmukha Sreenivas Potti <
>> shanmu...@utexas.edu> wrote:
>>
>> ​​
>> Hi Moon,
>>
>> Thanks for responding. Exporting Spark_home is exactly where I have a
>> problem. I'm using Zeppelin notebook with Spark on EMR clusters from an AWS
>> account on cloud. I'm not the master account holder for that AWS account
>> but I'm guessing I'm a client account with limited access probably. Can I
>> still do it?
>>
>> If yes, can you explain where and how should I do that shell scripting to
>> export the variable? Can I do this in the notebook itself by starting the
>> paragraph with sh% or do I need to do something else?
>> If you can share any video that would be great. I would like to let you
>> know that I'm a novice user just getting to explore Big Data.
>>
>> Sharing more info for better context.
>>
>> Here's my AWS account detail type:
>> assumed-role/ConduitAccessClientRole-DO-NOT-DELETE/shan
>>
>> Spark Interpreter config in Zeppelin:
>> [image: image.png]
>>
>> Thanks for your help.
>>
>> Shan
>>
>> On Sat, Mar 18, 2017 at 8:39 AM, 

Re: How do I configure R interpreter in Zeppelin?

2017-03-19 Thread moon soo Lee
Easiest way to figure out what your environment needs is,

1. run SPARK_HOME/bin/sparkR in your shell and make sure it works in the
same host where Zeppelin going to run.
2. try use %spark.r in Zeppelin with SPARK_HOME configured. Normally it
should work when 1) works without problem, otherwise take a look error
message and error log to get more informations.

Thanks,
moon

On Sat, Mar 18, 2017 at 8:47 PM Shanmukha Sreenivas Potti <
shanmu...@utexas.edu> wrote:

> I'm not 100% sure as I haven't set it up but it looks like I'm using
> Zeppelin preconfigured with Spark and I've also taken a snapshot of the
> Spark Interpreter configuration that I have access to/using in Zeppelin.
> This interpreter comes with SQL and Python integration and I'm figuring out
> how do I get to use R.
>
> On Sat, Mar 18, 2017 at 8:06 PM, moon soo Lee  wrote:
>
> AFAIK, Amazon EMR service has an option that launches Zeppelin
> (preconfigured) with Spark. Do you use Zeppelin provided by EMR or are you
> setting up Zeppelin separately?
>
> Thanks,
> moon
>
> On Sat, Mar 18, 2017 at 4:13 PM Shanmukha Sreenivas Potti <
> shanmu...@utexas.edu> wrote:
>
> ​​
> Hi Moon,
>
> Thanks for responding. Exporting Spark_home is exactly where I have a
> problem. I'm using Zeppelin notebook with Spark on EMR clusters from an AWS
> account on cloud. I'm not the master account holder for that AWS account
> but I'm guessing I'm a client account with limited access probably. Can I
> still do it?
>
> If yes, can you explain where and how should I do that shell scripting to
> export the variable? Can I do this in the notebook itself by starting the
> paragraph with sh% or do I need to do something else?
> If you can share any video that would be great. I would like to let you
> know that I'm a novice user just getting to explore Big Data.
>
> Sharing more info for better context.
>
> Here's my AWS account detail type:
> assumed-role/ConduitAccessClientRole-DO-NOT-DELETE/shan
>
> Spark Interpreter config in Zeppelin:
> [image: image.png]
>
> Thanks for your help.
>
> Shan
>
> On Sat, Mar 18, 2017 at 8:39 AM, moon soo Lee  wrote:
>
> If you don't have spark cluster, then you don't need to do 2).
> After 1) %spark.r interpreter should work.
>
> If you do have spark cluster, export SPARK_HOME env variable in
> conf/zeppelin-env.sh, that should be enough make it work.
>
> Hope this helps.
>
> Thanks,
> moon
>
> On Fri, Mar 17, 2017 at 2:41 PM Shanmukha Sreenivas Potti <
> shanmu...@utexas.edu> wrote:
>
> Hello Group!
>
> I'm trying to leverage various R functions in Zeppelin but am having
> challenges in figuring out how to configure the Spark interpreter/
> SPARK_HOME variable.
>
> I'm going by this
>  documentation
> for now, and specifically have issues with the following steps:
>
>1.
>
>To run R code and visualize plots in Apache Zeppelin, you will need R
>on your master node (or your dev laptop).
>
>For Centos: yum install R R-devel libcurl-devel openssl-devel For
>Ubuntu: apt-get install r-base
>
> How do I figure out the master node and install the R interpreter? Novice
> user here.
>
>
> 2. To run Zeppelin with the R Interpreter, the SPARK_HOME environment
> variable must be set. The best way to do this is by editing
> conf/zeppelin-env.sh. If it is not set, the R Interpreter will not be able
> to interface with Spark. You should also copy
> conf/zeppelin-site.xml.template to conf/zeppelin-site.xml. That will ensure
> that Zeppelin sees the R Interpreter the first time it starts up.
>
> No idea as to how to do step 2 either.
>
> Appreciate your help. If there is a video that you can point me to that
> talks about these steps, that would be fantabulous.
>
> Thanks! Shan
>
> --
> Shan S. Potti,
>
>
>
>
> --
> Shan S. Potti,
> 737-333-1952 <(737)%20333-1952>
> https://www.linkedin.com/in/shanmukhasreenivas
>
> On Sat, Mar 18, 2017 at 8:39 AM, moon soo Lee  wrote:
>
> If you don't have spark cluster, then you don't need to do 2).
> After 1) %spark.r interpreter should work.
>
> If you do have spark cluster, export SPARK_HOME env variable in
> conf/zeppelin-env.sh, that should be enough make it work.
>
> Hope this helps.
>
> Thanks,
> moon
>
> On Fri, Mar 17, 2017 at 2:41 PM Shanmukha Sreenivas Potti <
> shanmu...@utexas.edu> wrote:
>
> Hello Group!
>
> I'm trying to leverage various R functions in Zeppelin but am having
> challenges in figuring out how to configure the Spark interpreter/
> SPARK_HOME variable.
>
> I'm going by this
>  documentation
> for now, and specifically have issues with the following steps:
>
>1.
>
>To run R code and visualize plots in Apache Zeppelin, you will need R
>on your master node (or your dev laptop).
>
>For Centos: yum install R R-devel libcurl-devel openssl-devel For
>Ubuntu: apt-get install r-base
>
> How do I figure out the master node and install the R interpret

Re: How do I configure R interpreter in Zeppelin?

2017-03-18 Thread Shanmukha Sreenivas Potti
I'm not 100% sure as I haven't set it up but it looks like I'm using
Zeppelin preconfigured with Spark and I've also taken a snapshot of the
Spark Interpreter configuration that I have access to/using in Zeppelin.
This interpreter comes with SQL and Python integration and I'm figuring out
how do I get to use R.

On Sat, Mar 18, 2017 at 8:06 PM, moon soo Lee  wrote:

> AFAIK, Amazon EMR service has an option that launches Zeppelin
> (preconfigured) with Spark. Do you use Zeppelin provided by EMR or are you
> setting up Zeppelin separately?
>
> Thanks,
> moon
>
> On Sat, Mar 18, 2017 at 4:13 PM Shanmukha Sreenivas Potti <
> shanmu...@utexas.edu> wrote:
>
>> ​​
>> Hi Moon,
>>
>> Thanks for responding. Exporting Spark_home is exactly where I have a
>> problem. I'm using Zeppelin notebook with Spark on EMR clusters from an AWS
>> account on cloud. I'm not the master account holder for that AWS account
>> but I'm guessing I'm a client account with limited access probably. Can I
>> still do it?
>>
>> If yes, can you explain where and how should I do that shell scripting to
>> export the variable? Can I do this in the notebook itself by starting the
>> paragraph with sh% or do I need to do something else?
>> If you can share any video that would be great. I would like to let you
>> know that I'm a novice user just getting to explore Big Data.
>>
>> Sharing more info for better context.
>>
>> Here's my AWS account detail type:
>> assumed-role/ConduitAccessClientRole-DO-NOT-DELETE/shan
>>
>> Spark Interpreter config in Zeppelin:
>> [image: image.png]
>>
>> Thanks for your help.
>>
>> Shan
>>
>> On Sat, Mar 18, 2017 at 8:39 AM, moon soo Lee  wrote:
>>
>> If you don't have spark cluster, then you don't need to do 2).
>> After 1) %spark.r interpreter should work.
>>
>> If you do have spark cluster, export SPARK_HOME env variable in
>> conf/zeppelin-env.sh, that should be enough make it work.
>>
>> Hope this helps.
>>
>> Thanks,
>> moon
>>
>> On Fri, Mar 17, 2017 at 2:41 PM Shanmukha Sreenivas Potti <
>> shanmu...@utexas.edu> wrote:
>>
>> Hello Group!
>>
>> I'm trying to leverage various R functions in Zeppelin but am having
>> challenges in figuring out how to configure the Spark interpreter/
>> SPARK_HOME variable.
>>
>> I'm going by this
>>  documentation
>> for now, and specifically have issues with the following steps:
>>
>>1.
>>
>>To run R code and visualize plots in Apache Zeppelin, you will need R
>>on your master node (or your dev laptop).
>>
>>For Centos: yum install R R-devel libcurl-devel openssl-devel For
>>Ubuntu: apt-get install r-base
>>
>> How do I figure out the master node and install the R interpreter? Novice
>> user here.
>>
>>
>> 2. To run Zeppelin with the R Interpreter, the SPARK_HOME environment
>> variable must be set. The best way to do this is by editing
>> conf/zeppelin-env.sh. If it is not set, the R Interpreter will not be able
>> to interface with Spark. You should also copy conf/zeppelin-site.xml.template
>> to conf/zeppelin-site.xml. That will ensure that Zeppelin sees the R
>> Interpreter the first time it starts up.
>>
>> No idea as to how to do step 2 either.
>>
>> Appreciate your help. If there is a video that you can point me to that
>> talks about these steps, that would be fantabulous.
>>
>> Thanks! Shan
>>
>> --
>> Shan S. Potti,
>>
>>
>>
>>
>> --
>> Shan S. Potti,
>> 737-333-1952 <(737)%20333-1952>
>> https://www.linkedin.com/in/shanmukhasreenivas
>>
>> On Sat, Mar 18, 2017 at 8:39 AM, moon soo Lee  wrote:
>>
>> If you don't have spark cluster, then you don't need to do 2).
>> After 1) %spark.r interpreter should work.
>>
>> If you do have spark cluster, export SPARK_HOME env variable in
>> conf/zeppelin-env.sh, that should be enough make it work.
>>
>> Hope this helps.
>>
>> Thanks,
>> moon
>>
>> On Fri, Mar 17, 2017 at 2:41 PM Shanmukha Sreenivas Potti <
>> shanmu...@utexas.edu> wrote:
>>
>> Hello Group!
>>
>> I'm trying to leverage various R functions in Zeppelin but am having
>> challenges in figuring out how to configure the Spark interpreter/
>> SPARK_HOME variable.
>>
>> I'm going by this
>> 
>> documentation for now, and specifically have issues with the following
>> steps:
>>
>>1.
>>
>>To run R code and visualize plots in Apache Zeppelin, you will need R
>>on your master node (or your dev laptop).
>>
>>For Centos: yum install R R-devel libcurl-devel openssl-devel For
>>Ubuntu: apt-get install r-base
>>
>> How do I figure out the master node and install the R interpreter? Novice
>> user here.
>>
>>
>> 2. To run Zeppelin with the R Interpreter, the SPARK_HOME environment
>> variable must be set. The best way to do this is by editing
>> conf/zeppelin-env.sh. If it is not set, the R Interpreter will not be able
>> to interface with Spark. You should also copy conf/zeppelin-site.xml.template
>> to conf/zeppelin-site.xm

Re: How do I configure R interpreter in Zeppelin?

2017-03-18 Thread moon soo Lee
AFAIK, Amazon EMR service has an option that launches Zeppelin
(preconfigured) with Spark. Do you use Zeppelin provided by EMR or are you
setting up Zeppelin separately?

Thanks,
moon

On Sat, Mar 18, 2017 at 4:13 PM Shanmukha Sreenivas Potti <
shanmu...@utexas.edu> wrote:

> ​​
> Hi Moon,
>
> Thanks for responding. Exporting Spark_home is exactly where I have a
> problem. I'm using Zeppelin notebook with Spark on EMR clusters from an AWS
> account on cloud. I'm not the master account holder for that AWS account
> but I'm guessing I'm a client account with limited access probably. Can I
> still do it?
>
> If yes, can you explain where and how should I do that shell scripting to
> export the variable? Can I do this in the notebook itself by starting the
> paragraph with sh% or do I need to do something else?
> If you can share any video that would be great. I would like to let you
> know that I'm a novice user just getting to explore Big Data.
>
> Sharing more info for better context.
>
> Here's my AWS account detail type:
> assumed-role/ConduitAccessClientRole-DO-NOT-DELETE/shan
>
> Spark Interpreter config in Zeppelin:
> [image: image.png]
>
> Thanks for your help.
>
> Shan
>
> On Sat, Mar 18, 2017 at 8:39 AM, moon soo Lee  wrote:
>
> If you don't have spark cluster, then you don't need to do 2).
> After 1) %spark.r interpreter should work.
>
> If you do have spark cluster, export SPARK_HOME env variable in
> conf/zeppelin-env.sh, that should be enough make it work.
>
> Hope this helps.
>
> Thanks,
> moon
>
> On Fri, Mar 17, 2017 at 2:41 PM Shanmukha Sreenivas Potti <
> shanmu...@utexas.edu> wrote:
>
> Hello Group!
>
> I'm trying to leverage various R functions in Zeppelin but am having
> challenges in figuring out how to configure the Spark interpreter/
> SPARK_HOME variable.
>
> I'm going by this
>  documentation
> for now, and specifically have issues with the following steps:
>
>1.
>
>To run R code and visualize plots in Apache Zeppelin, you will need R
>on your master node (or your dev laptop).
>
>For Centos: yum install R R-devel libcurl-devel openssl-devel For
>Ubuntu: apt-get install r-base
>
> How do I figure out the master node and install the R interpreter? Novice
> user here.
>
>
> 2. To run Zeppelin with the R Interpreter, the SPARK_HOME environment
> variable must be set. The best way to do this is by editing
> conf/zeppelin-env.sh. If it is not set, the R Interpreter will not be able
> to interface with Spark. You should also copy
> conf/zeppelin-site.xml.template to conf/zeppelin-site.xml. That will ensure
> that Zeppelin sees the R Interpreter the first time it starts up.
>
> No idea as to how to do step 2 either.
>
> Appreciate your help. If there is a video that you can point me to that
> talks about these steps, that would be fantabulous.
>
> Thanks! Shan
>
> --
> Shan S. Potti,
>
>
>
>
> --
> Shan S. Potti,
> 737-333-1952 <(737)%20333-1952>
> https://www.linkedin.com/in/shanmukhasreenivas
>
> On Sat, Mar 18, 2017 at 8:39 AM, moon soo Lee  wrote:
>
> If you don't have spark cluster, then you don't need to do 2).
> After 1) %spark.r interpreter should work.
>
> If you do have spark cluster, export SPARK_HOME env variable in
> conf/zeppelin-env.sh, that should be enough make it work.
>
> Hope this helps.
>
> Thanks,
> moon
>
> On Fri, Mar 17, 2017 at 2:41 PM Shanmukha Sreenivas Potti <
> shanmu...@utexas.edu> wrote:
>
> Hello Group!
>
> I'm trying to leverage various R functions in Zeppelin but am having
> challenges in figuring out how to configure the Spark interpreter/
> SPARK_HOME variable.
>
> I'm going by this
>  documentation
> for now, and specifically have issues with the following steps:
>
>1.
>
>To run R code and visualize plots in Apache Zeppelin, you will need R
>on your master node (or your dev laptop).
>
>For Centos: yum install R R-devel libcurl-devel openssl-devel For
>Ubuntu: apt-get install r-base
>
> How do I figure out the master node and install the R interpreter? Novice
> user here.
>
>
> 2. To run Zeppelin with the R Interpreter, the SPARK_HOME environment
> variable must be set. The best way to do this is by editing
> conf/zeppelin-env.sh. If it is not set, the R Interpreter will not be able
> to interface with Spark. You should also copy
> conf/zeppelin-site.xml.template to conf/zeppelin-site.xml. That will ensure
> that Zeppelin sees the R Interpreter the first time it starts up.
>
> No idea as to how to do step 2 either.
>
> Appreciate your help. If there is a video that you can point me to that
> talks about these steps, that would be fantabulous.
>
> Thanks! Shan
>
> --
> Shan S. Potti,
>
>
>
>
> --
> Shan S. Potti,
> 737-333-1952 <(737)%20333-1952>
> https://www.linkedin.com/in/shanmukhasreenivas
>


Re: How do I configure R interpreter in Zeppelin?

2017-03-18 Thread Shanmukha Sreenivas Potti
​​
Hi Moon,

Thanks for responding. Exporting Spark_home is exactly where I have a
problem. I'm using Zeppelin notebook with Spark on EMR clusters from an AWS
account on cloud. I'm not the master account holder for that AWS account
but I'm guessing I'm a client account with limited access probably. Can I
still do it?

If yes, can you explain where and how should I do that shell scripting to
export the variable? Can I do this in the notebook itself by starting the
paragraph with sh% or do I need to do something else?
If you can share any video that would be great. I would like to let you
know that I'm a novice user just getting to explore Big Data.

Sharing more info for better context.

Here's my AWS account detail type:
assumed-role/ConduitAccessClientRole-DO-NOT-DELETE/shan

Spark Interpreter config in Zeppelin:
[image: Inline image 1]

Thanks for your help.

Shan

On Sat, Mar 18, 2017 at 8:39 AM, moon soo Lee  wrote:

> If you don't have spark cluster, then you don't need to do 2).
> After 1) %spark.r interpreter should work.
>
> If you do have spark cluster, export SPARK_HOME env variable in
> conf/zeppelin-env.sh, that should be enough make it work.
>
> Hope this helps.
>
> Thanks,
> moon
>
> On Fri, Mar 17, 2017 at 2:41 PM Shanmukha Sreenivas Potti <
> shanmu...@utexas.edu> wrote:
>
>> Hello Group!
>>
>> I'm trying to leverage various R functions in Zeppelin but am having
>> challenges in figuring out how to configure the Spark interpreter/
>> SPARK_HOME variable.
>>
>> I'm going by this
>>  documentation
>> for now, and specifically have issues with the following steps:
>>
>>1.
>>
>>To run R code and visualize plots in Apache Zeppelin, you will need R
>>on your master node (or your dev laptop).
>>
>>For Centos: yum install R R-devel libcurl-devel openssl-devel For
>>Ubuntu: apt-get install r-base
>>
>> How do I figure out the master node and install the R interpreter? Novice
>> user here.
>>
>>
>> 2. To run Zeppelin with the R Interpreter, the SPARK_HOME environment
>> variable must be set. The best way to do this is by editing
>> conf/zeppelin-env.sh. If it is not set, the R Interpreter will not be able
>> to interface with Spark. You should also copy conf/zeppelin-site.xml.template
>> to conf/zeppelin-site.xml. That will ensure that Zeppelin sees the R
>> Interpreter the first time it starts up.
>>
>> No idea as to how to do step 2 either.
>>
>> Appreciate your help. If there is a video that you can point me to that
>> talks about these steps, that would be fantabulous.
>>
>> Thanks! Shan
>>
>> --
>> Shan S. Potti,
>>
>>


-- 
Shan S. Potti,
737-333-1952
https://www.linkedin.com/in/shanmukhasreenivas

On Sat, Mar 18, 2017 at 8:39 AM, moon soo Lee  wrote:

> If you don't have spark cluster, then you don't need to do 2).
> After 1) %spark.r interpreter should work.
>
> If you do have spark cluster, export SPARK_HOME env variable in
> conf/zeppelin-env.sh, that should be enough make it work.
>
> Hope this helps.
>
> Thanks,
> moon
>
> On Fri, Mar 17, 2017 at 2:41 PM Shanmukha Sreenivas Potti <
> shanmu...@utexas.edu> wrote:
>
>> Hello Group!
>>
>> I'm trying to leverage various R functions in Zeppelin but am having
>> challenges in figuring out how to configure the Spark interpreter/
>> SPARK_HOME variable.
>>
>> I'm going by this
>> 
>> documentation for now, and specifically have issues with the following
>> steps:
>>
>>1.
>>
>>To run R code and visualize plots in Apache Zeppelin, you will need R
>>on your master node (or your dev laptop).
>>
>>For Centos: yum install R R-devel libcurl-devel openssl-devel For
>>Ubuntu: apt-get install r-base
>>
>> How do I figure out the master node and install the R interpreter? Novice
>> user here.
>>
>>
>> 2. To run Zeppelin with the R Interpreter, the SPARK_HOME environment
>> variable must be set. The best way to do this is by editing
>> conf/zeppelin-env.sh. If it is not set, the R Interpreter will not be able
>> to interface with Spark. You should also copy conf/zeppelin-site.xml.template
>> to conf/zeppelin-site.xml. That will ensure that Zeppelin sees the R
>> Interpreter the first time it starts up.
>>
>> No idea as to how to do step 2 either.
>>
>> Appreciate your help. If there is a video that you can point me to that
>> talks about these steps, that would be fantabulous.
>>
>> Thanks! Shan
>>
>> --
>> Shan S. Potti,
>>
>>


-- 
Shan S. Potti,
737-333-1952
https://www.linkedin.com/in/shanmukhasreenivas


Re: How do I configure R interpreter in Zeppelin?

2017-03-18 Thread moon soo Lee
If you don't have spark cluster, then you don't need to do 2).
After 1) %spark.r interpreter should work.

If you do have spark cluster, export SPARK_HOME env variable in
conf/zeppelin-env.sh, that should be enough make it work.

Hope this helps.

Thanks,
moon

On Fri, Mar 17, 2017 at 2:41 PM Shanmukha Sreenivas Potti <
shanmu...@utexas.edu> wrote:

> Hello Group!
>
> I'm trying to leverage various R functions in Zeppelin but am having
> challenges in figuring out how to configure the Spark interpreter/
> SPARK_HOME variable.
>
> I'm going by this
>  documentation
> for now, and specifically have issues with the following steps:
>
>1.
>
>To run R code and visualize plots in Apache Zeppelin, you will need R
>on your master node (or your dev laptop).
>
>For Centos: yum install R R-devel libcurl-devel openssl-devel For
>Ubuntu: apt-get install r-base
>
> How do I figure out the master node and install the R interpreter? Novice
> user here.
>
>
> 2. To run Zeppelin with the R Interpreter, the SPARK_HOME environment
> variable must be set. The best way to do this is by editing
> conf/zeppelin-env.sh. If it is not set, the R Interpreter will not be able
> to interface with Spark. You should also copy
> conf/zeppelin-site.xml.template to conf/zeppelin-site.xml. That will ensure
> that Zeppelin sees the R Interpreter the first time it starts up.
>
> No idea as to how to do step 2 either.
>
> Appreciate your help. If there is a video that you can point me to that
> talks about these steps, that would be fantabulous.
>
> Thanks! Shan
>
> --
> Shan S. Potti,
>
>


How do I configure R interpreter in Zeppelin?

2017-03-17 Thread Shanmukha Sreenivas Potti
Hello Group!

I'm trying to leverage various R functions in Zeppelin but am having
challenges in figuring out how to configure the Spark interpreter/
SPARK_HOME variable.

I'm going by this
 documentation
for now, and specifically have issues with the following steps:

   1.

   To run R code and visualize plots in Apache Zeppelin, you will need R on
   your master node (or your dev laptop).

   For Centos: yum install R R-devel libcurl-devel openssl-devel For
   Ubuntu: apt-get install r-base

How do I figure out the master node and install the R interpreter? Novice
user here.


2. To run Zeppelin with the R Interpreter, the SPARK_HOME environment
variable must be set. The best way to do this is by editing
conf/zeppelin-env.sh. If it is not set, the R Interpreter will not be able
to interface with Spark. You should also copy
conf/zeppelin-site.xml.template to conf/zeppelin-site.xml. That will ensure
that Zeppelin sees the R Interpreter the first time it starts up.

No idea as to how to do step 2 either.

Appreciate your help. If there is a video that you can point me to that
talks about these steps, that would be fantabulous.

Thanks! Shan

-- 
Shan S. Potti,