Thanks for feedback. %spark.r print("Hello World!") throws exception [2].
Understood - I'll try to remove -Pr and rebuild Zeppelin. Yep, I used a fresh master snapshot. ( I have't seen anything in maven build logs that could indicate a problem around R interpreter) Will update this email thread with result after rebuilding Zeppelin without -Pr [2] spark.r interpreter not found org.apache.zeppelin.interpreter.InterpreterException: spark.r interpreter not found at org.apache.zeppelin.interpreter.InterpreterFactory.getInterpreter(InterpreterFactory.java:417) at org.apache.zeppelin.notebook.Note.run(Note.java:620) at org.apache.zeppelin.socket.NotebookServer.persistAndExecuteSingleParagraph(NotebookServer.java:1781) at org.apache.zeppelin.socket.NotebookServer.runParagraph(NotebookServer.java:1741) at org.apache.zeppelin.socket.NotebookServer.onMessage(NotebookServer.java:288) at org.apache.zeppelin.socket.NotebookSocket.onWebSocketText(NotebookSocket.java:59) at org.eclipse.jetty.websocket.common.events.JettyListenerEventDriver.onTextMessage(JettyListenerEventDriver.java:128) at org.eclipse.jetty.websocket.common.message.SimpleTextMessage.messageComplete(SimpleTextMessage.java:69) at org.eclipse.jetty.websocket.common.events.AbstractEventDriver.appendMessage(AbstractEventDriver.java:65) at org.eclipse.jetty.websocket.common.events.JettyListenerEventDriver.onTextFrame(JettyListenerEventDriver.java:122) -- Ruslan Dautkhanov On Wed, Apr 26, 2017 at 2:13 PM, moon soo Lee <m...@apache.org> wrote: > Zeppelin includes two R interpreter implementations. > > One used to activated by -Psparkr the other -Pr. > Since https://github.com/apache/zeppelin/pull/2215, -Psparkr is activated > by default. And if you're trying to use sparkR, -Psparkr (activated by > default in master branch) is implementation you might be more interested. > > So you can just try use with %spark.r prefix. > Let me know if it works for you. > > Thanks, > moon > > On Wed, Apr 26, 2017 at 12:11 AM Ruslan Dautkhanov <dautkha...@gmail.com> > wrote: > >> Hi moon soo Lee, >> >> Cloudera's Spark doesn't have $SPARK_HOME/bin/sparkR >> Would Zeppelin still enable its sparkR interpreter then? >> >> Built Zeppelin using >> >> $ mvn clean package -DskipTests -Pspark-2.1 -Ppyspark >>> -Dhadoop.version=2.6.0-cdh5.10.1 -Phadoop-2.6 -Pyarn *-Pr* >>> -Pvendor-repo -Pscala-2.10 -pl '!...,!...' -e >> >> >> . . . >>> [INFO] Zeppelin: *R Interpreter* ............................ SUCCESS >>> [01:01 min] >>> [INFO] ------------------------------------------------------------ >>> ------------ >>> [INFO] BUILD SUCCESS >>> [INFO] ------------------------------------------------------------ >>> ------------ >>> [INFO] Total time: 11:28 min >> >> >> None of the R-related interpreters show up nevertheless. >> >> This is including latest Zeppelin snapshot and was the same on previous >> releases of Zeppelin. >> So something is missing on our side. >> >> R and R packages mentioned in http://zeppelin.apache.org/ >> docs/0.8.0-SNAPSHOT/interpreter/r.html >> are installed on the servers that runs Zeppelin (and Spark driver as it >> is yarn-client). >> >> I guess either above build options are wrong or there is another >> dependency I missed. >> conf/zeppelin-site.xml has R related interpreters mentioned - [1] but >> none of them >> show up once Zeppelin starts up. >> >> Any ideas? >> >> >> Thank you, >> Ruslan >> >> >> [1] >> >> <property> >>> <name>zeppelin.interpreters</name> >>> <value>org.apache.zeppelin.spark.PySparkInterpreter,org. >>> apache.zeppelin.spark.SparkInterpreter,org.apache.zeppelin. >>> *rinterpreter.RRepl*,org.apache.zeppelin.rinterpreter.*KnitR* >>> ,org.apache.zeppelin.spark.*SparkRInterpreter* >>> ,org.apache.zeppelin.spark.SparkSqlInterpreter,org. >>> apache.zeppelin.spark.DepInterpreter,org.apache. >>> zeppelin.markdown.Markdown,org.apache.zeppelin.angular. >>> AngularInterpreter,org.apache.zeppelin.shell. >>> ShellInterpreter,org.apache.zeppelin.file.HDFSFileInterpreter,org. >>> apache.zeppelin.flink.FlinkInterpreter,,org.apache.zeppelin.python. >>> PythonInterpreter,org.apache.zeppelin.lens.LensInterpreter, >>> org.apache.zeppelin.ignite.IgniteInterpreter,org.apache.zeppelin.ignite. >>> IgniteSqlInterpreter,org.apache.zeppelin.cassandra. >>> CassandraInterpreter,org.apache.zeppelin.geode.GeodeOqlInterpreter,org. >>> apache.zeppelin.postgresql.PostgreSqlInterpreter,org. >>> apache.zeppelin.jdbc.JDBCInterpreter,org.apache.zeppelin.kylin. >>> KylinInterpreter,org.apache.zeppelin.elasticsearch. >>> ElasticsearchInterpreter,org.apache.zeppelin.scalding. >>> ScaldingInterpreter,org.apache.zeppelin.alluxio. >>> AlluxioInterpreter,org.apache.zeppelin.hbase. >>> HbaseInterpreter,org.apache.zeppelin.livy.LivySparkInterpreter,org. >>> apache.zeppelin.livy.LivyPySparkInterpreter,org.apache.zeppelin.livy. >>> LivySparkRInterpreter,org.apache.zeppelin.livy. >>> LivySparkSQLInterpreter,org.apache.zeppelin.bigquery. >>> BigQueryInterpreter</value> >>> <description>Comma separated interpreter configurations. First >>> interpreter become a default</description> >>> </property> >> >> >> >> >> >> -- >> Ruslan Dautkhanov >> >> On Sun, Mar 19, 2017 at 1:07 PM, moon soo Lee <m...@apache.org> wrote: >> >>> Easiest way to figure out what your environment needs is, >>> >>> 1. run SPARK_HOME/bin/sparkR in your shell and make sure it works in the >>> same host where Zeppelin going to run. >>> 2. try use %spark.r in Zeppelin with SPARK_HOME configured. Normally it >>> should work when 1) works without problem, otherwise take a look error >>> message and error log to get more informations. >>> >>> Thanks, >>> moon >>> >>> >>> On Sat, Mar 18, 2017 at 8:47 PM Shanmukha Sreenivas Potti < >>> shanmu...@utexas.edu> wrote: >>> >>> I'm not 100% sure as I haven't set it up but it looks like I'm using >>>> Zeppelin preconfigured with Spark and I've also taken a snapshot of the >>>> Spark Interpreter configuration that I have access to/using in Zeppelin. >>>> This interpreter comes with SQL and Python integration and I'm figuring out >>>> how do I get to use R. >>>> >>>> On Sat, Mar 18, 2017 at 8:06 PM, moon soo Lee <m...@apache.org> wrote: >>>> >>>> AFAIK, Amazon EMR service has an option that launches Zeppelin >>>> (preconfigured) with Spark. Do you use Zeppelin provided by EMR or are you >>>> setting up Zeppelin separately? >>>> >>>> Thanks, >>>> moon >>>> >>>> On Sat, Mar 18, 2017 at 4:13 PM Shanmukha Sreenivas Potti < >>>> shanmu...@utexas.edu> wrote: >>>> >>>> >>>> Hi Moon, >>>> >>>> Thanks for responding. Exporting Spark_home is exactly where I have a >>>> problem. I'm using Zeppelin notebook with Spark on EMR clusters from an AWS >>>> account on cloud. I'm not the master account holder for that AWS account >>>> but I'm guessing I'm a client account with limited access probably. Can I >>>> still do it? >>>> >>>> If yes, can you explain where and how should I do that shell scripting >>>> to export the variable? Can I do this in the notebook itself by starting >>>> the paragraph with sh% or do I need to do something else? >>>> If you can share any video that would be great. I would like to let you >>>> know that I'm a novice user just getting to explore Big Data. >>>> >>>> Sharing more info for better context. >>>> >>>> Here's my AWS account detail type: >>>> assumed-role/ConduitAccessClientRole-DO-NOT-DELETE/shan >>>> >>>> Spark Interpreter config in Zeppelin: >>>> [image: image.png] >>>> >>>> Thanks for your help. >>>> >>>> Shan >>>> >>>> On Sat, Mar 18, 2017 at 8:39 AM, moon soo Lee <m...@apache.org> wrote: >>>> >>>> If you don't have spark cluster, then you don't need to do 2). >>>> After 1) %spark.r interpreter should work. >>>> >>>> If you do have spark cluster, export SPARK_HOME env variable in >>>> conf/zeppelin-env.sh, that should be enough make it work. >>>> >>>> Hope this helps. >>>> >>>> Thanks, >>>> moon >>>> >>>> On Fri, Mar 17, 2017 at 2:41 PM Shanmukha Sreenivas Potti < >>>> shanmu...@utexas.edu> wrote: >>>> >>>> Hello Group! >>>> >>>> I'm trying to leverage various R functions in Zeppelin but am having >>>> challenges in figuring out how to configure the Spark interpreter/ >>>> SPARK_HOME variable. >>>> >>>> I'm going by this >>>> <https://zeppelin.apache.org/docs/0.6.0/interpreter/r.html> documentation >>>> for now, and specifically have issues with the following steps: >>>> >>>> 1. >>>> >>>> To run R code and visualize plots in Apache Zeppelin, you will need >>>> R on your master node (or your dev laptop). >>>> >>>> For Centos: yum install R R-devel libcurl-devel openssl-devel For >>>> Ubuntu: apt-get install r-base >>>> >>>> How do I figure out the master node and install the R interpreter? >>>> Novice user here. >>>> >>>> >>>> 2. To run Zeppelin with the R Interpreter, the SPARK_HOME environment >>>> variable must be set. The best way to do this is by editing >>>> conf/zeppelin-env.sh. If it is not set, the R Interpreter will not be able >>>> to interface with Spark. You should also copy >>>> conf/zeppelin-site.xml.template >>>> to conf/zeppelin-site.xml. That will ensure that Zeppelin sees the R >>>> Interpreter the first time it starts up. >>>> >>>> No idea as to how to do step 2 either. >>>> >>>> Appreciate your help. If there is a video that you can point me to that >>>> talks about these steps, that would be fantabulous. >>>> >>>> Thanks! Shan >>>> >>>> -- >>>> Shan S. Potti, >>>> >>>> >>>> >>>> >>>> -- >>>> Shan S. Potti, >>>> 737-333-1952 <(737)%20333-1952> >>>> https://www.linkedin.com/in/shanmukhasreenivas >>>> >>>> On Sat, Mar 18, 2017 at 8:39 AM, moon soo Lee <m...@apache.org> wrote: >>>> >>>> If you don't have spark cluster, then you don't need to do 2). >>>> After 1) %spark.r interpreter should work. >>>> >>>> If you do have spark cluster, export SPARK_HOME env variable in >>>> conf/zeppelin-env.sh, that should be enough make it work. >>>> >>>> Hope this helps. >>>> >>>> Thanks, >>>> moon >>>> >>>> On Fri, Mar 17, 2017 at 2:41 PM Shanmukha Sreenivas Potti < >>>> shanmu...@utexas.edu> wrote: >>>> >>>> Hello Group! >>>> >>>> I'm trying to leverage various R functions in Zeppelin but am having >>>> challenges in figuring out how to configure the Spark interpreter/ >>>> SPARK_HOME variable. >>>> >>>> I'm going by this >>>> <https://zeppelin.apache.org/docs/0.6.0/interpreter/r.html> >>>> documentation for now, and specifically have issues with the following >>>> steps: >>>> >>>> 1. >>>> >>>> To run R code and visualize plots in Apache Zeppelin, you will need >>>> R on your master node (or your dev laptop). >>>> >>>> For Centos: yum install R R-devel libcurl-devel openssl-devel For >>>> Ubuntu: apt-get install r-base >>>> >>>> How do I figure out the master node and install the R interpreter? >>>> Novice user here. >>>> >>>> >>>> 2. To run Zeppelin with the R Interpreter, the SPARK_HOME environment >>>> variable must be set. The best way to do this is by editing >>>> conf/zeppelin-env.sh. If it is not set, the R Interpreter will not be able >>>> to interface with Spark. You should also copy >>>> conf/zeppelin-site.xml.template >>>> to conf/zeppelin-site.xml. That will ensure that Zeppelin sees the R >>>> Interpreter the first time it starts up. >>>> >>>> No idea as to how to do step 2 either. >>>> >>>> Appreciate your help. If there is a video that you can point me to that >>>> talks about these steps, that would be fantabulous. >>>> >>>> Thanks! Shan >>>> >>>> -- >>>> Shan S. Potti, >>>> >>>> >>>> >>>> >>>> -- >>>> Shan S. Potti, >>>> 737-333-1952 <(737)%20333-1952> >>>> https://www.linkedin.com/in/shanmukhasreenivas >>>> >>>> >>>> >>>> >>>> -- >>>> Shan S. Potti, >>>> 737-333-1952 <(737)%20333-1952> >>>> https://www.linkedin.com/in/shanmukhasreenivas >>>> >>> >>