Is spark installed on your zeppelin machine ? I would to try these
master yarn-client spark.home === SPARK INSTALLATION HOME directory on your zeppelin server. Looking at spark.yarn.jar , i see spark is installed at /usr/iop/current/spark-thriftserver/ . But why is it thirftserver (i do not know what is it). I have spark installed (unzip) on zeppelin machine at /usr/hdp/2.3.1.0-2574/spark/spark/ (can be any location) and have spark.yarn.jar to /usr/hdp/2.3.1.0-2574/spark/spark/lib/spark-assembly-1.4.1-hadoop2.6.0.jar. On Mon, Oct 5, 2015 at 10:20 AM, Sourav Mazumder < sourav.mazumde...@gmail.com> wrote: > Hi Deepu, > > Here u go. > > Regards, > Sourav > > > > > *Properties* name value args master yarn-cluster spark.app.name Zeppelin > spark.cores.max > spark.executor.memory 512m spark.home spark.yarn.jar > /usr/iop/current/spark-thriftserver/lib/spark-assembly.jar > zeppelin.dep.localrepo local-repo zeppelin.pyspark.python python > zeppelin.spark.concurrentSQL > false zeppelin.spark.maxResult 1000 zeppelin.spark.useHiveContext true > > On Mon, Oct 5, 2015 at 10:05 AM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> > wrote: > >> Can you share screen shot of your spark interpreter on zeppelin web >> interface. >> >> I have exact same deployment structure and it runs fine with right set of >> configurations. >> >> On Mon, Oct 5, 2015 at 7:56 AM, Sourav Mazumder < >> sourav.mazumde...@gmail.com> wrote: >> >>> Hi Moon, >>> >>> I'm using 0.6 SNAPSHOT which I built from latest git hub. >>> >>> I tried setting SPARK_HOME in zeppelin-env.sh. Also I could see that the >>> control goes to the appropriate IF-ELSE block in interpreter.sh by putting >>> some debug statement. >>> >>> But I get the same error as follows - >>> >>> org.apache.spark.SparkException: Detected yarn-cluster mode, but isn't >>> running on a cluster. Deployment to YARN is not supported directly by >>> SparkContext. Please use spark-submit. at >>> org.apache.spark.SparkContext.<init>(SparkContext.scala:378) at >>> org.apache.zeppelin.spark.SparkInterpreter.createSparkContext(SparkInterpreter.java:339) >>> at >>> org.apache.zeppelin.spark.SparkInterpreter.getSparkContext(SparkInterpreter.java:149) >>> at >>> org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:465) >>> at >>> org.apache.zeppelin.interpreter.ClassloaderInterpreter.open(ClassloaderInterpreter.java:74) >>> at >>> org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:68) >>> at >>> org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:92) >>> at >>> org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:276) >>> at org.apache.zeppelin.scheduler.Job.run(Job.java:170) at >>> org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:118) >>> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) >>> at java.util.concurrent.FutureTask.run(FutureTask.java:266) at >>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) >>> at >>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) >>> at >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) >>> at java.lang.Thread.run(Thread.java:745) >>> >>> Let me know if you need any other details to figure out what is going on. >>> >>> Regards, >>> Sourav >>> >>> On Wed, Sep 30, 2015 at 1:53 AM, moon soo Lee <m...@apache.org> wrote: >>> >>>> Which version of Zeppelin are you using? >>>> >>>> Master branch uses spark-submit command, when SPARK_HOME is defined in >>>> conf/zeppelin-env.sh >>>> >>>> If you're not on master branch, recommend try it with SPARK_HOME >>>> defined. >>>> >>>> Hope this helps, >>>> moon >>>> >>>> On Wed, Sep 23, 2015 at 10:21 PM Sourav Mazumder < >>>> sourav.mazumde...@gmail.com> wrote: >>>> >>>>> Hi, >>>>> >>>>> When I try to run Spark Interpreter in Yarn Cluster mode from a remote >>>>> machine I always get the error saying try spark-submit than using spark >>>>> context. >>>>> >>>>> Mu Zeppelin process runs in a separate machine remote to the YARN >>>>> cluster. >>>>> >>>>> Any idea why is this error ? >>>>> >>>>> Regards, >>>>> Sourav >>>>> >>>> >>> >> >> >> -- >> Deepak >> >> > -- Deepak