https://github.com/apache/zeppelin/pull/2577 pronounces yarn-cluster in it's title so I assume it's only yarn-cluster. Never used standalone-cluster myself.
Which distro of Hadoop do you use? Cloudera desupported standalone in CDH 5.5 and will remove in CDH 6. https://www.cloudera.com/documentation/enterprise/release-notes/topics/rg_deprecated.html -- Ruslan Dautkhanov On Tue, Mar 13, 2018 at 5:45 PM, Jhon Anderson Cardenas Diaz < jhonderson2...@gmail.com> wrote: > Does this new feature work only for yarn-cluster ?. Or for spark > standalone too ? > > El mar., 13 de mar. de 2018 18:34, Ruslan Dautkhanov <dautkha...@gmail.com> > escribió: > >> > Zeppelin version: 0.8.0 (merged at September 2017 version) >> >> https://issues.apache.org/jira/browse/ZEPPELIN-2898 was merged end of >> September so not sure if you have that. >> >> Check out https://medium.com/@zjffdu/zeppelin-0-8-0-new- >> features-ea53e8810235 how to set this up. >> >> >> >> -- >> Ruslan Dautkhanov >> >> On Tue, Mar 13, 2018 at 5:24 PM, Jhon Anderson Cardenas Diaz < >> jhonderson2...@gmail.com> wrote: >> >>> Hi zeppelin users ! >>> >>> I am working with zeppelin pointing to a spark in standalone. I am >>> trying to figure out a way to make zeppelin runs the spark driver outside >>> of client process that submits the application. >>> >>> According with the documentation (http://spark.apache.org/docs/ >>> 2.1.1/spark-standalone.html): >>> >>> *For standalone clusters, Spark currently supports two deploy modes. >>> In client mode, the driver is launched in the same process as the client >>> that submits the application. In cluster mode, however, the driver is >>> launched from one of the Worker processes inside the cluster, and the >>> client process exits as soon as it fulfills its responsibility of >>> submitting the application without waiting for the application to finish.* >>> >>> The problem is that, even when I set the properties for spark-standalone >>> cluster and deploy mode in cluster, the driver still run inside zeppelin >>> machine (according with spark UI/executors page). These are properties that >>> I am setting for the spark interpreter: >>> >>> master: spark://<master-name>:7077 >>> spark.submit.deployMode: cluster >>> spark.executor.memory: 16g >>> >>> Any ideas would be appreciated. >>> >>> Thank you >>> >>> Details: >>> Spark version: 2.1.1 >>> Zeppelin version: 0.8.0 (merged at September 2017 version) >>> >> >>