Does this new feature work only for yarn-cluster ?. Or for spark standalone too ?
El mar., 13 de mar. de 2018 18:34, Ruslan Dautkhanov <dautkha...@gmail.com> escribió: > > Zeppelin version: 0.8.0 (merged at September 2017 version) > > https://issues.apache.org/jira/browse/ZEPPELIN-2898 was merged end of > September so not sure if you have that. > > Check out > https://medium.com/@zjffdu/zeppelin-0-8-0-new-features-ea53e8810235 how > to set this up. > > > > -- > Ruslan Dautkhanov > > On Tue, Mar 13, 2018 at 5:24 PM, Jhon Anderson Cardenas Diaz < > jhonderson2...@gmail.com> wrote: > >> Hi zeppelin users ! >> >> I am working with zeppelin pointing to a spark in standalone. I am trying >> to figure out a way to make zeppelin runs the spark driver outside of >> client process that submits the application. >> >> According with the documentation ( >> http://spark.apache.org/docs/2.1.1/spark-standalone.html): >> >> *For standalone clusters, Spark currently supports two deploy modes. >> In client mode, the driver is launched in the same process as the client >> that submits the application. In cluster mode, however, the driver is >> launched from one of the Worker processes inside the cluster, and the >> client process exits as soon as it fulfills its responsibility of >> submitting the application without waiting for the application to finish.* >> >> The problem is that, even when I set the properties for spark-standalone >> cluster and deploy mode in cluster, the driver still run inside zeppelin >> machine (according with spark UI/executors page). These are properties that >> I am setting for the spark interpreter: >> >> master: spark://<master-name>:7077 >> spark.submit.deployMode: cluster >> spark.executor.memory: 16g >> >> Any ideas would be appreciated. >> >> Thank you >> >> Details: >> Spark version: 2.1.1 >> Zeppelin version: 0.8.0 (merged at September 2017 version) >> > >