Hi,

It's expected behavior in yarn-client mode. Zeppelin is not closing
SparkContext unless you have restart your interpreter, that's why the job
is in running state.

Hope this helps.

Thanks,
moon


On Fri, Jun 12, 2015 at 11:57 AM A Singh <[email protected]> wrote:

> Environment: Running Zeppelin on Spark 1.3.0 in yarn-client mode
>
> Problem: Spark-SQL query works perfectly and results are returned. Even
> after the results are returned - the Zeppelin job is in the running state
> and stuck at 10%.
>
> Is this  expected behavior - I would have imagined that after the SQL
> completes, the job should have ended and released the cluster resources?
>
> Interpreter Config:
> spark %spark , %pyspark , %sql , %dep  edit  restart  remove
> Properties
> name value
> args
> master yarn-client
> spark.app.name Zeppelin
> spark.cores.max 4
> spark.executor.memory 2048m
> spark.home
> spark.yarn.jar
> zeppelin.dep.localrepo local-repo
> zeppelin.pyspark.python python
> zeppelin.spark.concurrentSQL false
> zeppelin.spark.maxResult 1000
> zeppelin.spark.useHiveContext true
>
>
> Job Browser status:
> USER
> zeppelin
> STATUS
> RUNNING
> LOGS
>  Logs
> PROGRESS
> 10.0%
> DURATION
> 1m:54s
> ACTIONS
>
>
> Thanks,
> Anikate
>

Reply via email to