This is not by design.  Can you please file a JIRA?

On Wed, Nov 19, 2014 at 9:19 AM, Yana Kadiyska <yana.kadiy...@gmail.com>
wrote:

> Hi all, I am running HiveThriftServer2 and noticed that the process stays
> up even though there is no driver connected to the Spark master.
>
> I started the server via sbin/start-thriftserver and my namenodes are
> currently not operational. I can see from the log that there was an error
> on startup:
>
> 14/11/19 16:32:58 ERROR HiveThriftServer2: Error starting HiveThriftServer2
>
> and that the driver shut down as expected:
>
> 14/11/19 16:32:59 INFO SparkUI: Stopped Spark web UI at http://myip:4040
> 14/11/19 16:32:59 INFO DAGScheduler: Stopping DAGScheduler
> 14/11/19 16:32:59 INFO SparkDeploySchedulerBackend: Shutting down all 
> executors
> 14/11/19 16:32:59 INFO SparkDeploySchedulerBackend: Asking each executor to 
> shut down
> 14/11/19 16:33:00 INFO MapOutputTrackerMasterActor: MapOutputTrackerActor 
> stopped!
> 14/11/19 16:33:00 INFO MemoryStore: MemoryStore cleared
> 14/11/19 16:33:00 INFO BlockManager: BlockManager stopped
> 14/11/19 16:33:00 INFO BlockManagerMaster: BlockManagerMaster stopped
> 14/11/19 16:33:00 INFO SparkContext: Successfully stopped SparkContext
>
> ​
> However, when I try to run start-thriftserver.sh again I see an error
> message that the process is already running and indeed there is a process
> with that PID:
>
> root     32334     1  0 16:32 ?        00:00:00 /usr/local/bin/java 
> org.apache.spark.deploy.SparkSubmitDriverBootstrapper --class 
> org.apache.spark.sql.hive.thriftserver.HiveThriftServer2 --master 
> spark://myip:7077 --conf -spark.executor.extraJavaOptions=-verbose:gc 
> -XX:-PrintGCDetails -XX:+PrintGCTimeStamps spark-internal --hiveconf 
> hive.root.logger=INFO,console
>
> ​
>  Is this a bug or design decision -- I am upgrading from Shark and we had
> scripts that monitor the driver and restart on failure. Here it seems that
> we would not be able to restart even though the driver died?
>

Reply via email to