[ https://issues.apache.org/jira/browse/HIVE-16484?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15980076#comment-15980076 ]
Sahil Takiar commented on HIVE-16484: ------------------------------------- Only related failure is {{TestSparkClient.testRemoteClient}} The issue is what happens if {{SPARK_HOME}} is not set. The {{SparkClientImpl}} has some code to handle the case where {{SPARK_HOME}} is not set; if it isn't set, then it runs {{bin/java org.apache.spark.deploy.SparkSubmit}}. This patch deleted the code in {{SparkClientImpl}} that calls {{bin/java org.apache.spark.deploy.SparkSubmit}} since {{SparkLauncher}} is used for all Spark job submissions. There was only one test that actually invoked that code path ({{TestSparkClient.testRemoteClient}}) {{SparkLauncher}} requires {{SPARK_HOME}} to be set since it calls {{bin/spark-submit}}, it doesn't attempt to call {{org.apache.spark.deploy.SparkSubmit}} if {{SPARK_HOME}} is not present. So we could (1) modify {{SparkLauncher}} to not require {{sparkHome}} to be set, (2) modify this test so that {{SPARK_HOME}} is set, or (3) refactor the code so that it can still directly invoke {{bin/java org.apache.spark.deploy.SparkSubmit}} if {{SPARK_HOME}} isn't set. I'm leaning towards approach 2. [~vanzin] the code to run {{bin/java org.apache.spark.deploy.SparkSubmit}} if {{SPARK_HOME}} isn't set was added in HIVE-8528 - is there a use case for launching Spark jobs without {{SPARK_HOME}} being set, or was it just added for testing? > Investigate SparkLauncher for HoS as alternative to bin/spark-submit > -------------------------------------------------------------------- > > Key: HIVE-16484 > URL: https://issues.apache.org/jira/browse/HIVE-16484 > Project: Hive > Issue Type: Bug > Components: Spark > Reporter: Sahil Takiar > Assignee: Sahil Takiar > Attachments: HIVE-16484.1.patch, HIVE-16484.2.patch, > HIVE-16484.3.patch > > > The {{SparkClientImpl#startDriver}} currently looks for the {{SPARK_HOME}} > directory and invokes the {{bin/spark-submit}} script, which spawns a > separate process to run the Spark application. > {{SparkLauncher}} was added in SPARK-4924 and is a programatic way to launch > Spark applications. > I see a few advantages: > * No need to spawn a separate process to launch a HoS --> lower startup time > * Simplifies the code in {{SparkClientImpl}} --> easier to debug > * {{SparkLauncher#startApplication}} returns a {{SparkAppHandle}} which > contains some useful utilities for querying the state of the Spark job > ** It also allows the launcher to specify a list of job listeners -- This message was sent by Atlassian JIRA (v6.3.15#6346)