[ 
https://issues.apache.org/jira/browse/OOZIE-3228?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16451788#comment-16451788
 ] 

Tang Yan commented on OOZIE-3228:
---------------------------------

[~andras.piros] I checked the code difference of oozie spark action part 
(sharelib/spark/src/main/java/org/apache/oozie/action/hadoop) between the two 
versions I used, the only new commit in the new oozie version is  OOZIE-3112.

 For my job, I just used the Spark Copy example in Hue.

-> *workflow.xml*

<workflow-app name="Spark" xmlns="uri:oozie:workflow:0.5">
 <start to="spark-d909"/>
 <kill name="Kill">
 <message>Action failed, error 
message[${wf:errorMessage(wf:lastErrorNode())}]</message>
 </kill>
 <action name="spark-d909">
 <spark xmlns="uri:oozie:spark-action:0.1">
 <job-tracker>${jobTracker}</job-tracker>
 <name-node>${nameNode}</name-node>
 <master>yarn</master>
 <mode>client</mode>
 <name>MySpark</name>
 <class>org.apache.oozie.example.SparkFileCopy</class>
 <jar>oozie-examples.jar</jar>
 <arg>${input}</arg>
 <arg>${output}</arg>
 </spark>
 <ok to="End"/>
 <error to="Kill"/>
 </action>
 <end name="End"/>
</workflow-app>

-> */etc/spark/conf/spark-defaults.conf* 

spark.authenticate=false
spark.authenticate.enableSaslEncryption=false
spark.dynamicAllocation.enabled=true
spark.dynamicAllocation.executorIdleTimeout=60
spark.dynamicAllocation.minExecutors=0
spark.dynamicAllocation.schedulerBacklogTimeout=1
spark.eventLog.enabled=true
spark.serializer=org.apache.spark.serializer.KryoSerializer
spark.shuffle.encryption.enabled=false
spark.shuffle.service.enabled=true
spark.shuffle.service.port=7337
spark.ui.enabled=true
spark.ui.killEnabled=true
{color:#FF0000}spark.executor.extraClassPath=/etc/hbase/conf:/etc/hive/conf{color}
{color:#FF0000}spark.driver.extraClassPath=/etc/hbase/conf:/etc/hive/conf{color}
spark.master=yarn-client

 

But in the Oozie Spark job log --conf output, there's no defined extraClassPath.

*--conf spark.executor.extraClassPath=$PWD/** 

 *--conf spark.driver.extraClassPath=$PWD/**

While in the last oozie version, it can work well as below logs.

*--conf spark.executor.extraClassPath=/etc/hbase/conf:/etc/hive/conf:$PWD/** 

 *--conf spark.driver.extraClassPath=/etc/hbase/conf:/etc/hive/conf:$PWD/**

> Oozie Spark Action - the spark job can't load the properties in 
> spark-defaults.conf.
> ------------------------------------------------------------------------------------
>
>                 Key: OOZIE-3228
>                 URL: https://issues.apache.org/jira/browse/OOZIE-3228
>             Project: Oozie
>          Issue Type: Bug
>          Components: action
>    Affects Versions: 4.3.1
>            Reporter: Tang Yan
>            Priority: Major
>
> When I create a oozie workflow to launch a spark action, the spark job can't 
> load the configured properties in spark-defaults.conf. I've configured each 
> Nodemanager as the spark gateway role, so the spark-defaults.conf is 
> generated in /etc/spark/conf/ on each worker node.
> in spark-defaults.conf some configuration I've set into.
> spark.executor.extraClassPath=/etc/hbase/conf:/etc/hive/conf
> spark.driver.extraClassPath=/etc/hbase/conf:/etc/hive/conf
> But in the Oozie spark job, they're not loaded automatically.
> --conf spark.executor.extraClassPath=$PWD/*
> --conf spark.driver.extraClassPath=$PWD/*
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to