[
https://issues.apache.org/jira/browse/OOZIE-3228?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16451788#comment-16451788
]
Tang Yan commented on OOZIE-3228:
---------------------------------
[~andras.piros] I checked the code difference of oozie spark action part
(sharelib/spark/src/main/java/org/apache/oozie/action/hadoop) between the two
versions I used, the only new commit in the new oozie version is OOZIE-3112.
For my job, I just used the Spark Copy example in Hue.
-> *workflow.xml*
<workflow-app name="Spark" xmlns="uri:oozie:workflow:0.5">
<start to="spark-d909"/>
<kill name="Kill">
<message>Action failed, error
message[${wf:errorMessage(wf:lastErrorNode())}]</message>
</kill>
<action name="spark-d909">
<spark xmlns="uri:oozie:spark-action:0.1">
<job-tracker>${jobTracker}</job-tracker>
<name-node>${nameNode}</name-node>
<master>yarn</master>
<mode>client</mode>
<name>MySpark</name>
<class>org.apache.oozie.example.SparkFileCopy</class>
<jar>oozie-examples.jar</jar>
<arg>${input}</arg>
<arg>${output}</arg>
</spark>
<ok to="End"/>
<error to="Kill"/>
</action>
<end name="End"/>
</workflow-app>
-> */etc/spark/conf/spark-defaults.conf*
spark.authenticate=false
spark.authenticate.enableSaslEncryption=false
spark.dynamicAllocation.enabled=true
spark.dynamicAllocation.executorIdleTimeout=60
spark.dynamicAllocation.minExecutors=0
spark.dynamicAllocation.schedulerBacklogTimeout=1
spark.eventLog.enabled=true
spark.serializer=org.apache.spark.serializer.KryoSerializer
spark.shuffle.encryption.enabled=false
spark.shuffle.service.enabled=true
spark.shuffle.service.port=7337
spark.ui.enabled=true
spark.ui.killEnabled=true
{color:#FF0000}spark.executor.extraClassPath=/etc/hbase/conf:/etc/hive/conf{color}
{color:#FF0000}spark.driver.extraClassPath=/etc/hbase/conf:/etc/hive/conf{color}
spark.master=yarn-client
But in the Oozie Spark job log --conf output, there's no defined extraClassPath.
*--conf spark.executor.extraClassPath=$PWD/**
*--conf spark.driver.extraClassPath=$PWD/**
While in the last oozie version, it can work well as below logs.
*--conf spark.executor.extraClassPath=/etc/hbase/conf:/etc/hive/conf:$PWD/**
*--conf spark.driver.extraClassPath=/etc/hbase/conf:/etc/hive/conf:$PWD/**
> Oozie Spark Action - the spark job can't load the properties in
> spark-defaults.conf.
> ------------------------------------------------------------------------------------
>
> Key: OOZIE-3228
> URL: https://issues.apache.org/jira/browse/OOZIE-3228
> Project: Oozie
> Issue Type: Bug
> Components: action
> Affects Versions: 4.3.1
> Reporter: Tang Yan
> Priority: Major
>
> When I create a oozie workflow to launch a spark action, the spark job can't
> load the configured properties in spark-defaults.conf. I've configured each
> Nodemanager as the spark gateway role, so the spark-defaults.conf is
> generated in /etc/spark/conf/ on each worker node.
> in spark-defaults.conf some configuration I've set into.
> spark.executor.extraClassPath=/etc/hbase/conf:/etc/hive/conf
> spark.driver.extraClassPath=/etc/hbase/conf:/etc/hive/conf
> But in the Oozie spark job, they're not loaded automatically.
> --conf spark.executor.extraClassPath=$PWD/*
> --conf spark.driver.extraClassPath=$PWD/*
>
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)