> On Nov. 12, 2014, 8:17 a.m., Robert Kanter wrote:
> > core/src/main/java/org/apache/oozie/action/hadoop/JavaActionExecutor.java, 
> > line 1483
> > <https://reviews.apache.org/r/27614/diff/2/?file=757470#file757470line1483>
> >
> >     type should be the second argument.  i.e.
> >     
> >     LOG.debug("class for [{0}] Action: [{1}]", type, function);

Sure will fix it


> On Nov. 12, 2014, 8:17 a.m., Robert Kanter wrote:
> > core/src/main/java/org/apache/oozie/action/hadoop/SparkActionExecutor.java, 
> > line 32
> > <https://reviews.apache.org/r/27614/diff/2/?file=757471#file757471line32>
> >
> >     Can we name this something else?  It's easy to confuse it with 
> > LauncherMapper.CONF_OOZIE_ACTION_MAIN_CLASS

This is not required any more, i will call SparkSubmit directly instead of 
using reflection


> On Nov. 12, 2014, 8:17 a.m., Robert Kanter wrote:
> > core/src/main/java/org/apache/oozie/action/hadoop/SparkActionExecutor.java, 
> > line 34
> > <https://reviews.apache.org/r/27614/diff/2/?file=757471#file757471line34>
> >
> >     Do we actually need to use this?  In my experience, it causes nothing 
> > but trouble...

Yes it is needed due to some conflicts in dependencies


> On Nov. 12, 2014, 8:17 a.m., Robert Kanter wrote:
> > core/src/main/java/org/apache/oozie/action/hadoop/SparkActionExecutor.java, 
> > lines 35-41
> > <https://reviews.apache.org/r/27614/diff/2/?file=757471#file757471line35>
> >
> >     These should have more unique names (like we do in the other actions).  
> > This helps prevent collisions as they go into the actionConf.
> >     
> >     e.g.
> >     oozie.spark.spark
> >     oozie.spark.master
> >     oozie.spark.mode
> >     etc

Sure will do


> On Nov. 12, 2014, 8:17 a.m., Robert Kanter wrote:
> > core/src/main/java/org/apache/oozie/action/hadoop/SparkActionExecutor.java, 
> > line 78
> > <https://reviews.apache.org/r/27614/diff/2/?file=757471#file757471line78>
> >
> >     We have this for DistCp because the classes are different between 
> > different versions of MapReduce and there's also two versions of DistCp.  
> > I'm not familiar enough with Spark, but is there any reason for a user to 
> > use a different Spark class?  
> >     
> >     If not, then you can get rid of this to simplify things.  In fact, 
> > SparkMain should subclass LauncherMain instead of JavaMain, and you can 
> > call SparkSubmit.main(...) directly instead of using reflection.  Look at 
> > HiveMain or SqoopMain for examples.

I thought later somenbody can change the Spark job launcher class , so that we 
can take it from properties like distcp. Anyway i will remove it


> On Nov. 12, 2014, 8:17 a.m., Robert Kanter wrote:
> > core/src/main/java/org/apache/oozie/action/hadoop/SparkActionExecutor.java, 
> > line 85
> > <https://reviews.apache.org/r/27614/diff/2/?file=757471#file757471line85>
> >
> >     Do we actually need to use this property?  In my experience, it causes 
> > nothing but trouble...

Yes it is needed


> On Nov. 12, 2014, 8:17 a.m., Robert Kanter wrote:
> > sharelib/spark/pom.xml, line 41
> > <https://reviews.apache.org/r/27614/diff/2/?file=757475#file757475line41>
> >
> >     whitespace

will fix


> On Nov. 12, 2014, 8:17 a.m., Robert Kanter wrote:
> > sharelib/spark/pom.xml, line 45
> > <https://reviews.apache.org/r/27614/diff/2/?file=757475#file757475line45>
> >
> >     version numbers should go in the root pom

All other actions are using guava 11.0.2 , to run spark action we need 14.0.1 
for the reason i dont want to change in root pom


> On Nov. 12, 2014, 8:17 a.m., Robert Kanter wrote:
> > sharelib/spark/pom.xml, line 51
> > <https://reviews.apache.org/r/27614/diff/2/?file=757475#file757475line51>
> >
> >     versions numbers should go in the root pom

will do


> On Nov. 12, 2014, 8:17 a.m., Robert Kanter wrote:
> > sharelib/spark/src/main/java/org.apache.oozie.action.hadoop/SparkMain.java, 
> > line 29
> > <https://reviews.apache.org/r/27614/diff/2/?file=757476#file757476line29>
> >
> >     See earlier comment where I explain that this should subclass 
> > LauncherMain and other changes

Sure will change


> On Nov. 12, 2014, 8:17 a.m., Robert Kanter wrote:
> > sharelib/spark/src/main/java/org.apache.oozie.action.hadoop/SparkMain.java, 
> > line 100
> > <https://reviews.apache.org/r/27614/diff/2/?file=757476#file757476line100>
> >
> >     Most of the other actions have some additional logic to send 
> > information back to Oozie (e.g. launched child job ids, other stats).  I'm 
> > not sure but I believe Spark jobs have an ID, right?  Is it possible to 
> > return this back to Oozie as the child job?  Look at some of the other 
> > actions to see how we do this.

Yes spark jobs have application id when they are running in spark cluster , 
according to my knowledge i don't think there is way to get the application id 
from the cluster .


- pavan kumar


-----------------------------------------------------------
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/27614/#review60961
-----------------------------------------------------------


On Nov. 11, 2014, 6:35 a.m., pavan kumar kolamuri wrote:
> 
> -----------------------------------------------------------
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/27614/
> -----------------------------------------------------------
> 
> (Updated Nov. 11, 2014, 6:35 a.m.)
> 
> 
> Review request for oozie and shwethags.
> 
> 
> Bugs: OOZIE-1983
>     https://issues.apache.org/jira/browse/OOZIE-1983
> 
> 
> Repository: oozie-git
> 
> 
> Description
> -------
> 
> Add spark action executor in oozie. Spark jobs can be run using oozie 
> 
> 
> Diffs
> -----
> 
>   client/src/main/java/org/apache/oozie/cli/OozieCLI.java 9c2d14b 
>   client/src/main/resources/spark-action-0.1.xsd PRE-CREATION 
>   core/src/main/java/org/apache/oozie/action/hadoop/DistcpActionExecutor.java 
> 42f2965 
>   core/src/main/java/org/apache/oozie/action/hadoop/JavaActionExecutor.java 
> 7349d3f 
>   core/src/main/java/org/apache/oozie/action/hadoop/SparkActionExecutor.java 
> PRE-CREATION 
>   core/src/main/resources/oozie-default.xml 17155a1 
>   pom.xml 1e79186 
>   sharelib/pom.xml aa479a8 
>   sharelib/spark/pom.xml PRE-CREATION 
>   sharelib/spark/src/main/java/org.apache.oozie.action.hadoop/SparkMain.java 
> PRE-CREATION 
>   
> sharelib/spark/src/test/java/org/apache/oozie/action/hadoop/TestSparkActionExecutor.java
>  PRE-CREATION 
>   
> sharelib/spark/src/test/java/org/apache/oozie/action/hadoop/TestSparkMain.java
>  PRE-CREATION 
>   src/main/assemblies/sharelib.xml 4a46b90 
>   webapp/pom.xml 35776c5 
> 
> Diff: https://reviews.apache.org/r/27614/diff/
> 
> 
> Testing
> -------
> 
> Both unit testing and end to end testing done
> 
> 
> Thanks,
> 
> pavan kumar kolamuri
> 
>

Reply via email to