[jira] [Assigned] (NIFI-4946) nifi-spark-bundle : Adding support for pyfiles, file, jars options
[ https://issues.apache.org/jira/browse/NIFI-4946?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Matt Burgess reassigned NIFI-4946: -- Assignee: (was: Matt Burgess) > nifi-spark-bundle : Adding support for pyfiles, file, jars options > -- > > Key: NIFI-4946 > URL: https://issues.apache.org/jira/browse/NIFI-4946 > Project: Apache NiFi > Issue Type: New Feature > Components: Extensions > Environment: Ubuntu 16.04, IntelliJ >Reporter: Mageswaran >Priority: Major > Attachments: nifi-spark-options.png, nifi-spark.png > > Time Spent: 20m > Remaining Estimate: 0h > > Adding support for submitting PySpark based Sparks jobs (which is normally > structured as modules) over Livy on existing "ExecuteSparkInteractive" > processor. > This is done by reading file paths for pyfiles and file and an option from > user whether the processor should trigger a batch job or not. > [https://livy.incubator.apache.org/docs/latest/rest-api.html] > *Current Work flow Logic ( [https://github.com/apache/nifi/pull/2521 > )|https://github.com/apache/nifi/pull/2521]* > * Check whether the processor has to handle code or submit a Spark job > * Read incoming flow file > ** If batch == true > *** If flow file matches Livy `batches` JSON response through `wait` loop > Wait for Status Check Interval > Read the state > If state is `running` route it to `wait` or if it is `success` or > `dead` route it accordingly > *** Else > Ignore the flow file > Trigger the Spark job over Livy `batches` endpoint > Read the state of the submitted job > If state is `running` route it to `wait` or if it is `success` or > `dead` route it accordingly > ** Else: > *** Existing Logic to handle `Code` > > !nifi-spark-options.png! > !nifi-spark.png! > > Thanks. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (NIFI-4946) nifi-spark-bundle : Adding support for pyfiles, file, jars options
[ https://issues.apache.org/jira/browse/NIFI-4946?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Matt Burgess reassigned NIFI-4946: -- Assignee: Matt Burgess > nifi-spark-bundle : Adding support for pyfiles, file, jars options > -- > > Key: NIFI-4946 > URL: https://issues.apache.org/jira/browse/NIFI-4946 > Project: Apache NiFi > Issue Type: New Feature > Components: Extensions > Environment: Ubuntu 16.04, IntelliJ >Reporter: Mageswaran >Assignee: Matt Burgess >Priority: Major > Attachments: nifi-spark-options.png, nifi-spark.png > > Time Spent: 20m > Remaining Estimate: 0h > > Adding support for submitting PySpark based Sparks jobs (which is normally > structured as modules) over Livy on existing "ExecuteSparkInteractive" > processor. > This is done by reading file paths for pyfiles and file and an option from > user whether the processor should trigger a batch job or not. > [https://livy.incubator.apache.org/docs/latest/rest-api.html] > *Current Work flow Logic ( [https://github.com/apache/nifi/pull/2521 > )|https://github.com/apache/nifi/pull/2521]* > * Check whether the processor has to handle code or submit a Spark job > * Read incoming flow file > ** If batch == true > *** If flow file matches Livy `batches` JSON response through `wait` loop > Wait for Status Check Interval > Read the state > If state is `running` route it to `wait` or if it is `success` or > `dead` route it accordingly > *** Else > Ignore the flow file > Trigger the Spark job over Livy `batches` endpoint > Read the state of the submitted job > If state is `running` route it to `wait` or if it is `success` or > `dead` route it accordingly > ** Else: > *** Existing Logic to handle `Code` > > !nifi-spark-options.png! > !nifi-spark.png! > > Thanks. -- This message was sent by Atlassian Jira (v8.3.4#803005)