I do have similar question – as I have the Execute script using python to run code and it produces the output file (/tmp/test.xml) but not sure how to use that file to next processor without using additional flow file using GetFile processor to get a file produced by python execute script. I am very new to NiFi.
Following is need: 1. READ CSV file from HDFS 2. Execute python script – reads CSV file and produces XML output file – example /tmp/test.xml . 3. I need to process the /tmp/test.xml file using SplitXML processor 4. Put these into HDFS Thanks, Ram From: James Wing [mailto:[email protected]] Sent: Monday, August 29, 2016 12:47 AM To: [email protected] Subject: Re: ExecuteScript Processor - Control Flow Koustav, How are you running the Sqoop job? Can you share some code? Python is sequential by default, but your Sqoop job might run asynchronously. I believe the answer depends on your code (or library) not only starting the Sqoop job, but polling for it's status until it is complete. Thanks, James On Sun, Aug 28, 2016 at 8:24 PM, koustav choudhuri <[email protected]<mailto:[email protected]>> wrote: Hi All I have a python script running on a Nifi Server , whin in turn calls a Sqoop job on a different Server . The next step in the script is to use the flow file from the previous processor to continue to the next processor . So the python script is like : 1. call the sqoop job on server 2 2. get the flow file from the session and continue Question : Will step 2 wait till Step1 completes ? Or , As soon as the Sqoop job gets initiated through Step 1 , Step 2 Executes irrespective of whether Step 1 completes or not . Could be a dumb question , still asking . ************************************************************************* This e-mail may contain confidential or privileged information. If you are not the intended recipient, please notify the sender immediately and then delete it. TIAA *************************************************************************
