Hi Piotr, I have been doing the same process as you mentioned so far, now I am migrating the deployment process using AWS CDK and AWS Step Functions, kind of like the CICD process. I added a download step of jar and configs (1, 2, 3 and 4) from S3 using command-runner.jar (AWS Step); it loaded that into one of the Master nodes (out of 3). In the next step when I launched Flink Job it would not find build because Job is launched in some other yarn node.
I was hoping just like *Apache spark *where whatever files we provide in *--file*s are shipped to yarn (s3 to yarn workfirectory), Flink should also have a solution. Thanks, Vijay On Tue, May 25, 2021 at 12:50 AM Piotr Nowojski <pnowoj...@apache.org> wrote: > Hi Vijay, > > I'm not sure if I understand your question correctly. You have jar and > configs (1, 2, 3 and 4) on S3 and you want to start a Flink job using > those? Can you simply download those things (whole directory containing > those) to the machine that will be starting the Flink job? > > Best, Piotrek > > wt., 25 maj 2021 o 07:50 Vijayendra Yadav <contact....@gmail.com> > napisaĆ(a): > >> Hi Team, >> >> I am trying to find a way to ship files from aws s3 for a flink streaming >> job, I am running on AWS EMR. What i need to ship are following: >> 1) application jar >> 2) application property file >> 3) custom flink-conf.yaml >> 4) log4j application specific >> >> Please let me know options. >> >> Thanks, >> Vijay >> >