Hi I rechecked that I put all my Jars in the Lib folder . I have also noticed that it fails while loading my first Pojo class .
I start the cluster via Yarn using Flink 0.9.1 . Thanks , Mr Hanan Meyer On Thu, Sep 24, 2015 at 6:08 PM, Stephan Ewen <se...@apache.org> wrote: > My first guess would be that you did not put all jars into the lib folder. > > To help us understand this, do you start the cluster manually, or via YARN? > > On Thu, Sep 24, 2015 at 4:59 PM, Hanan Meyer <ha...@scalabill.it> wrote: > > > Hi > > Thanks for the fast response > > I Have tried the walk-around by excluding the Jars from the > > RemoteEnvironment's init line : > > ExecutionEnvironment env = > > ExecutionEnvironment.createRemoteEnvironment(FLINK_URL, FLINK_PORT); > > instrad of : > > ExecutionEnvironment env = > > ExecutionEnvironment.createRemoteEnvironment(FLINK_URL, FLINK_PORT, list > > of Jars ......); > > I copied the jars to the Flink's Lib folder and when I submit my job I > get > > the following exception which is caused because > > Flink can't find my Jars and Types : > > org.apache.flink.client.program.ProgramInvocationException: The program > > execution failed: Cannot initialize task 'CHAIN DataSource > > (at createInput(ExecutionEnvironment.java:502) > > (org.apache.flink.api.java.io.AvroInputFormat)) -> > > Filter (Filter at generateCsv(FlinkCSVProducer.java:51)) -> FlatMap > > (FlatMap at generateCsv(FlinkCSVProducer.java:78))': > > Deserializing the InputFormat (File Input > > (hdfs://localhost:9000/data/kpi/38fbbdef-d822-4e13-9031-faff907469df)) > > failed: > > Could not read the user code wrapper: com.scalabill.it.pa.event.Event > > at org.apache.flink.client.program.Client.run(Client.java:413) > > at org.apache.flink.client.program.Client.run(Client.java:356) > > at org.apache.flink.client.program.Client.run(Client.java:349) > > at > > > > > org.apache.flink.client.RemoteExecutor.executePlanWithJars(RemoteExecutor.java:89) > > at > > > org.apache.flink.client.RemoteExecutor.executePlan(RemoteExecutor.java:82) > > at > > > > > org.apache.flink.api.java.RemoteEnvironment.execute(RemoteEnvironment.java:71) > > at > > > > > org.apache.flink.api.java.ExecutionEnvironment.execute(ExecutionEnvironment.java:789) > > at org.apache.flink.api.java.DataSet.count(DataSet.java:391) > > at > > > > > com.scalabill.it.pa.core.FlinkCSVProducer.generateCsv(FlinkCSVProducer.java:70) > > at > > > > > com.scalabill.it.pa.core.FlinkDriver.generateChannelsCSVsforThisBackendServer(FlinkDriver.java:94) > > Have I been doing the walk-around currently ? > > Can you try to reproduce it in your environment ? > > Thanks for your attention! > > Hanan Meyer > > > > On Thu, Sep 24, 2015 at 4:58 PM, Till Rohrmann <till.rohrm...@gmail.com> > > wrote: > > > > > Hi Hanan, > > > > > > you're right that currently every time you submit a job to the Flink > > > cluster, all user code jars are uploaded and overwrite possibly > existing > > > files. This is not really necessary if they don't change. Maybe we > should > > > add a check that already existing files on the JobManager are not > > uploaded > > > again by the JobClient. This should improve the performance for your > use > > > case. > > > > > > The corresponding JIRA issue is > > > https://issues.apache.org/jira/browse/FLINK-2760. > > > > > > Cheers, > > > Till > > > > > > On Thu, Sep 24, 2015 at 1:31 PM, Hanan Meyer <ha...@scalabill.it> > wrote: > > > > > > > Hello All > > > > > > > > I use Flink in order to filter data from Hdfs and write it back as > CSV. > > > > > > > > I keep getting the "Checking and uploading JAR files" on every > DataSet > > > > filtering action or > > > > executionEnvironment execution. > > > > > > > > I use ExecutionEnvironment.createRemoteEnvironment(ip+jars..) > because I > > > > launch Flink from > > > > a J2EE Aplication Server . > > > > > > > > The Jars serialization and transportation takes a huge part of the > > > > execution time . > > > > Is there a way to force Flink to pass the Jars only once? > > > > > > > > Please advise > > > > > > > > Thanks, > > > > > > > > Hanan Meyer > > > > > > > > > >