At the moment, there is no concrete plan to introduce such a feature, because it cannot be guaranteed that you always have a distributed file system available. But we could maybe add it as a tool which we contribute to flink-contrib. Do you wanna take the lead?
Cheers, Till On Wed, Apr 27, 2016 at 10:12 AM, Theofilos Kakantousis <t...@kth.se> wrote: > Hi Till, > > Thank you for the quick reply. Do you think that would be a useful feature > in the future, for the Client to automatically download an job jar from > HDFS, or there are no plans to introduce it? > > Cheers, > Theofilos > > > On 2016-04-27 10:42, Till Rohrmann wrote: > > Hi Theofilos, > > I'm afraid, but that is currently not possible with Flink. Flink expects > the user code jar to be uploaded to its Blob server. That's what the client > does prior to submitting the job. You would have to upload the jar with the > BlobClient manually if you wanted to circumvent the Client. > > Cheers, > Till > On Apr 26, 2016 11:54 PM, "Theofilos Kakantousis" <t...@kth.se> wrote: > >> Hi everyone, >> >> Flink 0.10.1 >> Hadoop 2.4.0 >> >> Fairly new to Flink here, so my question might be simple but couldn't >> find something relevant in the docs. I am implementing a Flink client that >> submits jobs to a Flink Yarn cluster. At the moment I am using the Client >> and PackagedProgram classes which are working fine, however the latter >> expects the job jar to be available locally so that the Client can submit >> it to the Flink Yarn cluster. >> >> Is it possible to use a job jar that is already stored in the HDFS of the >> cluster where Flink is running on, without first copying it locally to >> where the Client is? >> >> Thank you, >> Theofilos >> > >