You mean putting the Flink-native S3 filesystem in the user jar or Hadoop in 
the user jar. The former wouldn't work, I think, because the FileSystems are 
being initialised before the user-jar is loaded. The latter might work but only 
if you don't have Hadoop in the classpath, i.e. not on YARN and only on a 
Hadoop-free cluster. Maybe...

> On 23. Feb 2018, at 13:32, Jamie Grier <jgr...@lyft.com> wrote:
> 
> Thanks, Aljoscha :)
> 
> So is it possible to continue to use the new "native' fllesystems along
> with the BucketingSink by including the Hadoop dependencies only in the
> user's uber jar? Or is that asking for trouble?  Has anyone tried that
> successfully?
> 
> -Jamie
> 
> 
> On Fri, Feb 23, 2018 at 12:39 AM, Aljoscha Krettek <aljos...@apache.org>
> wrote:
> 
>> Hi,
>> 
>> I'm afraid not, since the BucketingSink uses the Hadoop FileSystem
>> directly and not the Flink FileSystem abstraction. The flink-s3-fs-*
>> modules only provide Flink FileSystems.
>> 
>> One of the goals for 1.6 is to provide a BucketingSink that uses the Flink
>> FileSystem and also works well with eventually consistent file systems.
>> 
>> --
>> Aljoscha
>> 
>>> On 23. Feb 2018, at 06:31, Jamie Grier <jgr...@lyft.com> wrote:
>>> 
>>> Is the `flink-connector-filesystem` connector supposed to work with the
>>> latest hadoop-free Flink releases, say along with the
>> `flink-s3-fs-presto`
>>> filesystem implementation?
>>> 
>>> -Jamie
>> 
>> 

Reply via email to