There is a Pull Request to enable the new streaming sink for Hadoop < 2.7,
so it may become an option in the next release.

Thanks for bearing with us!

Best,
Stephan


On Sat, Sep 22, 2018 at 2:27 PM Paul Lam <paullin3...@gmail.com> wrote:

>
> Hi Stephan!
>
> It's bad that I'm using Hadoop 2.6, so I have to stick to the old
> bucketing sink. I made it by explicitly setting Hadoop conf for the
> bucketing sink in the user code.
>
> Thank you very much!
>
> Best,
> Paul Lam
>
>
> Stephan Ewen <se...@apache.org> 于2018年9月21日周五 下午6:30写道:
>
>> Hi!
>>
>> The old bucketing sink does not work with the Flink file systems, it only
>> works with Hadoop's direct file system support. IIRC it grabs the Flink
>> File System (which creates s3a) to get the Hadoop config etc and then
>> creates the Hadoop File System (s3a again).
>>
>> The new streaming file sink will use Flink Filesystem support, which is
>> important more efficient streaming fault tolerance. S3 support will be part
>> of Flink 1.7
>>
>> Best,
>> Stephan
>>
>>
>> On Fri, Sep 21, 2018 at 10:41 AM Paul Lam <paullin3...@gmail.com> wrote:
>>
>>> Hi Stefan, Stephan,
>>>
>>> Yes, the `hadoop.security.group.mapping` option is explicitly set
>>> to `org.apache.hadoop.security.LdapGroupsMapping`. Guess that was why the
>>> classloader found an unshaded class.
>>>
>>> I don’t have the permission to change the Hadoop cluster configurations
>>> so I modified the `core-default-shaded.xml` and marked the option as final
>>> to solve the problem, after which the class loading exceptions were gone.
>>>
>>> But anther problem came up (likely not related to the previous problem):
>>>
>>> In case of the old bucketing sink (version 1.5.3), it seems that the `
>>> org.apache.hadoop.fs.s3a.S3AFileSystem` is initiated twice before the
>>> task starts running. The first time is called by `
>>> org.apache.flink.fs.s3hadoop.S3FileSystemFactory` and works well, but
>>> the second time is called by bucketing sink itself, and fails to leverage
>>> the `s3.*` parameters like the access key and the secret key.
>>>
>>> The stack traces are as below:
>>>
>>> ```
>>>
>>> com.amazonaws.AmazonClientException: Unable to load AWS credentials from 
>>> any provider in the chain
>>>     at 
>>> com.amazonaws.auth.AWSCredentialsProviderChain.getCredentials(AWSCredentialsProviderChain.java:117)
>>>     at 
>>> com.amazonaws.services.s3.AmazonS3Client.invoke(AmazonS3Client.java:3521)
>>>     at 
>>> com.amazonaws.services.s3.AmazonS3Client.headBucket(AmazonS3Client.java:1031)
>>>     at 
>>> com.amazonaws.services.s3.AmazonS3Client.doesBucketExist(AmazonS3Client.java:994)
>>>     at 
>>> org.apache.hadoop.fs.s3a.S3AFileSystem.initialize(S3AFileSystem.java:297)
>>>     at 
>>> org.apache.flink.streaming.connectors.fs.bucketing.BucketingSink.createHadoopFileSystem(BucketingSink.java:1307)
>>>     at 
>>> org.apache.flink.streaming.connectors.fs.bucketing.BucketingSink.initFileSystem(BucketingSink.java:426)
>>>     at 
>>> org.apache.flink.streaming.connectors.fs.bucketing.BucketingSink.initializeState(BucketingSink.java:370)
>>>     at 
>>> org.apache.flink.streaming.util.functions.StreamingFunctionUtils.tryRestoreFunction(StreamingFunctionUtils.java:178)
>>>     at 
>>> org.apache.flink.streaming.util.functions.StreamingFunctionUtils.restoreFunctionState(StreamingFunctionUtils.java:160)
>>>     at 
>>> org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.initializeState(AbstractUdfStreamOperator.java:96)
>>>     at 
>>> org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:254)
>>>     at 
>>> org.apache.flink.streaming.runtime.tasks.StreamTask.initializeState(StreamTask.java:730)
>>>     at 
>>> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:295)
>>>     at org.apache.flink.runtime.taskmanager.Task.run(Task.java:712)
>>>     at java.lang.Thread.run(Thread.java:748)
>>>
>>> ```
>>>
>>> I haven’t figured out why the s3a filesystem needs to be initiated
>>> twice. And is it a bug that the bucketing sink does not use filesystem
>>> factories to create filesystem?
>>>
>>> Thank you very much!
>>>
>>> Best,
>>> Paul Lam
>>>
>>>
>>> 在 2018年9月20日,23:35,Stephan Ewen <se...@apache.org> 写道:
>>>
>>> Hi!
>>>
>>> A few questions to diagnose/fix this:
>>>
>>>  Do you explicitly configure the "hadoop.security.group.mapping"?
>>>
>>>   - If not, this setting may have leaked in from a Hadoop config in the
>>> classpath. We are fixing this in Flink 1.7, to make this insensitive to
>>> such settings leaking in.
>>>
>>>   - If yes, then please try setting the config variable to
>>> "hadoop.security.group.mapping: 
>>> org.apache.flink.fs.s3hadoop.shaded.org.apache.hadoop.security.LdapGroupsMapping"?
>>>
>>> Please let us know if that works!
>>>
>>>
>>>
>>> On Thu, Sep 20, 2018 at 1:40 PM, Stefan Richter <
>>> s.rich...@data-artisans.com> wrote:
>>> Hi,
>>>
>>> I could not find any open Jira for the problem you describe. Could you
>>> please open one?
>>>
>>> Best,
>>> Stefan
>>>
>>> > Am 19.09.2018 um 09:54 schrieb Paul Lam <paullin3...@gmail.com>:
>>> >
>>> > Hi,
>>> >
>>> > I’m using StreamingFileSink of 1.6.0 to write logs to S3 and encounter
>>> a classloader problem. It seems that there are conflicts in
>>> flink-shaded-hadoop2-uber-1.6.0.jar and flink-s3-fs-hadoop-1.6.0.jar, and
>>> maybe related to class loading orders.
>>> >
>>> > Did anyone meet this problem? Thanks a lot!
>>> >
>>> > The stack traces are as below:
>>> >
>>> > java.io.IOException: java.lang.RuntimeException: class
>>> org.apache.hadoop.security.LdapGroupsMapping not
>>> org.apache.flink.fs.s3hadoop.shaded.org.apache.hadoop.security.GroupMappingServiceProvider
>>> >       at
>>> org.apache.flink.fs.s3hadoop.shaded.org.apache.flink.runtime.fs.hdfs.AbstractFileSystemFactory.create(AbstractFileSystemFactory.java:62)
>>> >       at
>>> org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:395)
>>> >       at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:318)
>>> >       at
>>> org.apache.flink.streaming.api.functions.sink.filesystem.Buckets.<init>(Buckets.java:111)
>>> >       at
>>> org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink$RowFormatBuilder.createBuckets(StreamingFileSink.java:242)
>>> >       at
>>> org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink.initializeState(StreamingFileSink.java:327)
>>> >       at
>>> org.apache.flink.streaming.util.functions.StreamingFunctionUtils.tryRestoreFunction(StreamingFunctionUtils.java:178)
>>> >       at
>>> org.apache.flink.streaming.util.functions.StreamingFunctionUtils.restoreFunctionState(StreamingFunctionUtils.java:160)
>>> >       at
>>> org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.initializeState(AbstractUdfStreamOperator.java:96)
>>> >       at
>>> org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:254)
>>> >       at
>>> org.apache.flink.streaming.runtime.tasks.StreamTask.initializeState(StreamTask.java:738)
>>> >       at
>>> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:289)
>>> >       at org.apache.flink.runtime.taskmanager.Task.run(Task.java:711)
>>> >       at java.lang.Thread.run(Thread.java:748)
>>> > Caused by: java.lang.RuntimeException: java.lang.RuntimeException:
>>> class org.apache.hadoop.security.LdapGroupsMapping
>>> not 
>>> org.apache.flink.fs.s3hadoop.shaded.org.apache.hadoop.security.GroupMappingServiceProvider
>>> >       at
>>> org.apache.flink.fs.s3hadoop.shaded.org.apache.hadoop.conf.Configuration.getClass(Configuration.java:2246)
>>> >       at
>>> org.apache.flink.fs.s3hadoop.shaded.org.apache.hadoop.security.Groups.<init>(Groups.java:108)
>>> >       at
>>> org.apache.flink.fs.s3hadoop.shaded.org.apache.hadoop.security.Groups.<init>(Groups.java:102)
>>> >       at
>>> org.apache.flink.fs.s3hadoop.shaded.org.apache.hadoop.security.Groups.getUserToGroupsMappingService(Groups.java:450)
>>> >       at
>>> org.apache.flink.fs.s3hadoop.shaded.org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:309)
>>> >       at
>>> org.apache.flink.fs.s3hadoop.shaded.org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:276)
>>> >       at
>>> org.apache.flink.fs.s3hadoop.shaded.org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:832)
>>> >       at
>>> org.apache.flink.fs.s3hadoop.shaded.org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:802)
>>> >       at
>>> org.apache.flink.fs.s3hadoop.shaded.org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:675)
>>> >       at
>>> org.apache.flink.fs.s3hadoop.shaded.org.apache.hadoop.fs.s3a.S3AFileSystem.initialize(S3AFileSystem.java:177)
>>> >       at
>>> org.apache.flink.fs.s3hadoop.shaded.org.apache.flink.runtime.fs.hdfs.AbstractFileSystemFactory.create(AbstractFileSystemFactory.java:57)
>>> >       ... 13 more
>>> > Caused by: java.lang.RuntimeException: class
>>> org.apache.hadoop.security.LdapGroupsMapping not
>>> org.apache.flink.fs.s3hadoop.shaded.org.apache.hadoop.security.GroupMappingServiceProvider
>>> >       at
>>> org.apache.flink.fs.s3hadoop.shaded.org.apache.hadoop.conf.Configuration.getClass(Configuration.java:2240)
>>> >       ... 23 more
>>> >
>>> >
>>> > Best,
>>> > Paul Lam
>>> >
>>> >
>>> >
>>>
>>>
>>>
>>>

Reply via email to