[ 
https://issues.apache.org/jira/browse/FLINK-8794?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16383209#comment-16383209
 ] 

yanxiaobin commented on FLINK-8794:
-----------------------------------

hi, [~aljoscha] Thank you for your reply!
 
There are the following points:

1.What I described above is that there will be such a situation when there is 
no failure in this job.

2.This happens when a job has a failure(because one of the taskmanager nodes 
downtime) and recovery. Fault tolerance of a node in distributed computing is 
necessary.Because this is a problem in this case.

3.When recovery, the previous in-progress and pending files are not 
cleared,this causes the downstream processor to read excess dirty data.

5.I think we should first place data in computing nodes' local files, then 
upload them to the distributed file system after the local file is written 
completely, for example, S3, HDFS.

We are blocked of the problem at the moment. and because of this problem, we 
can't use this job.

 

> When using BucketingSink, it happens that one of the files is always in the 
> [.in-progress] state
> ------------------------------------------------------------------------------------------------
>
>                 Key: FLINK-8794
>                 URL: https://issues.apache.org/jira/browse/FLINK-8794
>             Project: Flink
>          Issue Type: Bug
>          Components: filesystem-connector
>    Affects Versions: 1.4.0, 1.4.1
>            Reporter: yanxiaobin
>            Priority: Major
>
> When using BucketingSink, it happens that one of the files is always in the 
> [.in-progress] state. And this state has never changed after that.  The 
> underlying use of S3 as storage.
>  
> {code:java}
> // code placeholder
> {code}
> 2018-02-28 11:58:42  147341619 {color:#d04437}_part-28-0.in-progress{color}
> 2018-02-28 12:06:27  147315059 part-0-0
> 2018-02-28 12:06:27  147462359 part-1-0
> 2018-02-28 12:06:27  147316006 part-10-0
> 2018-02-28 12:06:28  147349854 part-100-0
> 2018-02-28 12:06:27  147421625 part-101-0
> 2018-02-28 12:06:27  147443830 part-102-0
> 2018-02-28 12:06:27  147372801 part-103-0
> 2018-02-28 12:06:27  147343670 part-104-0
> ......



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to