Hi,
the issue is a bit more complicated and involves the Beam sink API and the
Flink runner.

I'll have to get a bit into how Beam sinks work. The base class for sinks
is Sink (TextIO.write gets translated to Write.to(new TextSink())).
Write.to normally gets translated to three ParDo operations that cooperate
to do the writing:

 - "Initialize": this does initial initialization of the Sink, this is run
only once, per sink, non-parallel.

 - "WriteBundles": this gets an initialized sink on a side-input and the
values to write on the main input. This runs in parallel, so for Flink, if
you set parallelism=6 you'll get 6 parallel instances of this operation at
runtime. This operation forwards information about where it writes to
downstream. This does not write to the final file location but an
intermediate staging location.

 - "Finalize": This gets the initialized sink on the main-input and and the
information about written files from "WriteBundles" as a side-input. This
also only runs once, non-parallel. Here we're writing the intermediate
files to a final location based on the sharding template.

The problem is that Write.to() and TextSink, as well as all other sinks,
are not aware of the number of shards. If you set "withoutSharding()" this
will set the shard template to "" (empty string) and the number of shards
to 1. "WriteBundles", however is not aware of this and will write 6
intermediate files if you set parallelism=6. In "Finalize" we will copy an
intermediate file to the same final location 6 times based on the sharding
template. The end result is that you only get one of the six result shards.

The reason why this does only occur in the Flink runner is that all other
runners have special overrides for TextIO.Write and AvroIO.Write that kick
in if sharding control is required. So, for the time being this is a Flink
runner bug and we might have to introduce special overrides as well until
this is solved in the general case.

Cheers,
Aljoscha

On Wed, 1 Jun 2016 at 07:37 Jean-Baptiste Onofré <[email protected]> wrote:

> Yes, just tested, it happens only with the flink runner.
>
> Agree to create a Jira.
>
> Regards
> JB
>
> On 06/01/2016 03:41 AM, Davor Bonaci wrote:
> > This will be a runner-specific issue. It would be the best to file a
> > JIRA issue for this.
> >
> > On Tue, May 31, 2016 at 9:46 AM, Jean-Baptiste Onofré <[email protected]
> > <mailto:[email protected]>> wrote:
> >
> >     Hi Pawel,
> >
> >     does it happen only with the Flink runner ? I bet it happens with
> >     any runner.
> >
> >     Let me take a look.
> >
> >     Regards
> >     JB
> >
> >     On 05/30/2016 01:38 AM, Pawel Szczur wrote:
> >
> >         Hi,
> >
> >         I'm running a pipeline with Flink backend, Beam bleeding edge,
> >         Oracle
> >         Java 1.8, maven 3.3.3, linux64.
> >
> >         The pipeline is run with --parallelism=6.
> >
> >         Adding .withoutSharding()causes a TextIO sink to write only one
> >         of the
> >         shards.
> >
> >         Example use:
> >
>  data.apply(TextIO.Write.named("write-debug-csv").to("/tmp/some-stats"));
> >         vs.
> >
>  
> data.apply(TextIO.Write.named("write-debug-csv").to("/tmp/some-stats")*.withoutSharding()*);
> >
> >         Result:
> >         Only part of data is written to file. After comparing to sharded
> >         output,
> >         it seems to be just one of shard files.
> >
> >         Cheers,
> >         Pawel
> >
> >
> >     --
> >     Jean-Baptiste Onofré
> >     [email protected] <mailto:[email protected]>
> >     http://blog.nanthrax.net
> >     Talend - http://www.talend.com
> >
> >
>
> --
> Jean-Baptiste Onofré
> [email protected]
> http://blog.nanthrax.net
> Talend - http://www.talend.com
>

Reply via email to