On Thu, Aug 3, 2017 at 6:08 PM, Eugene Kirpichov <kirpic...@google.com.invalid> wrote: > https://github.com/apache/beam/pull/3649 has landed. The main contribution > of this PR is deprecating PTransform.getDefaultOutputCoder(). > > Next steps are to get rid of all setCoder() calls in the SDK, and deprecate > setCoder(). > Nearly all setCoder() calls (perhaps simply all?) I found are on the output > of mapping transforms, such as ParDo, Map/FlatMapElements, WithKeys. > I think we should simply make these transforms optionally configurable with > an output coder: e.g. input.apply(ParDo.of(new > SomeFn<>()).withOutputCoder(SomeCoder.of())) > For multi-output ParDo this is a little more complex API-wise, but doable > too. > > (another minor next step is to say in PTransform Style Guide that the > transform must set a coder on all its outputs) > > Sounds reasonable?
+1 I'd like to do this in a way that lowers the burden for all PTransform authors. Can't think of a better way than a special subclass of PTransform that has the setters that one would subclass... > On Thu, Aug 3, 2017 at 5:34 AM Lukasz Cwik <lc...@google.com.invalid> wrote: > >> I'm for (1) and am not sure about the feasibility of (2) without having an >> escape hatch that allows a pipeline author to specify a coder to handle >> their special case. >> >> On Tue, Aug 1, 2017 at 2:15 PM, Reuven Lax <re...@google.com.invalid> >> wrote: >> >> > One interesting wrinkle: I'm about to propose a set of semantics for >> > snapshotting/in-place updating pipelines. Part of this proposal is the >> > ability to write pipelines to "upgrade" snapshots to make them compatible >> > with new graphs. This relies on the ability to have two separate coders >> for >> > the same type - the old coder and the new coder - in order to handle the >> > case where the user has changed coders in the new pipeline. >> > >> > On Tue, Aug 1, 2017 at 2:12 PM, Robert Bradshaw >> > <rober...@google.com.invalid >> > > wrote: >> > >> > > There are two concerns in this thread: >> > > >> > > (1) Getting rid of PCollection.setCoder(). Everyone seems in favor of >> > this >> > > (right?) >> > > >> > > (2) Deprecating specifying Coders in favor of specifying >> TypeDescriptors. >> > > I'm generally in favor, but it's unclear how far we can push this >> > through. >> > > >> > > Let's at least do (1), and separately state a preference for (2), >> seeing >> > > how fare we can push it. >> > > >> > > On Thu, Jul 27, 2017 at 9:13 PM, Kenneth Knowles >> <k...@google.com.invalid >> > > >> > > wrote: >> > > >> > > > Another thought on this: setting a custom coder to support a special >> > data >> > > > distribution is likely often a property of the input to the pipeline. >> > So >> > > > setting a coder during pipeline construction - more generally, when >> > > writing >> > > > a composite transform for reuse - you might not actually have the >> > needed >> > > > information. But setting up a special indicator type descriptor lets >> > your >> > > > users map that type descriptor to a coder that works well for their >> > data. >> > > > >> > > > But Robert's example of RawUnionValue seems like a deal breaker for >> all >> > > > approaches. It really requires .getCoder() during expand() and >> > explicitly >> > > > building coders encoding information that is cumbersome to get into a >> > > > TypeDescriptor. While making up new type languages is a comfortable >> > > > activity for me :-) I don't think we should head down that path, for >> > our >> > > > users' sake. So I'll stop hoping we can eliminate this pain point for >> > > now. >> > > > >> > > > Kenn >> > > > >> > > > On Thu, Jul 27, 2017 at 8:48 PM, Kenneth Knowles <k...@google.com> >> > wrote: >> > > > >> > > > > On Thu, Jul 27, 2017 at 11:18 AM, Thomas Groh >> > <tg...@google.com.invalid >> > > > >> > > > > wrote: >> > > > > >> > > > >> introduce a >> > > > >> new, specialized type to represent the restricted >> > > > >> (alternatively-distributed?) data. The TypeDescriptor for this >> type >> > > can >> > > > >> map >> > > > >> to the specialized coder, without having to perform a significant >> > > degree >> > > > >> of >> > > > >> potentially wasted encoding work, plus it includes the assumptions >> > > that >> > > > >> are >> > > > >> being made about the distribution of data. >> > > > >> >> > > > > >> > > > > This is a very cool idea, in theory :-) >> > > > > >> > > > > For complex types with a few allocations involved and/or nontrivial >> > > > > deserialization, or when a pipeline does a lot of real work, I >> think >> > > the >> > > > > wrapper cost won't be perceptible. >> > > > > >> > > > > But for more primitive types in pipelines that don't really do >> much >> > > > > computation but just move data around, I think it could matter. >> > > Certainly >> > > > > there are languages with constructs to allow type wrappers at zero >> > cost >> > > > > (Haskell's `newtype`). >> > > > > >> > > > > This is all just speculation until we measure, like most of this >> > > thread. >> > > > > >> > > > > Kenn >> > > > > >> > > > > >> > > > >> > On Thu, Jul 27, 2017 at 11:00 AM, Thomas Groh >> > > > <tg...@google.com.invalid >> > > > >> > >> > > > >> > wrote: >> > > > >> > >> > > > >> > > +1 on getting rid of setCoder; just from a Java SDK >> perspective, >> > > my >> > > > >> ideal >> > > > >> > > world contains PCollections which don't have a user-visible >> way >> > to >> > > > >> mutate >> > > > >> > > them. >> > > > >> > > >> > > > >> > > My preference would be to use TypeDescriptors everywhere >> within >> > > > >> Pipeline >> > > > >> > > construction (where possible), and utilize the CoderRegistry >> > > > >> everywhere >> > > > >> > to >> > > > >> > > actually extract the appropriate type. The unfortunate >> > difficulty >> > > of >> > > > >> > having >> > > > >> > > to encode a union type and the lack of variable-length >> generics >> > > does >> > > > >> > > complicate that. We could consider some way of constructing >> > coders >> > > > in >> > > > >> the >> > > > >> > > registry from a collection of type descriptors (which should >> be >> > > > >> > accessible >> > > > >> > > from the point the union-type is being constructed), e.g. >> > > something >> > > > >> like >> > > > >> > > `getCoder(TypeDescriptor output, TypeDescriptor... >> components)` >> > - >> > > > that >> > > > >> > does >> > > > >> > > only permit a single flat level (but since this is being >> invoked >> > > by >> > > > >> the >> > > > >> > SDK >> > > > >> > > during construction it could also pass Coder...). >> > > > >> > > >> > > > >> > > >> > > > >> > > >> > > > >> > > On Thu, Jul 27, 2017 at 10:22 AM, Robert Bradshaw < >> > > > >> > > rober...@google.com.invalid> wrote: >> > > > >> > > >> > > > >> > > > On Thu, Jul 27, 2017 at 10:04 AM, Kenneth Knowles >> > > > >> > > > <k...@google.com.invalid> wrote: >> > > > >> > > > > On Thu, Jul 27, 2017 at 2:22 AM, Lukasz Cwik >> > > > >> > <lc...@google.com.invalid >> > > > >> > > > >> > > > >> > > > > wrote: >> > > > >> > > > >> >> > > > >> > > > >> Ken/Robert, I believe users will want the ability to set >> > the >> > > > >> output >> > > > >> > > > coder >> > > > >> > > > >> because coders may have intrinsic properties where the >> type >> > > > isn't >> > > > >> > > enough >> > > > >> > > > >> information to fully specify what I want as a user. Some >> > > cases >> > > > I >> > > > >> can >> > > > >> > > see >> > > > >> > > > >> are: >> > > > >> > > > >> 1) I have a cheap and fast non-deterministic coder but a >> > > > >> different >> > > > >> > > > slower >> > > > >> > > > >> coder when I want to use it as the key to a GBK, For >> > example >> > > > >> with a >> > > > >> > > set >> > > > >> > > > >> coder, it would need to consistently order the values of >> > the >> > > > set >> > > > >> > when >> > > > >> > > > used >> > > > >> > > > >> as the key. >> > > > >> > > > >> 2) I know a property of the data which allows me to have >> a >> > > > >> cheaper >> > > > >> > > > >> encoding. Imagine I know that all the strings have a >> common >> > > > >> prefix >> > > > >> > or >> > > > >> > > > >> integers that are in a certain range, or that a matrix is >> > > > >> > > sparse/dense. >> > > > >> > > > Not >> > > > >> > > > >> all PCollections of strings / integers / matrices in the >> > > > pipeline >> > > > >> > will >> > > > >> > > > have >> > > > >> > > > >> this property, just some. >> > > > >> > > > >> 3) Sorting comes up occasionally, traditionally in Google >> > > this >> > > > >> was >> > > > >> > > done >> > > > >> > > > by >> > > > >> > > > >> sorting the encoded version of the object >> lexicographically >> > > > >> during a >> > > > >> > > > GBK. >> > > > >> > > > >> There are good lexicographical byte representations for >> > ASCII >> > > > >> > strings, >> > > > >> > > > >> integers, and for some IEEE number representations which >> > > could >> > > > be >> > > > >> > done >> > > > >> > > > by >> > > > >> > > > >> the use of a special coder. >> > > > >> > > > >> >> > > > >> > > > > >> > > > >> > > > > Items (1) and (3) do not require special knowledge from >> the >> > > > user. >> > > > >> > They >> > > > >> > > > are >> > > > >> > > > > easily observed properties of a pipeline. My proposal >> > included >> > > > >> full >> > > > >> > > > > automation for both. The suggestion is new methods >> > > > >> > > > > .getDeterministicCoder(TypeDescriptor) and >> > > > >> > > > > .getLexicographicCoder(TypeDescriptor). >> > > > >> > > > >> > > > >> > > > Completely agree--usecases (1) and (3) are an indirect use >> of >> > > > Coders >> > > > >> > > > that are used to achieve an effect that would be better >> > > expressed >> > > > >> > > > directly. >> > > > >> > > > >> > > > >> > > > > (2) is an interesting hypothetical for massive scale where >> > > tiny >> > > > >> > > > incremental >> > > > >> > > > > optimization represents a lot of cost _and_ your data has >> > > > >> sufficient >> > > > >> > > > > structure to realize a benefit _and_ it needs to be >> > pinpointed >> > > > to >> > > > >> > just >> > > > >> > > > some >> > > > >> > > > > PCollections. I think our experience with coders so far is >> > > that >> > > > >> their >> > > > >> > > > > existence is almost entirely negative. It would be nice to >> > > > support >> > > > >> > this >> > > > >> > > > > vanishingly rare case without inflicting a terrible pain >> > point >> > > > on >> > > > >> the >> > > > >> > > > model >> > > > >> > > > > and all other users. >> > > > >> > > > >> > > > >> > > > (2) is not just about cheapness, sometimes there's other >> > > structure >> > > > >> in >> > > > >> > > > the data we can leverage. Consider the UnionCoder used in >> > > > >> > > > CoGBK--RawUnionValue has an integer value that specifies >> > > indicates >> > > > >> the >> > > > >> > > > type of it's raw Object field. Unless we want to extend the >> > type >> > > > >> > > > language, there's not a sufficient type descriptor that can >> be >> > > > used >> > > > >> to >> > > > >> > > > infer the coder. I'm dubious going down the road of adding >> > > special >> > > > >> > > > cases is the right thing here. >> > > > >> > > > >> > > > >> > > > > For example, in those cases you could encode in your >> > > > >> > > > > DoFn so the type descriptor would just be byte[]. >> > > > >> > > > >> > > > >> > > > As well as being an extremely cumbersome API, this would >> incur >> > > the >> > > > >> > > > cost of coding/decoding at that DoFn boundary even if it is >> > > fused >> > > > >> > > > away. >> > > > >> > > > >> > > > >> > > > >> On Thu, Jul 27, 2017 at 1:34 AM, Jean-Baptiste Onofré < >> > > > >> > > j...@nanthrax.net> >> > > > >> > > > >> wrote: >> > > > >> > > > >> >> > > > >> > > > >> > Hi, >> > > > >> > > > >> > >> > > > >> > > > >> > That's an interesting thread and I was wondering the >> > > > >> relationship >> > > > >> > > > between >> > > > >> > > > >> > type descriptor and coder for a while ;) >> > > > >> > > > >> > >> > > > >> > > > >> > Today, in a PCollection, we can set the coder and we >> also >> > > > have >> > > > >> a >> > > > >> > > > >> > getTypeDescriptor(). It sounds weird to me: it should >> be >> > > one >> > > > or >> > > > >> > the >> > > > >> > > > >> other. >> > > > >> > > > >> > >> > > > >> > > > >> > Basically, if the Coder is not used to define the type, >> > > > than, I >> > > > >> > > fully >> > > > >> > > > >> > agree with Eugene. >> > > > >> > > > >> > >> > > > >> > > > >> > Basically, the PCollection should define only the type >> > > > >> descriptor, >> > > > >> > > not >> > > > >> > > > >> the >> > > > >> > > > >> > coder by itself: the coder can be found using the type >> > > > >> descriptor. >> > > > >> > > > >> > >> > > > >> > > > >> > With both coder and type descriptor on the PCollection, >> > it >> > > > >> sounds >> > > > >> > a >> > > > >> > > > big >> > > > >> > > > >> > "decoupled" to me and it would be possible to have a >> > coder >> > > on >> > > > >> the >> > > > >> > > > >> > PCollection that doesn't match the type descriptor. >> > > > >> > > > >> > >> > > > >> > > > >> > I think PCollection type descriptor should be defined, >> > and >> > > > the >> > > > >> > coder >> > > > >> > > > >> > should be implicit based on this type descriptor. >> > > > >> > > > >> > >> > > > >> > > > >> > Thoughts ? >> > > > >> > > > >> > >> > > > >> > > > >> > Regards >> > > > >> > > > >> > JB >> > > > >> > > > >> > >> > > > >> > > > >> > >> > > > >> > > > >> > On 07/26/2017 05:25 AM, Eugene Kirpichov wrote: >> > > > >> > > > >> > >> > > > >> > > > >> >> Hello, >> > > > >> > > > >> >> >> > > > >> > > > >> >> I've worked on a few different things recently and ran >> > > > >> repeatedly >> > > > >> > > > into >> > > > >> > > > >> the >> > > > >> > > > >> >> same issue: that we do not have clear guidance on who >> > > should >> > > > >> set >> > > > >> > > the >> > > > >> > > > >> Coder >> > > > >> > > > >> >> on a PCollection: is it responsibility of the >> PTransform >> > > > that >> > > > >> > > outputs >> > > > >> > > > >> it, >> > > > >> > > > >> >> or is it responsibility of the user, or is it >> sometimes >> > > one >> > > > >> and >> > > > >> > > > >> sometimes >> > > > >> > > > >> >> the other? >> > > > >> > > > >> >> >> > > > >> > > > >> >> I believe that the answer is "it's responsibility of >> the >> > > > >> > transform" >> > > > >> > > > and >> > > > >> > > > >> >> moreover that ideally PCollection.setCoder() should >> not >> > > > >> exist. >> > > > >> > > > Instead: >> > > > >> > > > >> >> >> > > > >> > > > >> >> - Require that all transforms set a Coder on the >> > > > PCollection's >> > > > >> > they >> > > > >> > > > >> >> produce >> > > > >> > > > >> >> - i.e. it should never be responsibility of the user >> to >> > > "fix >> > > > >> up" >> > > > >> > a >> > > > >> > > > coder >> > > > >> > > > >> >> on >> > > > >> > > > >> >> a PCollection produced by a transform. >> > > > >> > > > >> >> >> > > > >> > > > >> >> - Since all transforms are composed of primitive >> > > transforms, >> > > > >> > saying >> > > > >> > > > >> >> "transforms must set a Coder" means simply that all >> > > > >> *primitive* >> > > > >> > > > >> transforms >> > > > >> > > > >> >> must set a Coder on their output. >> > > > >> > > > >> >> >> > > > >> > > > >> >> - In some cases, a primitive PTransform currently >> > doesn't >> > > > have >> > > > >> > > enough >> > > > >> > > > >> >> information to infer a coder for its output >> collection - >> > > > e.g. >> > > > >> > > > >> >> ParDo.of(DoFn<InputT, OutputT>) might be unable to >> > infer a >> > > > >> coder >> > > > >> > > for >> > > > >> > > > >> >> OutputT. In that case such transforms should allow the >> > > user >> > > > to >> > > > >> > > > provide a >> > > > >> > > > >> >> coder: ParDo.of(DoFn).withOutputCoder(...) [note that >> > > this >> > > > >> > differs >> > > > >> > > > from >> > > > >> > > > >> >> requiring the user to set a coder on the resulting >> > > > collection] >> > > > >> > > > >> >> >> > > > >> > > > >> >> - Corollary: composite transforms need to only >> configure >> > > > their >> > > > >> > > > primitive >> > > > >> > > > >> >> transforms (and composite sub-transforms) properly, >> and >> > > give >> > > > >> > them a >> > > > >> > > > >> Coder >> > > > >> > > > >> >> if needed. >> > > > >> > > > >> >> >> > > > >> > > > >> >> - Corollary: a PTransform with type parameters <FooT, >> > > BarT, >> > > > >> ...> >> > > > >> > > > needs >> > > > >> > > > >> to >> > > > >> > > > >> >> be configurable with coders for all of these, because >> > the >> > > > >> > > > implementation >> > > > >> > > > >> >> of >> > > > >> > > > >> >> the transform may change and it may introduce >> > intermediate >> > > > >> > > > collections >> > > > >> > > > >> >> involving these types. However, in many cases, some of >> > > these >> > > > >> type >> > > > >> > > > >> >> parameters appear in the type of the transform's >> input, >> > > > e.g. a >> > > > >> > > > >> >> PTransform<PCollection<KV<FooT, BarT>>, >> > > PCollection<MooT>> >> > > > >> will >> > > > >> > > > always >> > > > >> > > > >> be >> > > > >> > > > >> >> able to extract the coders for FooT and BarT from the >> > > input >> > > > >> > > > PCollection, >> > > > >> > > > >> >> so >> > > > >> > > > >> >> the user does not need to provide them. However, a >> coder >> > > for >> > > > >> BarT >> > > > >> > > > must >> > > > >> > > > >> be >> > > > >> > > > >> >> provided. I think in most cases the transform needs to >> > be >> > > > >> > > > configurable >> > > > >> > > > >> >> only >> > > > >> > > > >> >> with coders for its output. >> > > > >> > > > >> >> >> > > > >> > > > >> >> Here's a smooth migration path to accomplish the >> above: >> > > > >> > > > >> >> - Make PCollection.createPrimitiveOutputInternal() >> > take a >> > > > >> Coder. >> > > > >> > > > >> >> - Make all primitive transforms optionally >> configurable >> > > > with a >> > > > >> > > coder >> > > > >> > > > for >> > > > >> > > > >> >> their outputs, such as ParDo.of(DoFn). >> > withOutputCoder(). >> > > > >> > > > >> >> - By using the above, make all composite transforms >> > > shipped >> > > > >> with >> > > > >> > > the >> > > > >> > > > SDK >> > > > >> > > > >> >> set a Coder on the collections they produce; in some >> > > cases, >> > > > >> this >> > > > >> > > will >> > > > >> > > > >> >> require adding a withSomethingCoder() option to the >> > > > transform >> > > > >> and >> > > > >> > > > >> >> propagating that coder to its sub-transforms. If the >> > > option >> > > > is >> > > > >> > > unset, >> > > > >> > > > >> >> that's fine for now. >> > > > >> > > > >> >> - As a result of the above, get rid of all setCoder() >> > > calls >> > > > in >> > > > >> > the >> > > > >> > > > Beam >> > > > >> > > > >> >> repo. The call will still be there, but it will just >> not >> > > be >> > > > >> used >> > > > >> > > > >> anywhere >> > > > >> > > > >> >> in the SDK or examples, and we can mark it deprecated. >> > > > >> > > > >> >> - Add guidance to PTransform Style Guide in line with >> > the >> > > > >> above >> > > > >> > > > >> >> >> > > > >> > > > >> >> Does this sound like a good idea? I'm not sure how >> > urgent >> > > it >> > > > >> > would >> > > > >> > > > be to >> > > > >> > > > >> >> actually do this, but I'd like to know whether people >> > > agree >> > > > >> that >> > > > >> > > this >> > > > >> > > > >> is a >> > > > >> > > > >> >> good goal in general. >> > > > >> > > > >> >> >> > > > >> > > > >> >> >> > > > >> > > > >> > -- >> > > > >> > > > >> > Jean-Baptiste Onofré >> > > > >> > > > >> > jbono...@apache.org >> > > > >> > > > >> > http://blog.nanthrax.net >> > > > >> > > > >> > Talend - http://www.talend.com >> > > > >> > > > >> > >> > > > >> > > > >> >> > > > >> > > > >> > > > >> > > >> > > > >> > >> > > > >> >> > > > > >> > > > > >> > > > >> > > >> > >>