+1. Thanks, Saisai! The impact of SPARK-24495 is large. We should release Spark 2.3.2 ASAP.
Thanks, Xiao 2018-06-27 23:28 GMT-07:00 Takeshi Yamamuro <[email protected]>: > +1, I heard some Spark users have skipped v2.3.1 because of these bugs. > > On Thu, Jun 28, 2018 at 3:09 PM Xingbo Jiang <[email protected]> > wrote: > >> +1 >> >> Wenchen Fan <[email protected]>于2018年6月28日 周四下午2:06写道: >> >>> Hi Saisai, that's great! please go ahead! >>> >>> On Thu, Jun 28, 2018 at 12:56 PM Saisai Shao <[email protected]> >>> wrote: >>> >>>> +1, like mentioned by Marcelo, these issues seems quite severe. >>>> >>>> I can work on the release if short of hands :). >>>> >>>> Thanks >>>> Jerry >>>> >>>> >>>> Marcelo Vanzin <[email protected]> 于2018年6月28日周四 上午11:40写道: >>>> >>>>> +1. SPARK-24589 / SPARK-24552 are kinda nasty and we should get fixes >>>>> for those out. >>>>> >>>>> (Those are what delayed 2.2.2 and 2.1.3 for those watching...) >>>>> >>>>> On Wed, Jun 27, 2018 at 7:59 PM, Wenchen Fan <[email protected]> >>>>> wrote: >>>>> > Hi all, >>>>> > >>>>> > Spark 2.3.1 was released just a while ago, but unfortunately we >>>>> discovered >>>>> > and fixed some critical issues afterward. >>>>> > >>>>> > SPARK-24495: SortMergeJoin may produce wrong result. >>>>> > This is a serious correctness bug, and is easy to hit: have >>>>> duplicated join >>>>> > key from the left table, e.g. `WHERE t1.a = t2.b AND t1.a = t2.c`, >>>>> and the >>>>> > join is a sort merge join. This bug is only present in Spark 2.3. >>>>> > >>>>> > SPARK-24588: stream-stream join may produce wrong result >>>>> > This is a correctness bug in a new feature of Spark 2.3: the >>>>> stream-stream >>>>> > join. Users can hit this bug if one of the join side is partitioned >>>>> by a >>>>> > subset of the join keys. >>>>> > >>>>> > SPARK-24552: Task attempt numbers are reused when stages are retried >>>>> > This is a long-standing bug in the output committer that may >>>>> introduce data >>>>> > corruption. >>>>> > >>>>> > SPARK-24542: UDFXPathXXXX allow users to pass carefully crafted XML >>>>> to >>>>> > access arbitrary files >>>>> > This is a potential security issue if users build access control >>>>> module upon >>>>> > Spark. >>>>> > >>>>> > I think we need a Spark 2.3.2 to address these issues(especially the >>>>> > correctness bugs) ASAP. Any thoughts? >>>>> > >>>>> > Thanks, >>>>> > Wenchen >>>>> >>>>> >>>>> >>>>> -- >>>>> Marcelo >>>>> >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe e-mail: [email protected] >>>>> >>>>> > > -- > --- > Takeshi Yamamuro >
