+1 too, I'd consider also to include SPARK-24208 if we can solve it timely...
2018-06-28 8:28 GMT+02:00 Takeshi Yamamuro <linguin....@gmail.com>: > +1, I heard some Spark users have skipped v2.3.1 because of these bugs. > > On Thu, Jun 28, 2018 at 3:09 PM Xingbo Jiang <jiangxb1...@gmail.com> > wrote: > >> +1 >> >> Wenchen Fan <cloud0...@gmail.com>于2018年6月28日 周四下午2:06写道: >> >>> Hi Saisai, that's great! please go ahead! >>> >>> On Thu, Jun 28, 2018 at 12:56 PM Saisai Shao <sai.sai.s...@gmail.com> >>> wrote: >>> >>>> +1, like mentioned by Marcelo, these issues seems quite severe. >>>> >>>> I can work on the release if short of hands :). >>>> >>>> Thanks >>>> Jerry >>>> >>>> >>>> Marcelo Vanzin <van...@cloudera.com.invalid> 于2018年6月28日周四 上午11:40写道: >>>> >>>>> +1. SPARK-24589 / SPARK-24552 are kinda nasty and we should get fixes >>>>> for those out. >>>>> >>>>> (Those are what delayed 2.2.2 and 2.1.3 for those watching...) >>>>> >>>>> On Wed, Jun 27, 2018 at 7:59 PM, Wenchen Fan <cloud0...@gmail.com> >>>>> wrote: >>>>> > Hi all, >>>>> > >>>>> > Spark 2.3.1 was released just a while ago, but unfortunately we >>>>> discovered >>>>> > and fixed some critical issues afterward. >>>>> > >>>>> > SPARK-24495: SortMergeJoin may produce wrong result. >>>>> > This is a serious correctness bug, and is easy to hit: have >>>>> duplicated join >>>>> > key from the left table, e.g. `WHERE t1.a = t2.b AND t1.a = t2.c`, >>>>> and the >>>>> > join is a sort merge join. This bug is only present in Spark 2.3. >>>>> > >>>>> > SPARK-24588: stream-stream join may produce wrong result >>>>> > This is a correctness bug in a new feature of Spark 2.3: the >>>>> stream-stream >>>>> > join. Users can hit this bug if one of the join side is partitioned >>>>> by a >>>>> > subset of the join keys. >>>>> > >>>>> > SPARK-24552: Task attempt numbers are reused when stages are retried >>>>> > This is a long-standing bug in the output committer that may >>>>> introduce data >>>>> > corruption. >>>>> > >>>>> > SPARK-24542: UDFXPathXXXX allow users to pass carefully crafted XML >>>>> to >>>>> > access arbitrary files >>>>> > This is a potential security issue if users build access control >>>>> module upon >>>>> > Spark. >>>>> > >>>>> > I think we need a Spark 2.3.2 to address these issues(especially the >>>>> > correctness bugs) ASAP. Any thoughts? >>>>> > >>>>> > Thanks, >>>>> > Wenchen >>>>> >>>>> >>>>> >>>>> -- >>>>> Marcelo >>>>> >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >>>>> >>>>> > > -- > --- > Takeshi Yamamuro >