I've checked the Jenkins log and It seems the commit from
https://github.com/apache/spark/pull/29404 caused the failure.


On Sat, Aug 15, 2020 at 10:43 PM Koert Kuipers <ko...@tresata.com> wrote:

> i noticed commit today that seems to prepare for 3.0.1-rc1:
> commit 05144a5c10cd37ebdbb55fde37d677def49af11f
> Author: Ruifeng Zheng <ruife...@apache.org>
> Date:   Sat Aug 15 01:37:47 2020 +0000
>
>     Preparing Spark release v3.0.1-rc1
>
> so i tried to build spark on that commit and i get failure in sql:
>
> 09:36:57.371 ERROR org.apache.spark.scheduler.TaskSetManager: Task 0 in
> stage 77.0 failed 1 times; aborting job
> [info] - SPARK-28224: Aggregate sum big decimal overflow *** FAILED ***
> (306 milliseconds)
> [info]   org.apache.spark.SparkException: Job aborted due to stage
> failure: Task 0 in stage 77.0 failed 1 times, most recent failure: Lost
> task 0.0 in stage 77.0 (TID 197, 192.168.11.17, executor driver):
> java.lang.ArithmeticException:
> Decimal(expanded,111111111111111111110.246000000000000000,39,18}) cannot be
> represented as Decimal(38, 18).
> [info] at org.apache.spark.sql.types.Decimal.toPrecision(Decimal.scala:369)
> [info] at
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage2.agg_doAggregate_sum_0$(Unknown
> Source)
> [info] at
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage2.agg_doConsume_0$(Unknown
> Source)
> [info] at
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage2.agg_doAggregateWithoutKey_0$(Unknown
> Source)
> [info] at
> org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage2.processNext(Unknown
> Source)
> [info] at
> org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
> [info] at
> org.apache.spark.sql.execution.WholeStageCodegenExec$$anon$1.hasNext(WholeStageCodegenExec.scala:729)
> [info] at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:458)
> [info] at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:458)
> [info] at org.apache.spark.util.Utils$.getIteratorSize(Utils.scala:1804)
> [info] at org.apache.spark.rdd.RDD.$anonfun$count$1(RDD.scala:1227)
> [info] at org.apache.spark.rdd.RDD.$anonfun$count$1$adapted(RDD.scala:1227)
> [info] at
> org.apache.spark.SparkContext.$anonfun$runJob$5(SparkContext.scala:2138)
> [info] at
> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
> [info] at org.apache.spark.scheduler.Task.run(Task.scala:127)
> [info] at
> org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:446)
> [info] at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1377)
> [info] at
> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:449)
> [info] at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> [info] at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> [info] at java.lang.Thread.run(Thread.java:748)
>
> [error] Failed tests:
> [error] org.apache.spark.sql.DataFrameSuite
>
> On Thu, Aug 13, 2020 at 8:19 PM Jason Moore
> <jason.mo...@quantium.com.au.invalid> wrote:
>
>> Thank you so much!  Any update on getting the RC1 up for vote?
>>
>> Jason.
>>
>>
>> ------------------------------
>> *From:* 郑瑞峰 <ruife...@foxmail.com>
>> *Sent:* Wednesday, 5 August 2020 12:54 PM
>> *To:* Jason Moore <jason.mo...@quantium.com.au.INVALID>; Spark dev list <
>> dev@spark.apache.org>
>> *Subject:* 回复: [DISCUSS] Apache Spark 3.0.1 Release
>>
>> Hi all,
>> I am going to prepare the realease of 3.0.1 RC1, with the help of Wenchen.
>>
>>
>> ------------------ 原始邮件 ------------------
>> *发件人:* "Jason Moore" <jason.mo...@quantium.com.au.INVALID>;
>> *发送时间:* 2020年7月30日(星期四) 上午10:35
>> *收件人:* "dev"<dev@spark.apache.org>;
>> *主题:* Re: [DISCUSS] Apache Spark 3.0.1 Release
>>
>> Hi all,
>>
>>
>>
>> Discussion around 3.0.1 seems to have trickled away.  What was blocking
>> the release process kicking off?  I can see some unresolved bugs raised
>> against 3.0.0, but conversely there were quite a few critical correctness
>> fixes waiting to be released.
>>
>>
>>
>> Cheers,
>>
>> Jason.
>>
>>
>>
>> *From: *Takeshi Yamamuro <linguin....@gmail.com>
>> *Date: *Wednesday, 15 July 2020 at 9:00 am
>> *To: *Shivaram Venkataraman <shiva...@eecs.berkeley.edu>
>> *Cc: *"dev@spark.apache.org" <dev@spark.apache.org>
>> *Subject: *Re: [DISCUSS] Apache Spark 3.0.1 Release
>>
>>
>>
>> > Just wanted to check if there are any blockers that we are still
>> waiting for to start the new release process.
>>
>> I don't see any on-going blocker in my area.
>>
>> Thanks for the notification.
>>
>>
>>
>> Bests,
>>
>> Tkaeshi
>>
>>
>>
>> On Wed, Jul 15, 2020 at 4:03 AM Dongjoon Hyun <dongjoon.h...@gmail.com>
>> wrote:
>>
>> Hi, Yi.
>>
>>
>>
>> Could you explain why you think that is a blocker? For the given example
>> from the JIRA description,
>>
>>
>>
>> spark.udf.register("key", udf((m: Map[String, String]) => m.keys.head.toInt))
>>
>> Seq(Map("1" -> "one", "2" -> "two")).toDF("a").createOrReplaceTempView("t")
>>
>> checkAnswer(sql("SELECT key(a) AS k FROM t GROUP BY key(a)"), Row(1) :: Nil)
>>
>>
>>
>> Apache Spark 3.0.0 seems to work like the following.
>>
>>
>>
>> scala> spark.version
>>
>> res0: String = 3.0.0
>>
>>
>>
>> scala> spark.udf.register("key", udf((m: Map[String, String]) =>
>> m.keys.head.toInt))
>>
>> res1: org.apache.spark.sql.expressions.UserDefinedFunction =
>> SparkUserDefinedFunction($Lambda$1958/948653928@5d6bed7b,IntegerType,List(Some(class[value[0]:
>> map<string,string>])),None,false,true)
>>
>>
>>
>> scala> Seq(Map("1" -> "one", "2" ->
>> "two")).toDF("a").createOrReplaceTempView("t")
>>
>>
>>
>> scala> sql("SELECT key(a) AS k FROM t GROUP BY key(a)").collect
>>
>> res3: Array[org.apache.spark.sql.Row] = Array([1])
>>
>>
>>
>> Could you provide a reproducible example?
>>
>>
>>
>> Bests,
>>
>> Dongjoon.
>>
>>
>>
>>
>>
>> On Tue, Jul 14, 2020 at 10:04 AM Yi Wu <yi...@databricks.com> wrote:
>>
>> This probably be a blocker:
>> https://issues.apache.org/jira/browse/SPARK-32307
>>
>>
>>
>> On Tue, Jul 14, 2020 at 11:13 PM Sean Owen <sro...@gmail.com> wrote:
>>
>> https://issues.apache.org/jira/browse/SPARK-32234 ?
>>
>> On Tue, Jul 14, 2020 at 9:57 AM Shivaram Venkataraman
>> <shiva...@eecs.berkeley.edu> wrote:
>> >
>> > Hi all
>> >
>> > Just wanted to check if there are any blockers that we are still
>> waiting for to start the new release process.
>> >
>> > Thanks
>> > Shivaram
>> >
>>
>>
>>
>>
>> --
>>
>> ---
>> Takeshi Yamamuro
>>
>

-- 
---
Takeshi Yamamuro

Reply via email to