[
https://issues.apache.org/jira/browse/SPARK-37135?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17435133#comment-17435133
]
Apache Spark commented on SPARK-37135:
--------------------------------------
User 'LuciferYang' has created a pull request for this issue:
https://github.com/apache/spark/pull/34409
> Fix some mirco-benchmarks run failed
> -------------------------------------
>
> Key: SPARK-37135
> URL: https://issues.apache.org/jira/browse/SPARK-37135
> Project: Spark
> Issue Type: Bug
> Components: Spark Core, SQL
> Affects Versions: 3.3.0
> Reporter: Yang Jie
> Priority: Minor
>
> 2 mirco-benchmarks run failed:
>
> org.apache.spark.serializer.KryoSerializerBenchmark
> {code:java}
> Running org.apache.spark.serializer.KryoSerializerBenchmark:Running
> org.apache.spark.serializer.KryoSerializerBenchmark:Running benchmark:
> Benchmark KryoPool vs old"pool of 1" implementation Running case:
> KryoPool:true21/10/27 16:09:26 ERROR SparkContext: Error initializing
> SparkContext.java.lang.AssertionError: assertion failed: spark.test.home is
> not set! at scala.Predef$.assert(Predef.scala:223) at
> org.apache.spark.deploy.worker.Worker.<init>(Worker.scala:148) at
> org.apache.spark.deploy.worker.Worker$.startRpcEnvAndEndpoint(Worker.scala:954)
> at
> org.apache.spark.deploy.LocalSparkCluster.$anonfun$start$2(LocalSparkCluster.scala:71)
> at
> org.apache.spark.deploy.LocalSparkCluster.$anonfun$start$2$adapted(LocalSparkCluster.scala:65)
> at scala.collection.immutable.Range.foreach(Range.scala:158) at
> org.apache.spark.deploy.LocalSparkCluster.start(LocalSparkCluster.scala:65)
> at
> org.apache.spark.SparkContext$.org$apache$spark$SparkContext$$createTaskScheduler(SparkContext.scala:2971)
> at org.apache.spark.SparkContext.<init>(SparkContext.scala:562) at
> org.apache.spark.SparkContext.<init>(SparkContext.scala:138) at
> org.apache.spark.serializer.KryoSerializerBenchmark$.createSparkContext(KryoSerializerBenchmark.scala:86)
> at
> org.apache.spark.serializer.KryoSerializerBenchmark$.sc$lzycompute$1(KryoSerializerBenchmark.scala:58)
> at
> org.apache.spark.serializer.KryoSerializerBenchmark$.sc$1(KryoSerializerBenchmark.scala:58)
> at
> org.apache.spark.serializer.KryoSerializerBenchmark$.$anonfun$run$3(KryoSerializerBenchmark.scala:63)
> at scala.runtime.java8.JFunction0$mcJ$sp.apply(JFunction0$mcJ$sp.java:23) at
> scala.concurrent.Future$.$anonfun$apply$1(Future.scala:659) at
> scala.util.Success.$anonfun$map$1(Try.scala:255) at
> scala.util.Success.map(Try.scala:213) at
> scala.concurrent.Future.$anonfun$map$1(Future.scala:292) at
> scala.concurrent.impl.Promise.liftedTree1$1(Promise.scala:33) at
> scala.concurrent.impl.Promise.$anonfun$transform$1(Promise.scala:33) at
> scala.concurrent.impl.CallbackRunnable.run(Promise.scala:64) at
> java.base/java.util.concurrent.ForkJoinTask$RunnableExecuteAction.exec(ForkJoinTask.java:1426)
> at java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:290)
> at
> java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1020)
> at java.base/java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1656)
> at
> java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1594)
> at
> java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:183){code}
> org.apache.spark.sql.execution.benchmark.DateTimeBenchmark
> {code:java}
> Exception in thread "main"
> org.apache.spark.sql.catalyst.parser.ParseException: Exception in thread
> "main" org.apache.spark.sql.catalyst.parser.ParseException: Cannot mix
> year-month and day-time fields: interval 1 month 2 day(line 1, pos 38)
> == SQL ==cast(timestamp_seconds(id) as date) + interval 1 month 2
> day--------------------------------------^^^
> at
> org.apache.spark.sql.errors.QueryParsingErrors$.mixedIntervalUnitsError(QueryParsingErrors.scala:214)
> at
> org.apache.spark.sql.catalyst.parser.AstBuilder.constructMultiUnitsIntervalLiteral(AstBuilder.scala:2435)
> at
> org.apache.spark.sql.catalyst.parser.AstBuilder.$anonfun$visitInterval$1(AstBuilder.scala:2479)
> at
> org.apache.spark.sql.catalyst.parser.ParserUtils$.withOrigin(ParserUtils.scala:133)
> at
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitInterval(AstBuilder.scala:2454)
> at
> org.apache.spark.sql.catalyst.parser.AstBuilder.visitInterval(AstBuilder.scala:57)
> at
> org.apache.spark.sql.catalyst.parser.SqlBaseParser$IntervalContext.accept(SqlBaseParser.java:17681)
> {code}
>
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]