This is an automated email from the ASF dual-hosted git repository.
gurwls223 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push:
new ae1f6a26ed3 [SPARK-39856][SQL][TESTS][FOLLOW-UP] Increase the number
of partitions in TPC-DS build to avoid out-of-memory
ae1f6a26ed3 is described below
commit ae1f6a26ed39b297ace8d6c9420b72a3c01a3291
Author: yangjie01 <[email protected]>
AuthorDate: Mon Jul 25 22:08:11 2022 +0900
[SPARK-39856][SQL][TESTS][FOLLOW-UP] Increase the number of partitions in
TPC-DS build to avoid out-of-memory
### What changes were proposed in this pull request?
This PR increases the number of partitions further more (see also
https://github.com/apache/spark/pull/37270)
### Why are the changes needed?
To make the build pass.
### Does this PR introduce _any_ user-facing change?
No, test and dev-only.
### How was this patch tested?
It's tested in
https://github.com/LuciferYang/spark/runs/7497163716?check_suite_focus=true
Closes #37273 from LuciferYang/SPARK-39856-FOLLOWUP.
Authored-by: yangjie01 <[email protected]>
Signed-off-by: Hyukjin Kwon <[email protected]>
---
sql/core/src/test/scala/org/apache/spark/sql/TPCDSQueryTestSuite.scala | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git
a/sql/core/src/test/scala/org/apache/spark/sql/TPCDSQueryTestSuite.scala
b/sql/core/src/test/scala/org/apache/spark/sql/TPCDSQueryTestSuite.scala
index 92cf574781f..f3eaa898e59 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/TPCDSQueryTestSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/TPCDSQueryTestSuite.scala
@@ -62,7 +62,7 @@ class TPCDSQueryTestSuite extends QueryTest with TPCDSBase
with SQLQueryTestHelp
// To make output results deterministic
override protected def sparkConf: SparkConf = super.sparkConf
- .set(SQLConf.SHUFFLE_PARTITIONS.key, 4.toString)
+ .set(SQLConf.SHUFFLE_PARTITIONS.key, 16.toString)
protected override def createSparkSession: TestSparkSession = {
new TestSparkSession(new SparkContext("local[1]",
this.getClass.getSimpleName, sparkConf))
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]