srielau commented on code in PR #38531:
URL: https://github.com/apache/spark/pull/38531#discussion_r1022941891
##
core/src/main/resources/error/error-classes.json:
##
@@ -290,6 +290,46 @@
"Null typed values cannot be used as arguments of ."
]
},
+
srielau commented on code in PR #38531:
URL: https://github.com/apache/spark/pull/38531#discussion_r1022941891
##
core/src/main/resources/error/error-classes.json:
##
@@ -290,6 +290,46 @@
"Null typed values cannot be used as arguments of ."
]
},
+
tgravescs commented on code in PR #38622:
URL: https://github.com/apache/spark/pull/38622#discussion_r1022918575
##
resource-managers/yarn/src/main/scala/org/apache/spark/deploy/yarn/ApplicationMaster.scala:
##
@@ -815,6 +815,7 @@ private[spark] class ApplicationMaster(
c
Yikun commented on PR #23:
URL: https://github.com/apache/spark-docker/pull/23#issuecomment-1315319292
@HyukjinKwon @martin-g Thanks! Merge to master. Looks like no more feedback
about publishing. Maybe let's publish after docker official image review
completed.
--
This is an automated m
Yikun closed pull request #23: [SPARK-40519] Add "Publish" workflow to help
release apache/spark image
URL: https://github.com/apache/spark-docker/pull/23
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to g
peter-toth commented on PR #38640:
URL: https://github.com/apache/spark/pull/38640#issuecomment-1315310978
> Ah I just realized that there is no way to use v2 parquet table today.
Shall we support it first before benchmarking it?
I'm ok with switching this benchmark to parquet v2 when
cloud-fan commented on PR #38640:
URL: https://github.com/apache/spark/pull/38640#issuecomment-1315299552
Ah I just realized that there is no way to use v2 parquet table today. Shall
we support it first before benchmarking it?
--
This is an automated message from the Apache Git Service.
T
peter-toth commented on PR #38640:
URL: https://github.com/apache/spark/pull/38640#issuecomment-1315293892
> @peter-toth there is an easy way to enable parquet v2: set
`spark.sql.sources.useV1SourceList` to empty.
I thought that config only controls sources when `spark.read.parquet(..
Ngone51 commented on PR #38467:
URL: https://github.com/apache/spark/pull/38467#issuecomment-1315276964
> BlockInfoManager#blockInfoWrappers block info and lock not removed.
Can't we catch the exception from `BlockManager#removeBlockInternal` and
release the lock when caught the execp
cloud-fan closed pull request #38627: [SPARK-40875] [CONNECT] [FOLLOW] Retain
Group expressions in aggregate.
URL: https://github.com/apache/spark/pull/38627
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above t
cloud-fan commented on PR #38627:
URL: https://github.com/apache/spark/pull/38627#issuecomment-1315270191
thanks, merging to master!
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific c
cloud-fan commented on code in PR #38627:
URL: https://github.com/apache/spark/pull/38627#discussion_r1022752267
##
connector/connect/src/main/scala/org/apache/spark/sql/connect/planner/SparkConnectPlanner.scala:
##
@@ -441,11 +441,14 @@ class SparkConnectPlanner(session: SparkS
grundprinzip commented on code in PR #38605:
URL: https://github.com/apache/spark/pull/38605#discussion_r1022745612
##
connector/connect/docs/adding-proto-messages.md:
##
@@ -0,0 +1,41 @@
+# Required, Optional and default values
+
+Connect adopts proto3, which does not support `
cloud-fan commented on PR #38640:
URL: https://github.com/apache/spark/pull/38640#issuecomment-1315256115
@peter-toth there is an easy way to enable parquet v2: set
`spark.sql.sources.useV1SourceList` to empty.
--
This is an automated message from the Apache Git Service.
To respond to the
Ngone51 commented on code in PR #38441:
URL: https://github.com/apache/spark/pull/38441#discussion_r1022732173
##
core/src/main/scala/org/apache/spark/internal/config/package.scala:
##
@@ -2024,6 +2024,16 @@ package object config {
.stringConf
.createOptional
+
Ngone51 commented on code in PR #38441:
URL: https://github.com/apache/spark/pull/38441#discussion_r1022731167
##
core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala:
##
@@ -2193,9 +2193,11 @@ private[spark] class DAGScheduler(
* Return true when:
* 1. Wai
Ngone51 commented on PR #38441:
URL: https://github.com/apache/spark/pull/38441#issuecomment-1315242107
Seem like this PR is addressing my concern at
https://github.com/apache/spark/pull/37924#discussion_r990925620. I actually
think the original PR won't work in most cases without this PR.
Ngone51 commented on code in PR #38441:
URL: https://github.com/apache/spark/pull/38441#discussion_r1022725414
##
core/src/main/scala/org/apache/spark/internal/config/package.scala:
##
@@ -2024,6 +2024,16 @@ package object config {
.stringConf
.createOptional
+
AmplabJenkins commented on PR #38663:
URL: https://github.com/apache/spark/pull/38663#issuecomment-1315236539
Can one of the admins verify this patch?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go
MaxGekk commented on code in PR #38531:
URL: https://github.com/apache/spark/pull/38531#discussion_r1022706881
##
core/src/main/resources/error/error-classes.json:
##
@@ -290,6 +290,46 @@
"Null typed values cannot be used as arguments of ."
]
},
+
peter-toth commented on code in PR #38640:
URL: https://github.com/apache/spark/pull/38640#discussion_r1022701848
##
sql/core/src/test/scala/org/apache/spark/sql/PlanStabilitySuite.scala:
##
@@ -351,6 +353,62 @@ class TPCDSModifiedPlanStabilityWithStatsSuite extends
PlanStabili
MaxGekk closed pull request #38656: [SPARK-41140][SQL] Rename the error class
`_LEGACY_ERROR_TEMP_2440` to `INVALID_WHERE_CONDITION`
URL: https://github.com/apache/spark/pull/38656
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub
MaxGekk commented on PR #38656:
URL: https://github.com/apache/spark/pull/38656#issuecomment-1315205891
Merging to master. Thank you, @LuciferYang @cloud-fan @srielau @itholic for
review.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on
grundprinzip commented on code in PR #38631:
URL: https://github.com/apache/spark/pull/38631#discussion_r1022651386
##
python/pyspark/sql/connect/dataframe.py:
##
@@ -44,7 +44,7 @@
from pyspark.sql.connect.typing import ColumnOrString, ExpressionOrString
from pyspark.s
grundprinzip commented on code in PR #38631:
URL: https://github.com/apache/spark/pull/38631#discussion_r1022651011
##
python/pyspark/sql/connect/dataframe.py:
##
@@ -44,7 +44,7 @@
from pyspark.sql.connect.typing import ColumnOrString, ExpressionOrString
from pyspark.s
itholic commented on PR #38664:
URL: https://github.com/apache/spark/pull/38664#issuecomment-1315140725
Sure, let me integrate into `DATATYPE_MISMATCH.WRONG_NUM_ARGS`
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
fred-db commented on PR #38497:
URL: https://github.com/apache/spark/pull/38497#issuecomment-1315136114
@allisonwang-db Incorporated all the changes requested, lmk what you think!
:)
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to G
MaxGekk commented on code in PR #38531:
URL: https://github.com/apache/spark/pull/38531#discussion_r1022613029
##
core/src/main/resources/error/error-classes.json:
##
@@ -290,6 +290,46 @@
"Null typed values cannot be used as arguments of ."
]
},
+
zhengruifeng commented on PR #38653:
URL: https://github.com/apache/spark/pull/38653#issuecomment-1315103593
merged into master, thank you guys
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
zhengruifeng closed pull request #38653: [SPARK-41128][CONNECT][PYTHON]
Implement `DataFrame.fillna ` and `DataFrame.na.fill `
URL: https://github.com/apache/spark/pull/38653
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and us
WweiL commented on code in PR #38503:
URL: https://github.com/apache/spark/pull/38503#discussion_r1022251949
##
sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationSuite.scala:
##
@@ -190,20 +190,25 @@ class StreamingDeduplicationSuite extends
StateStor
ulysses-you commented on PR #38619:
URL: https://github.com/apache/spark/pull/38619#issuecomment-1315058947
thank you @cloud-fan @dongjoon-hyun
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to th
cloud-fan commented on code in PR #38605:
URL: https://github.com/apache/spark/pull/38605#discussion_r1022545314
##
connector/connect/docs/adding-proto-messages.md:
##
@@ -0,0 +1,41 @@
+# Required, Optional and default values
+
+Connect adopts proto3, which does not support `req
zhengruifeng commented on code in PR #38653:
URL: https://github.com/apache/spark/pull/38653#discussion_r1022538668
##
connector/connect/src/main/scala/org/apache/spark/sql/connect/dsl/package.scala:
##
@@ -226,6 +226,58 @@ package object dsl {
}
}
+implicit cl
WweiL commented on code in PR #38503:
URL: https://github.com/apache/spark/pull/38503#discussion_r1021870148
##
sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingDeduplicationSuite.scala:
##
@@ -190,20 +190,25 @@ class StreamingDeduplicationSuite extends
StateStor
cloud-fan commented on code in PR #38653:
URL: https://github.com/apache/spark/pull/38653#discussion_r1022527827
##
connector/connect/src/main/scala/org/apache/spark/sql/connect/dsl/package.scala:
##
@@ -226,6 +226,58 @@ package object dsl {
}
}
+implicit class
cloud-fan commented on code in PR #38495:
URL: https://github.com/apache/spark/pull/38495#discussion_r1022525354
##
sql/hive/src/main/scala/org/apache/spark/sql/hive/client/HiveClientImpl.scala:
##
@@ -609,6 +609,17 @@ private[hive] class HiveClientImpl(
shim.alterTable(cli
cloud-fan commented on code in PR #38595:
URL: https://github.com/apache/spark/pull/38595#discussion_r1022521452
##
sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryParsingErrors.scala:
##
@@ -542,11 +542,11 @@ private[sql] object QueryParsingErrors extends
QueryErr
cloud-fan commented on code in PR #38595:
URL: https://github.com/apache/spark/pull/38595#discussion_r1022520366
##
sql/core/src/main/scala/org/apache/spark/sql/Dataset.scala:
##
@@ -3799,13 +3798,21 @@ class Dataset[T] private[sql](
global: Boolean): CreateViewCommand =
wankunde commented on code in PR #38495:
URL: https://github.com/apache/spark/pull/38495#discussion_r1022516593
##
sql/hive/src/test/scala/org/apache/spark/sql/hive/InsertSuite.scala:
##
@@ -894,12 +895,14 @@ class InsertSuite extends QueryTest with
TestHiveSingleton with Befor
cloud-fan commented on code in PR #38595:
URL: https://github.com/apache/spark/pull/38595#discussion_r1022515695
##
core/src/main/resources/error/error-classes.json:
##
@@ -933,6 +933,11 @@
],
"sqlState" : "42000"
},
+ "TEMP_VIEW_NAME_CONTAINS_UNSUPPORTED_NAME_PART
wankunde commented on code in PR #38495:
URL: https://github.com/apache/spark/pull/38495#discussion_r1022515586
##
sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveExternalCatalog.scala:
##
@@ -722,18 +722,15 @@ private[spark] class HiveExternalCatalog(conf: SparkConf,
had
wankunde commented on code in PR #38495:
URL: https://github.com/apache/spark/pull/38495#discussion_r1022515184
##
sql/hive/src/main/scala/org/apache/spark/sql/hive/client/HiveClient.scala:
##
@@ -127,6 +127,9 @@ private[hive] trait HiveClient {
*/
def alterTable(dbName:
cloud-fan commented on code in PR #38531:
URL: https://github.com/apache/spark/pull/38531#discussion_r1022513821
##
core/src/main/resources/error/error-classes.json:
##
@@ -290,6 +290,46 @@
"Null typed values cannot be used as arguments of ."
]
},
+
cloud-fan closed pull request #38404: [SPARK-40956] SQL Equivalent for
Dataframe overwrite command
URL: https://github.com/apache/spark/pull/38404
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
cloud-fan commented on PR #38404:
URL: https://github.com/apache/spark/pull/38404#issuecomment-1314988679
The failed test job is flaky (network issue) and it has passed in the
previous commit. Given the last commit is comment only, I'm merging this PR to
master. Thanks!
--
This is an aut
grundprinzip commented on code in PR #38631:
URL: https://github.com/apache/spark/pull/38631#discussion_r1022503915
##
python/pyspark/sql/connect/dataframe.py:
##
@@ -44,7 +44,7 @@
from pyspark.sql.connect.typing import ColumnOrString, ExpressionOrString
from pyspark.s
grundprinzip commented on code in PR #38631:
URL: https://github.com/apache/spark/pull/38631#discussion_r1022503590
##
python/pyspark/sql/connect/dataframe.py:
##
@@ -44,7 +44,7 @@
from pyspark.sql.connect.typing import ColumnOrString, ExpressionOrString
from pyspark.s
cloud-fan closed pull request #38662: [SPARK-41144][SQL] Unresolved hint should
not cause query failure
URL: https://github.com/apache/spark/pull/38662
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go t
cloud-fan commented on PR #38662:
URL: https://github.com/apache/spark/pull/38662#issuecomment-1314981448
thanks, merging to master/3.3!
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specif
cloud-fan closed pull request #38619: [SPARK-41112][SQL] RuntimeFilter should
apply ColumnPruning eagerly with in-subquery filter
URL: https://github.com/apache/spark/pull/38619
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and
grundprinzip commented on code in PR #38631:
URL: https://github.com/apache/spark/pull/38631#discussion_r1022498105
##
python/pyspark/sql/connect/column.py:
##
@@ -82,6 +82,73 @@ def to_plan(self, session: "RemoteSparkSession") ->
"proto.Expression":
def __str__(self) -> s
grundprinzip commented on code in PR #38631:
URL: https://github.com/apache/spark/pull/38631#discussion_r1022495692
##
python/pyspark/sql/tests/connect/test_connect_basic.py:
##
@@ -248,6 +248,20 @@ def test_simple_datasource_read(self) -> None:
actualResult = panda
grundprinzip commented on code in PR #38631:
URL: https://github.com/apache/spark/pull/38631#discussion_r1022492699
##
python/pyspark/sql/tests/connect/test_connect_column_expressions.py:
##
@@ -134,6 +134,16 @@ def test_list_to_literal(self):
lit_list_plan = fun.lit([f
101 - 154 of 154 matches
Mail list logo