zhengruifeng commented on PR #40322:
URL: https://github.com/apache/spark/pull/40322#issuecomment-1459625070
merged into master/branch-3.4
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
zhengruifeng closed pull request #40322: [SPARK-41775][PYTHON][FOLLOW-UP]
Updating error message for training using PyTorch functions
URL: https://github.com/apache/spark/pull/40322
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to
mridulm commented on PR #40307:
URL: https://github.com/apache/spark/pull/40307#issuecomment-1459618430
@jerqi the basic issue here is, `getPreferredLocations` in `ShuffledRowRDD`
should return `Nil` at the very beginning in case
`spark.shuffle.reduceLocality.enabled = false`
We
WeichenXu123 commented on code in PR #40297:
URL: https://github.com/apache/spark/pull/40297#discussion_r1127834899
##
connector/connect/server/src/main/scala/org/apache/spark/sql/connect/ml/Serializer.scala:
##
@@ -0,0 +1,87 @@
+/*
+ * Licensed to the Apache Software
MaxGekk commented on code in PR #40126:
URL: https://github.com/apache/spark/pull/40126#discussion_r1129047990
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/Analyzer.scala:
##
@@ -465,7 +465,20 @@ class Analyzer(override val catalogManager:
LuciferYang commented on PR #40317:
URL: https://github.com/apache/spark/pull/40317#issuecomment-1459646333
Thanks @HyukjinKwon @hvanhovell @dongjoon-hyun @beliefer
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
xinrong-meng opened a new pull request, #40330:
URL: https://github.com/apache/spark/pull/40330
### What changes were proposed in this pull request?
Improve docstring of mapInPandas and mapInArrow
### Why are the changes needed?
For readability. We call out they are not scalar
xinrong-meng commented on PR #40329:
URL: https://github.com/apache/spark/pull/40329#issuecomment-1459636445
CC @HyukjinKwon @hvanhovell
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
yaooqinn commented on PR #40258:
URL: https://github.com/apache/spark/pull/40258#issuecomment-1459654900
You first defined a case-sensitive data set, then queried in a
case-insensitive way, I guess the error is expected.
--
This is an automated message from the Apache Git Service.
To
shrprasa commented on PR #40258:
URL: https://github.com/apache/spark/pull/40258#issuecomment-1459668923
> You first defined a case-sensitive data set, then queried in a
case-insensitive way, I guess the error is expected.
In the physical plan, both id and ID columns are projected to
zhengruifeng commented on code in PR #40297:
URL: https://github.com/apache/spark/pull/40297#discussion_r1129079073
##
connector/connect/server/src/main/scala/org/apache/spark/sql/connect/ml/MLUtils.scala:
##
@@ -0,0 +1,113 @@
+/*
+ * Licensed to the Apache Software Foundation
zhengruifeng opened a new pull request, #40331:
URL: https://github.com/apache/spark/pull/40331
### What changes were proposed in this pull request?
Add '__getattr__' and '__getitem__' of DataFrame and Column to API reference
### Why are the changes needed?
'__getattr__'
shrprasa commented on PR #40258:
URL: https://github.com/apache/spark/pull/40258#issuecomment-1459652942
> Can you try `set spark.sql.caseSensitive=true`?
Yes, I have tried it. With caseSensitive set to false, it will work as then
id and ID will be treated as separate columns.
WeichenXu123 commented on code in PR #40297:
URL: https://github.com/apache/spark/pull/40297#discussion_r1129073675
##
connector/connect/common/src/main/protobuf/spark/connect/ml.proto:
##
@@ -0,0 +1,135 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or
WeichenXu123 commented on code in PR #40297:
URL: https://github.com/apache/spark/pull/40297#discussion_r1129072923
##
connector/connect/server/src/main/scala/org/apache/spark/sql/connect/ml/MLUtils.scala:
##
@@ -0,0 +1,113 @@
+/*
+ * Licensed to the Apache Software Foundation
WeichenXu123 commented on code in PR #40297:
URL: https://github.com/apache/spark/pull/40297#discussion_r1129073121
##
connector/connect/common/src/main/protobuf/spark/connect/ml.proto:
##
@@ -0,0 +1,135 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or
jerqi commented on PR #40307:
URL: https://github.com/apache/spark/pull/40307#issuecomment-1459630519
>
Could I raise another pr to fix this issue?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above
jerqi commented on PR #40307:
URL: https://github.com/apache/spark/pull/40307#issuecomment-1459630906
> @jerqi the basic issue here is, `getPreferredLocations` in
`ShuffledRowRDD` should return `Nil` at the very beginning in case
`spark.shuffle.reduceLocality.enabled = false`
mridulm commented on PR #40307:
URL: https://github.com/apache/spark/pull/40307#issuecomment-1459652457
Sure ! Please go ahead :-)
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
WeichenXu123 commented on code in PR #40297:
URL: https://github.com/apache/spark/pull/40297#discussion_r1129073881
##
connector/connect/common/src/main/protobuf/spark/connect/ml.proto:
##
@@ -0,0 +1,135 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or
xinrong-meng commented on PR #40330:
URL: https://github.com/apache/spark/pull/40330#issuecomment-1459635902
CC @HyukjinKwon @hvanhovell
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
yaooqinn commented on PR #40258:
URL: https://github.com/apache/spark/pull/40258#issuecomment-1459648316
Can you try `set spark.sql.caseSensitive=true`?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to
peter-toth commented on code in PR #40268:
URL: https://github.com/apache/spark/pull/40268#discussion_r1127512831
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/optimizer/expressions.scala:
##
@@ -112,16 +113,13 @@ object ConstantFolding extends Rule[LogicalPlan]
yaooqinn opened a new pull request, #40313:
URL: https://github.com/apache/spark/pull/40313
### What changes were proposed in this pull request?
Fix /api/v1/applications to return total uptime instead of 0 for duration
### Why are the changes needed?
Fix
AngersZh opened a new pull request, #40315:
URL: https://github.com/apache/spark/pull/40315
### What changes were proposed in this pull request?
Since in https://github.com/apache/spark/pull/35594 we support pass a exit
code to AM,
when SparkConnectServer exit with -1, need pass
AngersZh commented on PR #40315:
URL: https://github.com/apache/spark/pull/40315#issuecomment-1457984092
ping @HyukjinKwon
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
jerqi commented on PR #40307:
URL: https://github.com/apache/spark/pull/40307#issuecomment-1457723960
> @jerqi locality may still have benefits when RSS works in hybrid
deployments, besides, there is a dedicated configuration for that
`spark.shuffle.reduceLocality.enabled`
AngersZh commented on PR #40314:
URL: https://github.com/apache/spark/pull/40314#issuecomment-1457967882
ping @cloud-fan @dongjoon-hyun @HyukjinKwon
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above
zhengruifeng commented on PR #40097:
URL: https://github.com/apache/spark/pull/40097#issuecomment-1457733081
@WeichenXu123 I think it is ready for review
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above
alkis commented on PR #40302:
URL: https://github.com/apache/spark/pull/40302#issuecomment-1457984961
> Mind retriggering https://github.com/alkis/spark/runs/11797022157?
Done.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to
peter-toth commented on code in PR #40268:
URL: https://github.com/apache/spark/pull/40268#discussion_r1127511901
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/optimizer/expressions.scala:
##
@@ -138,56 +136,53 @@ object ConstantPropagation extends
FurcyPin commented on code in PR #40271:
URL: https://github.com/apache/spark/pull/40271#discussion_r1127557744
##
python/pyspark/sql/tests/test_functions.py:
##
@@ -1268,6 +1268,12 @@ def test_bucket(self):
message_parameters={"arg_name": "numBuckets", "arg_type":
AngersZh opened a new pull request, #40314:
URL: https://github.com/apache/spark/pull/40314
### What changes were proposed in this pull request?
Currently when we run client mode SparkSubmit, when we catch exception
during `runMain()`
It just calls `sc.stop()`, then AM still exit
xingchaozh opened a new pull request, #40312:
URL: https://github.com/apache/spark/pull/40312
### What changes were proposed in this pull request?
We could handle the steam side skew of BroadcastHashJoin to improve the join
performance
Before | After
-- | --
HyukjinKwon commented on code in PR #40270:
URL: https://github.com/apache/spark/pull/40270#discussion_r1127644938
##
connector/connect/common/src/main/protobuf/spark/connect/relations.proto:
##
@@ -781,3 +782,10 @@ message FrameMap {
CommonInlineUserDefinedFunction func =
HyukjinKwon commented on PR #40311:
URL: https://github.com/apache/spark/pull/40311#issuecomment-1457730273
Merged to master and branch-3.4.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
HyukjinKwon closed pull request #40311:
[SPARK-42559][CONNECT][TESTS][FOLLOW-UP] Disable ANSI in several tests at
DataFrameNaFunctionSuite.scala
URL: https://github.com/apache/spark/pull/40311
--
This is an automated message from the Apache Git Service.
To respond to the message, please log
zhengruifeng commented on code in PR #40097:
URL: https://github.com/apache/spark/pull/40097#discussion_r1127499837
##
mllib/common/pom.xml:
##
@@ -0,0 +1,109 @@
+
+
+
+http://maven.apache.org/POM/4.0.0;
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance;
HyukjinKwon commented on code in PR #40270:
URL: https://github.com/apache/spark/pull/40270#discussion_r1127647073
##
connector/connect/common/src/main/protobuf/spark/connect/relations.proto:
##
@@ -781,3 +782,10 @@ message FrameMap {
CommonInlineUserDefinedFunction func =
HyukjinKwon commented on PR #40302:
URL: https://github.com/apache/spark/pull/40302#issuecomment-1457930310
Mind retriggering https://github.com/alkis/spark/runs/11797022157?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub
HeartSaVioR commented on code in PR #40215:
URL: https://github.com/apache/spark/pull/40215#discussion_r1127778137
##
docs/structured-streaming-programming-guide.md:
##
@@ -1848,12 +1848,137 @@ Additional details on supported joins:
- As of Spark 2.4, you can use joins only
HeartSaVioR commented on code in PR #40215:
URL: https://github.com/apache/spark/pull/40215#discussion_r1127779011
##
docs/structured-streaming-programming-guide.md:
##
@@ -1848,12 +1848,137 @@ Additional details on supported joins:
- As of Spark 2.4, you can use joins only
itholic commented on PR #40280:
URL: https://github.com/apache/spark/pull/40280#issuecomment-1458111230
Awesome!! Let me take a look at your PR.
Thanks!
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above
LuciferYang commented on code in PR #40305:
URL: https://github.com/apache/spark/pull/40305#discussion_r1127825622
##
repl/src/main/scala-2.12/org/apache/spark/repl/Main.scala:
##
@@ -121,6 +121,11 @@ object Main extends Logging {
sparkContext = sparkSession.sparkContext
LuciferYang commented on PR #40317:
URL: https://github.com/apache/spark/pull/40317#issuecomment-1458156395
> this is the 101st we have broken the maven build in the last month alone.
We don't test with it, but we do feel comfortable to release with it. Are we
sure the dual build setup is
shrprasa commented on PR #37880:
URL: https://github.com/apache/spark/pull/37880#issuecomment-1458162368
@holdenk Thanks for approving the PR. Can you please merge this PR or tag
someone who can do it?
--
This is an automated message from the Apache Git Service.
To respond to the
hvanhovell commented on code in PR #40319:
URL: https://github.com/apache/spark/pull/40319#discussion_r1128015075
##
connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/Dataset.scala:
##
@@ -2777,7 +2778,11 @@ class Dataset[T] private[sql] (
}
def toJSON:
panbingkun opened a new pull request, #40316:
URL: https://github.com/apache/spark/pull/40316
### What changes were proposed in this pull request?
### Why are the changes needed?
### Does this PR introduce _any_ user-facing change?
No.
### How was this patch
LuciferYang commented on code in PR #40318:
URL: https://github.com/apache/spark/pull/40318#discussion_r1127827370
##
repl/src/main/scala-2.13/org/apache/spark/repl/Main.scala:
##
@@ -129,6 +129,11 @@ object Main extends Logging {
sparkContext = sparkSession.sparkContext
shrprasa commented on PR #40258:
URL: https://github.com/apache/spark/pull/40258#issuecomment-1458159825
> I'm not sure about the change, not sure I'm qualified to review it. I
think at best the error message should change; I am not clear that the result
is 'wrong'
Thanks for
hvanhovell commented on PR #40276:
URL: https://github.com/apache/spark/pull/40276#issuecomment-1458274759
At the end of the day it is an optimization. However I do it is a sound one
to have.
--
This is an automated message from the Apache Git Service.
To respond to the message, please
cloud-fan commented on code in PR #40190:
URL: https://github.com/apache/spark/pull/40190#discussion_r1127955295
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/optimizer/UnwrapCastInBinaryComparison.scala:
##
@@ -350,7 +384,7 @@ object UnwrapCastInBinaryComparison
cloud-fan commented on code in PR #40294:
URL: https://github.com/apache/spark/pull/40294#discussion_r1127959943
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/optimizer/UnwrapCastInBinaryComparison.scala:
##
@@ -133,6 +133,11 @@ object
LuciferYang opened a new pull request, #40319:
URL: https://github.com/apache/spark/pull/40319
### What changes were proposed in this pull request?
### Why are the changes needed?
### Does this PR introduce _any_ user-facing change?
###
hvanhovell closed pull request #40318: [SPARK-42656][SPARK
SHELL][CONNECT][FOLLOWUP] Add same `ClassNotFoundException` catch to
`repl.Main` for Scala 2.13
URL: https://github.com/apache/spark/pull/40318
--
This is an automated message from the Apache Git Service.
To respond to the message,
hvanhovell commented on code in PR #40319:
URL: https://github.com/apache/spark/pull/40319#discussion_r1128052249
##
connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/Dataset.scala:
##
@@ -2777,7 +2778,11 @@ class Dataset[T] private[sql] (
}
def toJSON:
hvanhovell commented on code in PR #40291:
URL: https://github.com/apache/spark/pull/40291#discussion_r1128061371
##
connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/DataFrameWriter.scala:
##
@@ -345,6 +345,48 @@ final class DataFrameWriter[T] private[sql] (ds:
panbingkun commented on PR #40280:
URL: https://github.com/apache/spark/pull/40280#issuecomment-1458081353
> Thanks, @panbingkun ! By the way, I think this issue has a pretty high
priority since the default nullability of a schema is `False`.
>
> ```python
> >>> sdf =
waitinfuture commented on PR #40307:
URL: https://github.com/apache/spark/pull/40307#issuecomment-1458107057
> This is still WIP, but want to get early feedback. +CC @Ngone51, @otterc,
@waitinfuture
Hi @mridulm , thanks for the work and it really simplifies the usage of
Apache
srowen commented on PR #40258:
URL: https://github.com/apache/spark/pull/40258#issuecomment-1458138622
I'm not sure about the change, not sure I'm qualified to review it. I think
at best the error message should change; I am not clear that the result is
'wrong'
--
This is an automated
LuciferYang opened a new pull request, #40318:
URL: https://github.com/apache/spark/pull/40318
### What changes were proposed in this pull request?
This pr add the same `ClassNotFoundException` catch to `repl.Main` for
Scala 2.13 as https://github.com/apache/spark/pull/40305 due
WeichenXu123 commented on code in PR #40297:
URL: https://github.com/apache/spark/pull/40297#discussion_r1127830764
##
connector/connect/common/src/main/protobuf/spark/connect/relations.proto:
##
@@ -81,13 +82,50 @@ message Relation {
// Catalog API (experimental /
cloud-fan commented on PR #40276:
URL: https://github.com/apache/spark/pull/40276#issuecomment-1458245647
does it mean every spark connect client must implement a data type parser in
its language? This seems a bit overkill. Can we revisit all the places that
need to parse data type at
cloud-fan commented on PR #38358:
URL: https://github.com/apache/spark/pull/38358#issuecomment-1458251751
@wangyum do you know why it's a problem only in 3.2?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL
LuciferYang commented on code in PR #40319:
URL: https://github.com/apache/spark/pull/40319#discussion_r1128066536
##
connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/Dataset.scala:
##
@@ -2777,7 +2778,11 @@ class Dataset[T] private[sql] (
}
def toJSON:
beliefer commented on PR #40287:
URL: https://github.com/apache/spark/pull/40287#issuecomment-1458109080
> @beliefer here is the thing. When this was designed it was mainly aimed at
sql, and there we definitely do not generate unique names in lambda functions
either. This is all done in
WeichenXu123 commented on code in PR #40297:
URL: https://github.com/apache/spark/pull/40297#discussion_r1127841115
##
connector/connect/server/src/main/scala/org/apache/spark/sql/connect/ml/AlgorithmRegisty.scala:
##
@@ -0,0 +1,104 @@
+/*
+ * Licensed to the Apache Software
WeichenXu123 commented on code in PR #40297:
URL: https://github.com/apache/spark/pull/40297#discussion_r1127841115
##
connector/connect/server/src/main/scala/org/apache/spark/sql/connect/ml/AlgorithmRegisty.scala:
##
@@ -0,0 +1,104 @@
+/*
+ * Licensed to the Apache Software
beliefer commented on PR #40287:
URL: https://github.com/apache/spark/pull/40287#issuecomment-1458193681
> E... SQL/scala/Python all use the analyzer; they are all just
frontends to the same thing.
I found the reason. Although the scala API use analyzer too. `object
justaparth opened a new pull request, #40320:
URL: https://github.com/apache/spark/pull/40320
### What changes were proposed in this pull request?
I was reviewing this markdown document about proto parsing, and found that
the formatting of code blocks looked incorrect:
some
hvanhovell commented on PR #40305:
URL: https://github.com/apache/spark/pull/40305#issuecomment-1458096673
Merging.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To
hvanhovell closed pull request #40305: [SPARK-42656][CONNECT][Followup] Spark
Connect Shell
URL: https://github.com/apache/spark/pull/40305
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
zhengruifeng commented on code in PR #40297:
URL: https://github.com/apache/spark/pull/40297#discussion_r1127788041
##
connector/connect/server/src/main/scala/org/apache/spark/sql/connect/ml/MLUtils.scala:
##
@@ -0,0 +1,113 @@
+/*
+ * Licensed to the Apache Software Foundation
hvanhovell commented on PR #40287:
URL: https://github.com/apache/spark/pull/40287#issuecomment-1458116337
E... SQL/scala/Python all use the analyzer; they are all just frontends
to the same thing.
--
This is an automated message from the Apache Git Service.
To respond to the
LuciferYang opened a new pull request, #40317:
URL: https://github.com/apache/spark/pull/40317
### What changes were proposed in this pull request?
Run the following commands
```
build/mvn clean install -DskipTests -pl connector/connect/server -am
build/mvn test -pl
hvanhovell commented on code in PR #40315:
URL: https://github.com/apache/spark/pull/40315#discussion_r1127848561
##
sql/core/src/main/scala/org/apache/spark/sql/SparkSession.scala:
##
@@ -736,13 +736,15 @@ class SparkSession private(
}
// scalastyle:on
+ def stop():
cloud-fan commented on code in PR #40308:
URL: https://github.com/apache/spark/pull/40308#discussion_r1127909983
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/Analyzer.scala:
##
@@ -3344,43 +3345,6 @@ class Analyzer(override val catalogManager:
panbingkun commented on PR #40316:
URL: https://github.com/apache/spark/pull/40316#issuecomment-1458241313
https://user-images.githubusercontent.com/15246973/223446693-3c296b56-f9aa-4b70-9eb3-5bc9059ba631.png;>
--
This is an automated message from the Apache Git Service.
To respond to
LuciferYang commented on code in PR #40319:
URL: https://github.com/apache/spark/pull/40319#discussion_r1128042664
##
connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/Dataset.scala:
##
@@ -2777,7 +2778,11 @@ class Dataset[T] private[sql] (
}
def toJSON:
mridulm commented on code in PR #40286:
URL: https://github.com/apache/spark/pull/40286#discussion_r1128353043
##
core/src/test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala:
##
@@ -4572,6 +4572,48 @@ class DAGSchedulerSuite extends SparkFunSuite with
mridulm commented on code in PR #40286:
URL: https://github.com/apache/spark/pull/40286#discussion_r1128353043
##
core/src/test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala:
##
@@ -4572,6 +4572,48 @@ class DAGSchedulerSuite extends SparkFunSuite with
hvanhovell commented on code in PR #40291:
URL: https://github.com/apache/spark/pull/40291#discussion_r1128152776
##
connector/connect/common/src/main/protobuf/spark/connect/commands.proto:
##
@@ -116,6 +116,7 @@ message WriteOperation {
TABLE_SAVE_METHOD_UNSPECIFIED =
mridulm commented on code in PR #40286:
URL: https://github.com/apache/spark/pull/40286#discussion_r1128351289
##
core/src/test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala:
##
@@ -4572,6 +4572,48 @@ class DAGSchedulerSuite extends SparkFunSuite with
hvanhovell commented on code in PR #40277:
URL: https://github.com/apache/spark/pull/40277#discussion_r1128154890
##
connector/connect/common/src/main/protobuf/spark/connect/relations.proto:
##
@@ -140,6 +140,11 @@ message Read {
// (Optional) A list of path for
amaliujia commented on PR #40319:
URL: https://github.com/apache/spark/pull/40319#issuecomment-1458469794
LGTM
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To
otterc commented on code in PR #40307:
URL: https://github.com/apache/spark/pull/40307#discussion_r1128337921
##
core/src/main/scala/org/apache/spark/ExecutorAllocationManager.scala:
##
@@ -203,7 +205,8 @@ private[spark] class ExecutorAllocationManager(
throw new
mridulm commented on code in PR #40313:
URL: https://github.com/apache/spark/pull/40313#discussion_r1128332164
##
core/src/main/scala/org/apache/spark/ui/SparkUI.scala:
##
@@ -167,7 +167,7 @@ private[spark] class SparkUI private (
attemptId = None,
startTime =
sunchao commented on code in PR #40190:
URL: https://github.com/apache/spark/pull/40190#discussion_r1128241190
##
sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/optimizer/UnwrapCastInBinaryComparisonSuite.scala:
##
@@ -368,6 +370,61 @@ class
ryan-johnson-databricks opened a new pull request, #40321:
URL: https://github.com/apache/spark/pull/40321
### What changes were proposed in this pull request?
The `AddMetadataColumns` analyzer rule is designed to resolve metadata
columns using
mridulm commented on code in PR #40313:
URL: https://github.com/apache/spark/pull/40313#discussion_r1128332164
##
core/src/main/scala/org/apache/spark/ui/SparkUI.scala:
##
@@ -167,7 +167,7 @@ private[spark] class SparkUI private (
attemptId = None,
startTime =
LuciferYang commented on PR #40323:
URL: https://github.com/apache/spark/pull/40323#issuecomment-1459158853
Is there a similar case on Scala connect client ?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
jerqi commented on PR #40307:
URL: https://github.com/apache/spark/pull/40307#issuecomment-1459167441
> @jerqi Agree that we should have a way to specify locality preference for
disaggregated shuffle implementations to spark scheduler - so that shuffle
tasks are closer to the data.
>
yaooqinn commented on code in PR #40313:
URL: https://github.com/apache/spark/pull/40313#discussion_r1128885696
##
core/src/main/scala/org/apache/spark/ui/SparkUI.scala:
##
@@ -167,7 +167,7 @@ private[spark] class SparkUI private (
attemptId = None,
startTime
AngersZh commented on code in PR #40315:
URL: https://github.com/apache/spark/pull/40315#discussion_r112640
##
sql/core/src/main/scala/org/apache/spark/sql/SparkSession.scala:
##
@@ -736,13 +736,15 @@ class SparkSession private(
}
// scalastyle:on
+ def stop():
panbingkun commented on PR #40316:
URL: https://github.com/apache/spark/pull/40316#issuecomment-1459182974
cc @itholic
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To
ueshin commented on PR #40323:
URL: https://github.com/apache/spark/pull/40323#issuecomment-1459184767
> Is there a similar case on Scala connect client ?
I haven't tried Scala client, but yes, it would happen, and this will fix
both.
--
This is an automated message from the
AngersZh commented on code in PR #40314:
URL: https://github.com/apache/spark/pull/40314#discussion_r1128899260
##
core/src/main/scala/org/apache/spark/deploy/SparkSubmit.scala:
##
@@ -1005,17 +1005,20 @@ private[spark] class SparkSubmit extends Logging {
e
}
AngersZh commented on code in PR #40314:
URL: https://github.com/apache/spark/pull/40314#discussion_r1128902772
##
core/src/main/scala/org/apache/spark/deploy/SparkSubmit.scala:
##
@@ -1005,17 +1005,20 @@ private[spark] class SparkSubmit extends Logging {
e
}
allanf-db opened a new pull request, #40324:
URL: https://github.com/apache/spark/pull/40324
### What changes were proposed in this pull request?
Adding a Spark Connect overview page to the Spark 3.4 documentation.
### Why are the changes needed?
The first
itholic commented on code in PR #40316:
URL: https://github.com/apache/spark/pull/40316#discussion_r1128906598
##
python/pyspark/sql/tests/connect/test_connect_basic.py:
##
@@ -2876,6 +2876,13 @@ def test_unsupported_io_functions(self):
with
1 - 100 of 184 matches
Mail list logo